Parsing with Pretrained Language Models, Multiple Datasets, and Dataset Embeddings

12/07/2021
by   Rob van der Goot, et al.
0

With an increase of dataset availability, the potential for learning from a variety of data sources has increased. One particular method to improve learning from multiple data sources is to embed the data source during training. This allows the model to learn generalizable features as well as distinguishing features between datasets. However, these dataset embeddings have mostly been used before contextualized transformer-based embeddings were introduced in the field of Natural Language Processing. In this work, we compare two methods to embed datasets in a transformer-based multilingual dependency parser, and perform an extensive evaluation. We show that: 1) embedding the dataset is still beneficial with these models 2) performance increases are highest when embedding the dataset at the encoder level 3) unsurprisingly, we confirm that performance increases are highest for small datasets and datasets with a low baseline score. 4) we show that training on the combination of all datasets performs similarly to designing smaller clusters based on language-relatedness.

READ FULL TEXT

page 4

page 6

research
03/01/2021

On the Effectiveness of Dataset Embeddings in Mono-lingual,Multi-lingual and Zero-shot Conditions

Recent complementary strands of research have shown that leveraging info...
research
08/17/2022

Neural Embeddings for Text

We propose a new kind of embedding for natural language text that deeply...
research
02/04/2016

Many Languages, One Parser

We train one multilingual model for dependency parsing and use it to par...
research
01/21/2021

Distilling Large Language Models into Tiny and Effective Students using pQRNN

Large pre-trained multilingual models like mBERT, XLM-R achieve state of...
research
10/15/2022

A Simple and Strong Baseline for End-to-End Neural RST-style Discourse Parsing

To promote and further develop RST-style discourse parsing models, we ne...
research
07/05/2023

Deductive Additivity for Planning of Natural Language Proofs

Current natural language systems designed for multi-step claim validatio...
research
10/01/2015

A Generative Model of Words and Relationships from Multiple Sources

Neural language models are a powerful tool to embed words into semantic ...

Please sign up or login with your details

Forgot password? Click here to reset