Multilingual Constituency Parsing with Self-Attention and Pre-Training

12/31/2018
by   Nikita Kitaev, et al.
0

We extend our previous work on constituency parsing (Kitaev and Klein, 2018) by incorporating pre-training for ten additional languages, and compare the benefits of no pre-training, ELMo (Peters et al., 2018), and BERT (Devlin et al., 2018). Pre-training is effective across all languages evaluated, and BERT outperforms ELMo in large part due to the benefits of increased model capacity. Our parser obtains new state-of-the-art results for 11 languages, including English (95.8 F1) and Chinese (91.8 F1).

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/19/2020

An Empirical Study for Vietnamese Constituency Parsing with Pre-training

In this work, we use a span-based approach for Vietnamese constituency p...
research
09/30/2021

Multilingual AMR Parsing with Noisy Knowledge Distillation

We study multilingual AMR parsing from the perspective of knowledge dist...
research
11/11/2019

Attending to Entities for Better Text Understanding

Recent progress in NLP witnessed the development of large-scale pre-trai...
research
05/02/2018

Constituency Parsing with a Self-Attentive Encoder

We demonstrate that replacing an LSTM encoder with a self-attentive arch...
research
04/02/2021

TAPAS at SemEval-2021 Task 9: Reasoning over tables with intermediate pre-training

We present the TAPAS contribution to the Shared Task on Statement Verifi...
research
11/10/2019

Rethinking Self-Attention: An Interpretable Self-Attentive Encoder-Decoder Parser

Attention mechanisms have improved the performance of NLP tasks while pr...
research
10/01/2020

Understanding tables with intermediate pre-training

Table entailment, the binary classification task of finding if a sentenc...

Please sign up or login with your details

Forgot password? Click here to reset