Deep Contextualized Self-training for Low Resource Dependency Parsing

11/11/2019
by   Guy Rotman, et al.
0

Neural dependency parsing has proven very effective, achieving state-of-the-art results on numerous domains and languages. Unfortunately, it requires large amounts of labeled data, that is costly and laborious to create. In this paper we propose a self-training algorithm that alleviates this annotation bottleneck by training a parser on its own output. Our Deep Contextualized Self-training (DCST) algorithm utilizes representation models trained on sequence labeling tasks that are derived from the parser's output when applied to unlabeled data, and integrates these models with the base parser through a gating mechanism. We conduct experiments across multiple languages, both in low resource in-domain and in cross-domain setups, and demonstrate that DCST substantially outperforms traditional self-training as well as recent semi-supervised training methods.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/18/2017

Universal Dependencies Parsing for Colloquial Singaporean English

Singlish can be interesting to the ACL community both linguistically as ...
research
11/12/2020

Exploiting Cross-Dialectal Gold Syntax for Low-Resource Historical Languages: Towards a Generic Parser for Pre-Modern Slavic

This paper explores the possibility of improving the performance of spec...
research
09/19/2019

Low-Resource Parsing with Crosslingual Contextualized Representations

Despite advances in dependency parsing, languages with small treebanks s...
research
11/02/2020

Semi-supervised Autoencoding Projective Dependency Parsing

We describe two end-to-end autoencoding models for semi-supervised graph...
research
08/08/2019

Neural Network based Deep Transfer Learning for Cross-domain Dependency Parsing

In this paper, we describe the details of the neural dependency parser s...
research
05/27/2020

Self-Training for Unsupervised Parsing with PRPN

Neural unsupervised parsing (UP) models learn to parse without access to...
research
08/17/2021

Not All Linearizations Are Equally Data-Hungry in Sequence Labeling Parsing

Different linearizations have been proposed to cast dependency parsing a...

Please sign up or login with your details

Forgot password? Click here to reset