Graph-Based Universal Dependency Parsing in the Age of the Transformer: What Works, and What Doesn't

10/23/2020
by   Stefan Grünewald, et al.
0

Current state-of-the-art graph-based dependency parsers differ on various dimensions. Among others, these include (a) the choice of pre-trained word embeddings or language models used for representing token, (b) training setups performing only parsing or additional tasks such as part-of-speech-tagging, and (c) their mechanism of constructing trees or graphs from edge scores. Because of this, it is difficult to estimate the impact of these architectural decisions when comparing parsers. In this paper, we perform a series of experiments on STEPS, a new modular graph-based parser for basic and enhanced Universal Dependencies, analyzing the effects of architectural configurations. We find that pre-trained embeddings have by far the greatest and most clear-cut impact on parser performance. The choice of factorized vs. unfactorized architectures and a multi-task training setup affect parsing accuracy in more subtle ways, depending on target language and output representation (trees vs. graphs). Our parser achieves new state-of-the-art results for a wide range of languages on both basic as well as enhanced Universal Dependencies, using a unified and comparatively simple architecture for both parsing tasks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/11/2018

An improved neural network model for joint POS tagging and dependency parsing

We propose a novel neural network model for joint part-of-speech (POS) t...
research
07/06/2021

Enhanced Universal Dependency Parsing with Automated Concatenation of Embeddings

This paper describes the system used in submission from SHANGHAITECH tea...
research
07/11/2018

UniParse: A universal graph-based parsing toolkit

This paper describes the design and use of the graph-based parsing frame...
research
06/10/2022

Sort by Structure: Language Model Ranking as Dependency Probing

Making an informed choice of pre-trained language model (LM) is critical...
research
07/14/2021

TGIF: Tree-Graph Integrated-Format Parser for Enhanced UD with Two-Stage Generic- to Individual-Language Finetuning

We present our contribution to the IWPT 2021 shared task on parsing into...
research
06/21/2023

A Semi-Autoregressive Graph Generative Model for Dependency Graph Parsing

Recent years have witnessed the impressive progress in Neural Dependency...
research
11/07/2018

IMS at the PolEval 2018: A Bulky Ensemble Depedency Parser meets 12 Simple Rules for Predicting Enhanced Dependencies in Polish

This paper presents the IMS contribution to the PolEval 2018 Shared Task...

Please sign up or login with your details

Forgot password? Click here to reset