Improving AMR Parsing with Sequence-to-Sequence Pre-training

10/05/2020
by   Dongqin Xu, et al.
0

In the literature, the research on abstract meaning representation (AMR) parsing is much restricted by the size of human-curated dataset which is critical to build an AMR parser with good performance. To alleviate such data size restriction, pre-trained models have been drawing more and more attention in AMR parsing. However, previous pre-trained models, like BERT, are implemented for general purpose which may not work as expected for the specific task of AMR parsing. In this paper, we focus on sequence-to-sequence (seq2seq) AMR parsing and propose a seq2seq pre-training approach to build pre-trained models in both single and joint way on three relevant tasks, i.e., machine translation, syntactic parsing, and AMR parsing itself. Moreover, we extend the vanilla fine-tuning method to a multi-task learning fine-tuning method that optimizes for the performance of AMR parsing while endeavors to preserve the response of pre-trained models. Extensive experimental results on two English benchmark datasets show that both the single and joint pre-trained models significantly improve the performance (e.g., from 71.5 to 80.2 on AMR 2.0), which reaches the state of the art. The result is very encouraging since we achieve this with seq2seq models rather than complex models. We make our code and model available at https://github.com/xdqkid/S2S-AMR-Parser.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/29/2021

Structure-aware Fine-tuning of Sequence-to-sequence Transformers for Transition-based AMR Parsing

Predicting linearized Abstract Meaning Representation (AMR) graphs using...
research
09/18/2023

Parameter-Efficient Long-Tailed Recognition

The "pre-training and fine-tuning" paradigm in addressing long-tailed re...
research
08/07/2021

Tiny Neural Models for Seq2Seq

Semantic parsing models with applications in task oriented dialog system...
research
11/02/2022

CODEP: Grammatical Seq2Seq Model for General-Purpose Code Generation

General-purpose code generation (GPCG) aims to automatically convert the...
research
10/20/2020

Pushing the Limits of AMR Parsing with Self-Learning

Abstract Meaning Representation (AMR) parsing has experienced a notable ...
research
05/05/2020

Multi-task pre-training of deep neural networks

In this work, we investigate multi-task learning as a way of pre-trainin...
research
05/21/2019

AMR Parsing as Sequence-to-Graph Transduction

We propose an attention-based model that treats AMR parsing as sequence-...

Please sign up or login with your details

Forgot password? Click here to reset