Log In Sign Up

Graph-Based Decoding for Task Oriented Semantic Parsing

by   Jeremy R. Cole, et al.

The dominant paradigm for semantic parsing in recent years is to formulate parsing as a sequence-to-sequence task, generating predictions with auto-regressive sequence decoders. In this work, we explore an alternative paradigm. We formulate semantic parsing as a dependency parsing task, applying graph-based decoding techniques developed for syntactic parsing. We compare various decoding techniques given the same pre-trained Transformer encoder on the TOP dataset, including settings where training data is limited or contains only partially-annotated examples. We find that our graph-based approach is competitive with sequence decoders on the standard setting, and offers significant improvements in data efficiency and settings where partially-annotated data is available.


page 1

page 2

page 3

page 4


AMR Parsing as Sequence-to-Graph Transduction

We propose an attention-based model that treats AMR parsing as sequence-...

Structure-aware Fine-tuning of Sequence-to-sequence Transformers for Transition-based AMR Parsing

Predicting linearized Abstract Meaning Representation (AMR) graphs using...

Sequence Labeling Parsing by Learning Across Representations

We use parsing as sequence labeling as a common framework to learn acros...

Evaluating Induced CCG Parsers on Grounded Semantic Parsing

We compare the effectiveness of four different syntactic CCG parsers for...

Transfer Learning for Neural Semantic Parsing

The goal of semantic parsing is to map natural language to a machine int...

Schema-Free Dependency Parsing via Sequence Generation

Dependency parsing aims to extract syntactic dependency structure or sem...

Maximum Spanning Trees Are Invariant to Temperature Scaling in Graph-based Dependency Parsing

Modern graph-based syntactic dependency parsers operate by predicting, f...