Sequence-to-Sequence Data Augmentation for Dialogue Language Understanding

07/04/2018
by   Yutai Hou, et al.
0

In this paper, we study the problem of data augmentation for language understanding in task-oriented dialogue system. In contrast to previous work which augments an utterance without considering its relation with other utterances, we propose a sequence-to-sequence generation based data augmentation framework that leverages one utterance's same semantic alternatives in the training data. A novel diversity rank is incorporated into the utterance representation to make the model produce diverse utterances and these diversely augmented utterances help to improve the language understanding module. Experimental results on the Airline Travel Information System dataset and a newly created semantic frame annotation on Stanford Multi-turn, Multidomain Dialogue Dataset show that our framework achieves significant improvements of 6.38 and 10.04 F-scores respectively when only a training set of hundreds utterances is represented. Case studies also confirm that our method generates diverse utterances.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/29/2020

Data Augmentation for Spoken Language Understanding via Pretrained Models

The training of spoken language understanding (SLU) models often faces t...
research
08/28/2019

Data Augmentation with Atomic Templates for Spoken Language Understanding

Spoken Language Understanding (SLU) converts user utterances into struct...
research
12/13/2020

C2C-GenDA: Cluster-to-Cluster Generation for Data Augmentation of Slot Filling

Slot filling, a fundamental module of spoken language understanding, oft...
research
07/22/2016

CFGs-2-NLU: Sequence-to-Sequence Learning for Mapping Utterances to Semantics and Pragmatics

In this paper, we present a novel approach to natural language understan...
research
11/24/2019

Enhancing Out-Of-Domain Utterance Detection with Data Augmentation Based on Word Embeddings

For most intelligent assistant systems, it is essential to have a mechan...
research
09/30/2020

Learning from Mistakes: Combining Ontologies via Self-Training for Dialogue Generation

Natural language generators (NLGs) for task-oriented dialogue typically ...
research
07/26/2022

Controllable User Dialogue Act Augmentation for Dialogue State Tracking

Prior work has demonstrated that data augmentation is useful for improvi...

Please sign up or login with your details

Forgot password? Click here to reset