Data Augmentation for Spoken Language Understanding via Joint Variational Generation

09/07/2018
by   Kang Min Yoo, et al.
0

Data scarcity is one of the main obstacles of domain adaptation in spoken language understanding (SLU) due to the high cost of creating manually tagged SLU datasets. Recent works in neural text generative models, particularly latent variable models such as variational autoencoder (VAE), have shown promising results in regards to generating plausible and natural sentences. In this paper, we propose a novel generative architecture which leverages the generative power of latent variable models to jointly synthesize fully annotated utterances. Our experiments show that existing SLU models trained on the additional synthetic examples achieve performance gains. Our approach not only helps alleviate the data scarcity issue in the SLU task for many datasets but also indiscriminately improves language understanding performances for various SLU models, supported by extensive experiments and rigorous statistical testing.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/29/2020

Data Augmentation for Spoken Language Understanding via Pretrained Models

The training of spoken language understanding (SLU) models often faces t...
research
08/19/2021

Augmenting Slot Values and Contexts for Spoken Language Understanding with Pretrained Models

Spoken Language Understanding (SLU) is one essential step in building a ...
research
07/25/2021

A Joint and Domain-Adaptive Approach to Spoken Language Understanding

Spoken Language Understanding (SLU) is composed of two subtasks: intent ...
research
08/28/2019

Data Augmentation with Atomic Templates for Spoken Language Understanding

Spoken Language Understanding (SLU) converts user utterances into struct...
research
04/21/2021

Lossless Compression with Latent Variable Models

We develop a simple and elegant method for lossless compression using la...
research
07/01/2022

Vers la compréhension automatique de la parole bout-en-bout à moindre effort

Recent advances in spoken language understanding benefited from Self-Sup...
research
06/20/2019

One-vs-All Models for Asynchronous Training: An Empirical Analysis

Any given classification problem can be modeled using multi-class or One...

Please sign up or login with your details

Forgot password? Click here to reset