Jointly Improving Language Understanding and Generation with Quality-Weighted Weak Supervision of Automatic Labeling

02/06/2021
by   Ernie Chang, et al.
0

Neural natural language generation (NLG) and understanding (NLU) models are data-hungry and require massive amounts of annotated data to be competitive. Recent frameworks address this bottleneck with generative models that synthesize weak labels at scale, where a small amount of training labels are expert-curated and the rest of the data is automatically annotated. We follow that approach, by automatically constructing a large-scale weakly-labeled data with a fine-tuned GPT-2, and employ a semi-supervised framework to jointly train the NLG and NLU models. The proposed framework adapts the parameter updates to the models according to the estimated label-quality. On both the E2E and Weather benchmarks, we show that this weakly supervised training paradigm is an effective approach under low resource scenarios and outperforming benchmark systems on both datasets when 100

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/28/2021

WALNUT: A Benchmark on Weakly Supervised Learning for Natural Language Understanding

Building quality machine learning models for natural language understand...
research
11/08/2017

Fidelity-Weighted Learning

Training deep neural networks requires many training samples, but in pra...
research
01/30/2019

Learning Fast Matching Models from Weak Annotations

This paper proposes a novel training scheme for fast matching models in ...
research
09/29/2019

Semi-Supervised Neural Text Generation by Joint Learning of Natural Language Generation and Natural Language Understanding Models

In Natural Language Generation (NLG), End-to-End (E2E) systems trained t...
research
03/02/2017

Learning the Structure of Generative Models without Labeled Data

Curating labeled training data has become the primary bottleneck in mach...
research
11/01/2017

Avoiding Your Teacher's Mistakes: Training Neural Networks with Controlled Weak Supervision

Training deep neural networks requires massive amounts of training data,...
research
06/21/2018

Learning to Rank from Samples of Variable Quality

Training deep neural networks requires many training samples, but in pra...

Please sign up or login with your details

Forgot password? Click here to reset