Quantifying Exposure Bias for Neural Language Generation

05/25/2019
by   Tianxing He, et al.
0

The exposure bias problem refers to the training-inference discrepancy caused by teacher forcing in maximum likelihood estimation (MLE) training for recurrent neural network language models (RNNLM). It has been regarded as a central problem for natural language generation (NLG) model training. Although a lot of algorithms have been proposed to avoid teacher forcing and therefore to remove exposure bias, there is little work showing how serious the exposure bias problem is. In this work, starting from the definition of exposure bias, we propose two simple and intuitive approaches to quantify exposure bias for MLE-trained language models. Experiments are conducted on both synthetic and real data-sets. Surprisingly, our results indicate that either exposure bias is trivial (i.e. indistinguishable from the mismatch between model and data distribution), or is not as significant as it is presumed to be (with a measured performance gap of 3 viewpoint that teacher forcing or exposure bias is a major drawback of MLE training.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/24/2021

Reducing Exposure Bias in Training Recurrent Neural Network Transducers

When recurrent neural network transducers (RNNTs) are trained using the ...
research
10/01/2019

Generalization in Generation: A closer look at Exposure Bias

Exposure bias refers to the train-test discrepancy that seemingly arises...
research
10/13/2019

Rethinking Exposure Bias In Language Modeling

Exposure bias describes the phenomenon that a language model trained und...
research
09/17/2021

Relating Neural Text Degeneration to Exposure Bias

This work focuses on relating two mysteries in neural-based text generat...
research
11/06/2018

Language GANs Falling Short

Generating high-quality text with sufficient diversity is essential for ...
research
09/15/2018

Neural Networks and Quantifier Conservativity: Does Data Distribution Affect Learnability?

All known natural language determiners are conservative. Psycholinguisti...
research
04/03/2022

Why Exposure Bias Matters: An Imitation Learning Perspective of Error Accumulation in Language Generation

Current language generation models suffer from issues such as repetition...

Please sign up or login with your details

Forgot password? Click here to reset