The Curious Case of Neural Text Degeneration

04/22/2019
by   Ari Holtzman, et al.
0

Despite considerable advancements with deep neural language models, the enigma of neural text degeneration persists when these models are tested as text generators. The counter-intuitive empirical observation is that even though the use of likelihood as training objective leads to high quality models for a broad range of language understanding tasks, using likelihood as a decoding objective leads to text that is bland and strangely repetitive. In this paper, we reveal surprising distributional differences between human text and machine text. In addition, we find that decoding strategies alone can dramatically effect the quality of machine text, even when generated from exactly the same neural language model. Our findings motivate Nucleus Sampling, a simple but effective method to draw the best out of neural generation. By sampling text from the dynamic nucleus of the probability distribution, which allows for diversity while effectively truncating the less reliable tail of the distribution, the resulting text better demonstrates the quality of human text, yielding enhanced diversity without sacrificing fluency and coherence.

READ FULL TEXT
research
01/02/2021

On-the-Fly Attention Modularization for Neural Generation

Despite considerable advancements with deep neural language models (LMs)...
research
07/29/2020

Mirostat: A Perplexity-Controlled Neural Text Decoding Algorithm

Neural text decoding is important for generating high-quality texts usin...
research
04/22/2020

Trading Off Diversity and Quality in Natural Language Generation

For open-ended language generation tasks such as storytelling and dialog...
research
11/09/2019

How Decoding Strategies Affect the Verifiability of Generated Text

Language models are of considerable importance. They are used for pretra...
research
06/16/2023

FALL-E: A Foley Sound Synthesis Model and Strategies

This paper introduces FALL-E, a foley synthesis system and its training/...
research
03/13/2021

Improving Diversity of Neural Text Generation via Inverse Probability Weighting

The neural network based text generation suffers from the text degenerat...
research
04/06/2020

Sparse Text Generation

Current state-of-the-art text generators build on powerful language mode...

Please sign up or login with your details

Forgot password? Click here to reset