Taming Repetition in Dialogue Generation

12/16/2021
by   Yadong Xi, et al.
0

The wave of pre-training language models has been continuously improving the quality of the machine-generated conversations, however, some of the generated responses still suffer from excessive repetition, sometimes repeating words from utterance, sometimes repeating words within self-generated responses, or both. Inappropriate repetition of words can significantly degrade the quality of the generated texts. Penalized sampling is one popular solution, reducing the sampling probability of existing words during inference, however, it is highly vulnerable to the inappropriate setting of the static weight. Setting it too high can yield strange and unrealistic sentences while setting it too low makes the task of suppressing repetition trivial. To remedy the shortcomings of the above methods, we design a context-aware classifier to explicitly decide when to allow repetition and when to employ penalized sampling. Such a classifier can be easily integrated with existing decoding methods, reducing repetitions where appropriate while preserving the diversity of the text. Experimental results demonstrate that our method can generate higher quality and more authentic dialogues.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/16/2019

Improving Background Based Conversation with Context-aware Knowledge Pre-selection

Background Based Conversations (BBCs) have been developed to make dialog...
research
04/16/2020

Generate, Delete and Rewrite: A Three-Stage Framework for Improving Persona Consistency of Dialogue Generation

Maintaining a consistent personality in conversations is quite natural f...
research
07/29/2020

Mirostat: A Perplexity-Controlled Neural Text Decoding Algorithm

Neural text decoding is important for generating high-quality texts usin...
research
08/27/2021

Lingxi: A Diversity-aware Chinese Modern Poetry Generation System

Poetry generation has been a difficult task in natural language processi...
research
07/03/2022

Generating Repetitions with Appropriate Repeated Words

A repetition is a response that repeats words in the previous speaker's ...
research
02/01/2022

Typical Decoding for Natural Language Generation

Despite achieving incredibly low perplexities on myriad natural language...

Please sign up or login with your details

Forgot password? Click here to reset