Multi-Referenced Training for Dialogue Response Generation

09/15/2020
by   Tianyu Zhao, et al.
0

In open-domain dialogue response generation, a dialogue context can be continued with diverse responses, and the dialogue models should capture such one-to-many relations. In this work, we first analyze the training objective of dialogue models from the view of Kullback-Leibler divergence (KLD) and show that the gap between the real world probability distribution and the single-referenced data's probability distribution prevents the model from learning the one-to-many relations efficiently. Then we explore approaches to multi-referenced training in two aspects. Data-wise, we generate diverse pseudo references from a powerful pretrained model to build multi-referenced data that provides a better approximation of the real-world distribution. Model-wise, we propose to equip variational models with an expressive prior, named linear Gaussian model (LGM). Experimental results of automated evaluation and human evaluation show that the methods yield significant improvements over baselines. We will release our code and data in https://github.com/ZHAOTING/dialog-processing.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/10/2020

Designing Precise and Robust Dialogue Response Evaluators

Automatic dialogue response evaluator has been proposed as an alternativ...
research
05/24/2023

Dior-CVAE: Diffusion Priors in Variational Dialog Generation

Conditional variational autoencoders (CVAEs) have been used recently for...
research
05/31/2021

Learning from Perturbations: Diverse and Informative Dialogue Generation with Inverse Adversarial Training

In this paper, we propose Inverse Adversarial Training (IAT) algorithm f...
research
12/18/2022

PVGRU: Generating Diverse and Relevant Dialogue Responses via Pseudo-Variational Mechanism

We investigate response generation for multi-turn dialogue in generative...
research
12/04/2022

Constructing Highly Inductive Contexts for Dialogue Safety through Controllable Reverse Generation

Large pretrained language models can easily produce toxic or biased cont...
research
08/27/2020

Improvement of a dedicated model for open domain persona-aware dialogue generation

This paper analyzes some speed and performance improvement methods of Tr...
research
03/02/2023

Less is More: Mitigate Spurious Correlations for Open-Domain Dialogue Response Generation Models by Causal Discovery

In this paper, we conduct the first study on spurious correlations for o...

Please sign up or login with your details

Forgot password? Click here to reset