DialogVED: A Pre-trained Latent Variable Encoder-Decoder Model for Dialog Response Generation

04/27/2022
by   Wei Chen, et al.
0

Dialog response generation in open domain is an important research topic where the main challenge is to generate relevant and diverse responses. In this paper, we propose a new dialog pre-training framework called DialogVED, which introduces continuous latent variables into the enhanced encoder-decoder pre-training framework to increase the relevance and diversity of responses. With the help of a large dialog corpus (Reddit), we pre-train the model using the following 4 tasks, used in training language models (LMs) and Variational Autoencoders (VAEs) literature: 1) masked language model; 2) response generation; 3) bag-of-words prediction; and 4) KL divergence reduction. We also add additional parameters to model the turn structure in dialogs to improve the performance of the pre-trained model. We conduct experiments on PersonaChat, DailyDialog, and DSTC7-AVSD benchmarks for response generation. Experimental results show that our model achieves the new state-of-the-art results on all these datasets.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/19/2022

Self-supervised Graph Masking Pre-training for Graph-to-Text Generation

Large-scale pre-trained language models (PLMs) have advanced Graph-to-Te...
research
05/24/2023

Dior-CVAE: Diffusion Priors in Variational Dialog Generation

Conditional variational autoencoders (CVAEs) have been used recently for...
research
04/30/2017

A Conditional Variational Framework for Dialog Generation

Deep latent variable models have been shown to facilitate the response g...
research
03/31/2017

Learning Discourse-level Diversity for Neural Dialog Models using Conditional Variational Autoencoders

While recent neural encoder-decoder models have shown great promise in m...
research
05/14/2019

Atom Responding Machine for Dialog Generation

Recently, improving the relevance and diversity of dialogue system has a...
research
04/22/2018

Unsupervised Discrete Sentence Representation Learning for Interpretable Neural Dialog Generation

The encoder-decoder dialog model is one of the most prominent methods us...
research
10/17/2019

PLATO: Pre-trained Dialogue Generation Model with Discrete Latent Variable

Pre-training models have been proved effective for a wide range of natur...

Please sign up or login with your details

Forgot password? Click here to reset