Generalized Conditioned Dialogue Generation Based on Pre-trained Language Model

10/21/2020
by   Yan Zeng, et al.
0

We investigate the general problem of conditioned dialogue, in which a condition label is used as input to designate the type of the target response such as a persona. A major challenge for conditioned dialogue generation is the lack of substantial dialogue data labeled with conditions. Thus, we propose to complement the labeled dialogue data with labeled non-dialogue text data, and fine-tune BERT based on them. Our fine-tuning approach utilizes BERT for both encoder and decoder via different input representations and self-attention masks in order to distinguish the source and target side. On the target (generation) side, we use a new attention routing mechanism to choose between generating a generic word or condition-related word at each position. Our model is instantiated to persona- and topic-related dialogue. Experimental results in both cases show that our approach can produce significantly better responses than the state-of-the-art baselines.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/08/2018

Polite Dialogue Generation Without Parallel Data

Stylistic dialogue response generation, with valuable applications in pe...
research
11/12/2019

A Pre-training Based Personalized Dialogue Generation Model with Persona-sparse Data

Endowing dialogue systems with personas is essential to deliver more hum...
research
07/03/2022

Generating Repetitions with Appropriate Repeated Words

A repetition is a response that repeats words in the previous speaker's ...
research
09/13/2021

Text is NOT Enough: Integrating Visual Impressions intoOpen-domain Dialogue Generation

Open-domain dialogue generation in natural language processing (NLP) is ...
research
10/24/2020

Open-Domain Dialogue Generation Based on Pre-trained Language Models

Pre-trained language models have been successfully used in response gene...
research
07/13/2023

In-context Autoencoder for Context Compression in a Large Language Model

We propose the In-context Autoencoder (ICAE) for context compression in ...
research
09/27/2020

Modeling Topical Relevance for Multi-Turn Dialogue Generation

Topic drift is a common phenomenon in multi-turn dialogue. Therefore, an...

Please sign up or login with your details

Forgot password? Click here to reset