Why Do Neural Response Generation Models Prefer Universal Replies?

08/28/2018
by   Bowen Wu, et al.
0

Recent advances in sequence-to-sequence learning reveal a purely data-driven approach to the response generation task. Despite its diverse applications, existing neural models are prone to producing short and generic replies, making it infeasible to tackle open-domain challenges. In this research, we analyze this critical issue in light of the model's optimization goal and the specific characteristics of the human-to-human dialog corpus. By decomposing the black box into parts, a detailed analysis of the probability limit was conducted to reveal the reason behind these universal replies. Based on these analyses, we propose a max-margin ranking regularization term to avoid the models leaning to these replies. Finally, empirical experiments on case studies and benchmarks with several metrics validate this approach.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/02/2018

Augmenting Neural Response Generation with Context-Aware Topical Attention

Sequence-to-Sequence (Seq2Seq) models have witnessed a notable success i...
research
06/12/2023

TrojPrompt: A Black-box Trojan Attack on Pre-trained Language Models

Prompt learning has been proven to be highly effective in improving pre-...
research
03/17/2020

Recent Advances and Challenges in Task-oriented Dialog System

Due to the significance and value in human-computer interaction and natu...
research
06/14/2019

Neural Response Generation with Meta-Words

We present open domain response generation with meta-words. A meta-word ...
research
07/05/2019

Deep Reinforcement Learning For Modeling Chit-Chat Dialog With Discrete Attributes

Open domain dialog systems face the challenge of being repetitive and pr...
research
05/25/2021

Empirical Error Modeling Improves Robustness of Noisy Neural Sequence Labeling

Despite recent advances, standard sequence labeling systems often fail w...

Please sign up or login with your details

Forgot password? Click here to reset