Towards Quantifiable Dialogue Coherence Evaluation

06/01/2021
by   Zheng Ye, et al.
0

Automatic dialogue coherence evaluation has attracted increasing attention and is crucial for developing promising dialogue systems. However, existing metrics have two major limitations: (a) they are mostly trained in a simplified two-level setting (coherent vs. incoherent), while humans give Likert-type multi-level coherence scores, dubbed as "quantifiable"; (b) their predicted coherence scores cannot align with the actual human rating standards due to the absence of human guidance during training. To address these limitations, we propose Quantifiable Dialogue Coherence Evaluation (QuantiDCE), a novel framework aiming to train a quantifiable dialogue coherence metric that can reflect the actual human rating standards. Specifically, QuantiDCE includes two training stages, Multi-Level Ranking (MLR) pre-training and Knowledge Distillation (KD) fine-tuning. During MLR pre-training, a new MLR loss is proposed for enabling the model to learn the coarse judgement of coherence degrees. Then, during KD fine-tuning, the pretrained model is further finetuned to learn the actual human rating standards with only very few human-annotated data. To advocate the generalizability even with limited fine-tuning data, a novel KD regularization is introduced to retain the knowledge learned at the pre-training stage. Experimental results show that the model trained by QuantiDCE presents stronger correlations with human judgements than the other state-of-the-art metrics.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/08/2020

GRADE: Automatic Graph-Enhanced Coherence Metric for Evaluating Open-Domain Dialogue Systems

Automatically evaluating dialogue coherence is a challenging but high-de...
research
11/28/2022

BJTU-WeChat's Systems for the WMT22 Chat Translation Task

This paper introduces the joint submission of the Beijing Jiaotong Unive...
research
01/27/2023

A Multi-task Multi-stage Transitional Training Framework for Neural Chat Translation

Neural chat translation (NCT) aims to translate a cross-lingual chat bet...
research
05/08/2022

Scheduled Multi-task Learning for Neural Chat Translation

Neural Chat Translation (NCT) aims to translate conversational text into...
research
06/17/2020

Is this Dialogue Coherent? Learning from Dialogue Acts and Entities

In this work, we investigate the human perception of coherence in open-d...
research
04/28/2022

RoBLEURT Submission for the WMT2021 Metrics Task

In this paper, we present our submission to Shared Metrics Task: RoBLEUR...
research
10/08/2021

CheerBots: Chatbots toward Empathy and Emotionusing Reinforcement Learning

Apart from the coherence and fluency of responses, an empathetic chatbot...

Please sign up or login with your details

Forgot password? Click here to reset