Understanding and Detecting Hallucinations in Neural Machine Translation via Model Introspection

01/18/2023
by   Weijia Xu, et al.
0

Neural sequence generation models are known to "hallucinate", by producing outputs that are unrelated to the source text. These hallucinations are potentially harmful, yet it remains unclear in what conditions they arise and how to mitigate their impact. In this work, we first identify internal model symptoms of hallucinations by analyzing the relative token contributions to the generation in contrastive hallucinated vs. non-hallucinated outputs generated via source perturbations. We then show that these symptoms are reliable indicators of natural hallucinations, by using them to design a lightweight hallucination detector which outperforms both model-free baselines and strong classifiers based on quality estimation or large pre-trained models on manually annotated English-Chinese and German-English translation test beds.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/17/2022

Better Datastore, Better Translation: Generating Datastores from Pre-Trained Models for Nearest Neural Machine Translation

Nearest Neighbor Machine Translation (kNNMT) is a simple and effective m...
research
07/14/2017

LIUM Machine Translation Systems for WMT17 News Translation Task

This paper describes LIUM submissions to WMT17 News Translation Task for...
research
08/05/2021

WeChat Neural Machine Translation Systems for WMT21

This paper introduces WeChat AI's participation in WMT 2021 shared news ...
research
09/03/2019

Multi-agent Learning for Neural Machine Translation

Conventional Neural Machine Translation (NMT) models benefit from the tr...
research
06/15/2019

Tagged Back-Translation

Recent work in Neural Machine Translation (NMT) has shown significant qu...
research
10/21/2020

Analyzing the Source and Target Contributions to Predictions in Neural Machine Translation

In Neural Machine Translation (and, more generally, conditional language...
research
03/03/2022

As Little as Possible, as Much as Necessary: Detecting Over- and Undertranslations with Contrastive Conditioning

Omission and addition of content is a typical issue in neural machine tr...

Please sign up or login with your details

Forgot password? Click here to reset