Looking for a Needle in a Haystack: A Comprehensive Study of Hallucinations in Neural Machine Translation

08/10/2022
by   Nuno M. Guerreiro, et al.
0

Although the problem of hallucinations in neural machine translation (NMT) has received some attention, research on this highly pathological phenomenon lacks solid ground. Previous work has been limited in several ways: it often resorts to artificial settings where the problem is amplified, it disregards some (common) types of hallucinations, and it does not validate adequacy of detection heuristics. In this paper, we set foundations for the study of NMT hallucinations. First, we work in a natural setting, i.e., in-domain data without artificial noise neither in training nor in inference. Next, we annotate a dataset of over 3.4k sentences indicating different kinds of critical errors and hallucinations. Then, we turn to detection methods and both revisit methods used previously and propose using glass-box uncertainty-based detectors. Overall, we show that for preventive settings, (i) previously used methods are largely inadequate, (ii) sequence log-probability works best and performs on par with reference-based methods. Finally, we propose DeHallucinator, a simple method for alleviating hallucinations at test time that significantly reduces the hallucinatory rate. To ease future research, we release our annotated dataset for WMT18 German-English data, along with the model, training data, and code.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/19/2019

Semantic Neural Machine Translation using AMR

It is intuitive that semantic representations can be useful for machine ...
research
04/23/2020

Multiple Segmentations of Thai Sentences for Neural Machine Translation

Thai is a low-resource language, so it is often the case that data is no...
research
03/27/2020

Towards Supervised and Unsupervised Neural Machine Translation Baselines for Nigerian Pidgin

Nigerian Pidgin is arguably the most widely spoken language in Nigeria. ...
research
09/27/2019

On the use of BERT for Neural Machine Translation

Exploiting large pretrained models for various NMT tasks have gained a l...
research
06/21/2021

Phrase-level Active Learning for Neural Machine Translation

Neural machine translation (NMT) is sensitive to domain shift. In this p...

Please sign up or login with your details

Forgot password? Click here to reset