The standard training algorithm in neural machine translation (NMT) suffers from exposure bias, and alternative algorithms have been proposed to mitigate this. However, the practical impact of exposure bias is under debate. In this paper, we link exposure bias to another well-known problem in NMT, namely the tendency to generate hallucinations under domain shift. In experiments on three datasets with multiple test domains, we show that exposure bias is partially to blame for hallucinations, and that training with Minimum Risk Training, which avoids exposure bias, can mitigate this. Our analysis explains why exposure bias is more problematic under domain shift, and also links exposure bias to the beam search problem, i.e. performance deterioration with increasing beam size. Our results provide a new justification for methods that reduce exposure bias: even if they do not increase performance on in-domain test sets, they can increase model robustness to domain shift.READ FULL TEXT VIEW PDF
We study two problems in neural machine translation (NMT). First, in bea...
The exposure bias problem refers to the training-inference discrepancy c...
Neural machine translation (NMT) models are usually trained with the
We report on search errors and model errors in neural machine translatio...
Recent studies have revealed a number of pathologies of neural machine
Training data for NLP tasks often exhibits gender bias in that fewer
We consider paths with low "exposure" to a polygonal domain, i.e., paths...
Neural Machine Translation (NMT) has advanced the state of the art in MT (Sutskever et al., 2014; Bahdanau et al., 2015; Vaswani et al., 2017), but is susceptible to domain shift. Koehn and Knowles (2017) consider out-of-domain translation one of the key challenges in NMT. Such translations may be fluent, but completely unrelated to the input (hallucinations), and their misleading nature makes them particularly problematic.
We hypothesise that exposure bias (Ranzato et al., 2016), a discrepancy between training and inference, makes this problem worse. Specifically, training with teacher forcing only exposes the model to gold history, while previous predictions during inference may be erroneous. Thus, the model trained with teacher forcing may over-rely on previously predicted words, which would exacerbate error propagation. Previous work has sought to reduce exposure bias in training (Bengio et al., 2015; Ranzato et al., 2016; Shen et al., 2016; Wiseman and Rush, 2016; Zhang et al., 2019). However, the relevance of error propagation is under debate: Wu et al. (2018) argue that its role is overstated in literature, and that linguistic features explain some of the accuracy drop at higher time steps.
Previous work has established a link between domain shift and hallucination in NMT (Koehn and Knowles, 2017; Müller et al., 2019). In this paper, we will aim to also establish an empirical link between hallucination and exposure bias. Such a link will deepen our understanding of the hallucination problem, but also has practical relevance, e.g. to help predicting in which settings the use of sequence-level objectives is likely to be helpful. We further empirically confirm the link between exposure bias and the ‘beam search problem’, i.e. the fact that translation quality does not increase consistently with beam size (Koehn and Knowles, 2017; Ott et al., 2018; Stahlberg and Byrne, 2019).
We base our experiments on GermanEnglish IWSLT’14, and two datasets used to investigate domain robustness by Müller et al. (2019): a selection of corpora from OPUS (Lison and Tiedemann, 2016) for GermanEnglish, and a low-resource GermanRomansh scenario. We experiment with Minimum Risk Training (MRT) (Och, 2003; Shen et al., 2016), a training objective which inherently avoids exposure bias.
Our experiments show that MRT indeed improves quality more in out-of-domain settings, and reduces the amount of hallucination. Our analysis of translation uncertainty also shows how the MLE baseline over-estimates the probability of random translations at all but the initial time steps, and how MRT mitigates this problem. Finally, we show that the beam search problem is reduced by MRT.
The de-facto standard training objective in NMT is to minimize the negative log-likelihood of the training data 111This is equivalent to maximizing the likelihood of the data, hence Maximum Likelihood Estimation (MLE).:
where and are the source and target sequence, respectively, is the tth token in , and denotes all previous tokens. MLE is typically performed with teacher forcing, where are ground-truth labels in training, which creates a mismatch to inference, where are model predictions.
Minimum Risk Training (MRT) is a sequence-level objective that avoids this problem. Specifically, the objective function of MRT is the expected loss (risk) with respect to the posterior distribution:
in which the loss indicates the discrepancy between the gold translation and the model prediction . Due to the intractable search space, the posterior distribution is approximated by a subspace by sampling a certain number of candidate translations, and normalizing:
is a hyperparameter to control the sharpness of the subspace. Based on preliminary results, we use random sampling to generate candidate translations, and followingEdunov et al. (2018), do not add the reference translation to the subspace.
To verify the effectiveness of our MRT implementation on top of a strong Transformer baseline (Vaswani et al., 2017), we first conduct experiments on IWSLT’14 GermanEnglish (DEEN) (Cettolo et al., 2014), which consists of sentence pairs. We follow previous work for data splits (Ranzato et al., 2016; Edunov et al., 2018).
For experiments with domain shift, we use data sets and preprocessing as Müller et al. (2019)222https://github.com/ZurichNLP/domain-robustness. For DEEN, data comes from OPUS (Lison and Tiedemann, 2016), and is comprised of five domains: medical, IT, law, koran and subtitles. We use medical for training and development, and report results on an in-domain test set and the four other domains (out-of-domain; OOD). GermanRomansh (DERM) is a low-resource language pair where robustness to domain shift is of practical relevance. The training data is from the Allegra corpus (Scherrer and Cartoni, 2012) (law domain) with sentence pairs. The test domain are blogs, using data from Convivenza333https://www.suedostschweiz.ch/blogs/convivenza. We have access to 2000 sentences for development and testing, respectively, in each domain.
We implement444Code available at https://github.com/zippotju/Exposure-Bias-Hallucination-Domain-Shift MRT in the Nematus toolkit (Sennrich et al., 2017). All our experiments use the Transformer architecture (Vaswani et al., 2017). Following Edunov et al. (2018), we use - (Lin and Och, 2004) as the MRT loss. Models are pre-trained with the token-level objective MLE and then fine-tuned with MRT. Hyperparameters mostly follow previous work (Edunov et al., 2018; Müller et al., 2019); for MRT, we conduct limited hyperparameter search on the IWSLT’14 development set, including learning rate, batch size, and the sharpness parameter . We set the number of candidate translations for MRT to 4 to balance effectiveness and efficiency. Detailed hyperparameters are reported in the Appendix.
For comparison to previous work, we report lowercased, tokenised BLEU (Papineni et al., 2002) with multi-bleu.perl for IWSLT’14, and cased, detokenised BLEU with SacreBLEU (Post, 2018)555Signature: BLEU+c.mixed+#.1+s.exp+tok.13a+v.1.4.2
otherwise. For settings with domain shift, we report average and standard deviation of 3 independent training runs to account for optimizer instability.
The manual evaluation was performed by two native speakers of German who completed bilingual (German/English) high school or University programs. We collected
3600 annotations in total, spread over 12 configurations. We ask annotators to evaluate translations according to fluency and adequacy. For fluency, the annotator classifies a translation as fluent, partially fluent or not fluent; for adequacy, as adequate, partially adequate or inadequate. We report kappa coefficient () (Carletta, 1996) for inter-annotator and intra-annotator agreement in Table 1, and assess statistical significance with Fisher’s exact test (two-tailed).
Table 2 shows results for IWSLT’14. We compare to results by Edunov et al. (2018), who use a convolutional architecture (Gehring et al., 2017), and Wu et al. (2019), who report results with Transfomer-base and dynamic convolution.
|ConvS2S (MLE) (Edunov et al., 2018)||32.2|
|ConvS2S (MRT) (Edunov et al., 2018)||32.8 (+0.6)|
|Transformer (MLE) (Wu et al., 2019)||34.4|
|DynamicConv (MLE) (Wu et al., 2019)||35.2|
|system||in-domain||average OOD||in-domain||average OOD|
|SMT Müller et al. (2019)||58.4||11.8||45.2||15.5|
|NMT Müller et al. (2019)||61.5||11.7||52.5||18.9|
|NMT+RC+SR+NC Müller et al. (2019)||60.8||13.1||52.4||20.7|
|MLE w/o LS||58.3 (0.53)||9.7 (0.25)||52.2 (0.19)||15.8 (0.39)|
|+MRT||58.4 (0.39)||10.2 (0.26)||52.1 (0.08)||15.9 (0.28)|
|MLE w/ LS||58.9 (0.45)||11.2 (0.16)||53.9 (0.16)||18.0 (0.17)|
|+MRT||58.8 (0.36)||12.0 (0.29)||53.9 (0.12)||18.7 (0.09)|
With 34.7 BLEU, our baseline is competitive. We observe an improvement of 0.5 BLEU from MRT, comparable to Edunov et al. (2018), although we start from a stronger baseline (+2.5 BLEU).
Table 3 shows results for data sets with domain shift. To explore the effect of label smoothing (Szegedy et al., 2016), we train baselines with and without label smoothing. MLE with label smoothing performs better by itself, and we also found MRT to be more effective on top of the initial model with label smoothing. For DEEN, MRT increases average OOD BLEU by 0.8 compared to the MLE baseline with label smoothing; for DERM the improvement is 0.7 BLEU. We note that MRT does not consistently improve in-domain performance, which is a first indicator that exposure bias may be more problematic under domain shift.
Our OOD results lag slightly behind those of Müller et al. (2019), but note that the techniques employed by them, namely reconstruction (Tu et al., 2017; Niu et al., 2019), subword regularization (Kudo, 2018), and noisy channel modelling (Li and Jurafsky, 2016) are orthogonal to MRT. We leave the combination of these approaches to future work.
BLEU results indicate that MRT can improve domain robustness. In this section, we report on additional experiments to establish more direct links between exposure bias and domain robustness, hallucination, and the beam search problem. Experiments are performed on DEEN OPUS data.
We manually evaluate the proportion of hallucinated translations on out-of-domain and in-domain test sets. We follow the definition and evaluation by Müller et al. (2019), considering a translation a hallucination if it is (partially) fluent, but unrelated in content to the source text (inadequate). We report the proportion of such hallucinations for each system.
|% hallucinations (BLEU)|
|MLE w/o LS||35% (9.7)||2% (58.3)|
|MLE w/ LS||33% (11.2)||1% (58.9)|
Results in Table 4 confirm that hallucinations are much more pronounced in out-of-domain test sets (33–35%) than in in-domain test sets (1–2%). MRT reduces the proportion of hallucinations on out-of-domain test sets (N=500 for each system; reductions statistically significant at ) and improves BLEU. Note that the two metrics do not correlate perfectly: MLE with label smoothing has higher BLEU (+1) than MRT based on MLE without label smoothing, but a similar proportion of hallucinations. This indicates that label smoothing increases translation quality in other aspects, while MRT has a clear effect on the number of hallucinations, reducing it by up to 21% (relative).
|source||Wir haben ihn gefunden.|
|reference||We found him.|
|MLE||Do not pass it.|
|MRT||We have found it.|
|source||So höre nicht auf die Ableugner.|
|reference||So hearken not to those who deny (the Truth).|
|MLE||Do not drive or use machines.|
|MRT||Do not apply to dleugner.|
A closer inspection of segments where the MLE system was found to hallucinate shows that some segments were scored higher in adequacy with MRT, others lower in fluency. One example for each case is shown in Table 5. Even the example where MRT was considered disfluent and inadequate actually shows an attempt to cover the source sentence: the source word ‘Ableugner’ (denier) is mistranslated into ‘dleugner’. We consider this preferable to producing a complete hallucination.
Inspired by Ott et al. (2018), we analyse the model’s uncertainty by computing the average probability at each time step across a set of sentences. Besides the reference translations, we also consider a set of ‘distractor’ translations, which are random sentences from the in-domain test set which match the corresponding reference translation in length.
In Figure 1, we show out-of-domain results for an MLE model and multiple checkpoints of MRT fine-tuning. The left two graphs show probabilities for references and distractors, respectively. The right-most graph shows a direct comparison of probabilities for references and distractors for the MLE baseline and the final MRT model. The MLE baseline assigns similar probabilities to tokens in the references and the distractors. Only for the first time steps is there a clear preference for the references over the (mostly random!) distractors. This shows that error propagation is a big risk: should the model make a wrong prediction initially, this is unlikely to be penalised in later time steps.
MRT tends to increase the model’s certainty at later time steps666The uncertainty of the baseline is due to label smoothing., but importantly, the increase is sharper for the reference translations than for the distractors. The direct comparison shows a widening gap in certainty between the reference and distractor sentences.777For intermediate checkpoints, see Appendix, Figure 2. In other words, producing a hallucination will incur a small penalty at each time step (compared to producing the reference), presumably due to a higher reliance on the source signal, lessening the risk of error propagation and hallucinations.
Our analysis shows similar trends on in-domain references. However, much higher probabilities are assigned to the first few tokens of the references than to the distractors. Hence, it is much less likely that a hallucination is kept in the beam, or will overtake a good translation in overall probability, reducing the practical impact of the model’s over-reliance on its history.888Figures are shown in the Appendix (Figure 3).
Figure 1 shows that with MLE, distractor sentences are assigned lower probabilities than the references at the first few time steps, but are assigned similar, potentially even higher probabilities at later time steps. This establishes a connection between exposure bias and the beam search problem, i.e. the problem that increasing the search space can lead to worse model performance.999The beam search problem has previously been linked to length bias (Yang et al., 2018; Murray and Chiang, 2018) and the copy mode (Ott et al., 2018). We consider hallucinations another result of using large search spaces with MLE models. With larger beam size, it is more likely that hallucinations survive pruning at the first few time steps, and with high probabilities assigned to them at later time steps, there is a chance that they become the top-scoring translation.
We investigate whether the beam search problem is mitigated by MRT. In Table 6, we report OOD BLEU and the proportion of hallucinations with beam sizes of 1, 4 and 50. While MRT does not eliminate the beam search problem, performance drops less steeply as beam size increases. With beam size 4, our MRT models outperform the MLE baseline by 0.5-0.8 BLEU; with beam size 50, this difference grows to 0.6-1.5 BLEU. Our manual evaluation (N=200 for each system for beam size 1 and 50) shows that the proportion of hallucinations increases with beam size, and that MRT consistently reduces the proportion by 11-21% (relative). For the system with label smoothing, the relative increase in hallucinations with increasing beam size is also smaller with MRT (+33%) than with MLE (+44%).
|BLEU (% hallucinations)|
|MLE w/o LS||8.9 (28%)||9.7 (35%)||9.3 (37%)|
|+MRT||9.1 (24%)||10.2 (29%)||9.9 (33%)|
|MLE w/ LS||10.6 (27%)||11.2 (33%)||9.4 (39%)|
|+MRT||11.3 (24%)||12.0 (26%)||10.9 (32%)|
Our results and analysis show a connection between the exposure bias due to MLE training with teacher forcing and several well-known problems in neural machine translation, namely poor performance under domain shift, hallucinated translations, and deteriorating performance with increasing beam size. We find that Minimum Risk Training, which does not suffer from exposure bias, can be useful even when it does not increase performance on an in-domain test set: it increases performance under domain shift, reduces the number of hallucinations substantially, and makes beam search with large beams more stable.
Our findings are pertinent to the academic debate how big of a problem exposure bias is in practice – we find that this can vary substantially depending on the dataset –, and they provide a new justification for sequence-level training objectives that reduce or eliminate exposure bias. Furthermore, we believe that a better understanding of the links between exposure bias and well-known translation problems will help practitioners decide when sequence-level training objectives are especially promising, for example in settings where the test domain is unknown, or where hallucinations are a common problem.
Chaojun Wang was supported by the UK Engineering and Physical Sciences Research Council (EPSRC) fellowship grant EP/S001271/1 (MTStretch). Rico Sennrich acknowledges support of the Swiss National Science Foundation (MUTAMUR; no. 176727). This project has received support from Samsung Electronics Polska sp. z o.o. - Samsung R&D Institute Poland.
Proceedings of the 34th International Conference on Machine Learning - Volume 70, ICML’17, pages 1243–1252. JMLR.org.
Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 3347–3353, Hong Kong, China. Association for Computational Linguistics.
Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, AAAI’17, pages 3097–3103. AAAI Press.