Data Rejuvenation: Exploiting Inactive Training Examples for Neural Machine Translation

10/06/2020
by   Wenxiang Jiao, et al.
0

Large-scale training datasets lie at the core of the recent success of neural machine translation (NMT) models. However, the complex patterns and potential noises in the large-scale data make training NMT models difficult. In this work, we explore to identify the inactive training examples which contribute less to the model performance, and show that the existence of inactive examples depends on the data distribution. We further introduce data rejuvenation to improve the training of NMT models on large-scale datasets by exploiting inactive examples. The proposed framework consists of three phases. First, we train an identification model on the original training data, and use it to distinguish inactive examples and active examples by their sentence-level output probabilities. Then, we train a rejuvenation model on the active examples, which is used to re-label the inactive examples with forward-translation. Finally, the rejuvenated examples and the active examples are combined to train the final NMT model. Experimental results on WMT14 English-German and English-French datasets show that the proposed data rejuvenation consistently and significantly improves performance for several strong NMT models. Extensive analyses reveal that our approach stabilizes and accelerates the training process of NMT models, resulting in final models with better generalization capability.

READ FULL TEXT
research
06/04/2019

Exploiting Sentential Context for Neural Machine Translation

In this work, we present novel approaches to exploit sentential context ...
research
07/26/2021

Revisiting Negation in Neural Machine Translation

In this paper, we evaluate the translation of negation both automaticall...
research
04/20/2017

Adversarial Neural Machine Translation

In this paper, we study a new learning paradigm for Neural Machine Trans...
research
08/31/2019

Improving Back-Translation with Uncertainty-based Confidence Estimation

While back-translation is simple and effective in exploiting abundant mo...
research
12/10/2020

Rewriter-Evaluator Framework for Neural Machine Translation

Encoder-decoder architecture has been widely used in neural machine tran...
research
04/28/2020

Assessing the Bilingual Knowledge Learned by Neural Machine Translation Models

Machine translation (MT) systems translate text between different langua...
research
04/05/2020

Understanding Learning Dynamics for Neural Machine Translation

Despite the great success of NMT, there still remains a severe challenge...

Please sign up or login with your details

Forgot password? Click here to reset