Token Drop mechanism for Neural Machine Translation

10/21/2020
by   Huaao Zhang, et al.
0

Neural machine translation with millions of parameters is vulnerable to unfamiliar inputs. We propose Token Drop to improve generalization and avoid overfitting for the NMT model. Similar to word dropout, whereas we replace dropped token with a special token instead of setting zero to words. We further introduce two self-supervised objectives: Replaced Token Detection and Dropped Token Prediction. Our method aims to force model generating target translation with less information, in this way the model can learn textual representation better. Experiments on Chinese-English and English-Romanian benchmark demonstrate the effectiveness of our approach and our model achieves significant improvements over a strong Transformer baseline.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/24/2021

Prevent the Language Model from being Overconfident in Neural Machine Translation

The Neural Machine Translation (NMT) model is essentially a joint langua...
research
07/21/2020

Neural Machine Translation with Error Correction

Neural machine translation (NMT) generates the next target token given a...
research
11/27/2019

Simultaneous Neural Machine Translation using Connectionist Temporal Classification

Simultaneous machine translation is a variant of machine translation tha...
research
05/23/2022

Towards Opening the Black Box of Neural Machine Translation: Source and Target Interpretations of the Transformer

In Neural Machine Translation (NMT), each token prediction is conditione...
research
02/28/2020

Modeling Future Cost for Neural Machine Translation

Existing neural machine translation (NMT) systems utilize sequence-to-se...
research
10/13/2022

Categorizing Semantic Representations for Neural Machine Translation

Modern neural machine translation (NMT) models have achieved competitive...
research
03/06/2022

Conditional Bilingual Mutual Information Based Adaptive Training for Neural Machine Translation

Token-level adaptive training approaches can alleviate the token imbalan...

Please sign up or login with your details

Forgot password? Click here to reset