Minimizing the Bag-of-Ngrams Difference for Non-Autoregressive Neural Machine Translation

11/21/2019
by   Chenze Shao, et al.
0

Non-Autoregressive Neural Machine Translation (NAT) achieves significant decoding speedup through generating target words independently and simultaneously. However, in the context of non-autoregressive translation, the word-level cross-entropy loss cannot model the target-side sequential dependency properly, leading to its weak correlation with the translation quality. As a result, NAT tends to generate influent translations with over-translation and under-translation errors. In this paper, we propose to train NAT to minimize the Bag-of-Ngrams (BoN) difference between the model output and the reference sentence. The bag-of-ngrams training objective is differentiable and can be efficiently calculated, which encourages NAT to capture the target-side sequential dependency and correlates well with the translation quality. We validate our approach on three translation tasks and show that our approach largely outperforms the NAT baseline by about 5.0 BLEU scores on WMT14 EnDe and about 2.5 BLEU scores on WMT16 EnRo.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/15/2021

Sequence-Level Training for Non-Autoregressive Neural Machine Translation

In recent years, Neural Machine Translation (NMT) has achieved notable r...
research
06/22/2019

Retrieving Sequential Information for Non-Autoregressive Neural Machine Translation

Non-Autoregressive Transformer (NAT) aims to accelerate the Transformer ...
research
05/28/2022

One Reference Is Not Enough: Diverse Distillation with Reference Selection for Non-Autoregressive Translation

Non-autoregressive neural machine translation (NAT) suffers from the mul...
research
05/13/2018

Bag-of-Words as Target for Neural Machine Translation

A sentence can be translated into more than one correct sentences. Howev...
research
10/18/2021

Monotonic Simultaneous Translation with Chunk-wise Reordering and Refinement

Recent work in simultaneous machine translation is often trained with co...
research
12/23/2018

Non-Autoregressive Neural Machine Translation with Enhanced Decoder Input

Non-autoregressive translation (NAT) models, which remove the dependence...
research
04/24/2021

Modeling Coverage for Non-Autoregressive Neural Machine Translation

Non-Autoregressive Neural Machine Translation (NAT) has achieved signifi...

Please sign up or login with your details

Forgot password? Click here to reset