Beyond Error Propagation in Neural Machine Translation: Characteristics of Language Also Matter

09/01/2018
by   Lijun Wu, et al.
0

Neural machine translation usually adopts autoregressive models and suffers from exposure bias as well as the consequent error propagation problem. Many previous works have discussed the relationship between error propagation and the accuracy drop (i.e., the left part of the translated sentence is often better than its right part in left-to-right decoding models) problem. In this paper, we conduct a series of analyses to deeply understand this problem and get several interesting findings. (1) The role of error propagation on accuracy drop is overstated in the literature, although it indeed contributes to the accuracy drop problem. (2) Characteristics of a language play a more important role in causing the accuracy drop: the left part of the translation result in a right-branching language (e.g., English) is more likely to be more accurate than its right part, while the right part is more accurate for a left-branching language (e.g., Japanese). Our discoveries are confirmed on different model structures including Transformer and RNN, and in other sequence generation tasks such as text summarization.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/23/2019

Sequence Generation: From Both Sides to the Middle

The encoder-decoder framework has achieved promising process for many se...
research
08/25/2019

Efficient Bidirectional Neural Machine Translation

The encoder-decoder based neural machine translation usually generates a...
research
11/05/2019

Improving Bidirectional Decoding with Dynamic Target Semantics in Neural Machine Translation

Generally, Neural Machine Translation models generate target words in a ...
research
02/08/2019

Insertion Transformer: Flexible Sequence Generation via Insertion Operations

We present the Insertion Transformer, an iterative, partially autoregres...
research
04/19/2019

Constant-Time Machine Translation with Conditional Masked Language Models

Most machine translation systems generate text autoregressively, by sequ...
research
06/17/2023

Persian Semantic Role Labeling Using Transfer Learning and BERT-Based Models

Semantic role labeling (SRL) is the process of detecting the predicate-a...
research
12/16/2021

Learning and Analyzing Generation Order for Undirected Sequence Models

Undirected neural sequence models have achieved performance competitive ...

Please sign up or login with your details

Forgot password? Click here to reset