Why Neural Machine Translation Prefers Empty Outputs

12/24/2020
by   Xing Shi, et al.
0

We investigate why neural machine translation (NMT) systems assign high probability to empty translations. We find two explanations. First, label smoothing makes correct-length translations less confident, making it easier for the empty translation to finally outscore them. Second, NMT systems use the same, high-frequency EoS word to end all target sentences, regardless of length. This creates an implicit smoothing that increases zero-length translations. Using different EoS types in target sentences of different lengths exposes and eliminates this implicit smoothing.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/08/2018

Debugging Neural Machine Translations

In this paper, we describe a tool for debugging the output and attention...
research
10/23/2019

Controlling the Output Length of Neural Machine Translation

The recent advances introduced by neural machine translation (NMT) are r...
research
03/06/2022

Focus on the Target's Vocabulary: Masked Label Smoothing for Machine Translation

Label smoothing and vocabulary sharing are two widely used techniques in...
research
06/25/2022

Probing Causes of Hallucinations in Neural Machine Translations

Hallucination, one kind of pathological translations that bothers Neural...
research
09/21/2020

Target Conditioning for One-to-Many Generation

Neural Machine Translation (NMT) models often lack diversity in their ge...
research
07/08/2021

Using CollGram to Compare Formulaic Language in Human and Neural Machine Translation

A comparison of formulaic sequences in human and neural machine translat...
research
03/23/2021

Repairing Pronouns in Translation with BERT-Based Post-Editing

Pronouns are important determinants of a text's meaning but difficult to...

Please sign up or login with your details

Forgot password? Click here to reset