Detecting and Understanding Generalization Barriers for Neural Machine Translation

04/05/2020
by   Guanlin Li, et al.
0

Generalization to unseen instances is our eternal pursuit for all data-driven models. However, for realistic task like machine translation, the traditional approach measuring generalization in an average sense provides poor understanding for the fine-grained generalization ability. As a remedy, this paper attempts to identify and understand generalization barrier words within an unseen input sentence that cause the degradation of fine-grained generalization. We propose a principled definition of generalization barrier words and a modified version which is tractable in computation. Based on the modified one, we propose three simple methods for barrier detection by the search-aware risk estimation through counterfactual generation. We then conduct extensive analyses on those detected generalization barrier words on both ZhEn NIST benchmarks from various perspectives. Potential usage of the detected barrier words is also discussed.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/30/2018

Fine-Grained Attention Mechanism for Neural Machine Translation

Neural machine translation (NMT) has been a new paradigm in machine tran...
research
05/31/2021

On Compositional Generalization of Neural Machine Translation

Modern neural machine translation (NMT) models have achieved competitive...
research
11/02/2020

The 2020s Political Economy of Machine Translation

This paper explores the hypothesis that the diversity of human languages...
research
05/31/2021

Beyond Noise: Mitigating the Impact of Fine-grained Semantic Divergences on Neural Machine Translation

While it has been shown that Neural Machine Translation (NMT) is highly ...
research
10/13/2022

Categorizing Semantic Representations for Neural Machine Translation

Modern neural machine translation (NMT) models have achieved competitive...
research
08/16/2023

It Ain't That Bad: Understanding the Mysterious Performance Drop in OOD Generalization for Generative Transformer Models

Generative Transformer-based models have achieved remarkable proficiency...

Please sign up or login with your details

Forgot password? Click here to reset