Measuring and Increasing Context Usage in Context-Aware Machine Translation

05/07/2021
by   Patrick Fernandes, et al.
8

Recent work in neural machine translation has demonstrated both the necessity and feasibility of using inter-sentential context – context from sentences other than those currently being translated. However, while many current methods present model architectures that theoretically can use this extra context, it is often not clear how much they do actually utilize it at translation time. In this paper, we introduce a new metric, conditional cross-mutual information, to quantify the usage of context by these models. Using this metric, we measure how much document-level machine translation systems use particular varieties of context. We find that target context is referenced more than source context, and that conditioning on a longer context has a diminishing effect on results. We then introduce a new, simple training method, context-aware word dropout, to increase the usage of context by context-aware models. Experiments show that our method increases context usage and that this reflects on the translation quality according to metrics such as BLEU and COMET, as well as performance on anaphoric pronoun resolution and lexical cohesion contrastive datasets.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/24/2020

Context-aware Decoder for Neural Machine Translation using a Target-side Document-Level Language Model

Although many context-aware neural machine translation models have been ...
research
05/25/2018

Context-Aware Neural Machine Translation Learns Anaphora Resolution

Standard machine translation systems process sentences in isolation and ...
research
09/07/2021

Revisiting Context Choices for Context-aware Machine Translation

One of the most popular methods for context-aware machine translation (M...
research
12/28/2020

Towards Fully Automated Manga Translation

We tackle the problem of machine translation of manga, Japanese comics. ...
research
10/04/2018

A Large-Scale Test Set for the Evaluation of Context-Aware Pronoun Translation in Neural Machine Translation

The translation of pronouns presents a special challenge to machine tran...
research
03/06/2022

Conditional Bilingual Mutual Information Based Adaptive Training for Neural Machine Translation

Token-level adaptive training approaches can alleviate the token imbalan...
research
01/31/2018

Paraphrase-Supervised Models of Compositionality

Compositional vector space models of meaning promise new solutions to st...

Please sign up or login with your details

Forgot password? Click here to reset