Understanding and Enhancing the Use of Context for Machine Translation

02/20/2021
by   Marzieh Fadaee, et al.
0

To understand and infer meaning in language, neural models have to learn complicated nuances. Discovering distinctive linguistic phenomena from data is not an easy task. For instance, lexical ambiguity is a fundamental feature of language which is challenging to learn. Even more prominently, inferring the meaning of rare and unseen lexical units is difficult with neural networks. Meaning is often determined from context. With context, languages allow meaning to be conveyed even when the specific words used are not known by the reader. To model this learning process, a system has to learn from a few instances in context and be able to generalize well to unseen cases. The learning process is hindered when training data is scarce for a task. Even with sufficient data, learning patterns for the long tail of the lexical distribution is challenging. In this thesis, we focus on understanding certain potentials of contexts in neural models and design augmentation models to benefit from them. We focus on machine translation as an important instance of the more general language understanding problem. To translate from a source language to a target language, a neural model has to understand the meaning of constituents in the provided context and generate constituents with the same meanings in the target language. This task accentuates the value of capturing nuances of language and the necessity of generalization from few observations. The main problem we study in this thesis is what neural machine translation models learn from data and how we can devise more focused contexts to enhance this learning. Looking more in-depth into the role of context and the impact of data on learning models is essential to advance the NLP field. Moreover, it helps highlight the vulnerabilities of current neural networks and provides insights into designing more robust models.

READ FULL TEXT

page 33

page 37

research
10/30/2019

Fill in the Blanks: Imputing Missing Sentences for Larger-Context Neural Machine Translation

Most neural machine translation systems still translate sentences in iso...
research
12/18/2019

A Survey on Document-level Machine Translation: Methods and Evaluation

Machine translation (MT) is an important task in natural language proces...
research
08/12/2021

The paradox of the compositionality of natural language: a neural machine translation case study

Moving towards human-like linguistic performance is often argued to requ...
research
07/29/2017

Learning Language Representations for Typology Prediction

One central mystery of neural NLP is what neural models "know" about the...
research
08/20/2017

Neural Machine Translation with Extended Context

We investigate the use of extended context in attention-based neural mac...
research
06/18/2019

Distilling Translations with Visual Awareness

Previous work on multimodal machine translation has shown that visual in...
research
10/02/2022

The boundaries of meaning: a case study in neural machine translation

The success of deep learning in natural language processing raises intri...

Please sign up or login with your details

Forgot password? Click here to reset