Freezing Subnetworks to Analyze Domain Adaptation in Neural Machine Translation

09/14/2018
by   Brian Thompson, et al.
0

To better understand the effectiveness of continued training, we analyze the major components of a neural machine translation system (the encoder, decoder, and each embedding space) and consider each component's contribution to, and capacity for, domain adaptation. We find that freezing any single component during continued training has minimal impact on performance, and that performance is surprisingly good when a single component is adapted while holding the rest of the model fixed. We also find that continued training does not move the model very far from the out-of-domain model, compared to a sensitivity analysis metric, suggesting that the out-of-domain model can provide a good generic initialization for the new domain.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/08/2019

Improving Domain Adaptation Translation with Domain Invariant and Specific Information

In domain adaptation for neural machine translation, translation perform...
research
05/14/2019

Curriculum Learning for Domain Adaptation in Neural Machine Translation

We introduce a curriculum learning approach to adapt generic neural mach...
research
03/05/2020

Distill, Adapt, Distill: Training Small, In-Domain Models for Neural Machine Translation

We explore best practices for training small, memory efficient machine t...
research
02/19/2021

Multi-Domain Adaptation in Neural Machine Translation Through Multidimensional Tagging

Many modern Neural Machine Translation (NMT) systems are trained on nonh...
research
03/25/2021

Pruning-then-Expanding Model for Domain Adaptation of Neural Machine Translation

Domain Adaptation is widely used in practical applications of neural mac...
research
04/15/2020

Building a Multi-domain Neural Machine Translation Model using Knowledge Distillation

Lack of specialized data makes building a multi-domain neural machine tr...
research
09/21/2021

The Trade-offs of Domain Adaptation for Neural Language Models

In this paper, we connect language model adaptation with concepts of mac...

Please sign up or login with your details

Forgot password? Click here to reset