Simple, Scalable Adaptation for Neural Machine Translation

09/18/2019
by   Ankur Bapna, et al.
0

Fine-tuning pre-trained Neural Machine Translation (NMT) models is the dominant approach for adapting to new languages and domains. However, fine-tuning requires adapting and maintaining a separate model for each target task. We propose a simple yet efficient approach for adaptation in NMT. Our proposed approach consists of injecting tiny task specific adapter layers into a pre-trained model. These lightweight adapters, with just a small fraction of the original model size, adapt the model to multiple individual tasks simultaneously. We evaluate our approach on two tasks: (i) Domain Adaptation and (ii) Massively Multilingual NMT. Experiments on domain adaptation demonstrate that our proposed approach is on par with full fine-tuning on various domains, dataset sizes and model capacities. On a massively multilingual dataset of 103 languages, our adaptation approach bridges the gap between individual bilingual models and one massively multilingual model for most language pairs, paving the way towards universal machine translation.

READ FULL TEXT

page 7

page 8

research
06/19/2019

Multilingual Multi-Domain Adaptation Approaches for Neural Machine Translation

In this paper, we propose two novel methods for domain adaptation for th...
research
10/26/2022

Robust Domain Adaptation for Pre-trained Multilingual Neural Machine Translation Models

Recent literature has demonstrated the potential of multilingual Neural ...
research
06/02/2021

Lightweight Adapter Tuning for Multilingual Speech Translation

Adapter modules were recently introduced as an efficient alternative to ...
research
09/07/2021

IndicBART: A Pre-trained Model for Natural Language Generation of Indic Languages

In this paper we present IndicBART, a multilingual, sequence-to-sequence...
research
10/18/2021

Multilingual Domain Adaptation for NMT: Decoupling Language and Domain Information with Adapters

Adapter layers are lightweight, learnable units inserted between transfo...
research
10/30/2019

Adapting Multilingual Neural Machine Translation to Unseen Languages

Multilingual Neural Machine Translation (MNMT) for low-resource language...
research
10/21/2022

m^4Adapter: Multilingual Multi-Domain Adaptation for Machine Translation with a Meta-Adapter

Multilingual neural machine translation models (MNMT) yield state-of-the...

Please sign up or login with your details

Forgot password? Click here to reset