A Universal Parent Model for Low-Resource Neural Machine Translation Transfer

09/14/2019
by   Mozhdeh Gheini, et al.
0

Transfer learning from a high-resource language pair `parent' has been proven to be an effective way to improve neural machine translation quality for low-resource language pairs `children.' However, previous approaches build a custom parent model or at least update an existing parent model's vocabulary for each child language pair they wish to train, in an effort to align parent and child vocabularies. This is not a practical solution. It is wasteful to devote the majority of training time for new language pairs to optimizing parameters on an unrelated data set. Further, this overhead reduces the utility of neural machine translation for deployment in humanitarian assistance scenarios, where extra time to deploy a new language pair can mean the difference between life and death. In this work, we present a `universal' pre-trained neural parent model with constant vocabulary that can be used as a starting point for training practically any new low-resource language to a fixed target language. We demonstrate that our approach, which leverages orthography unification and a broad-coverage approach to subword identification, generalizes well to several languages from a variety of families, and that translation systems built with our approach can be built more quickly than competing methods and with better quality as well.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/02/2018

Trivial Transfer Learning for Low-Resource Neural Machine Translation

Transfer learning has been proven as an effective technique for neural m...
research
08/08/2021

Machine Translation of Low-Resource Indo-European Languages

Transfer learning has been an important technique for low-resource neura...
research
08/31/2017

Transfer Learning across Low-Resource, Related Languages for Neural Machine Translation

We present a simple method to improve neural translation of a low-resour...
research
09/24/2019

Transfer Learning across Languages from Someone Else's NMT Model

Neural machine translation is demanding in terms of training time, hardw...
research
09/30/2020

On Romanization for Model Transfer Between Scripts in Neural Machine Translation

Transfer learning is a popular strategy to improve the quality of low-re...
research
01/06/2020

Exploring Benefits of Transfer Learning in Neural Machine Translation

Neural machine translation is known to require large numbers of parallel...
research
10/19/2018

Optimizing Segmentation Granularity for Neural Machine Translation

In neural machine translation (NMT), it is has become standard to transl...

Please sign up or login with your details

Forgot password? Click here to reset