Vocabulary Adaptation for Distant Domain Adaptation in Neural Machine Translation

04/30/2020
by   Shoetsu Sato, et al.
0

Neural machine translation (NMT) models do not work well in domains different from the training data. The standard approach to this problem is to build a small parallel data in the target domain and perform domain adaptation from a source domain where massive parallel data is available. However, domain adaptation between distant domains (e.g., subtitles and research papers) does not perform effectively because of mismatches in vocabulary; it will encounter many domain-specific unknown words (e.g., `angstrom') and words whose meanings shift across domains (e.g., `conductor'). In this study, aiming to solve these vocabulary mismatches in distant domain adaptation, we propose vocabulary adaptation, a simple method for effective fine-tuning that adapts embedding layers in a given pre-trained NMT model to the target domain. Prior to fine-tuning, our method replaces word embeddings in embedding layers of the NMT model, by projecting general word embeddings induced from monolingual data in the target domain onto the source-domain embedding space. Experimental results on distant domain adaptation for English-to-Japanese translation and German-to-English translation indicate that our vocabulary adaptation improves the performance of fine-tuning by 3.6 BLEU points.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/12/2017

An Empirical Comparison of Simple Domain Adaptation Methods for Neural Machine Translation

In this paper, we propose a novel domain adaptation method named "mixed ...
research
06/02/2019

Domain Adaptation of Neural Machine Translation by Lexicon Induction

It has been previously noted that neural machine translation (NMT) is ve...
research
10/28/2022

Domain Adaptation of Machine Translation with Crowdworkers

Although a machine translation model trained with a large in-domain para...
research
05/21/2019

Domain adaptation for part-of-speech tagging of noisy user-generated text

The performance of a Part-of-speech (POS) tagger is highly dependent on ...
research
03/01/2023

A Systematic Analysis of Vocabulary and BPE Settings for Optimal Fine-tuning of NMT: A Case Study of In-domain Translation

The effectiveness of Neural Machine Translation (NMT) models largely dep...
research
12/19/2020

Finding Sparse Structure for Domain Specific Neural Machine Translation

Fine-tuning is a major approach for domain adaptation in Neural Machine ...
research
06/16/2020

PERL: Pivot-based Domain Adaptation for Pre-trained Deep Contextualized Embedding Models

Pivot-based neural representation models have lead to significant progre...

Please sign up or login with your details

Forgot password? Click here to reset