Improving Language Model Integration for Neural Machine Translation

06/08/2023
by   Christian Herold, et al.
0

The integration of language models for neural machine translation has been extensively studied in the past. It has been shown that an external language model, trained on additional target-side monolingual data, can help improve translation quality. However, there has always been the assumption that the translation model also learns an implicit target-side language model during training, which interferes with the external language model at decoding time. Recently, some works on automatic speech recognition have demonstrated that, if the implicit language model is neutralized in decoding, further improvements can be gained when integrating an external language model. In this work, we transfer this concept to the task of machine translation and compare with the most prominent way of including additional monolingual data - namely back-translation. We find that accounting for the implicit language model significantly boosts the performance of language model fusion, although this approach is still outperformed by back-translation.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/11/2019

Dynamic Fusion: Attentional Language Model for Neural Machine Translation

Neural Machine Translation (NMT) can be used to generate fluent output. ...
research
04/30/2020

Language Model Prior for Low-Resource Neural Machine Translation

The scarcity of large parallel corpora is an important obstacle for neur...
research
09/01/2021

Masked Adversarial Generation for Neural Machine Translation

Attacking Neural Machine Translation models is an inherently combinatori...
research
03/12/2020

Hybrid Autoregressive Transducer (hat)

This paper proposes and evaluates the hybrid autoregressive transducer (...
research
08/18/2020

Complementary Language Model and Parallel Bi-LRNN for False Trigger Mitigation

False triggers in voice assistants are unintended invocations of the ass...
research
05/23/2023

Cascaded Beam Search: Plug-and-Play Terminology-Forcing For Neural Machine Translation

This paper presents a plug-and-play approach for translation with termin...
research
09/19/2023

MBR and QE Finetuning: Training-time Distillation of the Best and Most Expensive Decoding Methods

Recent research in decoding methods for Natural Language Generation (NLG...

Please sign up or login with your details

Forgot password? Click here to reset