Neural Machine Translation Models Can Learn to be Few-shot Learners

09/15/2023
by   Raphael Reinauer, et al.
0

The emergent ability of Large Language Models to use a small number of examples to learn to perform in novel domains and tasks, also called in-context learning (ICL). In this work, we show that a much smaller model can be trained to perform ICL by fine-tuning towards a specialized training objective, exemplified on the task of domain adaptation for neural machine translation. With this capacity for ICL, the model can take advantage of relevant few-shot examples to adapt its output towards the domain. We compare the quality of this domain adaptation to traditional supervised techniques and ICL with a 40B-parameter Large Language Model. Our approach allows efficient batch inference on a mix of domains and outperforms state-of-the-art baselines in terms of both translation quality and immediate adaptation rate, i.e. the ability to reproduce a specific term after being shown a single example.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/19/2019

Multilingual Multi-Domain Adaptation Approaches for Neural Machine Translation

In this paper, we propose two novel methods for domain adaptation for th...
research
11/05/2018

Compact Personalized Models for Neural Machine Translation

We propose and compare methods for gradient-based domain adaptation of s...
research
06/07/2019

Word-based Domain Adaptation for Neural Machine Translation

In this paper, we empirically investigate applying word-level weights to...
research
04/26/2022

Efficient Machine Translation Domain Adaptation

Machine translation models struggle when translating out-of-domain text,...
research
12/20/2022

Localising In-Domain Adaptation of Transformer-Based Biomedical Language Models

In the era of digital healthcare, the huge volumes of textual informatio...
research
05/05/2023

In-context Learning as Maintaining Coherency: A Study of On-the-fly Machine Translation Using Large Language Models

The phenomena of in-context learning has typically been thought of as "l...
research
11/29/2022

Soft Alignment Objectives for Robust Adaptation in Machine Translation

Domain adaptation allows generative language models to address specific ...

Please sign up or login with your details

Forgot password? Click here to reset