In-context Learning as Maintaining Coherency: A Study of On-the-fly Machine Translation Using Large Language Models

05/05/2023
by   Suzanna Sia, et al.
0

The phenomena of in-context learning has typically been thought of as "learning from examples". In this work which focuses on Machine Translation, we present a perspective of in-context learning as the desired generation task maintaining coherency with its context, i.e., the prompt examples. We first investigate randomly sampled prompts across 4 domains, and find that translation performance improves when shown in-domain prompts. Next, we investigate coherency for the in-domain setting, which uses prompt examples from a moving window. We study this with respect to other factors that have previously been identified in the literature such as length, surface similarity and sentence embedding similarity. Our results across 3 models (GPTNeo2.7B, Bloom3B, XGLM2.9B), and three translation directions (→{}) suggest that the long-term coherency of the prompts and the test sentence is a good indicator of downstream translation performance. In doing so, we demonstrate the efficacy of In-context Machine Translation for on-the-fly adaptation.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/05/2022

In-context Examples Selection for Machine Translation

Large-scale generative models show an impressive ability to perform a wi...
research
09/15/2023

Neural Machine Translation Models Can Learn to be Few-shot Learners

The emergent ability of Large Language Models to use a small number of e...
research
09/07/2021

Revisiting Context Choices for Context-aware Machine Translation

One of the most popular methods for context-aware machine translation (M...
research
05/23/2023

In-context Example Selection for Machine Translation Using Multiple Features

Large language models have demonstrated the capability to perform well o...
research
05/27/2023

Augmenting Large Language Model Translators via Translation Memories

Using translation memories (TMs) as prompts is a promising approach to i...
research
01/17/2023

Prompting Large Language Model for Machine Translation: A Case Study

Research on prompting has shown excellent performance with little or eve...
research
12/16/2022

Detecting and Mitigating Hallucinations in Machine Translation: Model Internal Workings Alone Do Well, Sentence Similarity Even Better

While the problem of hallucinations in neural machine translation has lo...

Please sign up or login with your details

Forgot password? Click here to reset