Chunk-based Nearest Neighbor Machine Translation

05/24/2022
by   Pedro Henrique Martins, et al.
1

Semi-parametric models, which augment generation with retrieval, have led to impressive results in language modeling and machine translation, due to their ability to leverage information retrieved from a datastore of examples. One of the most prominent approaches, kNN-MT, has an outstanding performance on domain adaptation by retrieving tokens from a domain-specific datastore <cit.>. However, kNN-MT requires retrieval for every single generated token, leading to a very low decoding speed (around 8 times slower than a parametric model). In this paper, we introduce a chunk-based kNN-MT model which retrieves chunks of tokens from the datastore, instead of a single token. We propose several strategies for incorporating the retrieved chunks into the generation process, and for selecting the steps at which the model needs to search for neighbors in the datastore. Experiments on machine translation in two settings, static domain adaptation and “on-the-fly” adaptation, show that the chunk-based kNN-MT model leads to a significant speed-up (up to 4 times) with only a small drop in translation quality.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/23/2023

Simple and Scalable Nearest Neighbor Machine Translation

kNN-MT is a straightforward yet powerful approach for fast domain adapta...
research
04/26/2022

Efficient Machine Translation Domain Adaptation

Machine translation models struggle when translating out-of-domain text,...
research
11/29/2022

Soft Alignment Objectives for Robust Adaptation in Machine Translation

Domain adaptation allows generative language models to address specific ...
research
09/14/2021

Non-Parametric Unsupervised Domain Adaptation for Neural Machine Translation

Recently, kNN-MT has shown the promising capability of directly incorpor...
research
11/08/2022

What Knowledge Is Needed? Towards Explainable Memory for kNN-MT Domain Adaptation

kNN-MT presents a new paradigm for domain adaptation by building an exte...
research
02/27/2023

kNN-BOX: A Unified Framework for Nearest Neighbor Generation

Augmenting the base neural model with a token-level symbolic datastore i...
research
11/15/2022

Adaptation Approaches for Nearest Neighbor Language Models

Semi-parametric Nearest Neighbor Language Models (kNN-LMs) have produced...

Please sign up or login with your details

Forgot password? Click here to reset