Context Models for OOV Word Translation in Low-Resource Languages

01/26/2018
by   Angli Liu, et al.
0

Out-of-vocabulary word translation is a major problem for the translation of low-resource languages that suffer from a lack of parallel training data. This paper evaluates the contributions of target-language context models towards the translation of OOV words, specifically in those cases where OOV translations are derived from external knowledge sources, such as dictionaries. We develop both neural and non-neural context models and evaluate them within both phrase-based and self-attention based neural machine translation systems. Our results show that neural language models that integrate additional context beyond the current sentence are the most effective in disambiguating possible OOV word translations. We present an efficient second-pass lattice-rescoring method for wide-context neural language models and demonstrate performance improvements over state-of-the-art self-attention based neural MT systems in five out of six low-resource language pairs.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/28/2021

A Preordered RNN Layer Boosts Neural Machine Translation in Low Resource Settings

Neural Machine Translation (NMT) models are strong enough to convey sema...
research
11/10/2020

Neural Machine Translation for Extremely Low-Resource African Languages: A Case Study on Bambara

Low-resource languages present unique challenges to (neural) machine tra...
research
02/15/2023

Dictionary-based Phrase-level Prompting of Large Language Models for Machine Translation

Large language models (LLMs) demonstrate remarkable machine translation ...
research
05/11/2023

Chain-of-Dictionary Prompting Elicits Translation in Large Language Models

Large language models (LLMs) have shown surprisingly good performance in...
research
04/19/2023

Low-resource Bilingual Dialect Lexicon Induction with Large Language Models

Bilingual word lexicons are crucial tools for multilingual natural langu...
research
01/06/2016

Incorporating Structural Alignment Biases into an Attentional Neural Translation Model

Neural encoder-decoder models of machine translation have achieved impre...
research
08/26/2023

Translate Meanings, Not Just Words: IdiomKB's Role in Optimizing Idiomatic Translation with Language Models

To translate well, machine translation (MT) systems and general-purposed...

Please sign up or login with your details

Forgot password? Click here to reset