Cache-based Document-level Neural Machine Translation

11/30/2017
by   Shaohui Kuang, et al.
0

Sentences in a well-formed text are connected to each other via various links to form the cohesive structure of the text. Current neural machine translation (NMT) systems translate a text in a conventional sentence-by-sentence fashion, ignoring such cross-sentence links and dependencies. This may lead to generate an incohesive and incoherent target text for a cohesive and coherent source text. In order to handle this issue, we propose a cache-based approach to document-level neural machine translation by capturing contextual information either from recently translated sentences or the entire document. Particularly, we explore two types of caches: a dynamic cache, which stores words from the best translation hypotheses of preceding sentences, and a topic cache, which maintains a set of target-side topical words that are semantically related to the document to be translated. On this basis, we build a new layer to score target words in these two caches with a cache-based neural model. Here the estimated probabilities from the cache-based neural model are combined with NMT probabilities into the final word prediction probabilities via a gating mechanism. Finally, the proposed cache-based neural model is trained jointly with a state-of-the-art neural machine translation system in an end-to-end manner. On several NIST Chinese-English translation tasks, our experiments demonstrate that the proposed cache-based model achieves substantial improvements over several state-of-the-art SMT and NMT baselines.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/12/2018

Fusing Recency into Neural Machine Translation with an Inter-Sentence Gate Model

Neural machine translation (NMT) systems are usually trained on a large ...
research
03/30/2020

Learning Contextualized Sentence Representations for Document-Level Neural Machine Translation

Document-level machine translation incorporates inter-sentential depende...
research
11/26/2017

Learning to Remember Translation History with a Continuous Cache

Existing neural machine translation (NMT) models generally translate sen...
research
03/11/2020

Capturing document context inside sentence-level neural machine translation models with self-training

Neural machine translation (NMT) has arguably achieved human level parit...
research
12/13/2019

Document Sub-structure in Neural Machine Translation

Current approaches to machine translation (MT) either translate sentence...
research
04/14/2017

Translation of Patent Sentences with a Large Vocabulary of Technical Terms Using Neural Machine Translation

Neural machine translation (NMT), a new approach to machine translation,...
research
07/19/2019

Structure-Invariant Testing for Machine Translation

In recent years, machine translation software has increasingly been inte...

Please sign up or login with your details

Forgot password? Click here to reset