Learning to Remember Translation History with a Continuous Cache

11/26/2017
by   Zhaopeng Tu, et al.
0

Existing neural machine translation (NMT) models generally translate sentences in isolation, missing the opportunity to take advantage of document-level information. In this work, we propose to augment NMT models with a very light-weight cache-like memory network, which stores recent hidden representations as translation history. The probability distribution over generated words is updated online depending on the translation history retrieved from the memory, endowing NMT models with the capability to dynamically adapt over time. Experiments on multiple domains with different topics and styles show the effectiveness of the proposed approach with negligible impact on the computational cost.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/31/2019

Document-level Neural Machine Translation with Inter-Sentence Attention

Standard neural machine translation (NMT) is on the assumption of docume...
research
11/30/2017

Cache-based Document-level Neural Machine Translation

Sentences in a well-formed text are connected to each other via various ...
research
04/16/2021

Context-Adaptive Document-Level Neural Machine Translation

Most existing document-level neural machine translation (NMT) models lev...
research
06/29/2016

Compression of Neural Machine Translation Models via Pruning

Neural Machine Translation (NMT), like many other deep learning domains,...
research
08/23/2019

Reference Network for Neural Machine Translation

Neural Machine Translation (NMT) has achieved notable success in recent ...
research
02/06/2018

Decoding-History-Based Adaptive Control of Attention for Neural Machine Translation

Attention-based sequence-to-sequence model has proved successful in Neur...
research
04/02/2021

Attention Forcing for Machine Translation

Auto-regressive sequence-to-sequence models with attention mechanisms ha...

Please sign up or login with your details

Forgot password? Click here to reset