Learn to Use Future Information in Simultaneous Translation

07/10/2020
by   Xueqing Wu, et al.
0

Simultaneous neural machine translation (briefly, NMT) has attracted much attention recently. In contrast to standard NMT, where the NMT system can utilize the full input sentence, simultaneous NMT is formulated as a prefix-to-prefix problem, where the system can only utilize the prefix of the input sentence and more uncertainty is introduced to decoding. Wait-k is a simple yet effective strategy for simultaneous NMT, where the decoder generates the output sequence k words behind the input words. We observed that training simultaneous NMT systems with future information (i.e., trained with a larger k) generally outperforms the standard ones (i.e., trained with the given k). Based on this observation, we propose a framework that automatically learns how much future information to use in training for simultaneous NMT. We first build a series of tasks where each one is associated with a different k, and then learn a model on these tasks guided by a controller. The controller is jointly trained with the translation model through bi-level optimization. We conduct experiments on four datasets to demonstrate the effectiveness of our method.

READ FULL TEXT
research
10/03/2016

Learning to Translate in Real-time with Neural Machine Translation

Translating in real-time, a.k.a. simultaneous translation, outputs trans...
research
05/03/2017

Chunk-Based Bi-Scale Decoder for Neural Machine Translation

In typical neural machine translation (NMT), the decoder generates a sen...
research
06/19/2018

Learning from Chunk-based Feedback in Neural Machine Translation

We empirically investigate learning from partial feedback in neural mach...
research
06/06/2022

Bi-SimCut: A Simple Strategy for Boosting Neural Machine Translation

We introduce Bi-SimCut: a simple but effective training strategy to boos...
research
04/07/2018

Guiding Neural Machine Translation with Retrieved Translation Pieces

One of the difficulties of neural machine translation (NMT) is the recal...
research
07/06/2016

Guided Alignment Training for Topic-Aware Neural Machine Translation

In this paper, we propose an effective way for biasing the attention mec...
research
05/05/2021

Full-Sentence Models Perform Better in Simultaneous Translation Using the Information Enhanced Decoding Strategy

Simultaneous translation, which starts translating each sentence after r...

Please sign up or login with your details

Forgot password? Click here to reset