Neural Machine Translation with Gumbel-Greedy Decoding

06/22/2017
by   Jiatao Gu, et al.
0

Previous neural machine translation models used some heuristic search algorithms (e.g., beam search) in order to avoid solving the maximum a posteriori problem over translation sentences at test time. In this paper, we propose the Gumbel-Greedy Decoding which trains a generative network to predict translation under a trained model. We solve such a problem using the Gumbel-Softmax reparameterization, which makes our generative network differentiable and trainable through standard stochastic gradient methods. We empirically demonstrate that our proposed model is effective for generating sequences of discrete words.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/08/2017

Trainable Greedy Decoding for Neural Machine Translation

Recent research in neural machine translation has largely focused on two...
research
06/07/2016

Can neural machine translation do simultaneous translation?

We investigate the potential of attention-based neural machine translati...
research
04/21/2018

A Stable and Effective Learning Strategy for Trainable Greedy Decoding

As a widely used approximate search strategy for neural network decoders...
research
07/27/2018

Auto-Encoding Variational Neural Machine Translation

We present a deep generative model of bilingual sentence pairs. The mode...
research
08/28/2018

Breaking the Beam Search Curse: A Study of (Re-)Scoring Methods and Stopping Criteria for Neural Machine Translation

Beam search is widely used in neural machine translation, and usually im...
research
09/09/2018

Speeding Up Neural Machine Translation Decoding by Cube Pruning

Although neural machine translation has achieved promising results, it s...
research
06/12/2018

Explaining and Generalizing Back-Translation through Wake-Sleep

Back-translation has become a commonly employed heuristic for semi-super...

Please sign up or login with your details

Forgot password? Click here to reset