End-to-End Training of Both Translation Models in the Back-Translation Framework

02/17/2022
by   DongNyeong Heo, et al.
0

Semi-supervised learning algorithms in neural machine translation (NMT) have significantly improved translation quality compared to the supervised learning algorithms by using additional monolingual corpora. Among them, back-translation is a theoretically well-structured and cutting-edge method. Given two pre-trained NMT models between source and target languages, one translates a monolingual sentence as a latent sentence, and the other reconstructs the monolingual input sentence given the latent sentence. Therefore, previous works tried to apply the variational auto-encoder's (VAE) training framework to the back-translation framework. However, the discrete property of the latent sentence made it impossible to use backpropagation in the framework. This paper proposes a categorical reparameterization trick that generates a differentiable sentence, with which we practically implement the VAE's training framework for the back-translation and train it by end-to-end backpropagation. In addition, we propose several regularization techniques that are especially advantageous to this framework. In our experiments, we demonstrate that our method makes backpropagation available through the latent sentences and improves the BLEU scores on the datasets of the WMT18 translation task.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/15/2016

Semi-Supervised Learning for Neural Machine Translation

While end-to-end neural machine translation (NMT) has made remarkable pr...
research
06/13/2018

Generative Neural Machine Translation

We introduce Generative Neural Machine Translation (GNMT), a latent vari...
research
04/02/2023

Semi-supervised Neural Machine Translation with Consistency Regularization for Low-Resource Languages

The advent of deep learning has led to a significant gain in machine tra...
research
09/17/2020

Code-switching pre-training for neural machine translation

This paper proposes a new pre-training method, called Code-Switching Pre...
research
09/01/2018

Simple Fusion: Return of the Language Model

Neural Machine Translation (NMT) typically leverages monolingual data in...
research
09/19/2018

Monolingual sentence matching for text simplification

This work improves monolingual sentence alignment for text simplificatio...
research
04/07/2020

Machine Translation with Unsupervised Length-Constraints

We have seen significant improvements in machine translation due to the ...

Please sign up or login with your details

Forgot password? Click here to reset