Duplex Sequence-to-Sequence Learning for Reversible Machine Translation

05/07/2021
by   Zaixiang Zheng, et al.
0

Sequence-to-sequence (seq2seq) problems such as machine translation are bidirectional, which naturally derive a pair of directional tasks and two directional learning signals. However, typical seq2seq neural networks are simplex that only model one unidirectional task, which cannot fully exploit the potential of bidirectional learning signals from parallel data. To address this issue, we propose a duplex seq2seq neural network, REDER (Reversible Duplex Transformer), and apply it to machine translation. The architecture of REDER has two ends, each of which specializes in a language so as to read and yield sequences in that language. As a result, REDER can simultaneously learn from the bidirectional signals, and enables reversible machine translation by simply flipping the input and output ends, Experiments on widely-used machine translation benchmarks verify that REDER achieves the first success of reversible machine translation, which helps obtain considerable gains over several strong baselines.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/22/2023

Duplex Diffusion Models Improve Speech-to-Speech Translation

Speech-to-speech translation is a typical sequence-to-sequence learning ...
research
09/22/2017

Neural Machine Translation

Draft of textbook chapter on neural machine translation. a comprehensive...
research
07/25/2018

"Bilingual Expert" Can Find Translation Errors

Recent advances in statistical machine translation via the adoption of n...
research
06/22/2015

A Deep Memory-based Architecture for Sequence-to-Sequence Learning

We propose DEEPMEMORY, a novel deep architecture for sequence-to-sequenc...
research
05/29/2022

The impact of memory on learning sequence-to-sequence tasks

The recent success of neural networks in machine translation and other f...
research
09/15/2021

Sequence Length is a Domain: Length-based Overfitting in Transformer Models

Transformer-based sequence-to-sequence architectures, while achieving st...
research
06/07/2021

Lexicon Learning for Few-Shot Neural Sequence Modeling

Sequence-to-sequence transduction is the core problem in language proces...

Please sign up or login with your details

Forgot password? Click here to reset