Simplifying Neural Machine Translation with Addition-Subtraction Twin-Gated Recurrent Networks

10/30/2018
by   Biao Zhang, et al.
0

In this paper, we propose an additionsubtraction twin-gated recurrent network (ATR) to simplify neural machine translation. The recurrent units of ATR are heavily simplified to have the smallest number of weight matrices among units of all existing gated RNNs. With the simple addition and subtraction operation, we introduce a twin-gated mechanism to build input and forget gates which are highly correlated. Despite this simplification, the essential non-linearities and capability of modeling long-distance dependencies are preserved. Additionally, the proposed ATR is more transparent than LSTM/GRU due to the simplification. Forward self-attention can be easily established in ATR, which makes the proposed network interpretable. Experiments on WMT14 translation tasks demonstrate that ATR-based neural machine translation can yield competitive performance on English- German and English-French language pairs in terms of both translation quality and speed. Further experiments on NIST Chinese-English translation, natural language inference and Chinese word segmentation verify the generality and applicability of ATR on different natural language processing tasks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/12/2020

Towards Machine Translation for the Kurdish Language

Machine translation is the task of translating texts from one language t...
research
07/29/2016

Recurrent Neural Machine Translation

The vanilla attention-based neural machine translation has achieved prom...
research
05/10/2018

Deep Neural Machine Translation with Weakly-Recurrent Units

Recurrent neural networks (RNNs) have represented for years the state of...
research
09/03/2019

Multi-agent Learning for Neural Machine Translation

Conventional Neural Machine Translation (NMT) models benefit from the tr...
research
10/27/2022

The Effect of Normalization for Bi-directional Amharic-English Neural Machine Translation

Machine translation (MT) is one of the main tasks in natural language pr...
research
10/21/2020

Multi-Unit Transformers for Neural Machine Translation

Transformer models achieve remarkable success in Neural Machine Translat...
research
08/27/2018

Why Self-Attention? A Targeted Evaluation of Neural Machine Translation Architectures

Recently, non-recurrent architectures (convolutional, self-attentional) ...

Please sign up or login with your details

Forgot password? Click here to reset