Hybrid Data-Model Parallel Training for Sequence-to-Sequence Recurrent Neural Network Machine Translation

09/02/2019
by   Junya Ono, et al.
0

Reduction of training time is an important issue in many tasks like patent translation involving neural networks. Data parallelism and model parallelism are two common approaches for reducing training time using multiple graphics processing units (GPUs) on one machine. In this paper, we propose a hybrid data-model parallel approach for sequence-to-sequence (Seq2Seq) recurrent neural network (RNN) machine translation. We apply a model parallel approach to the RNN encoder-decoder part of the Seq2Seq model and a data parallel approach to the attention-softmax part of the model. We achieved a speed-up of 4.13 to 4.20 times when using 4 GPUs compared with the training speed when using 1 GPU without affecting machine translation accuracy as measured in terms of BLEU scores.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/29/2018

Parallel Attention Mechanisms in Neural Machine Translation

Recent papers in neural machine translation have proposed the strict use...
research
06/03/2014

Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation

In this paper, we propose a novel neural network model called RNN Encode...
research
03/22/2017

Classification-based RNN machine translation using GRUs

We report the results of our classification-based machine translation mo...
research
04/01/2018

Training Tips for the Transformer Model

This article describes our experiments in neural machine translation usi...
research
06/06/2017

Retrosynthetic reaction prediction using neural sequence-to-sequence models

We describe a fully data driven model that learns to perform a retrosynt...
research
09/15/2020

Recurrent autoencoder with sequence-aware encoding

Recurrent Neural Networks (RNN) received a vast amount of attention last...
research
12/27/2017

CNN Is All You Need

The Convolution Neural Network (CNN) has demonstrated the unique advanta...

Please sign up or login with your details

Forgot password? Click here to reset