Curriculum Learning and Minibatch Bucketing in Neural Machine Translation

07/29/2017
by   Tom Kocmi, et al.
0

We examine the effects of particular orderings of sentence pairs on the on-line training of neural machine translation (NMT). We focus on two types of such orderings: (1) ensuring that each minibatch contains sentences similar in some aspect and (2) gradual inclusion of some sentence types as the training progresses (so called "curriculum learning"). In our English-to-Czech experiments, the internal homogeneity of minibatches has no effect on the training but some of our "curricula" achieve a small improvement over the baseline.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/14/2019

Curriculum Learning for Domain Adaptation in Neural Machine Translation

We introduce a curriculum learning approach to adapt generic neural mach...
research
06/03/2020

Norm-Based Curriculum Learning for Neural Machine Translation

A neural machine translation (NMT) system is expensive to train, especia...
research
02/28/2019

Reinforcement Learning based Curriculum Optimization for Neural Machine Translation

We consider the problem of making efficient use of heterogeneous trainin...
research
10/09/2020

Self-Paced Learning for Neural Machine Translation

Recent studies have proven that the training of neural machine translati...
research
05/31/2018

On the Impact of Various Types of Noise on Neural Machine Translation

We examine how various types of noise in the parallel training data impa...
research
05/10/2021

Self-Guided Curriculum Learning for Neural Machine Translation

In the field of machine learning, the well-trained model is assumed to b...
research
04/07/2020

Self-Induced Curriculum Learning in Neural Machine Translation

Self-supervised neural machine translation (SS-NMT) learns how to extrac...

Please sign up or login with your details

Forgot password? Click here to reset