TSO: Curriculum Generation using continuous optimization

06/16/2021
by   Dipankar Sarkar, et al.
0

The training of deep learning models poses vast challenges of including parameter tuning and ordering of training data. Significant research has been done in Curriculum learning for optimizing the sequence of training data. Recent works have focused on using complex reinforcement learning techniques to find the optimal data ordering strategy to maximize learning for a given network. In this paper, we present a simple and efficient technique based on continuous optimization. We call this new approach Training Sequence Optimization (TSO). There are three critical components in our proposed approach: (a) An encoder network maps/embeds training sequence into continuous space. (b) A predictor network uses the continuous representation of a strategy as input and predicts the accuracy for fixed network architecture. (c) A decoder further maps a continuous representation of a strategy to the ordered training dataset. The performance predictor and encoder enable us to perform gradient-based optimization in the continuous space to find the embedding of optimal training data ordering with potentially better accuracy. Experiments show that we can gain 2AP with our generated optimal curriculum strategy over the random strategy using the CIFAR-100 dataset and have better boosts than the state of the art CL algorithms. We do an ablation study varying the architecture, dataset and sample sizes showcasing our approach's robustness.

READ FULL TEXT
POST COMMENT

Comments

There are no comments yet.

Authors

page 1

page 2

page 3

page 4

08/22/2018

Neural Architecture Optimization

Automatic neural architecture design has shown its potential in discover...
09/23/2019

Data Ordering Patterns for Neural Machine Translation: An Empirical Study

Recent works show that ordering of the training data affects the model p...
02/28/2019

Reinforcement Learning based Curriculum Optimization for Neural Machine Translation

We consider the problem of making efficient use of heterogeneous trainin...
01/16/2019

Evolutionarily-Curated Curriculum Learning for Deep Reinforcement Learning Agents

In this paper we propose a new training loop for deep reinforcement lear...
12/05/2020

When Do Curricula Work?

Inspired by human learning, researchers have proposed ordering examples ...
02/07/2020

Accelerating Reinforcement Learning for Reaching using Continuous Curriculum Learning

Reinforcement learning has shown great promise in the training of robot ...
02/08/2019

Architecture Compression

In this paper we propose a novel approach to model compression termed Ar...
This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.