Automated Curriculum Learning for Neural Networks

04/10/2017
by   Alex Graves, et al.
0

We introduce a method for automatically selecting the path, or syllabus, that a neural network follows through a curriculum so as to maximise learning efficiency. A measure of the amount that the network learns from each data sample is provided as a reward signal to a nonstationary multi-armed bandit algorithm, which then determines a stochastic syllabus. We consider a range of signals derived from two distinct indicators of learning progress: rate of increase in prediction accuracy, and rate of increase in network complexity. Experimental results for LSTM networks on three curricula demonstrate that our approach can significantly accelerate learning, in some cases halving the time required to attain a satisfactory performance level.

READ FULL TEXT
research
10/08/2021

Deep Upper Confidence Bound Algorithm for Contextual Bandit Ranking of Information Selection

Contextual multi-armed bandits (CMAB) have been widely used for learning...
research
01/28/2022

Top-K Ranking Deep Contextual Bandits for Information Selection Systems

In today's technology environment, information is abundant, dynamic, and...
research
06/07/2021

Multi-armed Bandit Requiring Monotone Arm Sequences

In many online learning or multi-armed bandit problems, the taken action...
research
07/31/2018

Online Adaptative Curriculum Learning for GANs

Generative Adversarial Networks (GANs) can successfully learn a probabil...
research
04/21/2020

Algorithms for slate bandits with non-separable reward functions

In this paper, we study a slate bandit problem where the function that d...
research
08/14/2018

Multi-user Communication Networks: A Coordinated Multi-armed Bandit Approach

Communication networks shared by many users are a widespread challenge n...
research
08/14/2020

Mastering Rate based Curriculum Learning

Recent automatic curriculum learning algorithms, and in particular Teach...

Please sign up or login with your details

Forgot password? Click here to reset