Robust MAML: Prioritization task buffer with adaptive learning process for model-agnostic meta-learning

03/15/2021
by   Thanh Nguyen, et al.
0

Model agnostic meta-learning (MAML) is a popular state-of-the-art meta-learning algorithm that provides good weight initialization of a model given a variety of learning tasks. The model initialized by provided weight can be fine-tuned to an unseen task despite only using a small amount of samples and within a few adaptation steps. MAML is simple and versatile but requires costly learning rate tuning and careful design of the task distribution which affects its scalability and generalization. This paper proposes a more robust MAML based on an adaptive learning scheme and a prioritization task buffer(PTB) referred to as Robust MAML (RMAML) for improving scalability of training process and alleviating the problem of distribution mismatch. RMAML uses gradient-based hyper-parameter optimization to automatically find the optimal learning rate and uses the PTB to gradually adjust train-ing task distribution toward testing task distribution over the course of training. Experimental results on meta reinforcement learning environments demonstrate a substantial performance gain as well as being less sensitive to hyper-parameter choice and robust to distribution mismatch.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/09/2017

Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks

We propose an algorithm for meta-learning that is model-agnostic, in the...
research
11/13/2020

A Reweighted Meta Learning Framework for Robust Few Shot Learning

Model-Agnostic Meta-Learning (MAML) is a popular gradient-based meta-lea...
research
03/20/2020

Weighted Meta-Learning

Meta-learning leverages related source tasks to learn an initialization ...
research
01/01/2021

B-SMALL: A Bayesian Neural Network approach to Sparse Model-Agnostic Meta-Learning

There is a growing interest in the learning-to-learn paradigm, also know...
research
10/31/2020

Meta-Learning with Adaptive Hyperparameters

Despite its popularity, several recent works question the effectiveness ...
research
10/28/2019

HIDRA: Head Initialization across Dynamic targets for Robust Architectures

The performance of gradient-based optimization strategies depends heavil...
research
03/08/2018

Reptile: a Scalable Metalearning Algorithm

This paper considers metalearning problems, where there is a distributio...

Please sign up or login with your details

Forgot password? Click here to reset