Distribution-Agnostic Model-Agnostic Meta-Learning

02/12/2020
by   Liam Collins, et al.
21

The Model-Agnostic Meta-Learning (MAML) algorithm <cit.> has been celebrated for its efficiency and generality, as it has demonstrated success in quickly learning the parameters of an arbitrary learning model. However, MAML implicitly assumes that the tasks come from a particular distribution, and optimizes the expected (or sample average) loss over tasks drawn from this distribution. Here, we amend this limitation of MAML by reformulating the objective function as a min-max problem, where the maximization is over the set of possible distributions over tasks. Our proposed algorithm is the first distribution-agnostic and model-agnostic meta-learning method, and we show that it converges to an ϵ-accurate point at the rate of O(1/ϵ^2) in the convex setting and to an (ϵ, δ)-stationary point at the rate of O(max{1/ϵ^5, 1/δ^5}) in nonconvex settings. We also provide numerical experiments that demonstrate the worst-case superiority of our algorithm in comparison to MAML.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/06/2022

Is Bayesian Model-Agnostic Meta Learning Better than Model-Agnostic Meta Learning, Provably?

Meta learning aims at learning a model that can quickly adapt to unseen ...
research
09/27/2021

ST-MAML: A Stochastic-Task based Method for Task-Heterogeneous Meta-Learning

Optimization-based meta-learning typically assumes tasks are sampled fro...
research
12/03/2021

Learning to Broadcast for Ultra-Reliable Communication with Differential Quality of Service via the Conditional Value at Risk

Broadcast/multicast communication systems are typically designed to opti...
research
06/12/2020

Task-similarity Aware Meta-learning through Nonparametric Kernel Regression

Meta-learning refers to the process of abstracting a learning rule for a...
research
08/27/2019

On the Convergence Theory of Gradient-Based Model-Agnostic Meta-Learning Algorithms

In this paper, we study the convergence theory of a class of gradient-ba...
research
02/10/2020

Compositional ADAM: An Adaptive Compositional Solver

In this paper, we present C-ADAM, the first adaptive solver for composit...
research
08/05/2021

Multimodal Meta-Learning for Time Series Regression

Recent work has shown the efficiency of deep learning models such as Ful...

Please sign up or login with your details

Forgot password? Click here to reset