Provable Generalization of Overparameterized Meta-learning Trained with SGD

06/18/2022
by   Yu Huang, et al.
0

Despite the superior empirical success of deep meta-learning, theoretical understanding of overparameterized meta-learning is still limited. This paper studies the generalization of a widely used meta-learning approach, Model-Agnostic Meta-Learning (MAML), which aims to find a good initialization for fast adaptation to new tasks. Under a mixed linear regression model, we analyze the generalization properties of MAML trained with SGD in the overparameterized regime. We provide both upper and lower bounds for the excess risk of MAML, which captures how SGD dynamics affect these generalization bounds. With such sharp characterizations, we further explore how various learning parameters impact the generalization capability of overparameterized MAML, including explicitly identifying typical data and task distributions that can achieve diminishing generalization error with overparameterization, and characterizing the impact of adaptation learning rate on both excess risk and the early stopping time. Our theoretical findings are further validated by experiments.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/09/2023

Theoretical Characterization of the Generalization Performance of Overfitted Meta-Learning

Meta-learning has arisen as a successful method for improving training p...
research
03/06/2022

Is Bayesian Model-Agnostic Meta Learning Better than Model-Agnostic Meta Learning, Provably?

Meta learning aims at learning a model that can quickly adapt to unseen ...
research
10/31/2020

On Optimality of Meta-Learning in Fixed-Design Regression with Weighted Biased Regularization

We consider a fixed-design linear regression in the meta-learning model ...
research
06/08/2022

Sharp-MAML: Sharpness-Aware Model-Agnostic Meta Learning

Model-agnostic meta learning (MAML) is currently one of the dominating a...
research
07/26/2020

Don't Overlook the Support Set: Towards Improving Generalization in Meta-learning

Meta-learning has proven to be a powerful paradigm for transferring the ...
research
03/18/2022

Negative Inner-Loop Learning Rates Learn Universal Features

Model Agnostic Meta-Learning (MAML) consists of two optimization loops: ...
research
10/14/2020

Theoretical bounds on estimation error for meta-learning

Machine learning models have traditionally been developed under the assu...

Please sign up or login with your details

Forgot password? Click here to reset