Information-Theoretic Generalization Bounds for Meta-Learning and Applications

05/09/2020
by   Sharu Theresa Jose, et al.
0

Meta-learning, or "learning to learn", refers to techniques that infer an inductive bias from data corresponding to multiple related tasks with the goal of improving the sample efficiency for new, previously unobserved, tasks. A key performance measure for meta-learning is the meta-generalization gap, that is, the difference between the average loss measured on the meta-training data and on a new, randomly selected task. This paper presents novel information-theoretic upper bounds on the meta-generalization gap. Two broad classes of meta-learning algorithms are considered that uses either separate within-task training and test sets, like MAML, or joint within-task training and test sets, like Reptile. Extending the existing work for conventional learning, an upper bound on the meta-generalization gap is derived for the former class that depends on the mutual information (MI) between the output of the meta-learning algorithm and its input meta-training data. For the latter, the derived bound includes an additional MI between the output of the per-task learning procedure and corresponding data set to capture within-task uncertainty. Tighter bounds are then developed for the two classes via novel Individual Task MI (ITMI) bounds. Applications of the derived bounds are finally discussed, including a broad class of noisy iterative algorithms for meta-learning.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/21/2020

Conditional Mutual Information Bound for Meta Generalization Gap

Meta-learning infers an inductive bias—typically in the form of the hype...
research
09/29/2021

Generalization Bounds For Meta-Learning: An Information-Theoretic Analysis

We derive a novel information-theoretic analysis of the generalization p...
research
10/29/2020

Learning to Actively Learn: A Robust Approach

This work proposes a procedure for designing algorithms for specific ada...
research
06/01/2021

Information-Theoretic Analysis of Epistemic Uncertainty in Bayesian Meta-learning

The overall predictive uncertainty of a trained predictor can be decompo...
research
10/19/2021

BAMLD: Bayesian Active Meta-Learning by Disagreement

Data-efficient learning algorithms are essential in many practical appli...
research
10/14/2020

Theoretical bounds on estimation error for meta-learning

Machine learning models have traditionally been developed under the assu...
research
01/21/2021

An Information-Theoretic Analysis of the Impact of Task Similarity on Meta-Learning

Meta-learning aims at optimizing the hyperparameters of a model class or...

Please sign up or login with your details

Forgot password? Click here to reset