Generalization Bounds for Transfer Learning with Pretrained Classifiers

12/23/2022
by   Tomer Galanti, et al.
4

We study the ability of foundation models to learn representations for classification that are transferable to new, unseen classes. Recent results in the literature show that representations learned by a single classifier over many classes are competitive on few-shot learning problems with representations learned by special-purpose algorithms designed for such problems. We offer an explanation for this phenomenon based on the concept of class-features variability collapse, which refers to the training dynamics of deep classification networks where the feature embeddings of samples belonging to the same class tend to concentrate around their class means. More specifically, we examine the few-shot error of the learned feature map, which is the classification error of the nearest class-center classifier using centers learned from a small number of random samples from each class. Assuming that the classes appearing in the data are selected independently from a distribution, we show that the few-shot error generalizes from the training data to unseen test data, and we provide an upper bound on the expected few-shot error for new classes (selected from the same distribution) using the average few-shot error for the source classes. Additionally, we show that the few-shot error on the training data can be upper bounded using the degree of class-features variability collapse. This suggests that foundation models can provide feature maps that are transferable to new downstream tasks even with limited data available.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/30/2021

On the Role of Neural Collapse in Transfer Learning

We study the ability of foundation models to learn representations for c...
research
08/22/2020

Few-Shot Learning with Intra-Class Knowledge Transfer

We consider the few-shot classification task with an unbalanced dataset,...
research
06/23/2018

Dynamic Spectrum Matching with One-shot Learning

Convolutional neural networks (CNN) have been shown to provide a good so...
research
08/23/2022

The Value of Out-of-Distribution Data

More data helps us generalize to a task. But real datasets can contain o...
research
12/13/2022

A Statistical Model for Predicting Generalization in Few-Shot Classification

The estimation of the generalization error of classifiers often relies o...
research
05/29/2022

A Model of One-Shot Generalization

We provide a theoretical framework to study a phenomenon that we call on...
research
06/03/2019

One-Way Prototypical Networks

Few-shot models have become a popular topic of research in the past year...

Please sign up or login with your details

Forgot password? Click here to reset