Principled and Efficient Transfer Learning of Deep Models via Neural Collapse

12/23/2022
by   Xiao Li, et al.
0

With the ever-growing model size and the limited availability of labeled training data, transfer learning has become an increasingly popular approach in many science and engineering domains. For classification problems, this work delves into the mystery of transfer learning through an intriguing phenomenon termed neural collapse (NC), where the last-layer features and classifiers of learned deep networks satisfy: (i) the within-class variability of the features collapses to zero, and (ii) the between-class feature means are maximally and equally separated. Through the lens of NC, our findings for transfer learning are the following: (i) when pre-training models, preventing intra-class variability collapse (to a certain extent) better preserves the intrinsic structures of the input data, so that it leads to better model transferability; (ii) when fine-tuning models on downstream tasks, obtaining features with more NC on downstream data results in better test accuracy on the given task. The above results not only demystify many widely used heuristics in model pre-training (e.g., data augmentation, projection head, self-supervised learning), but also leads to more efficient and principled fine-tuning method on downstream tasks that we demonstrate through extensive experimental results.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/27/2023

The Role of Pre-training Data in Transfer Learning

The transfer learning paradigm of model pre-training and subsequent fine...
research
07/19/2021

Adaptive Transfer Learning on Graph Neural Networks

Graph neural networks (GNNs) is widely used to learn a powerful represen...
research
11/08/2022

When How to Transfer with Transfer Learning

In deep learning, transfer learning (TL) has become the de facto approac...
research
03/27/2023

Generalizable Local Feature Pre-training for Deformable Shape Analysis

Transfer learning is fundamental for addressing problems in settings wit...
research
05/26/2022

Understanding new tasks through the lens of training data via exponential tilting

Deploying machine learning models to new tasks is a major challenge desp...
research
05/20/2022

Pre-Train Your Loss: Easy Bayesian Transfer Learning with Informative Priors

Deep learning is increasingly moving towards a transfer learning paradig...
research
08/01/2023

EEG-based Cognitive Load Classification using Feature Masked Autoencoding and Emotion Transfer Learning

Cognitive load, the amount of mental effort required for task completion...

Please sign up or login with your details

Forgot password? Click here to reset