How robust are pre-trained models to distribution shift?

06/17/2022
by   Yuge Shi, et al.
0

The vulnerability of machine learning models to spurious correlations has mostly been discussed in the context of supervised learning (SL). However, there is a lack of insight on how spurious correlations affect the performance of popular self-supervised learning (SSL) and auto-encoder based models (AE). In this work, we shed light on this by evaluating the performance of these models on both real world and synthetic distribution shift datasets. Following observations that the linear head itself can be susceptible to spurious correlations, we develop a novel evaluation scheme with the linear head trained on out-of-distribution (OOD) data, to isolate the performance of the pre-trained models from a potential bias of the linear head used for evaluation. With this new methodology, we show that SSL models are consistently more robust to distribution shifts and thus better at OOD generalisation than AE and SL models.

READ FULL TEXT

page 4

page 5

page 11

research
10/01/2021

A Survey of Knowledge Enhanced Pre-trained Models

Pre-trained models learn contextualized word representations on large-sc...
research
03/14/2022

SUPERB-SG: Enhanced Speech processing Universal PERformance Benchmark for Semantic and Generative Capabilities

Transfer learning has proven to be crucial in advancing the state of spe...
research
10/12/2020

Measuring and Reducing Gendered Correlations in Pre-trained Models

Pre-trained models have revolutionized natural language understanding. H...
research
06/03/2023

Uncovering the Hidden Dynamics of Video Self-supervised Learning under Distribution Shifts

Video self-supervised learning (VSSL) has made significant progress in r...
research
01/28/2023

Deciphering the Projection Head: Representation Evaluation Self-supervised Learning

Self-supervised learning (SSL) aims to learn intrinsic features without ...
research
05/18/2023

A Survey on Time-Series Pre-Trained Models

Time-Series Mining (TSM) is an important research area since it shows gr...
research
04/08/2023

Application of Self-Supervised Learning to MICA Model for Reconstructing Imperfect 3D Facial Structures

In this study, we emphasize the integration of a pre-trained MICA model ...

Please sign up or login with your details

Forgot password? Click here to reset