Training on Test Data with Bayesian Adaptation for Covariate Shift

09/27/2021
by   Aurick Zhou, et al.
0

When faced with distribution shift at test time, deep neural networks often make inaccurate predictions with unreliable uncertainty estimates. While improving the robustness of neural networks is one promising approach to mitigate this issue, an appealing alternate to robustifying networks against all possible test-time shifts is to instead directly adapt them to unlabeled inputs from the particular distribution shift we encounter at test time. However, this poses a challenging question: in the standard Bayesian model for supervised learning, unlabeled inputs are conditionally independent of model parameters when the labels are unobserved, so what can unlabeled data tell us about the model parameters at test-time? In this paper, we derive a Bayesian model that provides for a well-defined relationship between unlabeled inputs under distributional shift and model parameters, and show how approximate inference in this model can be instantiated with a simple regularized entropy minimization procedure at test-time. We evaluate our method on a variety of distribution shifts for image classification, including image corruptions, natural distribution shifts, and domain adaptation settings, and show that our method improves both accuracy and uncertainty estimation.

READ FULL TEXT

page 8

page 18

page 19

research
06/18/2020

Fully Test-time Adaptation by Entropy Minimization

Faced with new and different data during testing, a model must adapt its...
research
10/18/2021

MEMO: Test Time Robustness via Adaptation and Augmentation

While deep neural networks can attain good accuracy on in-distribution t...
research
07/06/2020

Adaptive Risk Minimization: A Meta-Learning Approach for Tackling Group Shift

A fundamental assumption of most machine learning algorithms is that the...
research
05/15/2022

Parameter Adaptation for Joint Distribution Shifts

While different methods exist to tackle distinct types of distribution s...
research
10/19/2021

Test time Adaptation through Perturbation Robustness

Data samples generated by several real world processes are dynamic in na...
research
06/06/2020

Self-Supervised Dynamic Networks for Covariate Shift Robustness

As supervised learning still dominates most AI applications, test-time p...
research
06/16/2023

Neural Priming for Sample-Efficient Adaptation

We propose Neural Priming, a technique for adapting large pretrained mod...

Please sign up or login with your details

Forgot password? Click here to reset