On Pitfalls of Test-Time Adaptation

06/06/2023
by   Hao Zhao, et al.
5

Test-Time Adaptation (TTA) has recently emerged as a promising approach for tackling the robustness challenge under distribution shifts. However, the lack of consistent settings and systematic studies in prior literature hinders thorough assessments of existing methods. To address this issue, we present TTAB, a test-time adaptation benchmark that encompasses ten state-of-the-art algorithms, a diverse array of distribution shifts, and two evaluation protocols. Through extensive experiments, our benchmark reveals three common pitfalls in prior efforts. First, selecting appropriate hyper-parameters, especially for model selection, is exceedingly difficult due to online batch dependency. Second, the effectiveness of TTA varies greatly depending on the quality and properties of the model being adapted. Third, even under optimal algorithmic conditions, none of the existing methods are capable of addressing all common types of distribution shifts. Our findings underscore the need for future research in the field to conduct rigorous evaluations on a broader set of models and shifts, and to re-examine the assumptions behind the empirical success of TTA. Our code is available at <https://github.com/lins-lab/ttab>.

READ FULL TEXT

page 4

page 14

research
07/06/2023

Benchmarking Test-Time Adaptation against Distribution Shifts in Image Classification

Test-time adaptation (TTA) is a technique aimed at enhancing the general...
research
11/24/2022

Video Test-Time Adaptation for Action Recognition

Although action recognition systems can achieve top performance when eva...
research
07/20/2021

Characterizing Generalization under Out-Of-Distribution Shifts in Deep Metric Learning

Deep Metric Learning (DML) aims to find representations suitable for zer...
research
04/10/2023

Revisiting Test Time Adaptation under Online Evaluation

This paper proposes a novel online evaluation protocol for Test Time Ada...
research
04/13/2022

Distributionally Robust Models with Parametric Likelihood Ratios

As machine learning models are deployed ever more broadly, it becomes in...
research
11/02/2022

Continual Conscious Active Fine-Tuning to Robustify Online Machine Learning Models Against Data Distribution Shifts

Unlike their offline traditional counterpart, online machine learning mo...
research
06/01/2023

Universal Test-time Adaptation through Weight Ensembling, Diversity Weighting, and Prior Correction

Since distribution shifts are likely to occur during test-time and can d...

Please sign up or login with your details

Forgot password? Click here to reset