Test-Time Adaptation to Distribution Shift by Confidence Maximization and Input Transformation

06/28/2021
by   Chaithanya Kumar Mummadi, et al.
0

Deep neural networks often exhibit poor performance on data that is unlikely under the train-time data distribution, for instance data affected by corruptions. Previous works demonstrate that test-time adaptation to data shift, for instance using entropy minimization, effectively improves performance on such shifted distributions. This paper focuses on the fully test-time adaptation setting, where only unlabeled data from the target distribution is required. This allows adapting arbitrary pretrained networks. Specifically, we propose a novel loss that improves test-time adaptation by addressing both premature convergence and instability of entropy minimization. This is achieved by replacing the entropy by a non-saturating surrogate and adding a diversity regularizer based on batch-wise entropy maximization that prevents convergence to trivial collapsed solutions. Moreover, we propose to prepend an input transformation module to the network that can partially undo test-time distribution shifts. Surprisingly, this preprocessing can be learned solely using the fully test-time adaptation loss in an end-to-end fashion without any target domain labels or source domain data. We show that our approach outperforms previous work in improving the robustness of publicly available pretrained image classifiers to common corruptions on such challenging benchmarks as ImageNet-C.

READ FULL TEXT
research
06/18/2020

Fully Test-time Adaptation by Entropy Minimization

Faced with new and different data during testing, a model must adapt its...
research
06/16/2023

Neural Priming for Sample-Efficient Adaptation

We propose Neural Priming, a technique for adapting large pretrained mod...
research
05/20/2022

Test-time Batch Normalization

Deep neural networks often suffer the data distribution shift between tr...
research
03/10/2022

On-the-Fly Test-time Adaptation for Medical Image Segmentation

One major problem in deep learning-based solutions for medical imaging i...
research
08/14/2023

Towards Open-Set Test-Time Adaptation Utilizing the Wisdom of Crowds in Entropy Minimization

Test-time adaptation (TTA) methods, which generally rely on the model's ...
research
12/05/2022

Addressing Distribution Shift at Test Time in Pre-trained Language Models

State-of-the-art pre-trained language models (PLMs) outperform other mod...
research
06/08/2023

RDumb: A simple approach that questions our progress in continual test-time adaptation

Test-Time Adaptation (TTA) allows to update pretrained models to changin...

Please sign up or login with your details

Forgot password? Click here to reset