UMIX: Improving Importance Weighting for Subpopulation Shift via Uncertainty-Aware Mixup

09/19/2022
by   Zongbo Han, et al.
0

Subpopulation shift wildly exists in many real-world machine learning applications, referring to the training and test distributions containing the same subpopulation groups but varying in subpopulation frequencies. Importance reweighting is a normal way to handle the subpopulation shift issue by imposing constant or adaptive sampling weights on each sample in the training dataset. However, some recent studies have recognized that most of these approaches fail to improve the performance over empirical risk minimization especially when applied to over-parameterized neural networks. In this work, we propose a simple yet practical framework, called uncertainty-aware mixup (Umix), to mitigate the overfitting issue in over-parameterized models by reweighting the "mixed" samples according to the sample uncertainty. The training-trajectories-based uncertainty estimation is equipped in the proposed Umix for each sample to flexibly characterize the subpopulation distribution. We also provide insightful theoretical analysis to verify that Umix achieves better generalization bounds over prior works. Further, we conduct extensive empirical studies across a wide range of tasks to validate the effectiveness of our method both qualitatively and quantitatively.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/09/2023

Reweighted Mixup for Subpopulation Shift

Subpopulation shift exists widely in many real-world applications, which...
research
07/08/2020

A One-step Approach to Covariate Shift Adaptation

A default assumption in many machine learning scenarios is that the trai...
research
12/24/2021

Is Importance Weighting Incompatible with Interpolating Classifiers?

Importance weighting is a classic technique to handle distribution shift...
research
02/11/2023

Pushing the Accuracy-Group Robustness Frontier with Introspective Self-play

Standard empirical risk minimization (ERM) training can produce deep neu...
research
01/28/2022

Understanding Why Generalized Reweighting Does Not Improve Over ERM

Empirical risk minimization (ERM) is known in practice to be non-robust ...
research
05/31/2022

Mitigating Dataset Bias by Using Per-sample Gradient

The performance of deep neural networks is strongly influenced by the tr...
research
10/12/2022

Deep Combinatorial Aggregation

Neural networks are known to produce poor uncertainty estimations, and a...

Please sign up or login with your details

Forgot password? Click here to reset