Pushing the Accuracy-Group Robustness Frontier with Introspective Self-play

02/11/2023
by   Jeremiah Zhe Liu, et al.
0

Standard empirical risk minimization (ERM) training can produce deep neural network (DNN) models that are accurate on average but under-perform in under-represented population subgroups, especially when there are imbalanced group distributions in the long-tailed training data. Therefore, approaches that improve the accuracy-group robustness trade-off frontier of a DNN model (i.e. improving worst-group accuracy without sacrificing average accuracy, or vice versa) is of crucial importance. Uncertainty-based active learning (AL) can potentially improve the frontier by preferentially sampling underrepresented subgroups to create a more balanced training dataset. However, the quality of uncertainty estimates from modern DNNs tend to degrade in the presence of spurious correlations and dataset bias, compromising the effectiveness of AL for sampling tail groups. In this work, we propose Introspective Self-play (ISP), a simple approach to improve the uncertainty estimation of a deep neural network under dataset bias, by adding an auxiliary introspection task requiring a model to predict the bias for each data point in addition to the label. We show that ISP provably improves the bias-awareness of the model representation and the resulting uncertainty estimates. On two real-world tabular and language tasks, ISP serves as a simple "plug-in" for AL model training, consistently improving both the tail-group sampling rate and the final accuracy-fairness trade-off frontier of popular AL methods.

READ FULL TEXT

page 4

page 30

research
07/19/2021

Just Train Twice: Improving Group Robustness without Training Group Information

Standard training via empirical risk minimization (ERM) can produce mode...
research
04/20/2022

Improved Worst-Group Robustness via Classifier Retraining on Independent Splits

High-capacity deep neural networks (DNNs) trained with Empirical Risk Mi...
research
12/10/2019

To Balance or Not to Balance: An Embarrassingly Simple Approach for Learning with Long-Tailed Distributions

Real-world visual data often exhibits a long-tailed distribution, where ...
research
08/01/2023

Is Last Layer Re-Training Truly Sufficient for Robustness to Spurious Correlations?

Models trained with empirical risk minimization (ERM) are known to learn...
research
09/19/2022

UMIX: Improving Importance Weighting for Subpopulation Shift via Uncertainty-Aware Mixup

Subpopulation shift wildly exists in many real-world machine learning ap...
research
04/14/2021

Can Active Learning Preemptively Mitigate Fairness Issues?

Dataset bias is one of the prevailing causes of unfairness in machine le...
research
10/20/2021

Does Data Repair Lead to Fair Models? Curating Contextually Fair Data To Reduce Model Bias

Contextual information is a valuable cue for Deep Neural Networks (DNNs)...

Please sign up or login with your details

Forgot password? Click here to reset