Robust Learning with Progressive Data Expansion Against Spurious Correlation

06/08/2023
by   Yihe Deng, et al.
4

While deep learning models have shown remarkable performance in various tasks, they are susceptible to learning non-generalizable spurious features rather than the core features that are genuinely correlated to the true label. In this paper, beyond existing analyses of linear models, we theoretically examine the learning process of a two-layer nonlinear convolutional neural network in the presence of spurious features. Our analysis suggests that imbalanced data groups and easily learnable spurious features can lead to the dominance of spurious features during the learning process. In light of this, we propose a new training algorithm called PDE that efficiently enhances the model's robustness for a better worst-group performance. PDE begins with a group-balanced subset of training data and progressively expands it to facilitate the learning of the core features. Experiments on synthetic and real-world benchmark datasets confirm the superior performance of our method on models such as ResNets and Transformers. On average, our method achieves a 2.8 improvement in worst-group accuracy compared with the state-of-the-art method, while enjoying up to 10x faster training efficiency.

READ FULL TEXT
research
08/01/2023

Is Last Layer Re-Training Truly Sufficient for Robustness to Spurious Correlations?

Models trained with empirical risk minimization (ERM) are known to learn...
research
05/19/2018

GEN Model: An Alternative Approach to Deep Neural Network Models

In this paper, we introduce an alternative approach, namely GEN (Genetic...
research
10/06/2021

Focus on the Common Good: Group Distributional Robustness Follows

We consider the problem of training a classification model with group an...
research
12/12/2022

You Only Need a Good Embeddings Extractor to Fix Spurious Correlations

Spurious correlations in training data often lead to robustness issues s...
research
05/30/2023

Identifying Spurious Biases Early in Training through the Lens of Simplicity Bias

Neural networks trained with (stochastic) gradient descent have an induc...
research
09/19/2022

Importance Tempering: Group Robustness for Overparameterized Models

Although overparameterized models have shown their success on many machi...
research
10/20/2022

Freeze then Train: Towards Provable Representation Learning under Spurious Correlations and Feature Noise

The existence of spurious correlations such as image backgrounds in the ...

Please sign up or login with your details

Forgot password? Click here to reset