When Does Group Invariant Learning Survive Spurious Correlations?

06/29/2022
by   Yimeng Chen, et al.
0

By inferring latent groups in the training data, recent works introduce invariant learning to the case where environment annotations are unavailable. Typically, learning group invariance under a majority/minority split is empirically shown to be effective in improving out-of-distribution generalization on many datasets. However, theoretical guarantee for these methods on learning invariant mechanisms is lacking. In this paper, we reveal the insufficiency of existing group invariant learning methods in preventing classifiers from depending on spurious correlations in the training set. Specifically, we propose two criteria on judging such sufficiency. Theoretically and empirically, we show that existing methods can violate both criteria and thus fail in generalizing to spurious correlation shifts. Motivated by this, we design a new group invariant learning method, which constructs groups with statistical independence tests, and reweights samples by group label proportion to meet the criteria. Experiments on both synthetic and real data demonstrate that the new method significantly outperforms existing group invariant learning methods in generalizing to spurious correlation shifts.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/09/2021

Heterogeneous Risk Minimization

Machine learning algorithms with empirical risk minimization usually suf...
research
05/13/2021

Causally-motivated Shortcut Removal Using Auxiliary Labels

Robustness to certain distribution shifts is a key requirement in many M...
research
01/10/2022

Towards Group Robustness in the presence of Partial Group Labels

Learning invariant representations is an important requirement when trai...
research
07/26/2022

Repeated Environment Inference for Invariant Learning

We study the problem of invariant learning when the environment labels a...
research
05/09/2020

An Investigation of Why Overparameterization Exacerbates Spurious Correlations

We study why overparameterization – increasing model size well beyond th...
research
05/30/2021

On the benefits of representation regularization in invariance based domain generalization

A crucial aspect in reliable machine learning is to design a deployable ...
research
11/18/2015

Unitary-Group Invariant Kernels and Features from Transformed Unlabeled Data

The study of representations invariant to common transformations of the ...

Please sign up or login with your details

Forgot password? Click here to reset