Malign Overfitting: Interpolation Can Provably Preclude Invariance

11/28/2022
by   Yoav Wald, et al.
0

Learned classifiers should often possess certain invariance properties meant to encourage fairness, robustness, or out-of-distribution generalization. However, multiple recent works empirically demonstrate that common invariance-inducing regularizers are ineffective in the over-parameterized regime, in which classifiers perfectly fit (i.e. interpolate) the training data. This suggests that the phenomenon of “benign overfitting," in which models generalize well despite interpolating, might not favorably extend to settings in which robustness or fairness are desirable. In this work we provide a theoretical justification for these observations. We prove that – even in the simplest of settings – any interpolating learning rule (with arbitrarily small margin) will not satisfy these invariance properties. We then propose and analyze an algorithm that – in the same setting – successfully learns a non-interpolating classifier that is provably invariant. We validate our theoretical observations on simulated data and the Waterbirds dataset.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/07/2023

On genuine invariance learning without weight-tying

In this paper, we investigate properties and limitations of invariance l...
research
12/15/2021

Robust Neural Network Classification via Double Regularization

The presence of mislabeled observations in data is a notoriously challen...
research
05/24/2021

Learning Security Classifiers with Verified Global Robustness Properties

Recent works have proposed methods to train classifiers with local robus...
research
09/24/2022

A Simple Strategy to Provable Invariance via Orbit Mapping

Many applications require robustness, or ideally invariance, of neural n...
research
08/05/2021

Interpolation can hurt robust generalization even when there is no noise

Numerous recent works show that overparameterization implicitly reduces ...
research
06/18/2021

Training or Architecture? How to Incorporate Invariance in Neural Networks

Many applications require the robustness, or ideally the invariance, of ...
research
05/07/2019

Unified Adversarial Invariance

We present a unified invariance framework for supervised neural networks...

Please sign up or login with your details

Forgot password? Click here to reset