(Almost) Provable Error Bounds Under Distribution Shift via Disagreement Discrepancy

06/01/2023
by   Elan Rosenfeld, et al.
0

We derive an (almost) guaranteed upper bound on the error of deep neural networks under distribution shift using unlabeled test data. Prior methods either give bounds that are vacuous in practice or give estimates that are accurate on average but heavily underestimate error for a sizeable fraction of shifts. In particular, the latter only give guarantees based on complex continuous measures such as test calibration – which cannot be identified without labels – and are therefore unreliable. Instead, our bound requires a simple, intuitive condition which is well justified by prior empirical works and holds in practice effectively 100 ℋΔℋ-divergence but is easier to evaluate and substantially tighter, consistently providing non-vacuous guarantees. Estimating the bound requires optimizing one multiclass classifier to disagree with another, for which some prior works have used sub-optimal proxy losses; we devise a "disagreement loss" which is theoretically justified and performs better in practice. We expect this loss can serve as a drop-in replacement for future methods which require maximizing multiclass disagreement. Across a wide range of benchmarks, our method gives valid error bounds while achieving average accuracy comparable to competitive estimation baselines. Code is publicly available at https://github.com/erosenfeld/disagree_discrep .

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/02/2023

Empirical Analysis of the AdaBoost's Error Bound

Understanding the accuracy limits of machine learning algorithms is esse...
research
07/20/2022

Test-Time Adaptation via Conjugate Pseudo-labels

Test-time adaptation (TTA) refers to adapting neural networks to distrib...
research
06/08/2019

Provably Robust Boosted Decision Stumps and Trees against Adversarial Attacks

The problem of adversarial samples has been studied extensively for neur...
research
08/14/2023

Distance Matters For Improving Performance Estimation Under Covariate Shift

Performance estimation under covariate shift is a crucial component of s...
research
10/09/2022

Test-time Recalibration of Conformal Predictors Under Distribution Shift Based on Unlabeled Examples

Modern image classifiers achieve high predictive accuracy, but the predi...
research
10/22/2020

Posterior Re-calibration for Imbalanced Datasets

Neural Networks can perform poorly when the training label distribution ...

Please sign up or login with your details

Forgot password? Click here to reset