Diagnosing Model Performance Under Distribution Shift

03/03/2023
by   Tiffany Tianhui Cai, et al.
0

Prediction models can perform poorly when deployed to target distributions different from the training distribution. To understand these operational failure modes, we develop a method, called DIstribution Shift DEcomposition (DISDE), to attribute a drop in performance to different types of distribution shifts. Our approach decomposes the performance drop into terms for 1) an increase in harder but frequently seen examples from training, 2) changes in the relationship between features and outcomes, and 3) poor performance on examples infrequent or unseen during training. These terms are defined by fixing a distribution on X while varying the conditional distribution of Y | X between training and target, or by fixing the conditional distribution of Y | X while varying the distribution on X. In order to do this, we define a hypothetical distribution on X consisting of values common in both training and target, over which it is easy to compare Y | X and thus predictive performance. We estimate performance on this hypothetical distribution via reweighting methods. Empirically, we show how our method can 1) inform potential modeling improvements across distribution shifts for employment prediction on tabular census data, and 2) help to explain why certain domain adaptation methods fail to improve model performance for satellite image classification.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/06/2020

Estimating Generalization under Distribution Shifts via Domain-Invariant Representations

When machine learning models are deployed on a test distribution differe...
research
03/05/2023

Robustness, Evaluation and Adaptation of Machine Learning Models in the Wild

Our goal is to improve reliability of Machine Learning (ML) systems depl...
research
10/22/2022

Explanation Shift: Detecting distribution shifts on tabular data via the explanation space

As input data distributions evolve, the predictive performance of machin...
research
05/15/2022

Parameter Adaptation for Joint Distribution Shifts

While different methods exist to tackle distinct types of distribution s...
research
03/14/2023

Explanation Shift: Investigating Interactions between Models and Shifting Data Distributions

As input data distributions evolve, the predictive performance of machin...
research
07/09/2021

Accuracy on the Line: On the Strong Correlation Between Out-of-Distribution and In-Distribution Generalization

For machine learning systems to be reliable, we must understand their pe...
research
05/01/2020

Mind the Trade-off: Debiasing NLU Models without Degrading the In-distribution Performance

Models for natural language understanding (NLU) tasks often rely on the ...

Please sign up or login with your details

Forgot password? Click here to reset