Beyond calibration: estimating the grouping loss of modern neural networks

10/28/2022
by   Alexandre Perez-Lebel, et al.
0

Good decision making requires machine-learning models to provide trustworthy confidence scores. To this end, recent work has focused on miscalibration, i.e, the over or under confidence of model scores. Yet, contrary to widespread belief, calibration is not enough: even a classifier with the best possible accuracy and perfect calibration can have confidence scores far from the true posterior probabilities. This is due to the grouping loss, created by samples with the same confidence scores but different true posterior probabilities. Proper scoring rule theory shows that given the calibration loss, the missing piece to characterize individual errors is the grouping loss. While there are many estimators of the calibration loss, none exists for the grouping loss in standard settings. Here, we propose an estimator to approximate the grouping loss. We use it to study modern neural network architectures in vision and NLP. We find that the grouping loss varies markedly across architectures, and that it is a key model-comparison factor across the most accurate, calibrated, models. We also show that distribution shifts lead to high grouping loss.

READ FULL TEXT

page 16

page 17

page 31

page 33

page 34

page 36

page 38

research
05/15/2021

Calibrating sufficiently

When probabilistic classifiers are trained and calibrated, the so-called...
research
11/21/2022

AdaFocal: Calibration-aware Adaptive Focal Loss

Much recent work has been devoted to the problem of ensuring that a neur...
research
06/08/2023

Beyond Probability Partitions: Calibrating Neural Networks with Semantic Aware Grouping

Research has shown that deep networks tend to be overly optimistic about...
research
02/08/2023

On the Richness of Calibration

Probabilistic predictions can be evaluated through comparisons with obse...
research
09/15/2021

Making Heads and Tails of Models with Marginal Calibration for Sparse Tagsets

For interpreting the behavior of a probabilistic model, it is useful to ...
research
11/18/2020

On Focal Loss for Class-Posterior Probability Estimation: A Theoretical Perspective

The focal loss has demonstrated its effectiveness in many real-world app...
research
09/08/2021

Estimating Expected Calibration Errors

Uncertainty in probabilistic classifiers predictions is a key concern wh...

Please sign up or login with your details

Forgot password? Click here to reset