Top-label calibration

07/18/2021
by   Chirag Gupta, et al.
0

We study the problem of post-hoc calibration for multiclass classification, with an emphasis on histogram binning. Multiple works have focused on calibration with respect to the confidence of just the predicted class (or 'top-label'). We find that the popular notion of confidence calibration [Guo et al., 2017] is not sufficiently strong – there exist predictors that are not calibrated in any meaningful way but are perfectly confidence calibrated. We propose a closely related (but subtly different) notion, top-label calibration, that accurately captures the intuition and simplicity of confidence calibration, but addresses its drawbacks. We formalize a histogram binning (HB) algorithm that reduces top-label multiclass calibration to the binary case, prove that it has clean theoretical guarantees without distributional assumptions, and perform a methodical study of its practical performance. Some prediction tasks require stricter notions of multiclass calibration such as class-wise or canonical calibration. We formalize appropriate HB algorithms corresponding to each of these goals. In experiments with deep neural nets, we find that our principled versions of HB are often better than temperature scaling, for both top-label and class-wise calibration. Code for this work will be made publicly available at https://github.com/aigen/df-posthoc-calibration.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/07/2022

Class-wise and reduced calibration methods

For many applications of probabilistic classifiers it is important that ...
research
05/10/2021

Distribution-free calibration guarantees for histogram binning without sample splitting

We prove calibration guarantees for the popular histogram binning (also ...
research
03/25/2022

A Stitch in Time Saves Nine: A Train-Time Regularizing Loss for Improved Neural Network Calibration

Deep Neural Networks ( DNN s) are known to make overconfident mistakes, ...
research
08/06/2023

Two Sides of Miscalibration: Identifying Over and Under-Confidence Prediction for Network Calibration

Proper confidence calibration of deep neural networks is essential for r...
research
06/24/2020

Class-Similarity Based Label Smoothing for Generalized Confidence Calibration

Since modern neural networks are known to be overconfident, several tech...
research
07/13/2022

Estimating Classification Confidence Using Kernel Densities

This paper investigates the post-hoc calibration of confidence for "expl...
research
11/09/2020

Improving Classifier Confidence using Lossy Label-Invariant Transformations

Providing reliable model uncertainty estimates is imperative to enabling...

Please sign up or login with your details

Forgot password? Click here to reset