Two Sides of Miscalibration: Identifying Over and Under-Confidence Prediction for Network Calibration

08/06/2023
by   Shuang Ao, et al.
0

Proper confidence calibration of deep neural networks is essential for reliable predictions in safety-critical tasks. Miscalibration can lead to model over-confidence and/or under-confidence; i.e., the model's confidence in its prediction can be greater or less than the model's accuracy. Recent studies have highlighted the over-confidence issue by introducing calibration techniques and demonstrated success on various tasks. However, miscalibration through under-confidence has not yet to receive much attention. In this paper, we address the necessity of paying attention to the under-confidence issue. We first introduce a novel metric, a miscalibration score, to identify the overall and class-wise calibration status, including being over or under-confident. Our proposed metric reveals the pitfalls of existing calibration techniques, where they often overly calibrate the model and worsen under-confident predictions. Then we utilize the class-wise miscalibration score as a proxy to design a calibration technique that can tackle both over and under-confidence. We report extensive experiments that show our proposed methods substantially outperforming existing calibration techniques. We also validate our proposed calibration technique on an automatic failure detection task with a risk-coverage curve, reporting that our methods improve failure detection as well as trustworthiness of the model. The code are available at <https://github.com/AoShuang92/miscalibration_TS>.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/06/2023

Rethinking Confidence Calibration for Failure Prediction

Reliable confidence estimation for the predictions is important in many ...
research
08/06/2023

Empirical Optimal Risk to Quantify Model Trustworthiness for Failure Detection

Failure detection (FD) in AI systems is a crucial safeguard for the depl...
research
09/06/2023

Do We Still Need Non-Maximum Suppression? Accurate Confidence Estimates and Implicit Duplication Modeling with IoU-Aware Calibration

Object detectors are at the heart of many semi- and fully autonomous dec...
research
07/21/2023

Making Pre-trained Language Models both Task-solvers and Self-calibrators

Pre-trained language models (PLMs) serve as backbones for various real-w...
research
07/18/2021

Top-label calibration

We study the problem of post-hoc calibration for multiclass classificati...
research
10/07/2022

Class-wise and reduced calibration methods

For many applications of probabilistic classifiers it is important that ...
research
05/31/2023

Perception and Semantic Aware Regularization for Sequential Confidence Calibration

Deep sequence recognition (DSR) models receive increasing attention due ...

Please sign up or login with your details

Forgot password? Click here to reset