Annealing Double-Head: An Architecture for Online Calibration of Deep Neural Networks

12/27/2022
by   Erdong Guo, et al.
0

Model calibration, which is concerned with how frequently the model predicts correctly, not only plays a vital part in statistical model design, but also has substantial practical applications, such as optimal decision-making in the real world. However, it has been discovered that modern deep neural networks are generally poorly calibrated due to the overestimation (or underestimation) of predictive confidence, which is closely related to overfitting. In this paper, we propose Annealing Double-Head, a simple-to-implement but highly effective architecture for calibrating the DNN during training. To be precise, we construct an additional calibration head-a shallow neural network that typically has one latent layer-on top of the last latent layer in the normal model to map the logits to the aligned confidence. Furthermore, a simple Annealing technique that dynamically scales the logits by calibration head in training procedure is developed to improve its performance. Under both the in-distribution and distributional shift circumstances, we exhaustively evaluate our Annealing Double-Head architecture on multiple pairs of contemporary DNN architectures and vision and speech datasets. We demonstrate that our method achieves state-of-the-art model calibration performance without post-processing while simultaneously providing comparable predictive accuracy in comparison to other recently proposed calibration methods on a range of learning tasks.

READ FULL TEXT

page 7

page 15

page 16

page 17

research
03/19/2023

Calibration of Neural Networks

Neural networks solving real-world problems are often required not only ...
research
06/16/2020

Calibrating Deep Neural Network Classifiers on Out-of-Distribution Datasets

To increase the trustworthiness of deep neural network (DNN) classifiers...
research
06/17/2021

On the Dark Side of Calibration for Modern Neural Networks

Modern neural networks are highly uncalibrated. It poses a significant c...
research
02/13/2023

Calibrating a Deep Neural Network with Its Predecessors

Confidence calibration - the process to calibrate the output probability...
research
12/07/2022

A Simple Nadaraya-Watson Head can offer Explainable and Calibrated Classification

In this paper, we empirically analyze a simple, non-learnable, and nonpa...
research
09/06/2023

Multiclass Alignment of Confidence and Certainty for Network Calibration

Deep neural networks (DNNs) have made great strides in pushing the state...
research
03/27/2023

Meta-Calibration Regularized Neural Networks

Miscalibration-the mismatch between predicted probability and the true c...

Please sign up or login with your details

Forgot password? Click here to reset