A Uniform Confidence Phenomenon in Deep Learning and its Implications for Calibration

06/01/2023
by   Muthu Chidambaram, et al.
0

Despite the impressive generalization capabilities of deep neural networks, they have been repeatedly shown to poorly estimate their predictive uncertainty - in other words, they are frequently overconfident when they are wrong. Fixing this issue is known as model calibration, and has consequently received much attention in the form of modified training schemes and post-training calibration procedures. In this work, we present a significant hurdle to the calibration of modern models: deep neural networks have large neighborhoods of almost certain confidence around their training points. We demonstrate in our experiments that this phenomenon consistently arises (in the context of image classification) across many model and dataset pairs. Furthermore, we prove that when this phenomenon holds, for a large class of data distributions with overlaps between classes, it is not possible to obtain a model that is asymptotically better than random (with respect to calibration) even after applying the standard post-training calibration technique of temperature scaling. On the other hand, we also prove that it is possible to circumvent this defect by changing the training process to use a modified loss based on the Mixup data augmentation technique.

READ FULL TEXT

page 22

page 23

research
08/30/2019

Bin-wise Temperature Scaling (BTS): Improvement in Confidence Calibration Performance through Simple Scaling Techniques

The prediction reliability of neural networks is important in many appli...
research
07/30/2021

Soft Calibration Objectives for Neural Networks

Optimal decision making requires that classifiers produce uncertainty es...
research
07/31/2022

Adaptive Temperature Scaling for Robust Calibration of Deep Neural Networks

In this paper, we study the post-hoc calibration of modern neural networ...
research
05/27/2019

On Mixup Training: Improved Calibration and Predictive Uncertainty for Deep Neural Networks

Mixup zhang2017mixup is a recently proposed method for training deep neu...
research
03/05/2023

Expectation consistency for calibration of neural networks

Despite their incredible performance, it is well reported that deep neur...
research
12/16/2019

On-manifold Adversarial Data Augmentation Improves Uncertainty Calibration

Uncertainty estimates help to identify ambiguous, novel, or anomalous in...
research
12/01/2022

Investigating Deep Learning Model Calibration for Classification Problems in Mechanics

Recently, there has been a growing interest in applying machine learning...

Please sign up or login with your details

Forgot password? Click here to reset