Calibrating Deep Neural Networks using Explicit Regularisation and Dynamic Data Pruning

12/20/2022
by   Ramya Hebbalaguppe, et al.
7

Deep neural networks (DNN) are prone to miscalibrated predictions, often exhibiting a mismatch between the predicted output and the associated confidence scores. Contemporary model calibration techniques mitigate the problem of overconfident predictions by pushing down the confidence of the winning class while increasing the confidence of the remaining classes across all test samples. However, from a deployment perspective, an ideal model is desired to (i) generate well-calibrated predictions for high-confidence samples with predicted probability say >0.95, and (ii) generate a higher proportion of legitimate high-confidence samples. To this end, we propose a novel regularization technique that can be used with classification losses, leading to state-of-the-art calibrated predictions at test time; From a deployment standpoint in safety-critical applications, only high-confidence samples from a well-calibrated model are of interest, as the remaining samples have to undergo manual inspection. Predictive confidence reduction of these potentially “high-confidence samples” is a downside of existing calibration approaches. We mitigate this by proposing a dynamic train-time data pruning strategy that prunes low-confidence samples every few epochs, providing an increase in "confident yet calibrated samples". We demonstrate state-of-the-art calibration performance across image classification benchmarks, reducing training time without much compromise in accuracy. We provide insights into why our dynamic pruning strategy that prunes low-confidence training samples leads to an increase in high-confidence samples at test time.

READ FULL TEXT
research
09/06/2023

Multiclass Alignment of Confidence and Certainty for Network Calibration

Deep neural networks (DNNs) have made great strides in pushing the state...
research
02/23/2020

On the Role of Dataset Quality and Heterogeneity in Model Confidence

Safety-critical applications require machine learning models that output...
research
12/03/2019

Distance-Based Learning from Errors for Confidence Calibration

Deep neural networks (DNNs) are poorly-calibrated when trained in conven...
research
07/21/2017

Confidence estimation in Deep Neural networks via density modelling

State-of-the-art Deep Neural Networks can be easily fooled into providin...
research
06/08/2023

Conservative Prediction via Data-Driven Confidence Minimization

Errors of machine learning models are costly, especially in safety-criti...
research
08/09/2023

Expert load matters: operating networks at high accuracy and low manual effort

In human-AI collaboration systems for critical applications, in order to...
research
07/19/2018

Improving Simple Models with Confidence Profiles

In this paper, we propose a new method called ProfWeight for transferrin...

Please sign up or login with your details

Forgot password? Click here to reset