Calibrating a Deep Neural Network with Its Predecessors

02/13/2023
by   Linwei Tao, et al.
0

Confidence calibration - the process to calibrate the output probability distribution of neural networks - is essential for safety-critical applications of such networks. Recent works verify the link between mis-calibration and overfitting. However, early stopping, as a well-known technique to mitigate overfitting, fails to calibrate networks. In this work, we study the limitions of early stopping and comprehensively analyze the overfitting problem of a network considering each individual block. We then propose a novel regularization method, predecessor combination search (PCS), to improve calibration by searching a combination of best-fitting block predecessors, where block predecessors are the corresponding network blocks with weight parameters from earlier training stages. PCS achieves the state-of-the-art calibration performance on multiple datasets and architectures. In addition, PCS improves model robustness under dataset distribution shift.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/27/2023

Conformal inference is (almost) free for neural networks trained with early stopping

Early stopping based on hold-out data is a popular regularization techni...
research
09/06/2023

Multiclass Alignment of Confidence and Certainty for Network Calibration

Deep neural networks (DNNs) have made great strides in pushing the state...
research
12/27/2022

Annealing Double-Head: An Architecture for Online Calibration of Deep Neural Networks

Model calibration, which is concerned with how frequently the model pred...
research
06/17/2021

On the Dark Side of Calibration for Modern Neural Networks

Modern neural networks are highly uncalibrated. It poses a significant c...
research
02/25/2021

Confidence Calibration with Bounded Error Using Transformations

As machine learning techniques become widely adopted in new domains, esp...
research
08/01/2023

Regularization, early-stopping and dreaming: a Hopfield-like setup to address generalization and overfitting

In this work we approach attractor neural networks from a machine learni...
research
07/13/2018

Neural Networks Regularization Through Representation Learning

Neural network models and deep models are one of the leading and state o...

Please sign up or login with your details

Forgot password? Click here to reset