Meta-Calibration Regularized Neural Networks

03/27/2023
by   Cheng Wang, et al.
0

Miscalibration-the mismatch between predicted probability and the true correctness likelihood-has been frequently identified in modern deep neural networks. Recent work in the field aims to address this problem by training calibrated models directly by optimizing a proxy of the calibration error alongside the conventional objective. Recently, Meta-Calibration (MC) showed the effectiveness of using meta-learning for learning better calibrated models. In this work, we extend MC with two main components: (1) gamma network (gamma-net), a meta network to learn a sample-wise gamma at a continuous space for focal loss for optimizing backbone network; (2) smooth expected calibration error (SECE), a Gaussian-kernel based unbiased and differentiable ECE which aims to smoothly optimizing gamma-net. The proposed method regularizes neural network towards better calibration meanwhile retain predictive performance. Our experiments show that (a) learning sample-wise gamma at continuous space can effectively perform calibration; (b) SECE smoothly optimise gamma-net towards better robustness to binning schemes; (c) the combination of gamma-net and SECE achieve the best calibration performance across various calibration metrics and retain very competitive predictive performance as compared to multiple recently proposed methods on three datasets.

READ FULL TEXT

page 6

page 7

page 14

page 15

research
06/17/2021

Meta-Calibration: Meta-Learning of Model Calibration Using Differentiable Expected Calibration Error

Calibration of neural networks is a topical problem that is becoming inc...
research
10/13/2022

A Consistent and Differentiable Lp Canonical Calibration Error Estimator

Calibrated probabilistic classifiers are models whose predicted probabil...
research
12/09/2021

Obtaining Calibrated Probabilities with Personalized Ranking Models

For personalized ranking models, the well-calibrated probability of an i...
research
08/02/2021

Learning to Learn to Demodulate with Uncertainty Quantification via Bayesian Meta-Learning

Meta-learning, or learning to learn, offers a principled framework for f...
research
03/15/2023

On the Calibration and Uncertainty with Pólya-Gamma Augmentation for Dialog Retrieval Models

Deep neural retrieval models have amply demonstrated their power but est...
research
12/27/2022

Annealing Double-Head: An Architecture for Online Calibration of Deep Neural Networks

Model calibration, which is concerned with how frequently the model pred...
research
07/05/2022

Meta-Learning a Real-Time Tabular AutoML Method For Small Data

We present TabPFN, an AutoML method that is competitive with the state o...

Please sign up or login with your details

Forgot password? Click here to reset