When Does Optimizing a Proper Loss Yield Calibration?

05/30/2023
by   Jarosław Błasiok, et al.
0

Optimizing proper loss functions is popularly believed to yield predictors with good calibration properties; the intuition being that for such losses, the global optimum is to predict the ground-truth probabilities, which is indeed calibrated. However, typical machine learning models are trained to approximately minimize loss over restricted families of predictors, that are unlikely to contain the ground truth. Under what circumstances does optimizing proper loss over a restricted family yield calibrated models? What precise calibration guarantees does it give? In this work, we provide a rigorous answer to these questions. We replace the global optimality with a local optimality condition stipulating that the (proper) loss of the predictor cannot be reduced much by post-processing its predictions with a certain family of Lipschitz functions. We show that any predictor with this local optimality satisfies smooth calibration as defined in Kakade-Foster (2008), Błasiok et al. (2023). Local optimality is plausibly satisfied by well-trained DNNs, which suggests an explanation for why they are calibrated from proper loss minimization alone. Finally, we show that the connection between local optimality and calibration error goes both ways: nearly calibrated predictors are also nearly locally optimal.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/21/2023

A Unifying Perspective on Multi-Calibration: Unleashing Game Dynamics for Multi-Objective Learning

We provide a unifying framework for the design and analysis of multi-cal...
research
11/30/2022

A Unifying Theory of Distance from Calibration

We study the fundamental question of how to define and measure the dista...
research
05/20/2022

On Calibration of Ensemble-Based Credal Predictors

In recent years, several classification methods that intend to quantify ...
research
01/30/2019

On the Consistency of Top-k Surrogate Losses

The top-k error is often employed to evaluate performance for challengin...
research
10/16/2022

Loss Minimization through the Lens of Outcome Indistinguishability

We present a new perspective on loss minimization and the recent notion ...
research
05/10/2022

Calibrating for Class Weights by Modeling Machine Learning

A much studied issue is the extent to which the confidence scores provid...
research
11/04/2021

Scaffolding Sets

Predictors map individual instances in a population to the interval [0,1...

Please sign up or login with your details

Forgot password? Click here to reset