Calibrated Predictive Distributions via Diagnostics for Conditional Coverage

05/29/2022
by   Biprateep Dey, et al.
0

Uncertainty quantification is crucial for assessing the predictive ability of AI algorithms. A large body of work (including normalizing flows and Bayesian neural networks) has been devoted to describing the entire predictive distribution (PD) of a target variable Y given input features 𝐗. However, off-the-shelf PDs are usually far from being conditionally calibrated; i.e., the probability of occurrence of an event given input 𝐗 can be significantly different from the predicted probability. Most current research on predictive inference (such as conformal prediction) concerns constructing prediction sets, that do not only provide correct uncertainties on average over the entire population (that is, averaging over 𝐗), but that are also approximately conditionally calibrated with accurate uncertainties for individual instances. It is often believed that the problem of obtaining and assessing entire conditionally calibrated PDs is too challenging to approach. In this work, we show that recalibration as well as validation are indeed attainable goals in practice. Our proposed method relies on the idea of regressing probability integral transform (PIT) scores against 𝐗. This regression gives full diagnostics of conditional coverage across the entire feature space and can be used to recalibrate misspecified PDs. We benchmark our corrected prediction bands against oracle bands and state-of-the-art predictive inference algorithms for synthetic data, including settings with distributional shift and dependent high-dimensional sequence data. Finally, we demonstrate an application to the physical sciences in which we assess and produce calibrated PDs for measurements of galaxy distances using imaging data (i.e., photometric redshifts).

READ FULL TEXT

page 8

page 26

research
05/31/2022

Simulation-Based Inference with WALDO: Perfectly Calibrated Confidence Regions Using Any Prediction or Posterior Estimation Algorithm

The vast majority of modern machine learning targets prediction problems...
research
08/26/2019

Marginally-calibrated deep distributional regression

Deep neural network (DNN) regression models are widely used in applicati...
research
02/20/2021

Validating Conditional Density Models and Bayesian Inference Algorithms

Conditional density models f(y|x), where x represents a potentially high...
research
08/17/2021

Neural density estimation and uncertainty quantification for laser induced breakdown spectroscopy spectra

Constructing probability densities for inference in high-dimensional spe...
research
06/02/2022

Practical Adversarial Multivalid Conformal Prediction

We give a simple, generic conformal prediction method for sequential pre...
research
12/08/2018

Adaptive and Calibrated Ensemble Learning with Dependent Tail-free Process

Ensemble learning is a mainstay in modern data science practice. Convent...
research
11/25/2020

Bayesian Triplet Loss: Uncertainty Quantification in Image Retrieval

Uncertainty quantification in image retrieval is crucial for downstream ...

Please sign up or login with your details

Forgot password? Click here to reset