Distribution-Free, Risk-Controlling Prediction Sets

01/07/2021
by   Stephen Bates, et al.
11

While improving prediction accuracy has been the focus of machine learning in recent years, this alone does not suffice for reliable decision-making. Deploying learning systems in consequential settings also requires calibrating and communicating the uncertainty of predictions. To convey instance-wise uncertainty for prediction tasks, we show how to generate set-valued predictions from a black-box predictor that control the expected loss on future test points at a user-specified level. Our approach provides explicit finite-sample guarantees for any dataset by using a holdout set to calibrate the size of the prediction sets. This framework enables simple, distribution-free, rigorous error control for many tasks, and we demonstrate it in five large-scale machine learning problems: (1) classification problems where some mistakes are more costly than others; (2) multi-label classification, where each observation has multiple associated labels; (3) classification problems where the labels have a hierarchical structure; (4) image segmentation, where we wish to predict a set of pixels containing an object of interest; and (5) protein structure prediction. Lastly, we discuss extensions to uncertainty quantification for ranking, metric learning and distributionally robust learning.

READ FULL TEXT

page 2

page 14

page 16

page 17

page 18

research
07/15/2021

A Gentle Introduction to Conformal Prediction and Distribution-Free Uncertainty Quantification

Black-box machine learning learning methods are now routinely used in hi...
research
10/18/2022

Uncertainty in Extreme Multi-label Classification

Uncertainty quantification is one of the most crucial tasks to obtain tr...
research
10/03/2021

Learn then Test: Calibrating Predictive Algorithms to Achieve Risk Control

We introduce Learn then Test, a framework for calibrating machine learni...
research
03/04/2021

Distribution-free uncertainty quantification for classification under label shift

Trustworthy deployment of ML models requires a proper measure of uncerta...
research
06/12/2023

On the Expected Size of Conformal Prediction Sets

While conformal predictors reap the benefits of rigorous statistical gua...
research
02/07/2023

How to Trust Your Diffusion Model: A Convex Optimization Approach to Conformal Risk Control

Score-based generative modeling, informally referred to as diffusion mod...
research
02/11/2021

Private Prediction Sets

In real-world settings involving consequential decision-making, the depl...

Please sign up or login with your details

Forgot password? Click here to reset