PAC Prediction Sets Under Covariate Shift

06/17/2021
by   Sangdon Park, et al.
0

An important challenge facing modern machine learning is how to rigorously quantify the uncertainty of model predictions. Conveying uncertainty is especially important when there are changes to the underlying data distribution that might invalidate the predictive model. Yet, most existing uncertainty quantification algorithms break down in the presence of such shifts. We propose a novel approach that addresses this challenge by constructing probably approximately correct (PAC) prediction sets in the presence of covariate shift. Our approach focuses on the setting where there is a covariate shift from the source distribution (where we have labeled training examples) to the target distribution (for which we want to quantify uncertainty). Our algorithm assumes given importance weights that encode how the probabilities of the training examples change under the covariate shift. In practice, importance weights typically need to be estimated; thus, we extend our algorithm to the setting where we are given confidence intervals for the importance weights rather than their true value. We demonstrate the effectiveness of our approach on various covariate shifts designed based on the DomainNet and ImageNet datasets.

READ FULL TEXT

page 25

page 26

page 27

research
03/11/2022

Distribution-free Prediction Sets Adaptive to Unknown Covariate Shift

Predicting sets of outcomes – instead of unique outcomes – is a promisin...
research
02/29/2020

Calibrated Prediction with Covariate Shift via Unsupervised Domain Adaptation

Reliable uncertainty estimates are an important tool for helping autonom...
research
06/26/2020

Unlabelled Data Improves Bayesian Uncertainty Calibration under Covariate Shift

Modern neural networks have proven to be powerful function approximators...
research
07/10/2023

Automatic Debiased Machine Learning for Covariate Shifts

In this paper we address the problem of bias in machine learning of para...
research
06/21/2021

Stratified Learning: a general-purpose statistical method for improved learning under Covariate Shift

Covariate shift arises when the labelled training (source) data is not r...
research
10/28/2020

Evaluating Model Robustness to Dataset Shift

As the use of machine learning in safety-critical domains becomes widesp...
research
02/27/2023

Statistical Learning under Heterogenous Distribution Shift

This paper studies the prediction of a target 𝐳 from a pair of random va...

Please sign up or login with your details

Forgot password? Click here to reset