Counterfactually Comparing Abstaining Classifiers

05/17/2023
by   Yo Joong Choe, et al.
0

Abstaining classifiers have the option to abstain from making predictions on inputs that they are unsure about. These classifiers are becoming increasingly popular in high-stake decision-making problems, as they can withhold uncertain predictions to improve their reliability and safety. When evaluating black-box abstaining classifier(s), however, we lack a principled approach that accounts for what the classifier would have predicted on its abstentions. These missing predictions are crucial when, e.g., a radiologist is unsure of their diagnosis or when a driver is inattentive in a self-driving car. In this paper, we introduce a novel approach and perspective to the problem of evaluating and comparing abstaining classifiers by treating abstentions as missing data. Our evaluation approach is centered around defining the counterfactual score of an abstaining classifier, defined as the expected performance of the classifier had it not been allowed to abstain. We specify the conditions under which the counterfactual score is identifiable: if the abstentions are stochastic, and if the evaluation data is independent of the training data (ensuring that the predictions are missing at random), then the score is identifiable. Note that, if abstentions are deterministic, then the score is unidentifiable because the classifier can perform arbitrarily poorly on its abstentions. Leveraging tools from observational causal inference, we then develop nonparametric and doubly robust methods to efficiently estimate this quantity under identification. Our approach is examined in both simulated and real data experiments.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/11/2022

Causal and counterfactual views of missing data models

It is often said that the fundamental problem of causal inference is a m...
research
12/18/2020

Robustness to Spurious Correlations in Text Classification via Automatically Generated Counterfactuals

Spurious correlations threaten the validity of statistical classifiers. ...
research
10/15/2020

Double Robust Representation Learning for Counterfactual Prediction

Causal inference, or counterfactual prediction, is central to decision m...
research
06/24/2022

Analyzing the Effects of Classifier Lipschitzness on Explainers

Machine learning methods are getting increasingly better at making predi...
research
10/27/2021

VACA: Design of Variational Graph Autoencoders for Interventional and Counterfactual Queries

In this paper, we introduce VACA, a novel class of variational graph aut...
research
06/28/2022

Flexible text generation for counterfactual fairness probing

A common approach for testing fairness issues in text-based classifiers ...

Please sign up or login with your details

Forgot password? Click here to reset