How Many and Which Training Points Would Need to be Removed to Flip this Prediction?

02/04/2023
by   Jinghan Yang, et al.
0

We consider the problem of identifying a minimal subset of training data 𝒮_t such that if the instances comprising 𝒮_t had been removed prior to training, the categorization of a given test point x_t would have been different. Identifying such a set may be of interest for a few reasons. First, the cardinality of 𝒮_t provides a measure of robustness (if |𝒮_t| is small for x_t, we might be less confident in the corresponding prediction), which we show is correlated with but complementary to predicted probabilities. Second, interrogation of 𝒮_t may provide a novel mechanism for contesting a particular model prediction: If one can make the case that the points in 𝒮_t are wrongly labeled or irrelevant, this may argue for overturning the associated prediction. Identifying 𝒮_t via brute-force is intractable. We propose comparatively fast approximation methods to find 𝒮_t based on influence functions, and find that – for simple convex text classification models – these approaches can often successfully identify relatively small sets of training examples which, if removed, would flip the prediction.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/22/2023

Relabel Minimal Training Subset to Flip a Prediction

Yang et al. (2023) discovered that removing a mere 1 often lead to the f...
research
02/19/2020

Estimating Training Data Influence by Tracking Gradient Descent

We introduce a method called TrackIn that computes the influence of a tr...
research
11/23/2018

Representer Point Selection for Explaining Deep Neural Networks

We propose to explain the predictions of a deep neural network, by point...
research
07/13/2021

DIVINE: Diverse Influential Training Points for Data Visualization and Model Refinement

As the complexity of machine learning (ML) models increases, resulting i...
research
02/04/2020

Iterative Data Programming for Expanding Text Classification Corpora

Real-world text classification tasks often require many labeled training...
research
12/11/2020

When is Memorization of Irrelevant Training Data Necessary for High-Accuracy Learning?

Modern machine learning models are complex and frequently encode surpris...
research
12/03/2019

Less Is Better: Unweighted Data Subsampling via Influence Function

In the time of Big Data, training complex models on large-scale data set...

Please sign up or login with your details

Forgot password? Click here to reset