Generalized Adversarial Distances to Efficiently Discover Classifier Errors

02/25/2021
by   Walter Bennette, et al.
0

Given a black-box classification model and an unlabeled evaluation dataset from some application domain, efficient strategies need to be developed to evaluate the model. Random sampling allows a user to estimate metrics like accuracy, precision, and recall, but may not provide insight to high-confidence errors. High-confidence errors are rare events for which the model is highly confident in its prediction, but is wrong. Such errors can represent costly mistakes and should be explicitly searched for. In this paper we propose a generalization to the Adversarial Distance search that leverages concepts from adversarial machine learning to identify predictions for which a classifier may be overly confident. These predictions are useful instances to sample when looking for high-confidence errors because they are prone to a higher rate of error than expected. Our generalization allows Adversarial Distance to be applied to any classifier or data domain. Experimental results show that the generalized method finds errors at rates greater than expected given the confidence of the sampled predictions, and outperforms competing methods.

READ FULL TEXT
research
06/29/2020

Harnessing Adversarial Distances to Discover High-Confidence Errors

Given a deep neural network image classification model that we treat as ...
research
02/11/2021

Defuse: Harnessing Unrestricted Adversarial Examples for Debugging Models Beyond Test Accuracy

We typically compute aggregate statistics on held-out test data to asses...
research
04/17/2017

Adversarial and Clean Data Are Not Twins

Adversarial attack has cast a shadow on the massive success of deep neur...
research
05/10/2021

Accountable Error Characterization

Customers of machine learning systems demand accountability from the com...
research
09/10/2019

The Prevalence of Errors in Machine Learning Experiments

Context: Conducting experiments is central to research machine learning ...
research
04/09/2018

G-Distillation: Reducing Overconfident Errors on Novel Samples

Counter to the intuition that unfamiliarity should lead to lack of confi...
research
06/30/2020

Classification Confidence Estimation with Test-Time Data-Augmentation

Machine learning plays an increasingly significant role in many aspects ...

Please sign up or login with your details

Forgot password? Click here to reset