Measurable Counterfactual Local Explanations for Any Classifier

08/08/2019
by   Adam White, et al.
2

We propose a novel method for explaining the predictions of any classifier. In our approach, local explanations are expected to explain both the outcome of a prediction and how that prediction would change if 'things had been different'. Furthermore, we argue that satisfactory explanations cannot be dissociated from a notion and measure of fidelity, as advocated in the early days of neural networks' knowledge extraction. We introduce a definition of fidelity to the underlying classifier for local explanation models which is based on distances to a target decision boundary. A system called CLEAR: Counterfactual Local Explanations via Regression, is introduced and evaluated. CLEAR generates w-counterfactual explanations that state minimum changes necessary to flip a prediction's classification. CLEAR then builds local regression models, using the w-counterfactuals to measure and improve the fidelity of its regressions. By contrast, the popular LIME method, which also uses regression to generate local explanations, neither measures its own fidelity nor generates counterfactuals. CLEAR's regressions are found to have significantly higher fidelity than LIME's, averaging over 45 paper's four case studies.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/20/2022

A Symbolic Approach for Counterfactual Explanations

In this paper titled A Symbolic Approach for Counterfactual Explanations...
research
06/29/2021

Counterfactual Explanations for Arbitrary Regression Models

We present a new method for counterfactual explanations (CFEs) based on ...
research
10/16/2022

CLEAR: Generative Counterfactual Explanations on Graphs

Counterfactual explanations promote explainability in machine learning m...
research
06/28/2021

Contrastive Counterfactual Visual Explanations With Overdetermination

A novel explainable AI method called CLEAR Image is introduced in this p...
research
05/13/2022

DualCF: Efficient Model Extraction Attack from Counterfactual Explanations

Cloud service providers have launched Machine-Learning-as-a-Service (MLa...
research
07/02/2022

PhilaeX: Explaining the Failure and Success of AI Models in Malware Detection

The explanation to an AI model's prediction used to support decision mak...
research
04/16/2020

SCOUT: Self-aware Discriminant Counterfactual Explanations

The problem of counterfactual visual explanations is considered. A new f...

Please sign up or login with your details

Forgot password? Click here to reset