Counterfactual Explanations for Misclassified Images: How Human and Machine Explanations Differ

12/16/2022
by   Eoin Delaney, et al.
0

Counterfactual explanations have emerged as a popular solution for the eXplainable AI (XAI) problem of elucidating the predictions of black-box deep-learning systems due to their psychological validity, flexibility across problem domains and proposed legal compliance. While over 100 counterfactual methods exist, claiming to generate plausible explanations akin to those preferred by people, few have actually been tested on users (∼7%). So, the psychological validity of these counterfactual algorithms for effective XAI for image data is not established. This issue is addressed here using a novel methodology that (i) gathers ground truth human-generated counterfactual explanations for misclassified images, in two user studies and, then, (ii) compares these human-generated ground-truth explanations to computationally-generated explanations for the same misclassifications. Results indicate that humans do not "minimally edit" images when generating counterfactual explanations. Instead, they make larger, "meaningful" edits that better approximate prototypes in the counterfactual class.

READ FULL TEXT

page 15

page 18

page 26

page 32

research
10/21/2022

The privacy issue of counterfactual explanations: explanation linkage attacks

Black-box machine learning models are being used in more and more high-s...
research
09/10/2020

On Generating Plausible Counterfactual and Semi-Factual Explanations for Deep Learning

There is a growing concern that the recent progress made in AI, especial...
research
06/21/2019

Generating Counterfactual and Contrastive Explanations using SHAP

With the advent of GDPR, the domain of explainable AI and model interpre...
research
05/26/2023

Counterfactuals of Counterfactuals: a back-translation-inspired approach to analyse counterfactual editors

In the wake of responsible AI, interpretability methods, which attempt t...
research
07/20/2021

Uncertainty Estimation and Out-of-Distribution Detection for Counterfactual Explanations: Pitfalls and Solutions

Whilst an abundance of techniques have recently been proposed to generat...
research
09/28/2022

Causal Proxy Models for Concept-Based Model Explanations

Explainability methods for NLP systems encounter a version of the fundam...
research
09/28/2020

Instance-Based Counterfactual Explanations for Time Series Classification

In recent years there has been a cascade of research in attempting to ma...

Please sign up or login with your details

Forgot password? Click here to reset