The privacy issue of counterfactual explanations: explanation linkage attacks

10/21/2022
by   Sofie Goethals, et al.
0

Black-box machine learning models are being used in more and more high-stakes domains, which creates a growing need for Explainable AI (XAI). Unfortunately, the use of XAI in machine learning introduces new privacy risks, which currently remain largely unnoticed. We introduce the explanation linkage attack, which can occur when deploying instance-based strategies to find counterfactual explanations. To counter such an attack, we propose k-anonymous counterfactual explanations and introduce pureness as a new metric to evaluate the validity of these k-anonymous counterfactual explanations. Our results show that making the explanations, rather than the whole dataset, k- anonymous, is beneficial for the quality of the explanations.

READ FULL TEXT

page 8

page 13

research
11/15/2019

On the computation of counterfactual explanations – A survey

Due to the increasing use of machine learning in practice it becomes mor...
research
12/16/2022

Counterfactual Explanations for Misclassified Images: How Human and Machine Explanations Differ

Counterfactual explanations have emerged as a popular solution for the e...
research
04/25/2023

Disagreement amongst counterfactual explanations: How transparency can be deceptive

Counterfactual explanations are increasingly used as an Explainable Arti...
research
02/26/2021

If Only We Had Better Counterfactual Explanations: Five Key Deficits to Rectify in the Evaluation of Counterfactual XAI Techniques

In recent years, there has been an explosion of AI research on counterfa...
research
01/31/2022

Causal Explanations and XAI

Although standard Machine Learning models are optimized for making predi...
research
01/21/2020

Adequate and fair explanations

Explaining sophisticated machine-learning based systems is an important ...
research
05/04/2020

Evaluating Explainable AI: Which Algorithmic Explanations Help Users Predict Model Behavior?

Algorithmic approaches to interpreting machine learning models have prol...

Please sign up or login with your details

Forgot password? Click here to reset