A Series of Unfortunate Counterfactual Events: the Role of Time in Counterfactual Explanations

10/09/2020
by   Andrea Ferrario, et al.
0

Counterfactual explanations are a prominent example of post-hoc interpretability methods in the explainable Artificial Intelligence research domain. They provide individuals with alternative scenarios and a set of recommendations to achieve a sought-after machine learning model outcome. Recently, the literature has identified desiderata of counterfactual explanations, such as feasibility, actionability and sparsity that should support their applicability in real-world contexts. However, we show that the literature has neglected the problem of the time dependency of counterfactual explanations. We argue that, due to their time dependency and because of the provision of recommendations, even feasible, actionable and sparse counterfactual explanations may not be appropriate in real-world applications. This is due to the possible emergence of what we call "unfortunate counterfactual events." These events may occur due to the retraining of machine learning models whose outcomes have to be explained via counterfactual explanation. Series of unfortunate counterfactual events frustrate the efforts of those individuals who successfully implemented the recommendations of counterfactual explanations. This negatively affects people's trust in the ability of institutions to provide machine learning-supported decisions consistently. We introduce an approach to address the problem of the emergence of unfortunate counterfactual events that makes use of histories of counterfactual explanations. In the final part of the paper we propose an ethical analysis of two distinct strategies to cope with the challenge of unfortunate counterfactual events. We show that they respond to an ethically responsible imperative to preserve the trustworthiness of credit lending organizations, the decision models they employ, and the social-economic function of credit lending.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/19/2019

Explaining Machine Learning Classifiers through Diverse Counterfactual Explanations

Post-hoc explanations of machine learning models are crucial for people ...
research
11/08/2022

Motif-guided Time Series Counterfactual Explanations

With the rising need of interpretable machine learning methods, there is...
research
01/06/2023

Evaluating counterfactual explanations using Pearl's counterfactual method

Counterfactual explanations (CEs) are methods for generating an alternat...
research
12/08/2022

Real-Time Counterfactual Explanations For Robotic Systems With Multiple Continuous Outputs

Although many machine learning methods, especially from the field of dee...
research
12/10/2019

The Hidden Assumptions Behind Counterfactual Explanations and Principal Reasons

Counterfactual explanations are gaining prominence within technical, leg...
research
09/29/2022

Rethinking Counterfactual Explanations as Local and Regional Counterfactual Policies

Among the challenges not yet resolved for Counterfactual Explanations (C...
research
02/03/2021

Directive Explanations for Actionable Explainability in Machine Learning Applications

This paper investigates the prospects of using directive explanations to...

Please sign up or login with your details

Forgot password? Click here to reset