Counterfactual Explanation Algorithms for Behavioral and Textual Data

12/04/2019
by   Yanou Ramon, et al.
0

We study the interpretability of predictive systems that use high-dimensonal behavioral and textual data. Examples include predicting product interest based on online browsing data and detecting spam emails or objectionable web content. Recently, counterfactual explanations have been proposed for generating insight into model predictions, which focus on what is relevant to a particular instance. Conducting a complete search to compute counterfactuals is very time-consuming because of the huge dimensionality. To our knowledge, for behavioral and text data, only one model-agnostic heuristic algorithm (SEDC) for finding counterfactual explanations has been proposed in the literature. However, there may be better algorithms for finding counterfactuals quickly. This study aligns the recently proposed Linear Interpretable Model-agnostic Explainer (LIME) and Shapley Additive Explanations (SHAP) with the notion of counterfactual explanations, and empirically benchmarks their effectiveness and efficiency against SEDC using a collection of 13 data sets. Results show that LIME-Counterfactual (LIME-C) and SHAP-Counterfactual (SHAP-C) have low and stable computation times, but mostly, they are less efficient than SEDC. However, for certain instances on certain data sets, SEDC's run time is comparably large. With regard to effectiveness, LIME-C and SHAP-C find reasonable, if not always optimal, counterfactual explanations. SHAP-C, however, seems to have difficulties with highly unbalanced data. Because of its good overall performance, LIME-C seems to be a favorable alternative to SEDC, which failed for some nonlinear models to find counterfactuals because of the particular heuristic search algorithm it uses. A main upshot of this paper is that there is a good deal of room for further research. For example, we propose algorithmic adjustments that are direct upshots of the paper's findings.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/03/2019

Interpretable Counterfactual Explanations Guided by Prototypes

We propose a fast, model agnostic method for finding interpretable count...
research
04/24/2023

TIGTEC : Token Importance Guided TExt Counterfactuals

Counterfactual examples explain a prediction by highlighting changes of ...
research
11/01/2022

Anytime Generation of Counterfactual Explanations for Text Classification

In many machine learning applications, it is important for the user to u...
research
06/29/2021

Counterfactual Explanations for Arbitrary Regression Models

We present a new method for counterfactual explanations (CFEs) based on ...
research
07/19/2021

Interpreting Process Predictions using a Milestone-Aware Counterfactual Approach

Predictive process analytics often apply machine learning to predict the...
research
08/02/2021

CARLA: A Python Library to Benchmark Algorithmic Recourse and Counterfactual Explanation Algorithms

Counterfactual explanations provide means for prescriptive model explana...
research
04/15/2021

NICE: An Algorithm for Nearest Instance Counterfactual Explanations

In this paper we suggest NICE: a new algorithm to generate counterfactua...

Please sign up or login with your details

Forgot password? Click here to reset