On Shapley Credit Allocation for Interpretability

12/10/2020
by   Debraj Basu, et al.
0

We emphasize the importance of asking the right question when interpreting the decisions of a learning model. We discuss a natural extension of the theoretical machinery from Janzing et. al. 2020, which answers the question "Why did my model predict a person has cancer?" for answering a more involved question, "What caused my model to predict a person has cancer?" While the former quantifies the direct effects of variables on the model, the latter also accounts for indirect effects, thereby providing meaningful insights wherever human beings can reason in terms of cause and effect. We propose three broad categories for interpretations: observational, model-specific and causal each of which are significant in their own right. Furthermore, this paper quantifies feature relevance by weaving different natures of interpretations together with different measures as characteristic functions for Shapley symmetrization. Besides the widely used expected value of the model, we also discuss measures of statistical uncertainty and dispersion as informative candidates, and their merits in generating explanations for each data point, some of which are used in this context for the first time. These measures are not only useful for studying the influence of variables on the model output, but also on the predictive performance of the model, and for that we propose relevant characteristic functions that are also used for the first time.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/17/2018

Functional Mediation Analysis with an Application to Functional Magnetic Resonance Imaging Data

Causal mediation analysis is widely utilized to separate the causal effe...
research
12/22/2019

Direct and Indirect Effects – An Information Theoretic Perspective

Information theoretic (IT) approaches to quantifying causal influences h...
research
06/02/2023

Alternative Measures of Direct and Indirect Effects

There are a number of measures of direct and indirect effects in the lit...
research
10/27/2019

CXPlain: Causal Explanations for Model Interpretation under Uncertainty

Feature importance estimates that inform users about the degree to which...
research
06/18/2019

Model Explanations under Calibration

Explaining and interpreting the decisions of recommender systems are bec...
research
06/15/2021

Decomposition of Global Feature Importance into Direct and Associative Components (DEDACT)

Global model-agnostic feature importance measures either quantify whethe...
research
02/06/2018

Granger-causal Attentive Mixtures of Experts

Several methods have recently been proposed to detect salient input feat...

Please sign up or login with your details

Forgot password? Click here to reset