From Shapley Values to Generalized Additive Models and back

09/08/2022
by   Sebastian Bordt, et al.
0

In explainable machine learning, local post-hoc explanation algorithms and inherently interpretable models are often seen as competing approaches. In this work, offer a novel perspective on Shapley Values, a prominent post-hoc explanation technique, and show that it is strongly connected with Glassbox-GAMs, a popular class of interpretable models. We introduce n-Shapley Values, a natural extension of Shapley Values that explain individual predictions with interaction terms up to order n. As n increases, the n-Shapley Values converge towards the Shapley-GAM, a uniquely determined decomposition of the original function. From the Shapley-GAM, we can compute Shapley Values of arbitrary order, which gives precise insights into the limitations of these explanations. We then show that Shapley Values recover generalized additive models of order n, assuming that we allow for interaction terms up to order n in the explanations. This implies that the original Shapley Values recover Glassbox-GAMs. At the technical end, we show that there is a one-to-one correspondence between different ways to choose the value function and different functional decompositions of the original function. This provides a novel perspective on the question of how to choose the value function. We also present an empirical analysis of the degree of variable interaction that is present in various standard classifiers, and discuss the implications of our results for algorithmic explanations. A python package to compute n-Shapley Values and replicate the results in this paper is available at <https://github.com/tml-tuebingen/nshap>.

READ FULL TEXT

page 25

page 28

page 30

page 31

research
04/06/2021

Shapley Explanation Networks

Shapley values have become one of the most popular feature attribution e...
research
08/12/2022

Unifying local and global model explanations by functional decomposition of low dimensional structures

We consider a global explanation of a regression or classification funct...
research
02/03/2022

The Disagreement Problem in Explainable Machine Learning: A Practitioner's Perspective

As various post hoc explanation methods are increasingly being leveraged...
research
02/15/2023

Streamlining models with explanations in the learning loop

Several explainable AI methods allow a Machine Learning user to get insi...
research
03/02/2022

Faith-Shap: The Faithful Shapley Interaction Index

Shapley values, which were originally designed to assign attributions to...
research
06/02/2022

Which Explanation Should I Choose? A Function Approximation Perspective to Characterizing Post hoc Explanations

Despite the plethora of post hoc model explanation methods, the basic pr...
research
03/06/2013

Relevant Explanations: Allowing Disjunctive Assignments

Relevance-based explanation is a scheme in which partial assignments to ...

Please sign up or login with your details

Forgot password? Click here to reset