The many Shapley values for model explanation

08/22/2019
by   Mukund Sundararajan, et al.
0

The Shapley value has become a popular method to attribute the prediction of a machine-learning model on an input to its base features. The Shapley value [1] is known to be the unique method that satisfies certain desirable properties, and this motivates its use. Unfortunately, despite this uniqueness result, there are a multiplicity of Shapley values used in explaining a model's prediction. This is because there are many ways to apply the Shapley value that differ in how they reference the model, the training data, and the explanation context. In this paper, we study an approach that applies the Shapley value to conditional expectations (CES) of sets of features (cf. [2]) that subsumes several prior approaches within a common framework. We provide the first algorithm for the general version of CES. We show that CES can result in counterintuitive attributions in theory and in practice (we study a diabetes prediction task); for instance, CES can assign non-zero attributions to features that are not referenced by the model. In contrast, we show that an approach called the Baseline Shapley (BS) does not exhibit counterintuitive attributions; we support this claim with a uniqueness (axiomatic) result. We show that BS is a special case of CES, and CES with an independent feature distribution coincides with a randomized version of BS. Thus, BS fits into the CES framework, but does not suffer from many of CES's deficiencies.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/17/2019

The Explanation Game: Explaining Machine Learning Models with Cooperative Game Theory

Recently, a number of techniques have been proposed to explain a machine...
research
06/19/2023

Explaining the Model and Feature Dependencies by Decomposition of the Shapley Value

Shapley values have become one of the go-to methods to explain complex m...
research
07/23/2021

Joint Shapley values: a measure of joint feature importance

The Shapley value is one of the most widely used model-agnostic measures...
research
05/26/2022

Explaining Preferences with Shapley Values

While preference modelling is becoming one of the pillars of machine lea...
research
01/29/2021

Addendum to "Tilings problems on Baumslag-Solitar groups"

In our article in MCU'2013 we state the the Domino problem is undecidabl...
research
07/12/2020

Explaining the data or explaining a model? Shapley values that uncover non-linear dependencies

Shapley values have become increasingly popular in the machine learning ...
research
01/24/2020

Valid distribution-free inferential models for prediction

A fundamental problem in statistics and machine learning is that of usin...

Please sign up or login with your details

Forgot password? Click here to reset