WeightedSHAP: analyzing and improving Shapley based feature attributions

09/27/2022
by   Yongchan Kwon, et al.
0

Shapley value is a popular approach for measuring the influence of individual features. While Shapley feature attribution is built upon desiderata from game theory, some of its constraints may be less natural in certain machine learning settings, leading to unintuitive model interpretation. In particular, the Shapley value uses the same weight for all marginal contributions – i.e. it gives the same importance when a large number of other features are given versus when a small number of other features are given. This property can be problematic if larger feature sets are more or less informative than smaller feature sets. Our work performs a rigorous analysis of the potential limitations of Shapley feature attribution. We identify simple settings where the Shapley value is mathematically suboptimal by assigning larger attributions for less influential features. Motivated by this observation, we propose WeightedSHAP, which generalizes the Shapley value and learns which marginal contributions to focus directly from data. On several real-world datasets, we demonstrate that the influential features identified by WeightedSHAP are better able to recapitulate the model's predictions compared to the features identified by the Shapley value.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/22/2021

Interpreting Deep Learning Models with Marginal Attribution by Conditioning on Quantiles

A vastly growing literature on explaining deep learning models has emerg...
research
07/04/2023

Shapley Sets: Feature Attribution via Recursive Function Decomposition

Despite their ubiquitous use, Shapley value feature attributions can be ...
research
02/16/2023

On marginal feature attributions of tree-based models

Due to their power and ease of use, tree-based machine learning models h...
research
01/20/2019

Towards Aggregating Weighted Feature Attributions

Current approaches for explaining machine learning models fall into two ...
research
04/19/2021

Improving Attribution Methods by Learning Submodular Functions

This work explores the novel idea of learning a submodular scoring funct...
research
08/08/2018

L-Shapley and C-Shapley: Efficient Model Interpretation for Structured Data

We study instancewise feature importance scoring as a method for model i...
research
04/06/2023

Efficient SAGE Estimation via Causal Structure Learning

The Shapley Additive Global Importance (SAGE) value is a theoretically a...

Please sign up or login with your details

Forgot password? Click here to reset