RKHS-SHAP: Shapley Values for Kernel Methods

10/18/2021
by   Siu Lun Chau, et al.
0

Feature attribution for kernel methods is often heuristic and not individualised for each prediction. To address this, we turn to the concept of Shapley values, a coalition game theoretical framework that has previously been applied to different machine learning model interpretation tasks, such as linear models, tree ensembles and deep networks. By analysing Shapley values from a functional perspective, we propose RKHS-SHAP, an attribution method for kernel machines that can efficiently compute both Interventional and Observational Shapley values using kernel mean embeddings of distributions. We show theoretically that our method is robust with respect to local perturbations - a key yet often overlooked desideratum for interpretability. Further, we propose Shapley regulariser, applicable to a general empirical risk minimisation framework, allowing learning while controlling the level of specific feature's contributions to the model. We demonstrate that the Shapley regulariser enables learning which is robust to covariate shift of a given feature and fair learning which controls the Shapley values of sensitive features.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/16/2021

Exact Shapley Values for Local and Model-True Explanations of Decision Tree Ensembles

Additive feature explanations using Shapley values have become popular f...
research
06/29/2020

True to the Model or True to the Data?

A variety of recent papers discuss the application of Shapley values, a ...
research
02/12/2018

Consistent Individualized Feature Attribution for Tree Ensembles

Interpreting predictions from tree ensemble methods such as gradient boo...
research
06/19/2017

Consistent feature attribution for tree ensembles

It is critical in many applications to understand what features are impo...
research
07/04/2019

Fair Kernel Regression via Fair Feature Embedding in Kernel Space

In recent years, there have been significant efforts on mitigating uneth...
research
06/12/2023

On the Robustness of Removal-Based Feature Attributions

To explain complex models based on their inputs, many feature attributio...
research
11/03/2022

A k-additive Choquet integral-based approach to approximate the SHAP values for local interpretability in machine learning

Besides accuracy, recent studies on machine learning models have been ad...

Please sign up or login with your details

Forgot password? Click here to reset