Multicollinearity Correction and Combined Feature Effect in Shapley Values

11/03/2020
by   Indranil Basu, et al.
0

Model interpretability is one of the most intriguing problems in most of the Machine Learning models, particularly for those that are mathematically sophisticated. Computing Shapley Values are arguably the best approach so far to find the importance of each feature in a model, at the row level. In other words, Shapley values represent the importance of a feature for a particular row, especially for Classification or Regression problems. One of the biggest limitations of Shapley vales is that, Shapley value calculations assume all the features are uncorrelated (independent of each other), this assumption is often incorrect. To address this problem, we present a unified framework to calculate Shapley values with correlated features. To be more specific, we do an adjustment (Matrix formulation) of the features while calculating Independent Shapley values for the rows. Moreover, we have given a Mathematical proof against the said adjustments. With these adjustments, Shapley values (Importance) for the features become independent of the correlations existing between them. We have also enhanced this adjustment concept for more than features. As the Shapley values are additive, to calculate combined effect of two features, we just have to add their individual Shapley values. This is again not right if one or more of the features (used in the combination) are correlated with the other features (not in the combination). We have addressed this problem also by extending the correlation adjustment for one feature to multiple features in the said combination for which Shapley values are determined. Our implementation of this method proves that our method is computationally efficient also, compared to original Shapley method.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/23/2021

Joint Shapley values: a measure of joint feature importance

The Shapley value is one of the most widely used model-agnostic measures...
research
04/23/2021

Grouped Feature Importance and Combined Features Effect Plot

Interpretable machine learning has become a very active area of research...
research
03/10/2023

Feature Importance: A Closer Look at Shapley Values and LOCO

There is much interest lately in explainability in statistics and machin...
research
07/28/2023

Confident Feature Ranking

Interpretation of feature importance values often relies on the relative...
research
06/17/2020

Using Weighted P-Values in Fisher's Method

Fisher's method prescribes a way to combine p-values from multiple exper...
research
05/23/2019

Computationally Efficient Feature Significance and Importance for Machine Learning Models

We develop a simple and computationally efficient significance test for ...

Please sign up or login with your details

Forgot password? Click here to reset