Interpreting Deep Learning Models with Marginal Attribution by Conditioning on Quantiles

03/22/2021
by   M. Merz, et al.
0

A vastly growing literature on explaining deep learning models has emerged. This paper contributes to that literature by introducing a global gradient-based model-agnostic method, which we call Marginal Attribution by Conditioning on Quantiles (MACQ). Our approach is based on analyzing the marginal attribution of predictions (outputs) to individual features (inputs). Specificalllly, we consider variable importance by mixing (global) output levels and, thus, explain how features marginally contribute across different regions of the prediction space. Hence, MACQ can be seen as a marginal attribution counterpart to approaches such as accumulated local effects (ALE), which study the sensitivities of outputs by perturbing inputs. Furthermore, MACQ allows us to separate marginal attribution of individual features from interaction effect, and visually illustrate the 3-way relationship between marginal attribution, output level, and feature value.

READ FULL TEXT
research
04/13/2022

Baseline Computation for Attribution Methods Based on Interpolated Inputs

We discuss a way to find a well behaved baseline for attribution methods...
research
09/27/2022

WeightedSHAP: analyzing and improving Shapley based feature attributions

Shapley value is a popular approach for measuring the influence of indiv...
research
04/27/2019

Working women and caste in India: A study of social disadvantage using feature attribution

Women belonging to the socially disadvantaged caste-groups in India have...
research
02/16/2023

On marginal feature attributions of tree-based models

Due to their power and ease of use, tree-based machine learning models h...
research
05/04/2023

Distributing Synergy Functions: Unifying Game-Theoretic Interaction Methods for Machine-Learning Explainability

Deep learning has revolutionized many areas of machine learning, from co...
research
04/10/2023

Gradient-based Uncertainty Attribution for Explainable Bayesian Deep Learning

Predictions made by deep learning models are prone to data perturbations...
research
06/14/2022

Machines Explaining Linear Programs

There has been a recent push in making machine learning models more inte...

Please sign up or login with your details

Forgot password? Click here to reset