Evaluating and Aggregating Feature-based Model Explanations

05/01/2020
by   Umang Bhatt, et al.
14

A feature-based model explanation denotes how much each input feature contributes to a model's output for a given data point. As the number of proposed explanation functions grows, we lack quantitative evaluation criteria to help practitioners know when to use which explanation function. This paper proposes quantitative evaluation criteria for feature-based explanations: low sensitivity, high faithfulness, and low complexity. We devise a framework for aggregating explanation functions. We develop a procedure for learning an aggregate explanation function with lower complexity and then derive a new aggregate Shapley value explanation function that minimizes sensitivity.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/04/2022

Do Explanations Explain? Model Knows Best

It is a mystery which input features contribute to a neural network's ou...
research
02/14/2019

Which is the least complex explanation? Abduction and complexity

It may happen that for a certain abductive problem there are several pos...
research
06/05/2019

Evaluating Explainers via Perturbation

Due to high complexity of many modern machine learning models such as de...
research
01/27/2019

How Sensitive are Sensitivity-Based Explanations?

We propose a simple objective evaluation measure for explanations of a c...
research
06/21/2023

Evaluating the overall sensitivity of saliency-based explanation methods

We address the need to generate faithful explanations of "black box" Dee...
research
01/20/2019

Towards Aggregating Weighted Feature Attributions

Current approaches for explaining machine learning models fall into two ...
research
05/26/2022

Explaining Preferences with Shapley Values

While preference modelling is becoming one of the pillars of machine lea...

Please sign up or login with your details

Forgot password? Click here to reset