Evaluating the overall sensitivity of saliency-based explanation methods

06/21/2023
by   Harshinee Sriram, et al.
0

We address the need to generate faithful explanations of "black box" Deep Learning models. Several tests have been proposed to determine aspects of faithfulness of explanation methods, but they lack cross-domain applicability and a rigorous methodology. Hence, we select an existing test that is model agnostic and is well-suited for comparing one aspect of faithfulness (i.e., sensitivity) of multiple explanation methods, and extend it by specifying formal thresh-olds and building criteria to determine the over-all sensitivity of the explanation method. We present examples of how multiple explanation methods for Convolutional Neural Networks can be compared using this extended methodology. Finally, we discuss the relationship between sensitivity and faithfulness and consider how the test can be adapted to assess different explanation methods in other domains.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/27/2019

How Sensitive are Sensitivity-Based Explanations?

We propose a simple objective evaluation measure for explanations of a c...
research
05/01/2020

Evaluating and Aggregating Feature-based Model Explanations

A feature-based model explanation denotes how much each input feature co...
research
06/04/2023

Sanity Checks for Saliency Methods Explaining Object Detectors

Saliency methods are frequently used to explain Deep Neural Network-base...
research
08/29/2019

Human-grounded Evaluations of Explanation Methods for Text Classification

Due to the black-box nature of deep learning models, methods for explain...
research
02/01/2022

Framework for Evaluating Faithfulness of Local Explanations

We study the faithfulness of an explanation system to the underlying pre...
research
11/15/2022

Easy to Decide, Hard to Agree: Reducing Disagreements Between Saliency Methods

A popular approach to unveiling the black box of neural NLP models is to...
research
06/11/2018

A Note about: Local Explanation Methods for Deep Neural Networks lack Sensitivity to Parameter Values

Local explanation methods, also known as attribution methods, attribute ...

Please sign up or login with your details

Forgot password? Click here to reset