New Definitions and Evaluations for Saliency Methods: Staying Intrinsic, Complete and Sound

11/05/2022
by   Arushi Gupta, et al.
0

Saliency methods compute heat maps that highlight portions of an input that were most important for the label assigned to it by a deep net. Evaluations of saliency methods convert this heat map into a new masked input by retaining the k highest-ranked pixels of the original input and replacing the rest with “uninformative”pixels, and checking if the net's output is mostly unchanged. This is usually seen as an explanation of the output, but the current paper highlights reasons why this inference of causality may be suspect. Inspired by logic concepts of completeness & soundness, it observes that the above type of evaluation focuses on completeness of the explanation, but ignores soundness. New evaluation metrics are introduced to capture both notions, while staying in an intrinsic framework – i.e., using the dataset and the net, but no separately trained nets, human evaluations, etc. A simple saliency method is described that matches or outperforms prior methods in the evaluations. Experiments also suggest new intrinsic justifications, based on soundness, for popular heuristic tricks such as TV regularization and upsampling.

READ FULL TEXT

page 15

page 17

page 18

page 21

page 22

page 24

page 25

page 26

research
05/27/2019

A Simple Saliency Method That Passes the Sanity Checks

There is great interest in *saliency methods* (also called *attribution ...
research
11/29/2019

Sanity Checks for Saliency Metrics

Saliency maps are a popular approach to creating post-hoc explanations o...
research
05/13/2021

Sanity Simulations for Saliency Methods

Saliency methods are a popular class of feature attribution tools that a...
research
06/07/2022

Beyond Faithfulness: A Framework to Characterize and Compare Saliency Methods

Saliency methods calculate how important each input feature is to a mach...
research
10/31/2022

BOREx: Bayesian-Optimization–Based Refinement of Saliency Map for Image- and Video-Classification Models

Explaining a classification result produced by an image- and video-class...
research
06/16/2020

A generalizable saliency map-based interpretation of model outcome

One of the significant challenges of deep neural networks is that the co...
research
10/27/2021

Revisiting Sanity Checks for Saliency Maps

Saliency methods are a popular approach for model debugging and explaina...

Please sign up or login with your details

Forgot password? Click here to reset