Sanity Checks for Saliency Maps

10/08/2018
by   Julius Adebayo, et al.
2

Saliency methods have emerged as a popular tool to highlight features in an input deemed relevant for the prediction of a learned model. Several saliency methods have been proposed, often guided by visual appeal on image data. In this work, we propose an actionable methodology to evaluate what kinds of explanations a given method can and cannot provide. We find that reliance, solely, on visual assessment can be misleading. Through extensive experiments we show that some existing saliency methods are independent both of the model and of the data generating process. Consequently, methods that fail the proposed tests are inadequate for tasks that are sensitive to either data or model, such as, finding outliers in the data, explaining the relationship between inputs and outputs that the model learned, and debugging the model. We interpret our findings through an analogy with edge detection in images, a technique that requires neither training data nor model. Theory in the case of a linear model and a single-layer convolutional neural network supports our experimental findings.

READ FULL TEXT

page 15

page 17

page 22

page 23

page 24

page 25

page 26

page 27

research
08/22/2019

Saliency Methods for Explaining Adversarial Attacks

In this work, we aim to explain the classifications of adversary images ...
research
09/26/2022

Ablation Path Saliency

Various types of saliency methods have been proposed for explaining blac...
research
10/16/2019

Global Saliency: Aggregating Saliency Maps to Assess Dataset Artefact Bias

In high-stakes applications of machine learning models, interpretability...
research
06/27/2018

Learning a Saliency Evaluation Metric Using Crowdsourced Perceptual Judgments

In the area of human fixation prediction, dozens of computational salien...
research
06/15/2021

Explaining decision of model from its prediction

This document summarizes different visual explanations methods such as C...
research
08/03/2021

Where do Models go Wrong? Parameter-Space Saliency Maps for Explainability

Conventional saliency maps highlight input features to which neural netw...
research
09/22/2020

What Do You See? Evaluation of Explainable Artificial Intelligence (XAI) Interpretability through Neural Backdoors

EXplainable AI (XAI) methods have been proposed to interpret how a deep ...

Please sign up or login with your details

Forgot password? Click here to reset