Rethinking Positive Aggregation and Propagation of Gradients in Gradient-based Saliency Methods

12/01/2020
by   Ashkan Khakzar, et al.
8

Saliency methods interpret the prediction of a neural network by showing the importance of input elements for that prediction. A popular family of saliency methods utilize gradient information. In this work, we empirically show that two approaches for handling the gradient information, namely positive aggregation, and positive propagation, break these methods. Though these methods reflect visually salient information in the input, they do not explain the model prediction anymore as the generated saliency maps are insensitive to the predicted output and are insensitive to model parameter randomization. Specifically for methods that aggregate the gradients of a chosen layer such as GradCAM++ and FullGrad, exclusively aggregating positive gradients is detrimental. We further support this by proposing several variants of aggregation methods with positive handling of gradient information. For methods that backpropagate gradient information such as LRP, RectGrad, and Guided Backpropagation, we show the destructive effect of exclusively propagating positive gradient information.

READ FULL TEXT

page 2

page 3

page 5

page 6

research
04/06/2020

There and Back Again: Revisiting Backpropagation Saliency Methods

Saliency methods seek to explain the predictions of a model by producing...
research
11/10/2020

Removing Brightness Bias in Rectified Gradients

Interpretation and improvement of deep neural networks relies on better ...
research
06/29/2020

Scaling Symbolic Methods using Gradients for Neural Model Explanation

Symbolic techniques based on Satisfiability Modulo Theory (SMT) solvers ...
research
07/12/2022

Rethinking gradient weights' influence over saliency map estimation

Class activation map (CAM) helps to formulate saliency maps that aid in ...
research
10/12/2020

Gradient-based Analysis of NLP Models is Manipulable

Gradient-based analysis methods, such as saliency map visualizations and...
research
11/29/2021

Improving Deep Learning Interpretability by Saliency Guided Training

Saliency methods have been widely used to highlight important input feat...
research
07/06/2023

Generalizing Backpropagation for Gradient-Based Interpretability

Many popular feature-attribution methods for interpreting deep neural ne...

Please sign up or login with your details

Forgot password? Click here to reset