Towards Better Understanding Attribution Methods

05/20/2022
by   Sukrut Rao, et al.
32

Deep neural networks are very successful on many vision tasks, but hard to interpret due to their black box nature. To overcome this, various post-hoc attribution methods have been proposed to identify image regions most influential to the models' decisions. Evaluating such methods is challenging since no ground truth attributions exist. We thus propose three novel evaluation schemes to more reliably measure the faithfulness of those methods, to make comparisons between them more fair, and to make visual inspection more systematic. To address faithfulness, we propose a novel evaluation setting (DiFull) in which we carefully control which parts of the input can influence the output in order to distinguish possible from impossible attributions. To address fairness, we note that different methods are applied at different layers, which skews any comparison, and so evaluate all methods on the same layers (ML-Att) and discuss how this impacts their performance on quantitative metrics. For more systematic visualizations, we propose a scheme (AggAtt) to qualitatively evaluate the methods on complete datasets. We use these evaluation schemes to study strengths and shortcomings of some widely used attribution methods. Finally, we propose a post-processing smoothing step that significantly improves the performance of some attribution methods, and discuss its applicability.

READ FULL TEXT

page 13

page 14

page 15

page 16

page 17

page 18

page 20

page 21

research
03/21/2023

Better Understanding Differences in Attribution Methods via Systematic Evaluations

Deep neural networks are very successful on many vision tasks, but hard ...
research
11/16/2017

A unified view of gradient-based attribution methods for Deep Neural Networks

Understanding the flow of information in Deep Neural Networks is a chall...
research
07/01/2021

Towards Measuring Bias in Image Classification

Convolutional Neural Networks (CNN) have become de fact state-of-the-art...
research
09/01/2021

Spatio-Temporal Perturbations for Video Attribution

The attribution method provides a direction for interpreting opaque neur...
research
05/23/2023

Towards credible visual model interpretation with path attribution

Originally inspired by game-theory, path attribution framework stands ou...
research
11/14/2022

Explainer Divergence Scores (EDS): Some Post-Hoc Explanations May be Effective for Detecting Unknown Spurious Correlations

Recent work has suggested post-hoc explainers might be ineffective for d...
research
06/27/2021

Crowdsourcing Evaluation of Saliency-based XAI Methods

Understanding the reasons behind the predictions made by deep neural net...

Please sign up or login with your details

Forgot password? Click here to reset