What You See is What You Classify: Black Box Attributions

05/23/2022
by   Steven Stalder, et al.
0

An important step towards explaining deep image classifiers lies in the identification of image regions that contribute to individual class scores in the model's output. However, doing this accurately is a difficult task due to the black-box nature of such networks. Most existing approaches find such attributions either using activations and gradients or by repeatedly perturbing the input. We instead address this challenge by training a second deep network, the Explainer, to predict attributions for a pre-trained black-box classifier, the Explanandum. These attributions are in the form of masks that only show the classifier-relevant parts of an image, masking out the rest. Our approach produces sharper and more boundary-precise masks when compared to the saliency maps generated by other methods. Moreover, unlike most existing approaches, ours is capable of directly generating very distinct class-specific masks. Finally, the proposed method is very efficient for inference since it only takes a single forward pass through the Explainer to generate all class-specific masks. We show that our attributions are superior to established methods both visually and quantitatively, by evaluating them on the PASCAL VOC-2007 and Microsoft COCO-2014 datasets.

READ FULL TEXT

page 2

page 6

page 14

page 15

page 16

page 17

page 18

research
09/14/2023

Text-to-Image Models for Counterfactual Explanations: a Black-Box Approach

This paper addresses the challenge of generating Counterfactual Explanat...
research
05/22/2017

Real Time Image Saliency for Black Box Classifiers

In this work we develop a fast saliency detection method that can be app...
research
07/20/2018

Explaining Image Classifiers by Adaptive Dropout and Generative In-filling

Explanations of black-box classifiers often rely on saliency maps, which...
research
09/27/2019

Interpreting Undesirable Pixels for Image Classification on Black-Box Models

In an effort to interpret black-box models, researches for developing ex...
research
12/31/2020

iGOS++: Integrated Gradient Optimized Saliency by Bilateral Perturbations

The black-box nature of the deep networks makes the explanation for "why...
research
09/26/2022

Ablation Path Saliency

Various types of saliency methods have been proposed for explaining blac...
research
06/11/2018

Learning to Speed Up Structured Output Prediction

Predicting structured outputs can be computationally onerous due to the ...

Please sign up or login with your details

Forgot password? Click here to reset