Explaining Image Classifiers by Adaptive Dropout and Generative In-filling

07/20/2018
by   Chun-Hao Chang, et al.
0

Explanations of black-box classifiers often rely on saliency maps, which score the relevance of each input dimension to the resulting classification. Recent approaches compute saliency by optimizing regions of the input that maximally change the classification outcome when replaced by a reference value. These reference values are based on ad-hoc heuristics such as the input mean. In this work we marginalize out masked regions of the input, conditioning a generative model on the rest of the image. Our model-agnostic method produces realistic explanations, generating plausible inputs that would have caused the model to classify differently. When applied to image classification, our method produces more compact and relevant explanations, with fewer artifacts compared to previous methods.

READ FULL TEXT

page 2

page 5

page 6

page 7

page 8

page 10

page 11

research
06/24/2020

Generative causal explanations of black-box classifiers

We develop a method for generating causal post-hoc explanations of black...
research
05/23/2022

What You See is What You Classify: Black Box Attributions

An important step towards explaining deep image classifiers lies in the ...
research
07/15/2019

A study on the Interpretability of Neural Retrieval Models using DeepSHAP

A recent trend in IR has been the usage of neural networks to learn retr...
research
10/17/2019

Effect of Superpixel Aggregation on Explanations in LIME – A Case Study with Biological Data

End-to-end learning with deep neural networks, such as convolutional neu...
research
05/17/2022

Monotonicity Regularization: Improved Penalties and Novel Applications to Disentangled Representation Learning and Robust Classification

We study settings where gradient penalties are used alongside risk minim...
research
11/01/2019

Explanation by Progressive Exaggeration

As machine learning methods see greater adoption and implementation in h...
research
11/15/2022

Easy to Decide, Hard to Agree: Reducing Disagreements Between Saliency Methods

A popular approach to unveiling the black box of neural NLP models is to...

Please sign up or login with your details

Forgot password? Click here to reset