How Well do Feature Visualizations Support Causal Understanding of CNN Activations?

06/23/2021
by   Roland S. Zimmermann, et al.
26

One widely used approach towards understanding the inner workings of deep convolutional neural networks is to visualize unit responses via activation maximization. Feature visualizations via activation maximization are thought to provide humans with precise information about the image features that cause a unit to be activated. If this is indeed true, these synthetic images should enable humans to predict the effect of an intervention, such as whether occluding a certain patch of the image (say, a dog's head) changes a unit's activation. Here, we test this hypothesis by asking humans to predict which of two square occlusions causes a larger change to a unit's activation. Both a large-scale crowdsourced experiment and measurements with experts show that on average, the extremely activating feature visualizations by Olah et al. (2017) indeed help humans on this task (67 ± 4% accuracy; baseline performance without any visualizations is 60 ± 3%). However, they do not provide any significant advantage over other visualizations (such as e.g. dataset samples), which yield similar performance (66 ± 3% to 67 ± 3% accuracy). Taken together, we propose an objective psychophysical task to quantify the benefit of unit-level interpretability methods for humans, and find no evidence that feature visualizations provide humans with better "causal understanding" than simple alternative visualizations.

READ FULL TEXT

page 1

page 4

page 5

page 6

page 7

page 9

page 11

page 12

research
10/23/2020

Exemplary Natural Images Explain CNN Activations Better than Feature Visualizations

Feature visualizations such as synthetic maximally activating images are...
research
09/03/2019

Illuminated Decision Trees with Lucid

The Lucid methods described by Olah et al. (2018) provide a way to inspe...
research
06/07/2023

Don't trust your eyes: on the (un)reliability of feature visualizations

How do neural networks extract patterns from pixels? Feature visualizati...
research
07/21/2020

Inverting the Feature Visualization Process for Feedforward Neural Networks

This work sheds light on the invertibility of feature visualization in n...
research
12/07/2015

Visualizing Deep Convolutional Neural Networks Using Natural Pre-Images

Image representations, from SIFT and bag of visual words to Convolutiona...
research
06/11/2023

Unlocking Feature Visualization for Deeper Networks with MAgnitude Constrained Optimization

Feature visualization has gained substantial popularity, particularly af...
research
10/04/2021

AdjointBackMapV2: Precise Reconstruction of Arbitrary CNN Unit's Activation via Adjoint Operators

Adjoint operators have been found to be effective in the exploration of ...

Please sign up or login with your details

Forgot password? Click here to reset