Interpreting Deep Visual Representations via Network Dissection

11/15/2017
by   Bolei Zhou, et al.
0

The success of recent deep convolutional neural networks (CNNs) depends on learning hidden representations that can summarize the important factors of variation behind the data. However, CNNs often criticized as being black boxes that lack interpretability, since they have millions of unexplained model parameters. In this work, we describe Network Dissection, a method that interprets networks by providing labels for the units of their deep visual representations. The proposed method quantifies the interpretability of CNN representations by evaluating the alignment between individual hidden units and a set of visual semantic concepts. By identifying the best alignments, units are given human interpretable labels across a range of objects, parts, scenes, textures, materials, and colors. The method reveals that deep representations are more transparent and interpretable than expected: we find that representations are significantly more interpretable than they would be under a random equivalently powerful basis. We apply the method to interpret and compare the latent representations of various network architectures trained to solve different supervised and self-supervised training tasks. We then examine factors affecting the network interpretability such as the number of the training iterations, regularizations, different initializations, and the network depth and width. Finally we show that the interpreted units can be used to provide explicit explanations of a prediction given by a CNN for an image. Our results highlight that interpretability is an important property of deep neural networks that provides new insights into their hierarchical structure.

READ FULL TEXT

page 3

page 6

page 7

page 8

page 11

page 13

page 14

page 15

research
04/19/2017

Network Dissection: Quantifying Interpretability of Deep Visual Representations

We propose a general framework called Network Dissection for quantifying...
research
05/06/2019

Deep Visual City Recognition Visualization

Understanding how cities visually differ from each others is interesting...
research
03/19/2023

Unsupervised Interpretable Basis Extraction for Concept-Based Visual Explanations

An important line of research attempts to explain CNN image classifier p...
research
03/07/2022

Interpretable part-whole hierarchies and conceptual-semantic relationships in neural networks

Deep neural networks achieve outstanding results in a large variety of t...
research
02/18/2019

Discovery of Natural Language Concepts in Individual Units of CNNs

Although deep convolutional networks have achieved improved performance ...
research
04/10/2022

Explaining Deep Convolutional Neural Networks via Latent Visual-Semantic Filter Attention

Interpretability is an important property for visual models as it helps ...
research
01/08/2019

Interpretable BoW Networks for Adversarial Example Detection

The standard approach to providing interpretability to deep convolutiona...

Please sign up or login with your details

Forgot password? Click here to reset