Analyzing Representations inside Convolutional Neural Networks

12/23/2020
by   Uday Singh Saini, et al.
43

How can we discover and succinctly summarize the concepts that a neural network has learned? Such a task is of great importance in applications of networks in areas of inference that involve classification, like medical diagnosis based on fMRI/x-ray etc. In this work, we propose a framework to categorize the concepts a network learns based on the way it clusters a set of input examples, clusters neurons based on the examples they activate for, and input features all in the same latent space. This framework is unsupervised and can work without any labels for input features, it only needs access to internal activations of the network for each input example, thereby making it widely applicable. We extensively evaluate the proposed method and demonstrate that it produces human-understandable and coherent concepts that a ResNet-18 has learned on the CIFAR-100 dataset.

READ FULL TEXT

page 6

page 7

page 8

research
05/14/2021

Cause and Effect: Concept-based Explanation of Neural Networks

In many scenarios, human decisions are explained based on some high-leve...
research
02/07/2019

Automating Interpretability: Discovering and Testing Visual Concepts Learned by Neural Networks

Interpretability has become an important topic of research as more machi...
research
07/30/2017

Towards Visual Explanations for Convolutional Neural Networks via Input Resampling

The predictive power of neural networks often costs model interpretabili...
research
09/22/2022

Concept Activation Regions: A Generalized Framework For Concept-Based Explanations

Concept-based explanations permit to understand the predictions of a dee...
research
05/28/2018

Keep and Learn: Continual Learning by Constraining the Latent Space for Knowledge Preservation in Neural Networks

Data is one of the most important factors in machine learning. However, ...
research
03/30/2022

Interpretable Vertebral Fracture Diagnosis

Do black-box neural network models learn clinically relevant features fo...
research
09/03/2021

Topographic VAEs learn Equivariant Capsules

In this work we seek to bridge the concepts of topographic organization ...

Please sign up or login with your details

Forgot password? Click here to reset