Architecture Disentanglement for Deep Neural Networks

03/30/2020
by   Jie Hu, et al.
0

Deep Neural Networks (DNNs) are central to deep learning, and understanding their internal working mechanism is crucial if they are to be used for emerging applications in medical and industrial AI. To this end, the current line of research typically involves linking semantic concepts to a DNN's units or layers. However, this fails to capture the hierarchical inference procedure throughout the network. To address this issue, we introduce the novel concept of Neural Architecture Disentanglement (NAD) in this paper. Specifically, we disentangle a pre-trained network into hierarchical paths corresponding to specific concepts, forming the concept feature paths, i.e., the concept flows from the bottom to top layers of a DNN. Such paths further enable us to quantify the interpretability of DNNs according to the learned diversity of human concepts. We select four types of representative architectures ranging from handcrafted to autoML-based, and conduct extensive experiments on object-based and scene-based datasets. Our NAD sheds important light on the information flow of semantic concepts in DNNs, and provides a fundamental metric that will facilitate the design of interpretable network architectures. Code will be available at: https://github.com/hujiecpp/NAD.

READ FULL TEXT

page 1

page 6

page 7

page 8

research
09/12/2019

New Perspective of Interpretability of Deep Neural Networks

Deep neural networks (DNNs) are known as black-box models. In other word...
research
07/15/2020

Explaining Deep Neural Networks using Unsupervised Clustering

We propose a novel method to explain trained deep neural networks (DNNs)...
research
04/20/2023

Learning Bottleneck Concepts in Image Classification

Interpreting and explaining the behavior of deep neural networks is crit...
research
07/10/2023

Hierarchical Semantic Tree Concept Whitening for Interpretable Image Classification

With the popularity of deep neural networks (DNNs), model interpretabili...
research
03/25/2022

Concept Embedding Analysis: A Review

Deep neural networks (DNNs) have found their way into many applications ...
research
02/05/2018

ClassSim: Similarity between Classes Defined by Misclassification Ratios of Trained Classifiers

Deep neural networks (DNNs) have achieved exceptional performances in ma...
research
11/24/2021

Graph Modularity: Towards Understanding the Cross-Layer Transition of Feature Representations in Deep Neural Networks

There are good arguments to support the claim that feature representatio...

Please sign up or login with your details

Forgot password? Click here to reset