Explaining Deep Learning Hidden Neuron Activations using Concept Induction

01/23/2023
by   Abhilekha Dalal, et al.
0

One of the current key challenges in Explainable AI is in correctly interpreting activations of hidden neurons. It seems evident that accurate interpretations thereof would provide insights into the question what a deep learning system has internally detected as relevant on the input, thus lifting some of the black box character of deep learning systems. The state of the art on this front indicates that hidden node activations appear to be interpretable in a way that makes sense to humans, at least in some cases. Yet, systematic automated methods that would be able to first hypothesize an interpretation of hidden neuron activations, and then verify it, are mostly missing. In this paper, we provide such a method and demonstrate that it provides meaningful interpretations. It is based on using large-scale background knowledge – a class hierarchy of approx. 2 million classes curated from the Wikipedia Concept Hierarchy – together with a symbolic reasoning approach called concept induction based on description logics that was originally developed for applications in the Semantic Web field. Our results show that we can automatically attach meaningful labels from the background knowledge to individual neurons in the dense layer of a Convolutional Neural Network through a hypothesis and verification process.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/08/2023

Understanding CNN Hidden Neuron Activations Using Structured Background Knowledge and Deductive Reasoning

A major challenge in Explainable AI is in correctly interpreting activat...
research
09/27/2022

Towards Human-Compatible XAI: Explaining Data Differentials with Concept Induction over Background Knowledge

Concept induction, which is based on formal logical reasoning over descr...
research
12/13/2019

TopoAct: Exploring the Shape of Activations in Deep Learning

Deep neural networks such as GoogLeNet and ResNet have achieved superhum...
research
01/15/2022

Explainability Tools Enabling Deep Learning in Future In-Situ Real-Time Planetary Explorations

Deep learning (DL) has proven to be an effective machine learning and co...
research
03/16/2018

Dynamic-structured Semantic Propagation Network

Semantic concept hierarchy is still under-explored for semantic segmenta...
research
02/22/2021

Wider Vision: Enriching Convolutional Neural Networks via Alignment to External Knowledge Bases

Deep learning models suffer from opaqueness. For Convolutional Neural Ne...
research
07/25/2019

Interpretability Beyond Classification Output: Semantic Bottleneck Networks

Today's deep learning systems deliver high performance based on end-to-e...

Please sign up or login with your details

Forgot password? Click here to reset