Exploring Concept Contribution Spatially: Hidden Layer Interpretation with Spatial Activation Concept Vector

05/21/2022
by   Andong Wang, et al.
0

To interpret deep learning models, one mainstream is to explore the learned concepts by networks. Testing with Concept Activation Vector (TCAV) presents a powerful tool to quantify the contribution of query concepts (represented by user-defined guidance images) to a target class. For example, we can quantitatively evaluate whether and to what extent concept striped contributes to model prediction zebra with TCAV. Therefore, TCAV whitens the reasoning process of deep networks. And it has been applied to solve practical problems such as diagnosis. However, for some images where the target object only occupies a small fraction of the region, TCAV evaluation may be interfered with by redundant background features because TCAV calculates concept contribution to a target class based on a whole hidden layer. To tackle this problem, based on TCAV, we propose Spatial Activation Concept Vector (SACV) which identifies the relevant spatial locations to the query concept while evaluating their contributions to the model prediction of the target class. Experiment shows that SACV generates a more fine-grained explanation map for a hidden layer and quantifies concepts' contributions spatially. Moreover, it avoids interference from background features. The code is available on https://github.com/AntonotnaWang/Spatial-Activation-Concept-Vector.

READ FULL TEXT
research
06/02/2023

Probabilistic Concept Bottleneck Models

Interpretable models are designed to make decisions in a human-interpret...
research
02/10/2020

Adversarial TCAV – Robust and Effective Interpretation of Intermediate Layers in Neural Networks

Interpreting neural network decisions and the information learned in int...
research
05/19/2023

Generating Visual Spatial Description via Holistic 3D Scene Understanding

Visual spatial description (VSD) aims to generate texts that describe th...
research
02/07/2023

Towards a Deeper Understanding of Concept Bottleneck Models Through End-to-End Explanation

Concept Bottleneck Models (CBMs) first map raw input(s) to a vector of h...
research
03/30/2022

Eff the ineffable: on the uncommunicability of a conceptually simple contribution to HCI methodology

Given a simple concept that has the potential for a methodological contr...
research
08/31/2022

Concept Gradient: Concept-based Interpretation Without Linear Assumption

Concept-based interpretations of black-box models are often more intuiti...
research
02/07/2023

Concept Algebra for Text-Controlled Vision Models

This paper concerns the control of text-guided generative models, where ...

Please sign up or login with your details

Forgot password? Click here to reset