Towards a Deeper Understanding of Concept Bottleneck Models Through End-to-End Explanation

02/07/2023
by   Jack Furby, et al.
0

Concept Bottleneck Models (CBMs) first map raw input(s) to a vector of human-defined concepts, before using this vector to predict a final classification. We might therefore expect CBMs capable of predicting concepts based on distinct regions of an input. In doing so, this would support human interpretation when generating explanations of the model's outputs to visualise input features corresponding to concepts. The contribution of this paper is threefold: Firstly, we expand on existing literature by looking at relevance both from the input to the concept vector, confirming that relevance is distributed among the input features, and from the concept vector to the final classification where, for the most part, the final classification is made using concepts predicted as present. Secondly, we report a quantitative evaluation to measure the distance between the maximum input feature relevance and the ground truth location; we perform this with the techniques, Layer-wise Relevance Propagation (LRP), Integrated Gradients (IG) and a baseline gradient approach, finding LRP has a lower average distance than IG. Thirdly, we propose using the proportion of relevance as a measurement for explaining concept importance.

READ FULL TEXT

page 3

page 4

page 12

research
07/09/2020

Concept Bottleneck Models

We seek to learn models that we can interact with using high-level conce...
research
11/19/2019

Cross-Class Relevance Learning for Temporal Concept Localization

We present a novel Cross-Class Relevance Learning approach for the task ...
research
03/16/2020

Towards Ground Truth Evaluation of Visual Explanations

Several methods have been proposed to explain the decisions of neural ne...
research
12/14/2022

Interactive Concept Bottleneck Models

Concept bottleneck models (CBMs) (Koh et al. 2020) are interpretable neu...
research
05/21/2022

Exploring Concept Contribution Spatially: Hidden Layer Interpretation with Spatial Activation Concept Vector

To interpret deep learning models, one mainstream is to explore the lear...
research
11/07/2022

Towards learning to explain with concept bottleneck models: mitigating information leakage

Concept bottleneck models perform classification by first predicting whi...
research
11/17/2017

Understanding Graph and Understanding Map and their Potential Applications

Based on the previously proposed concept Understanding Tree, this paper ...

Please sign up or login with your details

Forgot password? Click here to reset