A Closer Look at the Intervention Procedure of Concept Bottleneck Models

02/28/2023
by   Sungbin Shin, et al.
0

Concept bottleneck models (CBMs) are a class of interpretable neural network models that predict the target response of a given input based on its high-level concepts. Unlike the standard end-to-end models, CBMs enable domain experts to intervene on the predicted concepts and rectify any mistakes at test time, so that more accurate task predictions can be made at the end. While such intervenability provides a powerful avenue of control, many aspects of the intervention procedure remain rather unexplored. In this work, we develop various ways of selecting intervening concepts to improve the intervention effectiveness and conduct an array of in-depth analyses as to how they evolve under different circumstances. Specifically, we find that an informed intervention strategy can reduce the task error more than ten times compared to the current baseline under the same amount of intervention counts in realistic settings, and yet, this can vary quite significantly when taking into account different intervention granularity. We verify our findings through comprehensive evaluations, not only on the standard real datasets, but also on synthetic datasets that we generate based on a set of different causal graphs. We further discover some major pitfalls of the current practices which, without a proper addressing, raise concerns on reliability and fairness of the intervention procedure.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/09/2020

Concept Bottleneck Models

We seek to learn models that we can interact with using high-level conce...
research
05/10/2021

Do Concept Bottleneck Models Learn as Intended?

Concept bottleneck models map from raw inputs to concepts, and then from...
research
06/21/2022

Automatic Concept Extraction for Concept Bottleneck-based Video Classification

Recent efforts in interpretable deep learning models have shown that con...
research
06/14/2023

Selective Concept Models: Permitting Stakeholder Customisation at Test-Time

Concept-based models perform prediction using a set of concepts that are...
research
11/21/2022

Learn to explain yourself, when you can: Equipping Concept Bottleneck Models with the ability to abstain on their concept predictions

The Concept Bottleneck Models (CBMs) of Koh et al. [2020] provide a mean...
research
11/07/2022

Towards learning to explain with concept bottleneck models: mitigating information leakage

Concept bottleneck models perform classification by first predicting whi...
research
12/09/2021

Latent Space Explanation by Intervention

The success of deep neural nets heavily relies on their ability to encod...

Please sign up or login with your details

Forgot password? Click here to reset