GlanceNets: Interpretabile, Leak-proof Concept-based Models

05/31/2022
by   Emanuele Marconato, et al.
0

There is growing interest in concept-based models (CBMs) that combine high-performance and interpretability by acquiring and reasoning with a vocabulary of high-level concepts. A key requirement is that the concepts be interpretable. Existing CBMs tackle this desideratum using a variety of heuristics based on unclear notions of interpretability, and fail to acquire concepts with the intended semantics. We address this by providing a clear definition of interpretability in terms of alignment between the model's representation and an underlying data generation process, and introduce GlanceNets, a new CBM that exploits techniques from disentangled representation learning and open-set recognition to achieve alignment, thus improving the interpretability of the learned concepts. We show that GlanceNets, paired with concept-level supervision, achieve better alignment than state-of-the-art approaches while preventing spurious information from unintendedly leaking into the learned concepts.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/14/2023

Interpretability is in the Mind of the Beholder: A Causal Framework for Human-interpretable Representation Learning

Focus in Explainable AI is shifting from explanations defined in terms o...
research
05/10/2021

Do Concept Bottleneck Models Learn as Intended?

Concept bottleneck models map from raw inputs to concepts, and then from...
research
04/27/2023

Interpretable Neural-Symbolic Concept Reasoning

Deep learning methods are highly accurate, yet their opaque decision pro...
research
03/22/2023

Neuro-Symbolic Reasoning Shortcuts: Mitigation Strategies and their Limitations

Neuro-symbolic predictors learn a mapping from sub-symbolic inputs to hi...
research
08/22/2022

Global Concept-Based Interpretability for Graph Neural Networks via Neuron Analysis

Graph neural networks (GNNs) are highly effective on a variety of graph-...
research
05/31/2023

Not All Neuro-Symbolic Concepts Are Created Equal: Analysis and Mitigation of Reasoning Shortcuts

Neuro-Symbolic (NeSy) predictive models hold the promise of improved com...
research
08/17/2022

A Concept and Argumentation based Interpretable Model in High Risk Domains

Interpretability has become an essential topic for artificial intelligen...

Please sign up or login with your details

Forgot password? Click here to reset