Concept Whitening for Interpretable Image Recognition

02/05/2020
by   Zhi Chen, et al.
5

What does a neural network encode about a concept as we traverse through the layers? Interpretability in machine learning is undoubtedly important, but the calculations of neural networks are very challenging to understand. Attempts to see inside their hidden layers can either be misleading, unusable, or rely on the latent space to possess properties that it may not have. In this work, rather than attempting to analyze a neural network posthoc, we introduce a mechanism, called concept whitening (CW), to alter a given layer of the network to allow us to better understand the computation leading up to that layer. When a concept whitening module is added to a CNN, the axes of the latent space can be aligned with concepts of interest. By experiment, we show that CW can provide us a much clearer understanding for how the network gradually learns concepts over layers without hurting predictive performance.

READ FULL TEXT

page 7

page 8

page 9

page 10

page 16

research
07/10/2023

Hierarchical Semantic Tree Concept Whitening for Interpretable Image Classification

With the popularity of deep neural networks (DNNs), model interpretabili...
research
07/17/2019

Differentiable Disentanglement Filter: an Application Agnostic Core Concept Discovery Probe

It has long been speculated that deep neural networks function by discov...
research
12/04/2021

Interactive Disentanglement: Learning Concepts by Interacting with their Prototype Representations

Learning visual concepts from raw images without strong supervision is a...
research
09/22/2022

Concept Activation Regions: A Generalized Framework For Concept-Based Explanations

Concept-based explanations permit to understand the predictions of a dee...
research
12/09/2021

Latent Space Explanation by Intervention

The success of deep neural nets heavily relies on their ability to encod...
research
05/04/2020

Explaining AI-based Decision Support Systems using Concept Localization Maps

Human-centric explainability of AI-based Decision Support Systems (DSS) ...
research
08/21/2023

Sparse Linear Concept Discovery Models

The recent mass adoption of DNNs, even in safety-critical scenarios, has...

Please sign up or login with your details

Forgot password? Click here to reset