From "Where" to "What": Towards Human-Understandable Explanations through Concept Relevance Propagation

06/07/2022
by   Reduan Achtibat, et al.
0

The emerging field of eXplainable Artificial Intelligence (XAI) aims to bring transparency to today's powerful but opaque deep learning models. While local XAI methods explain individual predictions in form of attribution maps, thereby identifying where important features occur (but not providing information about what they represent), global explanation techniques visualize what concepts a model has generally learned to encode. Both types of methods thus only provide partial insights and leave the burden of interpreting the model's reasoning to the user. Only few contemporary techniques aim at combining the principles behind both local and global XAI for obtaining more informative explanations. Those methods, however, are often limited to specific model architectures or impose additional requirements on training regimes or data and label availability, which renders the post-hoc application to arbitrarily pre-trained models practically impossible. In this work we introduce the Concept Relevance Propagation (CRP) approach, which combines the local and global perspectives of XAI and thus allows answering both the "where" and "what" questions for individual predictions, without additional constraints imposed. We further introduce the principle of Relevance Maximization for finding representative examples of encoded concepts based on their usefulness to the model. We thereby lift the dependency on the common practice of Activation Maximization and its limitations. We demonstrate the capabilities of our methods in various settings, showcasing that Concept Relevance Propagation and Relevance Maximization lead to more human interpretable explanations and provide deep insights into the model's representations and reasoning through concept atlases, concept composition analyses, and quantitative investigations of concept subspaces and their role in fine-grained decision making.

READ FULL TEXT

page 2

page 18

page 21

page 22

page 24

page 26

page 27

page 35

research
06/16/2021

Best of both worlds: local and global explanations with human-understandable concepts

Interpretability techniques aim to provide the rationale behind a model'...
research
12/30/2022

Disentangled Explanations of Neural Network Predictions by Finding Relevant Subspaces

Explainable AI transforms opaque decision strategies of ML models into e...
research
06/24/2021

Software for Dataset-wide XAI: From Local Explanations to Global Insights with Zennit, CoRelAy, and ViRelAy

Deep Neural Networks (DNNs) are known to be strong predictors, but their...
research
11/21/2022

Revealing Hidden Context Bias in Segmentation and Object Detection through Concept-specific Explanations

Applying traditional post-hoc attribution methods to segmentation or obj...
research
06/06/2023

Expanding Explainability Horizons: A Unified Concept-Based System for Local, Global, and Misclassification Explanations

Explainability of intelligent models has been garnering increasing atten...
research
01/27/2023

Multi-dimensional concept discovery (MCD): A unifying framework with completeness guarantees

The completeness axiom renders the explanation of a post-hoc XAI method ...
research
05/01/2021

A Peek Into the Reasoning of Neural Networks: Interpreting with Structural Visual Concepts

Despite substantial progress in applying neural networks (NN) to a wide ...

Please sign up or login with your details

Forgot password? Click here to reset