Automatic Concept Embedding Model (ACEM): No train-time concepts, No issue!

09/07/2023
by   Rishabh Jain, et al.
0

Interpretability and explainability of neural networks is continuously increasing in importance, especially within safety-critical domains and to provide the social right to explanation. Concept based explanations align well with how humans reason, proving to be a good way to explain models. Concept Embedding Models (CEMs) are one such concept based explanation architectures. These have shown to overcome the trade-off between explainability and performance. However, they have a key limitation – they require concept annotations for all their training data. For large datasets, this can be expensive and infeasible. Motivated by this, we propose Automatic Concept Embedding Models (ACEMs), which learn the concept annotations automatically.

READ FULL TEXT

page 4

page 7

research
06/06/2023

Expanding Explainability Horizons: A Unified Concept-Based System for Local, Global, and Misclassification Explanations

Explainability of intelligent models has been garnering increasing atten...
research
10/14/2021

Brittle interpretations: The Vulnerability of TCAV and Other Concept-based Explainability Tools to Adversarial Attack

Methods for model explainability have become increasingly critical for t...
research
12/07/2017

Network Analysis for Explanation

Safety critical systems strongly require the quality aspects of artifici...
research
08/22/2022

Global Concept-Based Interpretability for Graph Neural Networks via Neuron Analysis

Graph neural networks (GNNs) are highly effective on a variety of graph-...
research
10/03/2019

The Bouncer Problem: Challenges to Remote Explainability

The concept of explainability is envisioned to satisfy society's demands...
research
07/01/2020

Unifying Model Explainability and Robustness via Machine-Checkable Concepts

As deep neural networks (DNNs) get adopted in an ever-increasing number ...
research
03/06/2023

IFAN: An Explainability-Focused Interaction Framework for Humans and NLP Models

Interpretability and human oversight are fundamental pillars of deployin...

Please sign up or login with your details

Forgot password? Click here to reset