Class Introspection: A Novel Technique for Detecting Unlabeled Subclasses by Leveraging Classifier Explainability Methods

07/04/2021
by   Patrick Kage, et al.
8

Detecting latent structure within a dataset is a crucial step in performing analysis of a dataset. However, existing state-of-the-art techniques for subclass discovery are limited: either they are limited to detecting very small numbers of outliers or they lack the statistical power to deal with complex data such as image or audio. This paper proposes a solution to this subclass discovery problem: by leveraging instance explanation methods, an existing classifier can be extended to detect latent classes via differences in the classifier's internal decisions about each instance. This works not only with simple classification techniques but also with deep neural networks, allowing for a powerful and flexible approach to detecting latent structure within datasets. Effectively, this represents a projection of the dataset into the classifier's "explanation space," and preliminary results show that this technique outperforms the baseline for the detection of latent classes even with limited processing. This paper also contains a pipeline for analyzing classifiers automatically, and a web application for interactively exploring the results from this technique.

READ FULL TEXT

page 3

page 6

page 7

page 8

research
05/26/2017

Detecting and Explaining Crisis

Individuals on social media may reveal themselves to be in various state...
research
12/05/2022

Audio Latent Space Cartography

We explore the generation of visualisations of audio latent spaces using...
research
10/11/2022

Class-Specific Explainability for Deep Time Series Classifiers

Explainability helps users trust deep learning solutions for time series...
research
10/18/2019

Identifying the Most Explainable Classifier

We introduce the notion of pointwise coverage to measure the explainabil...
research
11/18/2019

NeuronInspect: Detecting Backdoors in Neural Networks via Output Explanations

Deep neural networks have achieved state-of-the-art performance on vario...
research
12/18/2019

Clusters in Explanation Space: Inferring disease subtypes from model explanations

Identification of disease subtypes and corresponding biomarkers can subs...
research
08/25/2021

Inducing Semantic Grouping of Latent Concepts for Explanations: An Ante-Hoc Approach

Self-explainable deep models are devised to represent the hidden concept...

Please sign up or login with your details

Forgot password? Click here to reset