Generative Counterfactual Introspection for Explainable Deep Learning

07/06/2019
by   Shusen Liu, et al.
7

In this work, we propose an introspection technique for deep neural networks that relies on a generative model to instigate salient editing of the input image for model interpretation. Such modification provides the fundamental interventional operation that allows us to obtain answers to counterfactual inquiries, i.e., what meaningful change can be made to the input image in order to alter the prediction. We demonstrate how to reveal interesting properties of the given classifiers by utilizing the proposed introspection approach on both the MNIST and the CelebA dataset.

READ FULL TEXT

page 4

page 5

page 6

page 7

research
03/25/2021

ECINN: Efficient Counterfactuals from Invertible Neural Networks

Counterfactual examples identify how inputs can be altered to change the...
research
01/25/2021

Conditional Generative Models for Counterfactual Explanations

Counterfactual instances offer human-interpretable insight into the loca...
research
01/15/2021

Counterfactual Generative Networks

Neural networks are prone to learning shortcuts – they often model simpl...
research
09/04/2023

SMPLitex: A Generative Model and Dataset for 3D Human Texture Estimation from Single Image

We propose SMPLitex, a method for estimating and manipulating the comple...
research
07/13/2020

Domain aware medical image classifier interpretation by counterfactual impact analysis

The success of machine learning methods for computer vision tasks has dr...
research
10/09/2019

Removing input features via a generative model to explain their attributions to classifier's decisions

Interpretability methods often measure the contribution of an input feat...
research
11/16/2015

Deep Kalman Filters

Kalman Filters are one of the most influential models of time-varying ph...

Please sign up or login with your details

Forgot password? Click here to reset