A Theory of Diagnostic Interpretation in Supervised Classification

06/26/2018
by   Anirban Mukhopadhyay, et al.
0

Interpretable deep learning is a fundamental building block towards safer AI, especially when the deployment possibilities of deep learning-based computer-aided medical diagnostic systems are so eminent. However, without a computational formulation of black-box interpretation, general interpretability research rely heavily on subjective bias. Clear decision structure of the medical diagnostics lets us approximate the decision process of a radiologist as a model - removed from subjective bias. We define the process of interpretation as a finite communication between a known model and a black-box model to optimally map the black box's decision process in the known model. Consequently, we define interpretability as maximal information gain over the initial uncertainty about the black-box's decision within finite communication. We relax this definition based on the observation that diagnostic interpretation is typically achieved by a process of minimal querying. We derive an algorithm to calculate diagnostic interpretability. The usual question of accuracy-interpretability tradeoff, i.e. whether a black-box model's prediction accuracy is dependent on its ability to be interpreted by a known source model, does not arise in this theory. With multiple example simulation experiments of various complexity levels, we demonstrate the working of such a theoretical model in synthetic supervised classification scenarios.

READ FULL TEXT
research
02/06/2018

A Survey Of Methods For Explaining Black Box Models

In the last years many accurate decision support systems have been const...
research
11/19/2016

Understanding Anatomy Classification Through Visualization

One of the main challenges for broad adoption of deep convolutional neur...
research
12/02/2019

Diagnostic Curves for Black Box Models

In safety-critical applications of machine learning, it is often necessa...
research
06/05/2019

Don't Paint It Black: White-Box Explanations for Deep Learning in Computer Security

Deep learning is increasingly used as a basic building block of security...
research
06/09/2022

Distillation Decision Tree

Black-box machine learning models are criticized as lacking interpretabi...
research
03/13/2023

Revisiting model self-interpretability in a decision-theoretic way for binary medical image classification

Interpretability is highly desired for deep neural network-based classif...
research
06/01/2018

Producing radiologist-quality reports for interpretable artificial intelligence

Current approaches to explaining the decisions of deep learning systems ...

Please sign up or login with your details

Forgot password? Click here to reset