A Model Explanation System: Latest Updates and Extensions

06/30/2016
by   Ryan Turner, et al.
0

We propose a general model explanation system (MES) for "explaining" the output of black box classifiers. This paper describes extensions to Turner (2015), which is referred to frequently in the text. We use the motivating example of a classifier trained to detect fraud in a credit card transaction history. The key aspect is that we provide explanations applicable to a single prediction, rather than provide an interpretable set of parameters. We focus on explaining positive predictions (alerts). However, the presented methodology is symmetrically applicable to negative predictions.

READ FULL TEXT
research
10/02/2019

Contextual Local Explanation for Black Box Classifiers

We introduce a new model-agnostic explanation technique which explains t...
research
11/13/2018

Interpretable Credit Application Predictions With Counterfactual Explanations

We predict credit applications with off-the-shelf, interchangeable black...
research
07/19/2019

Multi-Granular Text Encoding for Self-Explaining Categorization

Self-explaining text categorization requires a classifier to make a pred...
research
06/01/2021

Explanations for Monotonic Classifiers

In many classification tasks there is a requirement of monotonicity. Con...
research
10/04/2022

Explanation-by-Example Based on Item Response Theory

Intelligent systems that use Machine Learning classification algorithms ...
research
03/05/2021

Compositional Explanations for Image Classifiers

Existing algorithms for explaining the output of image classifiers perfo...
research
04/05/2022

A Set Membership Approach to Discovering Feature Relevance and Explaining Neural Classifier Decisions

Neural classifiers are non linear systems providing decisions on the cla...

Please sign up or login with your details

Forgot password? Click here to reset