Learning Global Transparent Models from Local Contrastive Explanations

02/19/2020
by   Tejaswini Pedapati, et al.
0

There is a rich and growing literature on producing local point wise contrastive/counterfactual explanations for complex models. These methods highlight what is important to justify the classification and/or produce a contrast point that alters the final classification. Other works try to build globally interpretable models like decision trees and rule lists directly by efficient model search using the data or by transferring information from a complex model using distillation-like methods. Although these interpretable global models can be useful, they may not be consistent with local explanations from a specific complex model of choice. In this work, we explore the question: Can we produce a transparent global model that is consistent with/derivable from local explanations? Based on a key insight we provide a novel method where every local contrastive/counterfactual explanation can be turned into a Boolean feature. These Boolean features are sparse conjunctions of binarized features. The dataset thus constructed is consistent with local explanations by design and one can train an interpretable model like a decision tree on it. We note that this approach strictly loses information due to reliance only on sparse local explanations, nonetheless, we demonstrate empirically that in many cases it can still be competitive with respect to the complex model's performance and also other methods that learn directly from the original dataset. Our approach also provides an avenue to benchmark local explanation methods in a quantitative manner.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/13/2020

Succinct Explanations With Cascading Decision Trees

Classic decision tree learning is a binary classification algorithm that...
research
06/19/2018

Contrastive Explanations with Local Foil Trees

Recent advances in interpretable Machine Learning (iML) and eXplainable ...
research
05/29/2023

Reason to explain: Interactive contrastive explanations (REASONX)

Many high-performing machine learning models are not interpretable. As t...
research
11/02/2020

A Learning Theoretic Perspective on Local Explainability

In this paper, we explore connections between interpretable machine lear...
research
07/13/2022

Policy Optimization with Sparse Global Contrastive Explanations

We develop a Reinforcement Learning (RL) framework for improving an exis...
research
05/09/2023

When a CBR in Hand is Better than Twins in the Bush

AI methods referred to as interpretable are often discredited as inaccur...
research
03/03/2020

Explaining Groups of Points in Low-Dimensional Representations

A common workflow in data exploration is to learn a low-dimensional repr...

Please sign up or login with your details

Forgot password? Click here to reset