An unexpected unity among methods for interpreting model predictions

11/22/2016
by   Scott Lundberg, et al.
0

Understanding why a model made a certain prediction is crucial in many data science fields. Interpretable predictions engender appropriate trust and provide insight into how the model may be improved. However, with large modern datasets the best accuracy is often achieved by complex models even experts struggle to interpret, which creates a tension between accuracy and interpretability. Recently, several methods have been proposed for interpreting predictions from complex models by estimating the importance of input features. Here, we present how a model-agnostic additive representation of the importance of input features unifies current methods. This representation is optimal, in the sense that it is the only set of additive values that satisfies important properties. We show how we can leverage these properties to create novel visual explanations of model predictions. The thread of unity that this representation weaves through the literature indicates that there are common principles to be learned about the interpretation of model predictions that apply in many scenarios.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/22/2017

A Unified Approach to Interpreting Model Predictions

Understanding why a model makes a certain prediction can be as crucial a...
research
12/22/2019

Interpreting Predictive Process Monitoring Benchmarks

Predictive process analytics has recently gained significant attention, ...
research
04/29/2019

Why should you trust my interpretation? Understanding uncertainty in LIME predictions

Methods for interpreting machine learning black-box models increase the ...
research
01/19/2021

Using StyleGAN for Visual Interpretability of Deep Learning Models on Medical Images

As AI-based medical devices are becoming more common in imaging fields l...
research
10/15/2020

Marginal Contribution Feature Importance – an Axiomatic Approach for The Natural Case

When training a predictive model over medical data, the goal is sometime...
research
01/28/2020

Statistical Exploration of Relationships Between Routine and Agnostic Features Towards Interpretable Risk Characterization

As is typical in other fields of application of high throughput systems,...
research
09/11/2023

Evaluating the Reliability of CNN Models on Classifying Traffic and Road Signs using LIME

The objective of this investigation is to evaluate and contrast the effe...

Please sign up or login with your details

Forgot password? Click here to reset