Explainable AI for Trees: From Local Explanations to Global Understanding

05/11/2019
by   Scott M. Lundberg, et al.
0

Tree-based machine learning models such as random forests, decision trees, and gradient boosted trees are the most popular non-linear predictive models used in practice today, yet comparatively little attention has been paid to explaining their predictions. Here we significantly improve the interpretability of tree-based models through three main contributions: 1) The first polynomial time algorithm to compute optimal explanations based on game theory. 2) A new type of explanation that directly measures local feature interaction effects. 3) A new set of tools for understanding global model structure based on combining many local explanations of each prediction. We apply these tools to three medical machine learning problems and show how combining many high-quality local explanations allows us to represent global structure while retaining local faithfulness to the original model. These tools enable us to i) identify high magnitude but low frequency non-linear mortality risk factors in the general US population, ii) highlight distinct population sub-groups with shared risk characteristics, iii) identify non-linear interaction effects among risk factors for chronic kidney disease, and iv) monitor a machine learning model deployed in a hospital by identifying which features are degrading the model's performance over time. Given the popularity of tree-based machine learning models, these improvements to their interpretability have implications across a broad set of domains.

READ FULL TEXT
research
07/09/2018

Supervised Local Modeling for Interpretability

Model interpretability is an increasingly important component of practic...
research
07/28/2023

Toward Transparent Sequence Models with Model-Based Tree Markov Model

In this study, we address the interpretability issue in complex, black-b...
research
04/14/2023

Grouping Shapley Value Feature Importances of Random Forests for explainable Yield Prediction

Explainability in yield prediction helps us fully explore the potential ...
research
08/13/2021

Data-driven advice for interpreting local and global model predictions in bioinformatics problems

Tree-based algorithms such as random forests and gradient boosted trees ...
research
04/30/2021

Explaining a Series of Models by Propagating Local Feature Attributions

Pipelines involving a series of several machine learning models (e.g., s...
research
09/16/2022

Linear TreeShap

Decision trees are well-known due to their ease of interpretability. To ...
research
08/12/2022

Unifying local and global model explanations by functional decomposition of low dimensional structures

We consider a global explanation of a regression or classification funct...

Please sign up or login with your details

Forgot password? Click here to reset