DeforestVis: Behavior Analysis of Machine Learning Models with Surrogate Decision Stumps

03/31/2023
by   Angelos Chatzimparmpas, et al.
0

As the complexity of machine learning (ML) models increases and the applications in different (and critical) domains grow, there is a strong demand for more interpretable and trustworthy ML. One straightforward and model-agnostic way to interpret complex ML models is to train surrogate models, such as rule sets and decision trees, that sufficiently approximate the original ones while being simpler and easier-to-explain. Yet, rule sets can become very lengthy, with many if-else statements, and decision tree depth grows rapidly when accurately emulating complex ML models. In such cases, both approaches can fail to meet their core goal, providing users with model interpretability. We tackle this by proposing DeforestVis, a visual analytics tool that offers user-friendly summarization of the behavior of complex ML models by providing surrogate decision stumps (one-level decision trees) generated with the adaptive boosting (AdaBoost) technique. Our solution helps users to explore the complexity vs fidelity trade-off by incrementally generating more stumps, creating attribute-based explanations with weighted stumps to justify decision making, and analyzing the impact of rule overriding on training instance allocation between one or more stumps. An independent test set allows users to monitor the effectiveness of manual rule changes and form hypotheses based on case-by-case investigations. We show the applicability and usefulness of DeforestVis with two use cases and expert interviews with data analysts and model developers.

READ FULL TEXT

page 1

page 5

page 6

page 7

page 8

research
12/01/2021

VisRuler: Visual Analytics for Extracting Decision Rules from Bagged and Boosted Decision Trees

Bagging and boosting are two popular ensemble methods in machine learnin...
research
06/10/2023

Interpretable Differencing of Machine Learning Models

Understanding the differences between machine learning (ML) models is of...
research
01/19/2022

Visual Exploration of Machine Learning Model Behavior with Hierarchical Surrogate Rule Sets

One of the potential solutions for model interpretation is to train a su...
research
04/25/2022

On Machine Learning-Driven Surrogates for Sound Transmission Loss Simulations

Surrogate models are data-based approximations of computationally expens...
research
02/08/2023

Decision trees compensate for model misspecification

The best-performing models in ML are not interpretable. If we can explai...
research
05/28/2023

Interactive Decision Tree Creation and Enhancement with Complete Visualization for Explainable Modeling

To increase the interpretability and prediction accuracy of the Machine ...
research
06/17/2019

Learning Interpretable Models Using an Oracle

As Machine Learning (ML) becomes pervasive in various real world systems...

Please sign up or login with your details

Forgot password? Click here to reset