Optimizing for Interpretability in Deep Neural Networks with Tree Regularization

08/14/2019
by   Mike Wu, et al.
0

Deep models have advanced prediction in many domains, but their lack of interpretability remains a key barrier to the adoption in many real world applications. There exists a large body of work aiming to help humans understand these black box functions to varying levels of granularity -- for example, through distillation, gradients, or adversarial examples. These methods however, all tackle interpretability as a separate process after training. In this work, we take a different approach and explicitly regularize deep models so that they are well-approximated by processes that humans can step-through in little time. Specifically, we train several families of deep neural networks to resemble compact, axis-aligned decision trees without significant compromises in accuracy. The resulting axis-aligned decision functions uniquely make tree regularized models easy for humans to interpret. Moreover, for situations in which a single, global tree is a poor estimator, we introduce a regional tree regularizer that encourages the deep model to resemble a compact, axis-aligned decision tree in predefined, human-interpretable contexts. Using intuitive toy examples as well as medical tasks for patients in critical care and with HIV, we demonstrate that this new family of tree regularizers yield models that are easier for humans to simulate than simpler L1 or L2 penalties without sacrificing predictive power.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/13/2019

Regional Tree Regularization for Interpretability in Black Box Models

The lack of interpretability remains a barrier to the adoption of deep n...
research
11/16/2017

Beyond Sparsity: Tree Regularization of Deep Models for Interpretability

The lack of interpretability remains a key barrier to the adoption of de...
research
05/26/2023

Improving Stability in Decision Tree Models

Owing to their inherently interpretable structure, decision trees are co...
research
04/10/2019

Enhancing Decision Tree based Interpretation of Deep Neural Networks through L1-Orthogonal Regularization

One obstacle that so far prevents the introduction of machine learning m...
research
12/28/2018

Improving the Interpretability of Deep Neural Networks with Knowledge Distillation

Deep Neural Networks have achieved huge success at a wide spectrum of ap...
research
06/09/2022

Distillation Decision Tree

Black-box machine learning models are criticized as lacking interpretabi...
research
10/26/2022

Convergence Rates of Oblique Regression Trees for Flexible Function Libraries

We develop a theoretical framework for the analysis of oblique decision ...

Please sign up or login with your details

Forgot password? Click here to reset