Transparent Model Distillation

01/26/2018
by   Sarah Tan, et al.
0

Model distillation was originally designed to distill knowledge from a large, complex teacher model to a faster, simpler student model without significant loss in prediction accuracy. We investigate model distillation for another goal -- transparency -- investigating if fully-connected neural networks can be distilled into models that are transparent or interpretable in some sense. Our teacher models are multilayer perceptrons, and we try two types of student models: (1) tree-based generalized additive models (GA2Ms), a type of boosted, short tree (2) gradient boosted trees (GBTs). More transparent student models are forthcoming. Our results are not yet conclusive. GA2Ms show some promise for distilling binary classification teachers, but not yet regression. GBTs are not "directly" interpretable but may be promising for regression teachers. GA2M models may provide a computationally viable alternative to additive decomposition methods for global function approximation.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/17/2017

Detecting Bias in Black-Box Models Using Transparent Model Distillation

Black-box risk scoring models permeate our lives, yet are typically prop...
research
11/07/2017

Moonshine: Distilling with Cheap Convolutions

Model distillation compresses a trained machine learning model, such as ...
research
01/21/2021

Distilling Interpretable Models into Human-Readable Code

The goal of model distillation is to faithfully transfer teacher model k...
research
11/22/2022

A Generic Approach for Reproducible Model Distillation

Model distillation has been a popular method for producing interpretable...
research
08/22/2018

Approximation Trees: Statistical Stability in Model Distillation

This paper examines the stability of learned explanations for black-box ...
research
04/28/2023

How to address monotonicity for model risk management?

In this paper, we study the problem of establishing the accountability a...
research
06/04/2021

Churn Reduction via Distillation

In real-world systems, models are frequently updated as more data become...

Please sign up or login with your details

Forgot password? Click here to reset