A Generic Approach for Reproducible Model Distillation

11/22/2022
by   Yunzhe Zhou, et al.
0

Model distillation has been a popular method for producing interpretable machine learning. It uses an interpretable "student" model to mimic the predictions made by the black box "teacher" model. However, when the student model is sensitive to the variability of the data sets used for training, the corresponded interpretation is not reliable. Existing strategies stabilize model distillation by checking whether a large enough corpus of pseudo-data is generated to reliably reproduce student models, but methods to do so have so far been developed for a specific student model. In this paper, we develop a generic approach for stable model distillation based on central limit theorem for the average loss. We start with a collection of candidate student models and search for candidates that reasonably agree with the teacher. Then we construct a multiple testing framework to select a corpus size such that the consistent student model would be selected under different pseudo sample. We demonstrate the application of our proposed approach on three commonly used intelligible models: decision trees, falling rule lists and symbolic regression. Finally, we conduct simulation experiments on Mammographic Mass and Breast Cancer datasets and illustrate the testing procedure throughout a theoretical analysis with Markov process.

READ FULL TEXT

page 19

page 20

page 21

page 22

research
08/22/2018

Approximation Trees: Statistical Stability in Model Distillation

This paper examines the stability of learned explanations for black-box ...
research
10/30/2019

Distilling Black-Box Travel Mode Choice Model for Behavioral Interpretation

Machine learning has proved to be very successful for making predictions...
research
01/26/2018

Transparent Model Distillation

Model distillation was originally designed to distill knowledge from a l...
research
06/07/2021

Zero-Shot Knowledge Distillation from a Decision-Based Black-Box Model

Knowledge distillation (KD) is a successful approach for deep neural net...
research
05/25/2023

Triplet Knowledge Distillation

In Knowledge Distillation, the teacher is generally much larger than the...
research
10/17/2017

Detecting Bias in Black-Box Models Using Transparent Model Distillation

Black-box risk scoring models permeate our lives, yet are typically prop...
research
01/21/2021

Distilling Interpretable Models into Human-Readable Code

The goal of model distillation is to faithfully transfer teacher model k...

Please sign up or login with your details

Forgot password? Click here to reset