Learning Credible Models

11/08/2017
by   Jiaxuan Wang, et al.
0

In many settings, it is important that a model be capable of providing reasons for its predictions (i.e., the model must be interpretable). However, the model's reasoning may not conform with well-established knowledge. In such cases, while interpretable, the model lacks credibility. In this work, we formally define credibility in the linear setting and focus on techniques for learning models that are both accurate and credible. In particular, we propose a regularization penalty, expert yielded estimates (EYE), that incorporates expert knowledge about well-known relationships among covariates and the outcome of interest. We give both theoretical and empirical results comparing our proposed method to several other regularization techniques. Across a range of settings, experiments on both synthetic and real data show that models learned using the EYE penalty are significantly more credible than those learned using other penalties. Applied to a large-scale patient risk stratification task, our proposed technique results in a model whose top features overlap significantly with known clinical risk factors, while still achieving good predictive performance.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/22/2021

Learning Predictive and Interpretable Timeseries Summaries from ICU Data

Machine learning models that utilize patient data across time (rather th...
research
06/03/2021

NODE-GAM: Neural Generalized Additive Model for Interpretable Deep Learning

Deployment of machine learning models in real high-risk settings (e.g. h...
research
02/26/2019

Human-in-the-loop Active Covariance Learning for Improving Prediction in Small Data Sets

Learning predictive models from small high-dimensional data sets is a ke...
research
10/31/2022

Prediction of Network Covariates Using Edge and Node Attributes

In this work we consider the setting where many networks are observed on...
research
01/28/2020

Statistical Exploration of Relationships Between Routine and Agnostic Features Towards Interpretable Risk Characterization

As is typical in other fields of application of high throughput systems,...
research
07/02/2020

2DNMR data inversion using locally adapted multi-penalty regularization

A crucial issue in two-dimensional Nuclear Magnetic Resonance (NMR) is t...
research
06/07/2019

Relaxed Weight Sharing: Effectively Modeling Time-Varying Relationships in Clinical Time-Series

Recurrent neural networks (RNNs) are commonly applied to clinical time-s...

Please sign up or login with your details

Forgot password? Click here to reset