DeepAI AI Chat
Log In Sign Up

Towards A Rigorous Science of Interpretable Machine Learning

02/28/2017
by   Finale Doshi-Velez, et al.
0

As machine learning systems become ubiquitous, there has been a surge of interest in interpretable machine learning: systems that provide explanation for their outputs. These explanations are often used to qualitatively assess other criteria such as safety or non-discrimination. However, despite the interest in interpretability, there is very little consensus on what interpretable machine learning is and how it should be measured. In this position paper, we first define interpretability and describe when interpretability is needed (and when it is not). Next, we suggest a taxonomy for rigorous evaluation and expose open questions towards a more rigorous science of interpretable machine learning.

READ FULL TEXT

page 1

page 2

page 3

page 4

11/27/2017

Proceedings of NIPS 2017 Symposium on Interpretable Machine Learning

This is the Proceedings of NIPS 2017 Symposium on Interpretable Machine ...
06/20/2018

Interpretable to Whom? A Role-based Model for Analyzing Interpretable Machine Learning Systems

Several researchers have argued that a machine learning system's interpr...
11/20/2017

"I know it when I see it". Visualization and Intuitive Interpretability

Most research on the interpretability of machine learning systems focuse...
01/31/2019

An Evaluation of the Human-Interpretability of Explanation

Recent years have seen a boom in interest in machine learning systems th...
03/28/2021

Explaining Representation by Mutual Information

Science is used to discover the law of world. Machine learning can be us...
04/24/2019

The Scientific Method in the Science of Machine Learning

In the quest to align deep learning with the sciences to address calls f...
11/11/2022

Rethinking Log Odds: Linear Probability Modelling and Expert Advice in Interpretable Machine Learning

We introduce a family of interpretable machine learning models, with two...