On quantitative aspects of model interpretability

07/15/2020
by   An-phi Nguyen, et al.
0

Despite the growing body of work in interpretable machine learning, it remains unclear how to evaluate different explainability methods without resorting to qualitative assessment and user-studies. While interpretability is an inherently subjective matter, previous works in cognitive science and epistemology have shown that good explanations do possess aspects that can be objectively judged apart from fidelity), such assimplicity and broadness. In this paper we propose a set of metrics to programmatically evaluate interpretability methods along these dimensions. In particular, we argue that the performance of methods along these dimensions can be orthogonally imputed to two conceptual parts, namely the feature extractor and the actual explainability method. We experimentally validate our metrics on different benchmark tasks and show how they can be used to guide a practitioner in the selection of the most appropriate method for the task at hand.

READ FULL TEXT

page 5

page 6

page 14

research
11/02/2021

Designing Inherently Interpretable Machine Learning Models

Interpretable machine learning (IML) becomes increasingly important in h...
research
04/17/2019

Explainability in Human-Agent Systems

This paper presents a taxonomy of explainability in Human-Agent Systems....
research
02/09/2019

Assessing the Local Interpretability of Machine Learning Models

The increasing adoption of machine learning tools has led to calls for a...
research
11/24/2019

A psychophysics approach for quantitative comparison of interpretable computer vision models

The field of transparent Machine Learning (ML) has contributed many nove...
research
11/20/2017

The Promise and Peril of Human Evaluation for Model Interpretability

Transparency, user trust, and human comprehension are popular ethical mo...
research
03/04/2022

A Typology to Explore and Guide Explanatory Interactive Machine Learning

Recently, more and more eXplanatory Interactive machine Learning (XIL) m...
research
04/21/2019

Beyond Explainability: Leveraging Interpretability for Improved Adversarial Learning

In this study, we propose the leveraging of interpretability for tasks b...

Please sign up or login with your details

Forgot password? Click here to reset