A Learning Theoretic Perspective on Local Explainability

11/02/2020
by   Jeffrey Li, et al.
0

In this paper, we explore connections between interpretable machine learning and learning theory through the lens of local approximation explanations. First, we tackle the traditional problem of performance generalization and bound the test-time accuracy of a model using a notion of how locally explainable it is. Second, we explore the novel problem of explanation generalization which is an important concern for a growing class of finite sample-based local approximation explanations. Finally, we validate our theoretical results empirically and show that they reflect what can be seen in practice.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/22/2023

TsSHAP: Robust model agnostic feature-based explainability for time series forecasting

A trustworthy machine learning model should be accurate as well as expla...
research
02/19/2020

Learning Global Transparent Models from Local Contrastive Explanations

There is a rich and growing literature on producing local point wise con...
research
01/21/2020

Adequate and fair explanations

Explaining sophisticated machine-learning based systems is an important ...
research
05/23/2020

Towards Analogy-Based Explanations in Machine Learning

Principles of analogical reasoning have recently been applied in the con...
research
02/01/2022

Framework for Evaluating Faithfulness of Local Explanations

We study the faithfulness of an explanation system to the underlying pre...
research
03/25/2023

Learning with Explanation Constraints

While supervised learning assumes the presence of labeled data, we may h...
research
06/24/2022

Analyzing the Effects of Classifier Lipschitzness on Explainers

Machine learning methods are getting increasingly better at making predi...

Please sign up or login with your details

Forgot password? Click here to reset