Margin-distancing for safe model explanation

02/23/2022
by   Tom Yan, et al.
0

The growing use of machine learning models in consequential settings has highlighted an important and seemingly irreconcilable tension between transparency and vulnerability to gaming. While this has sparked sizable debate in legal literature, there has been comparatively less technical study of this contention. In this work, we propose a clean-cut formulation of this tension and a way to make the tradeoff between transparency and gaming. We identify the source of gaming as being points close to the decision boundary of the model. And we initiate an investigation on how to provide example-based explanations that are expansive and yet consistent with a version space that is sufficiently uncertain with respect to the boundary points' labels. Finally, we furnish our theoretical results with empirical investigations of this tradeoff on real-world datasets.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/20/2017

The Promise and Peril of Human Evaluation for Model Interpretability

Transparency, user trust, and human comprehension are popular ethical mo...
research
05/18/2022

One Explanation to Rule them All – Ensemble Consistent Explanations

Transparency is a major requirement of modern AI based decision making s...
research
11/19/2018

On Human Predictions with Explanations and Predictions of Machine Learning Models: A Case Study on Deception Detection

Humans are the final decision makers in critical tasks that involve ethi...
research
02/08/2022

Machine Explanations and Human Understanding

Explanations are hypothesized to improve human understanding of machine ...
research
06/11/2023

On Minimizing the Impact of Dataset Shifts on Actionable Explanations

The Right to Explanation is an important regulatory principle that allow...
research
06/22/2020

Improving LIME Robustness with Smarter Locality Sampling

Explainability algorithms such as LIME have enabled machine learning sys...
research
06/14/2021

iNNformant: Boundary Samples as Telltale Watermarks

Boundary samples are special inputs to artificial neural networks crafte...

Please sign up or login with your details

Forgot password? Click here to reset