Explainers in the Wild: Making Surrogate Explainers Robust to Distortions through Perception

02/22/2021
by   Alexander Hepburn, et al.
2

Explaining the decisions of models is becoming pervasive in the image processing domain, whether it is by using post-hoc methods or by creating inherently interpretable models. While the widespread use of surrogate explainers is a welcome addition to inspect and understand black-box models, assessing the robustness and reliability of the explanations is key for their success. Additionally, whilst existing work in the explainability field proposes various strategies to address this problem, the challenges of working with data in the wild is often overlooked. For instance, in image classification, distortions to images can not only affect the predictions assigned by the model, but also the explanation. Given a clean and a distorted version of an image, even if the prediction probabilities are similar, the explanation may still be different. In this paper we propose a methodology to evaluate the effect of distortions in explanations by embedding perceptual distances that tailor the neighbourhoods used to training surrogate explainers. We also show that by operating in this way, we can make the explanations more robust to distortions. We generate explanations for images in the Imagenet-C dataset and demonstrate how using a perceptual distances in the surrogate explainer creates more coherent explanations for the distorted and reference images.

READ FULL TEXT

page 1

page 4

research
10/29/2019

bLIMEy: Surrogate Prediction Explanations Beyond LIME

Surrogate explainers of black-box machine learning predictions are of pa...
research
08/08/2022

Sampling Based On Natural Image Statistics Improves Local Surrogate Explainers

Many problems in computer vision have recently been tackled using models...
research
11/17/2021

Uncertainty Quantification of Surrogate Explanations: an Ordinal Consensus Approach

Explainability of black-box machine learning models is crucial, in parti...
research
12/17/2021

Global explainability in aligned image modalities

Deep learning (DL) models are very effective on many computer vision pro...
research
06/10/2021

On the overlooked issue of defining explanation objectives for local-surrogate explainers

Local surrogate approaches for explaining machine learning model predict...
research
08/27/2021

This looks more like that: Enhancing Self-Explaining Models by Prototypical Relevance Propagation

Current machine learning models have shown high efficiency in solving a ...

Please sign up or login with your details

Forgot password? Click here to reset