Evaluating Explainers via Perturbation

06/05/2019
by   Minh N. Vu, et al.
0

Due to high complexity of many modern machine learning models such as deep convolutional networks, understanding the cause of model's prediction is critical. Many explainers have been designed to give us more insights on the decision of complex classifiers. However, there is no common ground on evaluating the quality of different classification methods. Motivated by the needs for comprehensive evaluation, we introduce the c-Eval metric and the corresponding framework to quantify the explainer's quality on feature-based explainers of machine learning image classifiers. Given a prediction and the corresponding explanation on that prediction, c-Eval is the minimum-power perturbation that successfully alters the prediction while keeping the explanation's features unchanged. We also provide theoretical analysis linking the proposed parameter with the portion of predicted object covered by the explanation. Using a heuristic approach, we introduce the c-Eval plot, which not only displays a strong connection between c-Eval and explainers' quality, but also serves as a low-complexity approach of assessing explainers. We finally conduct extensive experiments of explainers on three different datasets in order to support the adoption of c-Eval in evaluating explainers' performance.

READ FULL TEXT

page 2

page 14

page 15

page 16

research
05/01/2020

Evaluating and Aggregating Feature-based Model Explanations

A feature-based model explanation denotes how much each input feature co...
research
06/16/2022

Constrained Submodular Optimization for Vaccine Design

Advances in machine learning have enabled the prediction of immune syste...
research
12/22/2020

Ordered Counterfactual Explanation by Mixed-Integer Linear Optimization

Post-hoc explanation methods for machine learning models have been widel...
research
09/23/2020

Explaining Chemical Toxicity using Missing Features

Chemical toxicity prediction using machine learning is important in drug...
research
07/26/2023

The Co-12 Recipe for Evaluating Interpretable Part-Prototype Image Classifiers

Interpretable part-prototype models are computer vision models that are ...
research
06/08/2023

AMEE: A Robust Framework for Explanation Evaluation in Time Series Classification

This paper aims to provide a framework to quantitatively evaluate and ra...

Please sign up or login with your details

Forgot password? Click here to reset