Analogies and Feature Attributions for Model Agnostic Explanation of Similarity Learners

Post-hoc explanations for black box models have been studied extensively in classification and regression settings. However, explanations for models that output similarity between two inputs have received comparatively lesser attention. In this paper, we provide model agnostic local explanations for similarity learners applicable to tabular and text data. We first propose a method that provides feature attributions to explain the similarity between a pair of inputs as determined by a black box similarity learner. We then propose analogies as a new form of explanation in machine learning. Here the goal is to identify diverse analogous pairs of examples that share the same level of similarity as the input pair and provide insight into (latent) factors underlying the model's prediction. The selection of analogies can optionally leverage feature attributions, thus connecting the two forms of explanation while still maintaining complementarity. We prove that our analogy objective function is submodular, making the search for good-quality analogies efficient. We apply the proposed approaches to explain similarities between sentences as predicted by a state-of-the-art sentence encoder, and between patients in a healthcare utilization application. Efficacy is measured through quantitative evaluations, a careful user study, and examples of explanations.

READ FULL TEXT

page 6

page 19

research
05/18/2023

BELLA: Black box model Explanations by Local Linear Approximations

In recent years, understanding the decision-making process of black-box ...
research
10/28/2021

Explaining Latent Representations with a Corpus of Examples

Modern machine learning models are complicated. Most of them rely on con...
research
03/11/2020

Building and Interpreting Deep Similarity Models

Many learning algorithms such as kernel machines, nearest neighbors, clu...
research
06/19/2023

Explaining the Model and Feature Dependencies by Decomposition of the Shapley Value

Shapley values have become one of the go-to methods to explain complex m...
research
10/20/2022

Toward Multiple Specialty Learners for Explaining GNNs via Online Knowledge Distillation

Graph Neural Networks (GNNs) have become increasingly ubiquitous in nume...
research
11/11/2020

Deja vu from the SVM Era: Example-based Explanations with Outlier Detection

Understanding the features that contributed to a prediction is important...
research
12/02/2019

EMAP: Explanation by Minimal Adversarial Perturbation

Modern instance-based model-agnostic explanation methods (LIME, SHAP, L2...

Please sign up or login with your details

Forgot password? Click here to reset