XPROAX-Local explanations for text classification with progressive neighborhood approximation

09/30/2021
by   Yi Cai, et al.
3

The importance of the neighborhood for training a local surrogate model to approximate the local decision boundary of a black box classifier has been already highlighted in the literature. Several attempts have been made to construct a better neighborhood for high dimensional data, like texts, by using generative autoencoders. However, existing approaches mainly generate neighbors by selecting purely at random from the latent space and struggle under the curse of dimensionality to learn a good local decision boundary. To overcome this problem, we propose a progressive approximation of the neighborhood using counterfactual instances as initial landmarks and a careful 2-stage sampling approach to refine counterfactuals and generate factuals in the neighborhood of the input instance to be explained. Our work focuses on textual data and our explanations consist of both word-level explanations from the original instance (intrinsic) and the neighborhood (extrinsic) and factual- and counterfactual-instances discovered during the neighborhood generation process that further reveal the effect of altering certain parts in the input text. Our experiments on real-world datasets demonstrate that our method outperforms the competitors in terms of usefulness and stability (for the qualitative part) and completeness, compactness and correctness (for the quantitative part).

READ FULL TEXT
research
02/11/2023

Explaining text classifiers through progressive neighborhood approximation with realistic samples

The importance of neighborhood construction in local explanation methods...
research
06/24/2021

Human-in-the-loop model explanation via verbatim boundary identification in generated neighborhoods

The black-box nature of machine learning models limits their use in case...
research
06/19/2018

Defining Locality for Surrogates in Post-hoc Interpretablity

Local surrogate models, to approximate the local decision boundary of a ...
research
10/22/2021

ReLACE: Reinforcement Learning Agent for Counterfactual Explanations of Arbitrary Predictive Models

The demand for explainable machine learning (ML) models has been growing...
research
07/04/2019

Explaining Predictions from Tree-based Boosting Ensembles

Understanding how "black-box" models arrive at their predictions has spa...
research
06/14/2020

Explaining Predictions by Approximating the Local Decision Boundary

Constructing accurate model-agnostic explanations for opaque machine lea...
research
05/19/2023

Latent Imitator: Generating Natural Individual Discriminatory Instances for Black-Box Fairness Testing

Machine learning (ML) systems have achieved remarkable performance acros...

Please sign up or login with your details

Forgot password? Click here to reset