Goodhart's Law Applies to NLP's Explanation Benchmarks

08/28/2023
by   Jennifer Hsia, et al.
0

Despite the rising popularity of saliency-based explanations, the research community remains at an impasse, facing doubts concerning their purpose, efficacy, and tendency to contradict each other. Seeking to unite the community's efforts around common goals, several recent works have proposed evaluation metrics. In this paper, we critically examine two sets of metrics: the ERASER metrics (comprehensiveness and sufficiency) and the EVAL-X metrics, focusing our inquiry on natural language processing. First, we show that we can inflate a model's comprehensiveness and sufficiency scores dramatically without altering its predictions or explanations on in-distribution test inputs. Our strategy exploits the tendency for extracted explanations and their complements to be "out-of-support" relative to each other and in-distribution inputs. Next, we demonstrate that the EVAL-X metrics can be inflated arbitrarily by a simple method that encodes the label, even though EVAL-X is precisely motivated to address such exploits. Our results raise doubts about the ability of current metrics to guide explainability research, underscoring the need for a broader reassessment of what precisely these metrics are intended to capture.

READ FULL TEXT
research
11/25/2022

Testing the effectiveness of saliency-based explainability in NLP using randomized survey-based experiments

As the applications of Natural Language Processing (NLP) in sensitive ar...
research
12/20/2022

BMX: Boosting Machine Translation Metrics with Explainability

State-of-the-art machine translation evaluation metrics are based on bla...
research
10/07/2022

Quantitative Metrics for Evaluating Explanations of Video DeepFake Detectors

The proliferation of DeepFake technology is a rising challenge in today'...
research
10/18/2019

Toward Metrics for Differentiating Out-of-Distribution Sets

Vanilla CNNs, as uncalibrated classifiers, suffer from classifying out-o...
research
09/02/2023

Explainability for Large Language Models: A Survey

Large language models (LLMs) have demonstrated impressive capabilities i...
research
06/07/2023

Dear XAI Community, We Need to Talk! Fundamental Misconceptions in Current XAI Research

Despite progress in the field, significant parts of current XAI research...
research
03/26/2023

An Evaluation of Memory Optimization Methods for Training Neural Networks

As models continue to grow in size, the development of memory optimizati...

Please sign up or login with your details

Forgot password? Click here to reset