Causally Estimating the Sensitivity of Neural NLP Models to Spurious Features

10/14/2021
by   Yunxiang Zhang, et al.
33

Recent work finds modern natural language processing (NLP) models relying on spurious features for prediction. Mitigating such effects is thus important. Despite this need, there is no quantitative measure to evaluate or compare the effects of different forms of spurious features in NLP. We address this gap in the literature by quantifying model sensitivity to spurious features with a causal estimand, dubbed CENT, which draws on the concept of average treatment effect from the causality literature. By conducting simulations with four prominent NLP models – TextRNN, BERT, RoBERTa and XLNet – we rank the models against their sensitivity to artificial injections of eight spurious features. We further hypothesize and validate that models that are more sensitive to a spurious feature will be less robust against perturbations with this feature during inference. Conversely, data augmentation with this feature improves robustness to similar perturbations. We find statistically significant inverse correlations between sensitivity and robustness, providing empirical support for our hypothesis.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/02/2021

Causal Inference in Natural Language Processing: Estimation, Prediction, Interpretation and Beyond

A fundamental goal of scientific research is to learn about causal relat...
research
05/19/2021

Balancing Robustness and Sensitivity using Feature Contrastive Learning

It is generally believed that robust training of extremely large network...
research
03/22/2022

Mouse Sensitivity Effects in First-Person Targeting Tasks

Despite billions of hours of play and copious discussion online, mouse s...
research
05/26/2023

Controlling Learned Effects to Reduce Spurious Correlations in Text Classifiers

To address the problem of NLP classifiers learning spurious correlations...
research
05/15/2023

Estimating the Causal Effects of Natural Logic Features in Neural NLI Models

Rigorous evaluation of the causal effects of semantic features on langua...
research
11/13/2022

Language Model Classifier Aligns Better with Physician Word Sensitivity than XGBoost on Readmission Prediction

Traditional evaluation metrics for classification in natural language pr...
research
10/14/2021

Identifying and Mitigating Spurious Correlations for Improving Robustness in NLP Models

Recently, NLP models have achieved remarkable progress across a variety ...

Please sign up or login with your details

Forgot password? Click here to reset