On the Limitations of Dataset Balancing: The Lost Battle Against Spurious Correlations

04/27/2022
by   Roy Schwartz, et al.
0

Recent work has shown that deep learning models in NLP are highly sensitive to low-level correlations between simple features and specific output labels, leading to overfitting and lack of generalization. To mitigate this problem, a common practice is to balance datasets by adding new instances or by filtering out "easy" instances (Sakaguchi et al., 2020), culminating in a recent proposal to eliminate single-word correlations altogether (Gardner et al., 2021). In this opinion paper, we identify that despite these efforts, increasingly-powerful models keep exploiting ever-smaller spurious correlations, and as a result even balancing all single-word features is insufficient for mitigating all of these correlations. In parallel, a truly balanced dataset may be bound to "throw the baby out with the bathwater" and miss important signal encoding common sense and world knowledge. We highlight several alternatives to dataset balancing, focusing on enhancing datasets with richer contexts, allowing models to abstain and interact with users, and turning from large-scale fine-tuning to zero- or few-shot setups.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/10/2021

Noise Stability Regularization for Improving BERT Fine-tuning

Fine-tuning pre-trained language models such as BERT has become a common...
research
06/21/2023

Towards Mitigating Spurious Correlations in the Wild: A Benchmark a more Realistic Dataset

Deep neural networks often exploit non-predictive features that are spur...
research
04/17/2021

Competency Problems: On Finding and Removing Artifacts in Language Data

Much recent work in NLP has documented dataset artifacts, bias, and spur...
research
05/11/2023

Exploring Zero and Few-shot Techniques for Intent Classification

Conversational NLU providers often need to scale to thousands of intent-...
research
04/13/2021

Detoxifying Language Models Risks Marginalizing Minority Voices

Language models (LMs) must be both safe and equitable to be responsibly ...
research
04/09/2022

Uninformative Input Features and Counterfactual Invariance: Two Perspectives on Spurious Correlations in Natural Language

Spurious correlations are a threat to the trustworthiness of natural lan...
research
11/07/2021

A Word on Machine Ethics: A Response to Jiang et al. (2021)

Ethics is one of the longest standing intellectual endeavors of humanity...

Please sign up or login with your details

Forgot password? Click here to reset