Square One Bias in NLP: Towards a Multi-Dimensional Exploration of the Research Manifold

06/20/2022
by   Sebastian Ruder, et al.
0

The prototypical NLP experiment trains a standard architecture on labeled English data and optimizes for accuracy, without accounting for other dimensions such as fairness, interpretability, or computational efficiency. We show through a manual classification of recent NLP research papers that this is indeed the case and refer to it as the square one experimental setup. We observe that NLP research often goes beyond the square one setup, e.g, focusing not only on accuracy, but also on fairness or interpretability, but typically only along a single dimension. Most work targeting multilinguality, for example, considers only accuracy; most work on fairness or interpretability considers only English; and so on. We show this through manual classification of recent NLP research papers and ACL Test-of-Time award recipients. Such one-dimensionality of most research means we are only exploring a fraction of the NLP research search space. We provide historical and recent examples of how the square one bias has led researchers to draw false conclusions or make unwise choices, point to promising yet unexplored directions on the research manifold, and make practical recommendations to enable more multi-dimensional research. We open-source the results of our annotations to enable further analysis at https://github.com/google-research/url-nlp

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/11/2023

Fair Enough: Standardizing Evaluation and Model Selection for Fairness Research in NLP

Modern NLP systems exhibit a range of biases, which a growing literature...
research
08/12/2020

The Language Interpretability Tool: Extensible, Interactive Visualizations and Analysis for NLP Models

We present the Language Interpretability Tool (LIT), an open-source plat...
research
11/08/2022

Bridging Fairness and Environmental Sustainability in Natural Language Processing

Fairness and environmental impact are important research directions for ...
research
02/25/2021

Are pre-trained text representations useful for multilingual and multi-dimensional language proficiency modeling?

Development of language proficiency models for non-native learners has b...
research
11/18/2018

Bayesian Modeling of Intersectional Fairness: The Variance of Bias

Intersectionality is a framework that analyzes how interlocking systems ...
research
06/21/2021

A Survey of Race, Racism, and Anti-Racism in NLP

Despite inextricable ties between race and language, little work has con...

Please sign up or login with your details

Forgot password? Click here to reset