Examining the Causal Effect of First Names on Language Models: The Case of Social Commonsense Reasoning

06/01/2023
by   Sullam Jeoung, et al.
0

As language models continue to be integrated into applications of personal and societal relevance, ensuring these models' trustworthiness is crucial, particularly with respect to producing consistent outputs regardless of sensitive attributes. Given that first names may serve as proxies for (intersectional) socio-demographic representations, it is imperative to examine the impact of first names on commonsense reasoning capabilities. In this paper, we study whether a model's reasoning given a specific input differs based on the first names provided. Our underlying assumption is that the reasoning about Alice should not differ from the reasoning about James. We propose and implement a controlled experimental framework to measure the causal effect of first names on commonsense reasoning, enabling us to distinguish between model predictions due to chance and caused by actual factors of interest. Our results indicate that the frequency of first names has a direct effect on model prediction, with less frequent names yielding divergent predictions compared to more frequent names. To gain insights into the internal mechanisms of models that are contributing to these behaviors, we also conduct an in-depth explainable analysis. Overall, our findings suggest that to ensure model robustness, it is essential to augment datasets with more diverse first names during the configuration stage.

READ FULL TEXT
research
10/01/2021

Low Frequency Names Exhibit Bias and Overfitting in Contextualizing Language Models

We use a dataset of U.S. first names with labels based on predominant ge...
research
05/26/2023

Nichelle and Nancy: The Influence of Demographic Attributes and Tokenization Length on First Name Biases

Through the use of first name substitution experiments, prior research h...
research
04/22/2023

Dialectical language model evaluation: An initial appraisal of the commonsense spatial reasoning abilities of LLMs

Language models have become very popular recently and many claims have b...
research
06/16/2022

PInKS: Preconditioned Commonsense Inference with Minimal Supervision

Reasoning with preconditions such as "glass can be used for drinking wat...
research
06/04/2023

Probing Physical Reasoning with Counter-Commonsense Context

In this study, we create a CConS (Counter-commonsense Contextual Size co...
research
10/21/2022

A Causal Framework to Quantify the Robustness of Mathematical Reasoning with Language Models

We have recently witnessed a number of impressive results on hard mathem...
research
10/27/2020

It's All in the Name: A Character Based Approach To Infer Religion

Demographic inference from text has received a surge of attention in the...

Please sign up or login with your details

Forgot password? Click here to reset