The Impact of Debiasing on the Performance of Language Models in Downstream Tasks is Underestimated

09/16/2023
by   Masahiro Kaneko, et al.
0

Pre-trained language models trained on large-scale data have learned serious levels of social biases. Consequently, various methods have been proposed to debias pre-trained models. Debiasing methods need to mitigate only discriminatory bias information from the pre-trained models, while retaining information that is useful for the downstream tasks. In previous research, whether useful information is retained has been confirmed by the performance of downstream tasks in debiased pre-trained models. On the other hand, it is not clear whether these benchmarks consist of data pertaining to social biases and are appropriate for investigating the impact of debiasing. For example in gender-related social biases, data containing female words (e.g. “she, female, woman”), male words (e.g. “he, male, man”), and stereotypical words (e.g. “nurse, doctor, professor”) are considered to be the most affected by debiasing. If there is not much data containing these words in a benchmark dataset for a target task, there is the possibility of erroneously evaluating the effects of debiasing. In this study, we compare the impact of debiasing on performance across multiple downstream tasks using a wide-range of benchmark datasets that containing female, male, and stereotypical words. Experiments show that the effects of debiasing are consistently underestimated across all tasks. Moreover, the effects of debiasing could be reliably evaluated by separately considering instances containing female, male, and stereotypical words than all of the instances in a benchmark dataset.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/07/2022

How Adversarial Robustness Transfers from Pre-training to Downstream Tasks

Given the rise of large-scale training regimes, adapting pre-trained mod...
research
12/11/2022

A Study of Slang Representation Methods

Warning: this paper contains content that may be offensive or upsetting....
research
06/08/2023

Bias Against 93 Stigmatized Groups in Masked Language Models and Downstream Sentiment Classification Tasks

The rapid deployment of artificial intelligence (AI) models demands a th...
research
09/02/2023

Studying the impacts of pre-training using ChatGPT-generated text on downstream tasks

In recent times, significant advancements have been witnessed in the fie...
research
10/06/2022

Debiasing isn't enough! – On the Effectiveness of Debiasing MLMs and their Social Biases in Downstream Tasks

We study the relationship between task-agnostic intrinsic and task-speci...
research
04/13/2023

Evaluation of Social Biases in Recent Large Pre-Trained Models

Large pre-trained language models are widely used in the community. Thes...
research
04/29/2021

MOROCCO: Model Resource Comparison Framework

The new generation of pre-trained NLP models push the SOTA to the new li...

Please sign up or login with your details

Forgot password? Click here to reset