Identifying and Measuring Token-Level Sentiment Bias in Pre-trained Language Models with Prompts

04/15/2022
by   Apoorv Garg, et al.
0

Due to the superior performance, large-scale pre-trained language models (PLMs) have been widely adopted in many aspects of human society. However, we still lack effective tools to understand the potential bias embedded in the black-box models. Recent advances in prompt tuning show the possibility to explore the internal mechanism of the PLMs. In this work, we propose two token-level sentiment tests: Sentiment Association Test (SAT) and Sentiment Shift Test (SST) which utilize the prompt as a probe to detect the latent bias in the PLMs. Our experiments on the collection of sentiment datasets show that both SAT and SST can identify sentiment bias in PLMs and SST is able to quantify the bias. The results also suggest that fine-tuning can possibly augment the existing bias in PLMs.

READ FULL TEXT
research
06/04/2023

Exposing Bias in Online Communities through Large-Scale Language Models

Progress in natural language generation research has been shaped by the ...
research
06/06/2023

An Empirical Analysis of Parameter-Efficient Methods for Debiasing Pre-Trained Language Models

The increasingly large size of modern pretrained language models not onl...
research
10/06/2020

On the Branching Bias of Syntax Extracted from Pre-trained Language Models

Many efforts have been devoted to extracting constituency trees from pre...
research
06/07/2023

Soft-prompt Tuning for Large Language Models to Evaluate Bias

Prompting large language models has gained immense popularity in recent ...
research
04/06/2020

"You are grounded!": Latent Name Artifacts in Pre-trained Language Models

Pre-trained language models (LMs) may perpetuate biases originating in t...
research
06/08/2023

Bias Against 93 Stigmatized Groups in Masked Language Models and Downstream Sentiment Classification Tasks

The rapid deployment of artificial intelligence (AI) models demands a th...
research
05/08/2023

Diffusion Theory as a Scalpel: Detecting and Purifying Poisonous Dimensions in Pre-trained Language Models Caused by Backdoor or Bias

Pre-trained Language Models (PLMs) may be poisonous with backdoors or bi...

Please sign up or login with your details

Forgot password? Click here to reset