Should We Attend More or Less? Modulating Attention for Fairness

05/22/2023
by   Abdelrahman Zayed, et al.
0

The abundance of annotated data in natural language processing (NLP) poses both opportunities and challenges. While it enables the development of high-performing models for a variety of tasks, it also poses the risk of models learning harmful biases from the data, such as gender stereotypes. In this work, we investigate the role of attention, a widely-used technique in current state-of-the-art NLP models, in the propagation of social biases. Specifically, we study the relationship between the entropy of the attention distribution and the model's performance and fairness. We then propose a novel method for modulating attention weights to improve model fairness after training. Since our method is only applied post-training and pre-inference, it is an intra-processing method and is, therefore, less computationally expensive than existing in-processing and pre-processing approaches. Our results show an increase in fairness and minimal performance loss on different text classification and generation tasks using language models of varying sizes. WARNING: This work uses language that is offensive.

READ FULL TEXT

page 7

page 14

research
07/28/2020

Defining and Evaluating Fair Natural Language Generation

Our work focuses on the biases that emerge in the natural language gener...
research
10/04/2020

Fairness in Machine Learning: A Survey

As Machine Learning technologies become increasingly used in contexts th...
research
11/20/2022

Deep Learning on a Healthy Data Diet: Finding Important Examples for Fairness

Data-driven predictive solutions predominant in commercial applications ...
research
02/28/2023

Automatic Scoring of Dream Reports' Emotional Content with Large Language Models

In the field of dream research, the study of dream content typically rel...
research
09/02/2023

Bias and Fairness in Large Language Models: A Survey

Rapid advancements of large language models (LLMs) have enabled the proc...
research
01/29/2021

Challenges in Automated Debiasing for Toxic Language Detection

Biased associations have been a challenge in the development of classifi...
research
05/25/2022

Perturbation Augmentation for Fairer NLP

Unwanted and often harmful social biases are becoming ever more salient ...

Please sign up or login with your details

Forgot password? Click here to reset