Defining and Evaluating Fair Natural Language Generation

07/28/2020
by   Catherine Yeo, et al.
0

Our work focuses on the biases that emerge in the natural language generation (NLG) task of sentence completion. In this paper, we introduce a framework of fairness for NLG followed by an evaluation of gender biases in two state-of-the-art language models. Our analysis provides a theoretical formulation for biases in NLG and empirical evidence that existing language generation models embed gender bias.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/12/2021

Evaluating Gender Bias in Natural Language Inference

Gender-bias stereotypes have recently raised significant ethical concern...
research
02/08/2021

How True is GPT-2? An Empirical Analysis of Intersectional Occupational Biases

The capabilities of natural language models trained on large-scale data ...
research
05/22/2023

Should We Attend More or Less? Modulating Attention for Fairness

The abundance of annotated data in natural language processing (NLP) pos...
research
05/18/2023

Exploiting Biased Models to De-bias Text: A Gender-Fair Rewriting Model

Natural language generation models reproduce and often amplify the biase...
research
09/11/2023

Detecting Natural Language Biases with Prompt-based Learning

In this project, we want to explore the newly emerging field of prompt e...
research
02/27/2023

Inseq: An Interpretability Toolkit for Sequence Generation Models

Past work in natural language processing interpretability focused mainly...
research
10/05/2022

GAPX: Generalized Autoregressive Paraphrase-Identification X

Paraphrase Identification is a fundamental task in Natural Language Proc...

Please sign up or login with your details

Forgot password? Click here to reset