Marked Attribute Bias in Natural Language Inference

09/28/2021
by   Hillary Dawkins, et al.
0

Reporting and providing test sets for harmful bias in NLP applications is essential for building a robust understanding of the current problem. We present a new observation of gender bias in a downstream NLP application: marked attribute bias in natural language inference. Bias in downstream applications can stem from training data, word embeddings, or be amplified by the model in use. However, focusing on biased word embeddings is potentially the most impactful first step due to their universal nature. Here we seek to understand how the intrinsic properties of word embeddings contribute to this observed marked attribute effect, and whether current post-processing methods address the bias successfully. An investigation of the current debiasing landscape reveals two open problems: none of the current debiased embeddings mitigate the marked attribute error, and none of the intrinsic bias measures are predictive of the marked attribute effect. By noticing that a new type of intrinsic bias measure correlates meaningfully with the marked attribute effect, we propose a new postprocessing debiasing scheme for static word embeddings. The proposed method applied to existing embeddings achieves new best results on the marked attribute bias test set. See https://github.com/hillary-dawkins/MAB.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/14/2021

[RE] Double-Hard Debias: Tailoring Word Embeddings for Gender Bias Mitigation

Despite widespread use in natural language processing (NLP) tasks, word ...
research
10/31/2020

Evaluating Bias In Dutch Word Embeddings

Recent research in Natural Language Processing has revealed that word em...
research
12/31/2020

Intrinsic Bias Metrics Do Not Correlate with Application Bias

Natural Language Processing (NLP) systems learn harmful societal biases ...
research
06/02/2020

Nurse is Closer to Woman than Surgeon? Mitigating Gender-Biased Proximities in Word Embeddings

Word embeddings are the standard model for semantic and syntactic repres...
research
08/25/2019

On Measuring and Mitigating Biased Inferences of Word Embeddings

Word embeddings carry stereotypical connotations from the text they are ...
research
11/10/2022

ADEPT: A DEbiasing PrompT Framework

Several works have proven that finetuning is an applicable approach for ...
research
12/12/2016

ConceptNet 5.5: An Open Multilingual Graph of General Knowledge

Machine learning about language can be improved by supplying it with spe...

Please sign up or login with your details

Forgot password? Click here to reset