Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings

07/21/2016
by   Tolga Bolukbasi, et al.
0

The blind application of machine learning runs the risk of amplifying biases present in data. Such a danger is facing us with word embedding, a popular framework to represent text data as vectors which has been used in many machine learning and natural language processing tasks. We show that even word embeddings trained on Google News articles exhibit female/male gender stereotypes to a disturbing extent. This raises concerns because their widespread use, as we describe, often tends to amplify these biases. Geometrically, gender bias is first shown to be captured by a direction in the word embedding. Second, gender neutral words are shown to be linearly separable from gender definition words in the word embedding. Using these properties, we provide a methodology for modifying an embedding to remove gender stereotypes, such as the association between between the words receptionist and female, while maintaining desired associations such as between the words queen and female. We define metrics to quantify both direct and indirect gender biases in embeddings, and develop algorithms to "debias" the embedding. Using crowd-worker evaluation as well as standard benchmarks, we empirically demonstrate that our algorithms significantly reduce gender bias in embeddings while preserving the its useful properties such as the ability to cluster related concepts and to solve analogy tasks. The resulting embeddings can be used in applications without amplifying gender bias.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/18/2019

Evaluating the Underlying Gender Bias in Contextualized Word Embeddings

Gender bias is highly impacting natural language processing applications...
research
06/20/2016

Quantifying and Reducing Stereotypes in Word Embeddings

Machine learning algorithms are optimized to model statistical propertie...
research
06/07/2022

Gender Bias in Word Embeddings: A Comprehensive Analysis of Frequency, Syntax, and Semantics

The statistical regularities in language corpora encode well-known socia...
research
06/06/2020

ValNorm: A New Word Embedding Intrinsic Evaluation Method Reveals Valence Biases are Consistent Across Languages and Over Decades

Word embeddings learn implicit biases from linguistic regularities captu...
research
11/22/2017

Word Embeddings Quantify 100 Years of Gender and Ethnic Stereotypes

Word embeddings use vectors to represent words such that the geometry be...
research
08/25/2016

Semantics derived automatically from language corpora contain human-like biases

Artificial intelligence and machine learning are in a period of astoundi...
research
05/28/2019

Algorithmic Bias and the Biases of the Bias Catchers

Concerns about gender bias have captured most of the attention in the AI...

Please sign up or login with your details

Forgot password? Click here to reset