What are the biases in my word embedding?

12/20/2018
by   Nathaniel Swinger, et al.
0

This paper presents an algorithm for enumerating biases in word embeddings. The algorithm exposes a large number of offensive associations related to sensitive features such as race and gender on publicly available embeddings, including a supposedly "debiased" embedding. These embedded biases are concerning in light of the widespread use of word embeddings. The associations are identified by geometric patterns in word embeddings that run parallel between people's names and common lower-case words and phrases. The algorithm is highly unsupervised: it does not even require the sensitive groups (such as gender or race) to be pre-specified. This is desirable because it may not always be easy to identify all vulnerable groups a priori, and because it makes it easier to identify biases against intersectional groups, which depend on combinations of sensitive features. The inputs to our algorithm are a list of target tokens, e.g. names, and a word embedding, and the outputs are a number of Word Embedding Association Tests (WEATs) that capture various biases present in the data. We illustrate the utility of our approach on publicly available word embeddings and lists of names, and evaluate its output using crowdsourcing. We also show how removing names may not remove potential proxy bias.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/14/2019

Conceptor Debiasing of Word Representations Evaluated on WEAT

Bias in word embeddings such as Word2Vec has been widely investigated, a...
research
03/01/2021

WordBias: An Interactive Visual Tool for Discovering Intersectional Biases Encoded in Word Embeddings

Intersectional bias is a bias caused by an overlap of multiple social fa...
research
06/15/2023

A Bayesian approach to uncertainty in word embedding bias estimation

Multiple measures, such as WEAT or MAC, attempt to quantify the magnitud...
research
04/17/2020

Wide range screening of algorithmic bias in word embedding models using large sentiment lexicons reveals underreported bias types

Concerns about gender bias in word embedding models have captured substa...
research
01/20/2022

Regional Negative Bias in Word Embeddings Predicts Racial Animus–but only via Name Frequency

The word embedding association test (WEAT) is an important method for me...
research
09/01/2018

Why is unsupervised alignment of English embeddings from different algorithms so hard?

This paper presents a challenge to the community: Generative adversarial...
research
08/18/2019

Understanding Undesirable Word Embedding Associations

Word embeddings are often criticized for capturing undesirable word asso...

Please sign up or login with your details

Forgot password? Click here to reset