OSCaR: Orthogonal Subspace Correction and Rectification of Biases in Word Embeddings

06/30/2020
by   Sunipa Dev, et al.
0

Language representations are known to carry stereotypical biases and, as a result, lead to biased predictions in downstream tasks. While existing methods are effective at mitigating biases by linear projection, such methods are too aggressive: they not only remove bias, but also erase valuable information from word embeddings. We develop new measures for evaluating specific information retention that demonstrate the tradeoff between bias removal and information retention. To address this challenge, we propose OSCaR (Orthogonal Subspace Correction and Rectification), a bias-mitigating method that focuses on disentangling biased associations between concepts instead of removing concepts wholesale. Our experiments on gender biases show that OSCaR is a well-balanced approach that ensures that semantic information is retained in the embeddings and bias is also effectively mitigated.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/21/2021

Using Adversarial Debiasing to Remove Bias from Word Embeddings

Word Embeddings have been shown to contain the societal biases present i...
research
10/26/2020

PowerTransformer: Unsupervised Controllable Revision for Biased Language Correction

Unconscious biases continue to be prevalent in modern text and media, ca...
research
10/31/2019

Probabilistic Bias Mitigation in Word Embeddings

It has been shown that word embeddings derived from large corpora tend t...
research
09/06/2018

Turning a Blind Eye: Explicit Removal of Biases and Variation from Deep Neural Network Embeddings

Neural networks achieve the state-of-the-art in image classification tas...
research
02/20/2020

Measuring Social Biases in Grounded Vision and Language Embeddings

We generalize the notion of social biases from language embeddings to gr...
research
05/23/2019

Fair is Better than Sensational:Man is to Doctor as Woman is to Doctor

Analogies such as man is to king as woman is to X are often used to illu...
research
05/17/2023

Shielded Representations: Protecting Sensitive Attributes Through Iterative Gradient-Based Projection

Natural language processing models tend to learn and encode social biase...

Please sign up or login with your details

Forgot password? Click here to reset