Towards classification parity across cohorts

05/16/2020
by   Aarsh Patel, et al.
17

Recently, there has been a lot of interest in ensuring algorithmic fairness in machine learning where the central question is how to prevent sensitive information (e.g. knowledge about the ethnic group of an individual) from adding "unfair" bias to a learning algorithm (Feldman et al. (2015), Zemel et al. (2013)). This has led to several debiasing algorithms on word embeddings (Qian et al. (2019) , Bolukbasi et al. (2016)), coreference resolution (Zhao et al. (2018a)), semantic role labeling (Zhao et al. (2017)), etc. Most of these existing work deals with explicit sensitive features such as gender, occupations or race which doesn't work with data where such features are not captured due to privacy concerns. In this research work, we aim to achieve classification parity across explicit as well as implicit sensitive features. We define explicit cohorts as groups of people based on explicit sensitive attributes provided in the data (age, gender, race) whereas implicit cohorts are defined as groups of people with similar language usage. We obtain implicit cohorts by clustering embeddings of each individual trained on the language generated by them using a language model. We achieve two primary objectives in this work : [1.] We experimented and discovered classification performance differences across cohorts based on implicit and explicit features , [2] We improved classification parity by introducing modification to the loss function aimed to minimize the range of model performances across cohorts.

READ FULL TEXT
research
08/07/2019

Debiasing Embeddings for Reduced Gender Bias in Text Classification

(Bolukbasi et al., 2016) demonstrated that pretrained word embeddings ca...
research
10/14/2022

Controlling Bias Exposure for Fair Interpretable Predictions

Recent work on reducing bias in NLP models usually focuses on protecting...
research
01/29/2019

Implicit Diversity in Image Summarization

Case studies, such as Kay et al., 2015 have shown that in image summariz...
research
09/20/2020

Exploring the Linear Subspace Hypothesis in Gender Bias Mitigation

Bolukbasi et al. (2016) presents one of the first gender bias mitigation...
research
10/22/2019

Grammatical Gender, Neo-Whorfianism, and Word Embeddings: A Data-Driven Approach to Linguistic Relativity

The relation between language and thought has occupied linguists for at ...
research
07/10/2020

A Causal Linear Model to Quantify Edge Unfairness for Unfair Edge Prioritization and Discrimination Removal

The dataset can be generated by an unfair mechanism in numerous settings...
research
05/21/2018

Implicit Probabilistic Integrators for ODEs

We introduce a family of implicit probabilistic integrators for initial ...

Please sign up or login with your details

Forgot password? Click here to reset