The Measure and Mismeasure of Fairness: A Critical Review of Fair Machine Learning

07/31/2018
by   Sam Corbett-Davies, et al.
0

In one broad class of supervised machine learning problems, researchers and practitioners seek to estimate the likelihood of a given event in a manner that is fair. For example, when estimating a loan applicant's risk of default, one might aim to create a statistical model that is free of race or gender bias. Over the last several years, three formal definitions of fairness have gained popularity in the computer science literature: (1) anti-classification, meaning that protected attributes---like race, gender, and their proxies---are not used to form predictions; (2) classification parity, meaning that common measures of predictive performance (e.g., false positive and false negative rates) are equal across groups defined by the protected attributes; and (3) calibration, meaning that conditional on predictions, outcomes are independent of protected attributes. Here we show that all three of these fairness definitions suffer from significant statistical limitations. In particular, requiring anti-classification or classification parity can, perversely, harm the historically disadvantaged groups they were designed to protect, while also violating common legal, economic, and social understandings of fairness; and calibration, while generally desirable, provides little guarantee that decisions are equitable. By highlighting these challenges in the foundation of fair machine learning, we hope to help researchers and practitioners productively advance the area.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/12/2023

Auditing ICU Readmission Rates in an Clinical Database: An Analysis of Risk Factors and Clinical Outcomes

This study presents a machine learning (ML) pipeline for clinical data c...
research
02/17/2023

Designing Equitable Algorithms

Predictive algorithms are now used to help distribute a large share of o...
research
10/15/2018

Neural Styling for Interpretable Fair Representations

We observe a rapid increase in machine learning models for learning data...
research
12/05/2020

FAIROD: Fairness-aware Outlier Detection

Fairness and Outlier Detection (OD) are closely related, as it is exactl...
research
06/18/2020

Towards Threshold Invariant Fair Classification

Effective machine learning models can automatically learn useful informa...
research
05/15/2021

Cohort Shapley value for algorithmic fairness

Cohort Shapley value is a model-free method of variable importance groun...
research
06/11/2019

ProPublica's COMPAS Data Revisited

In this paper I re-examine the COMPAS recidivism score and criminal hist...

Please sign up or login with your details

Forgot password? Click here to reset