It's COMPASlicated: The Messy Relationship between RAI Datasets and Algorithmic Fairness Benchmarks

06/10/2021
by   Michelle Bao, et al.
0

Risk assessment instrument (RAI) datasets, particularly ProPublica's COMPAS dataset, are commonly used in algorithmic fairness papers due to benchmarking practices of comparing algorithms on datasets used in prior work. In many cases, this data is used as a benchmark to demonstrate good performance without accounting for the complexities of criminal justice (CJ) processes. We show that pretrial RAI datasets contain numerous measurement biases and errors inherent to CJ pretrial evidence and due to disparities in discretion and deployment, are limited in making claims about real-world outcomes, making the datasets a poor fit for benchmarking under assumptions of ground truth and real-world impact. Conventional practices of simply replicating previous data experiments may implicitly inherit or edify normative positions without explicitly interrogating assumptions. With context of how interdisciplinary fields have engaged in CJ research, algorithmic fairness practices are misaligned for meaningful contribution in the context of CJ, and would benefit from transparent engagement with normative considerations and values related to fairness, justice, and equality. These factors prompt questions about whether benchmarks for intrinsically socio-technical systems like the CJ system can exist in a beneficial and ethical way.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/27/2021

A Sociotechnical View of Algorithmic Fairness

Algorithmic fairness has been framed as a newly emerging technology that...
research
02/03/2022

Algorithmic Fairness Datasets: the Story so Far

Data-driven algorithms are being studied and deployed in diverse domains...
research
02/26/2020

No computation without representation: Avoiding data and algorithm biases through diversity

The emergence and growth of research on issues of ethics in AI, and in p...
research
12/03/2020

Non-portability of Algorithmic Fairness in India

Conventional algorithmic fairness is Western in its sub-groups, values, ...
research
04/07/2023

About Voice: A Longitudinal Study of Speaker Recognition Dataset Dynamics

Like face recognition, speaker recognition is widely used for voice-base...
research
02/07/2023

From Utilitarian to Rawlsian Designs for Algorithmic Fairness

There is a lack of consensus within the literature as to how `fairness' ...
research
01/25/2021

Modeling Assumptions Clash with the Real World: Transparency, Equity, and Community Challenges for Student Assignment Algorithms

Across the United States, a growing number of school districts are turni...

Please sign up or login with your details

Forgot password? Click here to reset