Monitoring Algorithmic Fairness

05/25/2023
by   Thomas A. Henzinger, et al.
0

Machine-learned systems are in widespread use for making decisions about humans, and it is important that they are fair, i.e., not biased against individuals based on sensitive attributes. We present runtime verification of algorithmic fairness for systems whose models are unknown, but are assumed to have a Markov chain structure. We introduce a specification language that can model many common algorithmic fairness properties, such as demographic parity, equal opportunity, and social burden. We build monitors that observe a long sequence of events as generated by a given system, and output, after each observation, a quantitative estimate of how fair or biased the system was on that run until that point in time. The estimate is proven to be correct modulo a variable error bound and a given confidence level, where the error bound gets tighter as the observed sequence gets longer. Our monitors are of two types, and use, respectively, frequentist and Bayesian statistical inference techniques. While the frequentist monitors compute estimates that are objectively correct with respect to the ground truth, the Bayesian monitors compute estimates that are correct subject to a given prior belief about the system's model. Using a prototype implementation, we show how we can monitor if a bank is fair in giving loans to applicants from different social backgrounds, and if a college is fair in admitting students while maintaining a reasonable financial burden on the society. Although they exhibit different theoretical complexities in certain cases, in our experiments, both frequentist and Bayesian monitors took less than a millisecond to update their verdicts after each observation.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/01/2023

Monitoring Algorithmic Fairness under Partial Observations

As AI and machine-learned software are used increasingly for making deci...
research
11/01/2020

Making ML models fairer through explanations: the case of LimeOut

Algorithmic decisions are now being used on a daily basis, and based on ...
research
05/08/2023

Runtime Monitoring of Dynamic Fairness Properties

A machine-learned system that is fair in static decision-making tasks ma...
research
07/06/2022

Towards Substantive Conceptions of Algorithmic Fairness: Normative Guidance from Equal Opportunity Doctrines

In this work we use Equal Oppportunity (EO) doctrines from political phi...
research
11/21/2022

Bursting the Burden Bubble? An Assessment of Sharma et al.'s Counterfactual-based Fairness Metric

Machine learning has seen an increase in negative publicity in recent ye...
research
10/07/2020

Fairness Perception from a Network-Centric Perspective

Algorithmic fairness is a major concern in recent years as the influence...
research
02/18/2020

A Resolution in Algorithmic Fairness: Calibrated Scores for Fair Classifications

Calibration and equal error rates are fundamental conditions for algorit...

Please sign up or login with your details

Forgot password? Click here to reset