Measuring Fairness of Text Classifiers via Prediction Sensitivity

03/16/2022
by   Satyapriya Krishna, et al.
10

With the rapid growth in language processing applications, fairness has emerged as an important consideration in data-driven solutions. Although various fairness definitions have been explored in the recent literature, there is lack of consensus on which metrics most accurately reflect the fairness of a system. In this work, we propose a new formulation : ACCUMULATED PREDICTION SENSITIVITY, which measures fairness in machine learning models based on the model's prediction sensitivity to perturbations in input features. The metric attempts to quantify the extent to which a single prediction depends on a protected attribute, where the protected attribute encodes the membership status of an individual in a protected group. We show that the metric can be theoretically linked with a specific notion of group fairness (statistical parity) and individual fairness. It also correlates well with humans' perception of fairness. We conduct experiments on two text classification datasets : JIGSAW TOXICITY, and BIAS IN BIOS, and evaluate the correlations between metrics and manual annotations on whether the model produced a fair outcome. We observe that the proposed fairness metric based on prediction sensitivity is statistically significantly more correlated with human annotation than the existing counterfactual fairness metric.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/09/2022

Prediction Sensitivity: Continual Audit of Counterfactual Fairness in Deployed Classifiers

As AI-based systems increasingly impact many areas of our lives, auditin...
research
11/21/2022

Bursting the Burden Bubble? An Assessment of Sharma et al.'s Counterfactual-based Fairness Metric

Machine learning has seen an increase in negative publicity in recent ye...
research
09/28/2020

Towards a Measure of Individual Fairness for Deep Learning

Deep learning has produced big advances in artificial intelligence, but ...
research
11/30/2020

Towards Auditability for Fairness in Deep Learning

Group fairness metrics can detect when a deep learning model behaves dif...
research
03/03/2023

Model Explanation Disparities as a Fairness Diagnostic

In recent years, there has been a flurry of research focusing on the fai...
research
11/27/2020

Black Loans Matter: Distributionally Robust Fairness for Fighting Subgroup Discrimination

Algorithmic fairness in lending today relies on group fairness metrics f...
research
04/03/2020

FairALM: Augmented Lagrangian Method for Training Fair Models with Little Regret

Algorithmic decision making based on computer vision and machine learnin...

Please sign up or login with your details

Forgot password? Click here to reset