fairml: A Statistician's Take on Fair Machine Learning Modelling

05/03/2023
by   Marco Scutari, et al.
0

The adoption of machine learning in applications where it is crucial to ensure fairness and accountability has led to a large number of model proposals in the literature, largely formulated as optimisation problems with constraints reducing or eliminating the effect of sensitive attributes on the response. While this approach is very flexible from a theoretical perspective, the resulting models are somewhat black-box in nature: very little can be said about their statistical properties, what are the best practices in their applied use, and how they can be extended to problems other than those they were originally designed for. Furthermore, the estimation of each model requires a bespoke implementation involving an appropriate solver which is less than desirable from a software engineering perspective. In this paper, we describe the fairml R package which implements our previous work (Scutari, Panero, and Proissl 2022) and related models in the literature. fairml is designed around classical statistical models (generalised linear models) and penalised regression results (ridge regression) to produce fair models that are interpretable and whose properties are well-known. The constraint used to enforce fairness is orthogonal to model estimation, making it possible to mix-and-match the desired model family and fairness definition for each application. Furthermore, fairml provides facilities for model estimation, model selection and validation including diagnostic plots.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/26/2023

Insights From Insurance for Fair Machine Learning: Responsibility, Performativity and Aggregates

We argue that insurance can act as an analogon for the social situatedne...
research
05/08/2020

In Pursuit of Interpretable, Fair and Accurate Machine Learning for Criminal Recidivism Prediction

In recent years, academics and investigative journalists have criticized...
research
04/22/2022

Balancing Fairness and Accuracy in Sentiment Detection using Multiple Black Box Models

Sentiment detection is an important building block for multiple informat...
research
07/27/2022

Fairness and Randomness in Machine Learning: Statistical Independence and Relativization

Fair Machine Learning endeavors to prevent unfairness arising in the con...
research
06/18/2022

Fair Generalized Linear Models with a Convex Penalty

Despite recent advances in algorithmic fairness, methodologies for achie...
research
01/24/2019

Fairness risk measures

Ensuring that classifiers are non-discriminatory or fair with respect to...
research
05/18/2021

Achieving Fairness with a Simple Ridge Penalty

Estimating a fair linear regression model subject to a user-defined leve...

Please sign up or login with your details

Forgot password? Click here to reset