Fairness in Machine Learning with Tractable Models

05/16/2019
by   Michael Varley, et al.
0

Machine Learning techniques have become pervasive across a range of different applications, and are now widely used in areas as disparate as recidivism prediction, consumer credit-risk analysis and insurance pricing. The prevalence of machine learning techniques has raised concerns about the potential for learned algorithms to become biased against certain groups. Many definitions have been proposed in the literature, but the fundamental task of reasoning about probabilistic events is a challenging one, owing to the intractability of inference. The focus of this paper is taking steps towards the application of tractable models to fairness. Tractable probabilistic models have emerged that guarantee that conditional marginal can be computed in time linear in the size of the model. In particular, we show that sum product networks (SPNs) enable an effective technique for determining the statistical relationships between protected attributes and other training variables. If a subset of these training variables are found by the SPN to be independent of the training attribute then they can be considered `safe' variables, from which we can train a classification model without concern that the resulting classifier will result in disparate outcomes for different demographic groups. Our initial experiments on the `German Credit' data set indicate that this processing technique significantly reduces disparate treatment of male and female credit applicants, with a small reduction in classification accuracy compared to state of the art. We will also motivate the concept of "fairness through percentile equivalence", a new definition predicated on the notion that individuals at the same percentile of their respective distributions should be treated equivalently, and this prevents unfair penalisation of those individuals who lie at the extremities of their respective distributions.

READ FULL TEXT

page 17

page 18

page 19

page 23

page 24

research
05/20/2022

The Fairness of Credit Scoring Models

In credit markets, screening algorithms aim to discriminate between good...
research
01/16/2020

Fairness Measures for Regression via Probabilistic Classification

Algorithmic fairness involves expressing notions such as equity, or reas...
research
10/05/2019

The Impact of Data Preparation on the Fairness of Software Systems

Machine learning models are widely adopted in scenarios that directly af...
research
01/29/2020

Interventions and Counterfactuals in Tractable Probabilistic Models: Limitations of Contemporary Transformations

In recent years, there has been an increasing interest in studying causa...
research
08/01/2023

Fair Models in Credit: Intersectional Discrimination and the Amplification of Inequity

The increasing usage of new data sources and machine learning (ML) techn...
research
01/29/2020

On Constraint Definability in Tractable Probabilistic Models

Incorporating constraints is a major concern in probabilistic machine le...
research
03/06/2023

Defining and comparing SICR-events for classifying impaired loans under IFRS 9

The IFRS 9 accounting standard requires the prediction of credit deterio...

Please sign up or login with your details

Forgot password? Click here to reset