Fairness in Forecasting of Observations of Linear Dynamical Systems

09/12/2022
by   Quan Zhou, et al.
0

In machine learning, training data often capture the behaviour of multiple subgroups of some underlying human population. When the nature of training data for subgroups are not controlled carefully, under-representation bias arises. To counter this effect we introduce two natural notions of subgroup fairness and instantaneous fairness to address such under-representation bias in time-series forecasting problems. Here we show globally convergent methods for the fairness-constrained learning problems using hierarchies of convexifications of non-commutative polynomial optimisation problems. Our empirical results on a biased data set motivated by insurance applications and the well-known COMPAS data set demonstrate the efficacy of our methods. We also show that by exploiting sparsity in the convexifications, we can reduce the run time of our methods considerably.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/12/2020

Fairness in Forecasting and Learning Linear Dynamical Systems

As machine learning becomes more pervasive, the urgency of assuring its ...
research
12/02/2019

Recovering from Biased Data: Can Fairness Constraints Improve Accuracy?

Multiple fairness constraints have been proposed in the literature, moti...
research
02/11/2021

Fairness-Aware Learning from Corrupted Data

Addressing fairness concerns about machine learning models is a crucial ...
research
03/24/2022

Learning Spatiotemporal Chaos Using Next-Generation Reservoir Computing

Forecasting the behavior of high-dimensional dynamical systems using mac...
research
01/28/2019

Fairness in representation: quantifying stereotyping as a representational harm

While harms of allocation have been increasingly studied as part of the ...
research
05/31/2022

Inducing bias is simpler than you think

Machine learning may be oblivious to human bias but it is not immune to ...
research
05/31/2021

Using Pareto Simulated Annealing to Address Algorithmic Bias in Machine Learning

Algorithmic Bias can be due to bias in the training data or issues with ...

Please sign up or login with your details

Forgot password? Click here to reset