A survey on bias in machine learning research

Current research on bias in machine learning often focuses on fairness, while overlooking the roots or causes of bias. However, bias was originally defined as a "systematic error," often caused by humans at different stages of the research process. This article aims to bridge the gap between past literature on bias in research by providing taxonomy for potential sources of bias and errors in data and models. The paper focus on bias in machine learning pipelines. Survey analyses over forty potential sources of bias in the machine learning (ML) pipeline, providing clear examples for each. By understanding the sources and consequences of bias in machine learning, better methods can be developed for its detecting and mitigating, leading to fairer, more transparent, and more accurate ML models.

READ FULL TEXT

page 14

page 16

page 17

page 18

page 19

research
12/10/2021

A Framework for Fairness: A Systematic Review of Existing Fair AI Solutions

In a world of daily emerging scientific inquisition and discovery, the p...
research
04/01/2020

Bias in Machine Learning What is it Good (and Bad) for?

In public media as well as in scientific publications, the term bias is ...
research
08/18/2023

Data augmentation and explainability for bias discovery and mitigation in deep learning

This dissertation explores the impact of bias in deep neural networks an...
research
02/15/2022

Fairness Amidst Non-IID Graph Data: A Literature Review

Fairness in machine learning (ML), the process to understand and correct...
research
06/10/2020

System to Integrate Fairness Transparently: An Industry Approach

There have been significant research efforts to address the issue of uni...
research
07/19/2021

Introducing a Family of Synthetic Datasets for Research on Bias in Machine Learning

A significant impediment to progress in research on bias in machine lear...

Please sign up or login with your details

Forgot password? Click here to reset