Towards unraveling calibration biases in medical image analysis

In recent years the development of artificial intelligence (AI) systems for automated medical image analysis has gained enormous momentum. At the same time, a large body of work has shown that AI systems can systematically and unfairly discriminate against certain populations in various application scenarios. These two facts have motivated the emergence of algorithmic fairness studies in this field. Most research on healthcare algorithmic fairness to date has focused on the assessment of biases in terms of classical discrimination metrics such as AUC and accuracy. Potential biases in terms of model calibration, however, have only recently begun to be evaluated. This is especially important when working with clinical decision support systems, as predictive uncertainty is key for health professionals to optimally evaluate and combine multiple sources of information. In this work we study discrimination and calibration biases in models trained for automatic detection of malignant dermatological conditions from skin lesions images. Importantly, we show how several typically employed calibration metrics are systematically biased with respect to sample sizes, and how this can lead to erroneous fairness analysis if not taken into consideration. This is of particular relevance to fairness studies, where data imbalance results in drastic sample size differences between demographic sub-groups, which, if not taken into account, can act as confounders.

READ FULL TEXT

page 4

page 9

research
03/06/2023

Evaluating the Fairness of Deep Learning Uncertainty Estimates in Medical Image Analysis

Although deep learning (DL) models have shown great success in many medi...
research
03/07/2021

Estimating and Improving Fairness with Adversarial Learning

Fairness and accountability are two essential pillars for trustworthy Ar...
research
09/27/2022

A Survey of Fairness in Medical Image Analysis: Concepts, Algorithms, Evaluations, and Challenges

Fairness, a criterion focuses on evaluating algorithm performance on dif...
research
07/04/2023

Mitigating Calibration Bias Without Fixed Attribute Grouping for Improved Fairness in Medical Imaging Analysis

Trustworthy deployment of deep learning medical imaging models into real...
research
10/01/2021

Algorithm Fairness in AI for Medicine and Healthcare

In the current development and deployment of many artificial intelligenc...
research
08/08/2023

Learning Unbiased Image Segmentation: A Case Study with Plain Knee Radiographs

Automatic segmentation of knee bony anatomy is essential in orthopedics,...
research
05/20/2020

Risk of Training Diagnostic Algorithms on Data with Demographic Bias

One of the critical challenges in machine learning applications is to ha...

Please sign up or login with your details

Forgot password? Click here to reset