Understanding Prediction Discrepancies in Machine Learning Classifiers

04/12/2021
by   Xavier Renard, et al.
18

A multitude of classifiers can be trained on the same data to achieve similar performances during test time, while having learned significantly different classification patterns. This phenomenon, which we call prediction discrepancies, is often associated with the blind selection of one model instead of another with similar performances. When making a choice, the machine learning practitioner has no understanding on the differences between models, their limits, where they agree and where they don't. But his/her choice will result in concrete consequences for instances to be classified in the discrepancy zone, since the final decision will be based on the selected classification pattern. Besides the arbitrary nature of the result, a bad choice could have further negative consequences such as loss of opportunity or lack of fairness. This paper proposes to address this question by analyzing the prediction discrepancies in a pool of best-performing models trained on the same data. A model-agnostic algorithm, DIG, is proposed to capture and explain discrepancies locally, to enable the practitioner to make the best educated decision when selecting a model by anticipating its potential undesired consequences. All the code to reproduce the experiments is available.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/16/2018

Machine Decisions and Human Consequences

As we increasingly delegate decision-making to algorithms, whether direc...
research
05/18/2021

Quantifying sources of uncertainty in drug discovery predictions with probabilistic models

Knowing the uncertainty in a prediction is critical when making expensiv...
research
09/15/2023

Adaptive Priority Reweighing for Generalizing Fairness Improvement

With the increasing penetration of machine learning applications in crit...
research
06/02/2022

HEX: Human-in-the-loop Explainability via Deep Reinforcement Learning

The use of machine learning (ML) models in decision-making contexts, par...
research
06/23/2021

Speech is Silver, Silence is Golden: What do ASVspoof-trained Models Really Learn?

We present our analysis of a significant data artifact in the official 2...
research
02/21/2023

Does the evaluation stand up to evaluation? A first-principle approach to the evaluation of classifiers

How can one meaningfully make a measurement, if the meter does not confo...
research
07/17/2023

A General Framework for Learning under Corruption: Label Noise, Attribute Noise, and Beyond

Corruption is frequently observed in collected data and has been extensi...

Please sign up or login with your details

Forgot password? Click here to reset