Set Learning for Accurate and Calibrated Models

07/05/2023
by   Lukas Muttenthaler, et al.
0

Model overconfidence and poor calibration are common in machine learning and difficult to account for when applying standard empirical risk minimization. In this work, we propose a novel method to alleviate these problems that we call odd-k-out learning (OKO), which minimizes the cross-entropy error for sets rather than for single examples. This naturally allows the model to capture correlations across data examples and achieves both better accuracy and calibration, especially in limited training data and class-imbalanced regimes. Perhaps surprisingly, OKO often yields better calibration even when training with hard labels and dropping any additional calibration parameter tuning, such as temperature scaling. We provide theoretical justification, establishing that OKO naturally yields better calibration, and provide extensive experimental analyses that corroborate our theoretical findings. We emphasize that OKO is a general framework that can be easily adapted to many settings and the trained model can be applied to single examples at inference time, without introducing significant run-time overhead or architecture changes.

READ FULL TEXT
research
07/30/2021

Soft Calibration Objectives for Neural Networks

Optimal decision making requires that classifiers produce uncertainty es...
research
10/13/2020

Ensemble Distillation for Structured Prediction: Calibrated, Accurate, Fast—Choose Three

Modern neural networks do not always produce well-calibrated predictions...
research
03/17/2020

Calibration of Pre-trained Transformers

Pre-trained Transformers are now ubiquitous in natural language processi...
research
02/25/2021

Confidence Calibration with Bounded Error Using Transformations

As machine learning techniques become widely adopted in new domains, esp...
research
07/30/2020

From parameter calibration to parameter learning: Revolutionizing large-scale geoscientific modeling with big data

The behaviors and skills of models in many geoscientific domains strongl...
research
08/29/2018

Group calibration is a byproduct of unconstrained learning

Much recent work on fairness in machine learning has focused on how well...
research
06/01/2023

Conformal Prediction with Partially Labeled Data

While the predictions produced by conformal prediction are set-valued, t...

Please sign up or login with your details

Forgot password? Click here to reset