Learning to Defer to Multiple Experts: Consistent Surrogate Losses, Confidence Calibration, and Conformal Ensembles

10/30/2022
by   Rajeev Verma, et al.
0

We study the statistical properties of learning to defer (L2D) to multiple experts. In particular, we address the open problems of deriving a consistent surrogate loss, confidence calibration, and principled ensembling of experts. Firstly, we derive two consistent surrogates – one based on a softmax parameterization, the other on a one-vs-all (OvA) parameterization – that are analogous to the single expert losses proposed by Mozannar and Sontag (2020) and Verma and Nalisnick (2022), respectively. We then study the frameworks' ability to estimate P( m_j = y | x ), the probability that the jth expert will correctly predict the label for x. Theory shows the softmax-based loss causes mis-calibration to propagate between the estimates while the OvA-based loss does not (though in practice, we find there are trade offs). Lastly, we propose a conformal inference technique that chooses a subset of experts to query when the system defers. We perform empirical validation on tasks for galaxy, skin lesion, and hate speech classification.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/20/2016

Multiclass Classification Calibration Functions

In this paper we refine the process of computing calibration functions f...
research
02/08/2022

Calibrated Learning to Defer with One-vs-All Classifiers

The learning to defer (L2D) framework has the potential to make AI syste...
research
03/29/2021

von Mises-Fisher Loss: An Exploration of Embedding Geometries for Supervised Learning

Recent work has argued that classification losses utilizing softmax cros...
research
01/30/2019

On the Consistency of Top-k Surrogate Losses

The top-k error is often employed to evaluate performance for challengin...
research
01/30/2019

Doubly Sparse: Sparse Mixture of Sparse Experts for Efficient Softmax Inference

Computations for the softmax function are significantly expensive when t...
research
06/02/2020

Consistent Estimators for Learning to Defer to an Expert

Learning algorithms are often used in conjunction with expert decision m...
research
08/09/2023

Expert load matters: operating networks at high accuracy and low manual effort

In human-AI collaboration systems for critical applications, in order to...

Please sign up or login with your details

Forgot password? Click here to reset