Classification Under Ambiguity: When Is Average-K Better Than Top-K?

12/16/2021
by   Titouan Lorieul, et al.
14

When many labels are possible, choosing a single one can lead to low precision. A common alternative, referred to as top-K classification, is to choose some number K (commonly around 5) and to return the K labels with the highest scores. Unfortunately, for unambiguous cases, K>1 is too many and, for very ambiguous cases, K ≤ 5 (for example) can be too small. An alternative sensible strategy is to use an adaptive approach in which the number of labels returned varies as a function of the computed ambiguity, but must average to some particular K over all the samples. We denote this alternative average-K classification. This paper formally characterizes the ambiguity profile when average-K classification can achieve a lower error rate than a fixed top-K classification. Moreover, it provides natural estimation procedures for both the fixed-size and the adaptive classifier and proves their consistency. Finally, it reports experiments on real-world image data sets revealing the benefit of average-K classification over top-K in practice. Overall, when the ambiguity is known precisely, average-K is never worse than top-K, and, in our experiments, when it is estimated, this also holds.

READ FULL TEXT

page 2

page 38

research
03/31/2023

A two-head loss function for deep Average-K classification

Average-K classification is an alternative to top-K classification in wh...
research
06/06/2021

Embracing Ambiguity: Shifting the Training Target of NLI Models

Natural Language Inference (NLI) datasets contain examples with highly a...
research
07/13/2022

Is one annotation enough? A data-centric image classification benchmark for noisy and ambiguous label estimation

High-quality data is necessary for modern machine learning. However, the...
research
10/23/2019

High-Confidence Policy Optimization: Reshaping Ambiguity Sets in Robust MDPs

Robust MDPs are a promising framework for computing robust policies in r...
research
03/27/2023

Ambiguity-Resistant Semi-Supervised Learning for Dense Object Detection

With basic Semi-Supervised Object Detection (SSOD) techniques, one-stage...
research
11/07/2017

Challenges in Disentangling Independent Factors of Variation

We study the problem of building models that disentangle independent fac...

Please sign up or login with your details

Forgot password? Click here to reset