Being Properly Improper

06/18/2021
by   Richard Nock, et al.
0

In today's ML, data can be twisted (changed) in various ways, either for bad or good intent. Such twisted data challenges the founding theory of properness for supervised losses which form the basis for many popular losses for class probability estimation. Unfortunately, at its core, properness ensures that the optimal models also learn the twist. In this paper, we analyse such class probability-based losses when they are stripped off the mandatory properness; we define twist-proper losses as losses formally able to retrieve the optimum (untwisted) estimate off the twists, and show that a natural extension of a half-century old loss introduced by S. Arimoto is twist proper. We then turn to a theory that has provided some of the best off-the-shelf algorithms for proper losses, boosting. Boosting can require access to the derivative of the convex conjugate of a loss to compute examples weights. Such a function can be hard to get, for computational or mathematical reasons; this turns out to be the case for Arimoto's loss. We bypass this difficulty by inverting the problem as follows: suppose a blueprint boosting algorithm is implemented with a general weight update function. What are the losses for which boosting-compliant minimisation happens? Our answer comes as a general boosting algorithm which meets the optimal boosting dependence on the number of calls to the weak learner; when applied to Arimoto's loss, it leads to a simple optimisation algorithm whose performances are showcased on several domains and twists.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/19/2022

What killed the Convex Booster ?

A landmark negative result of Long and Servedio established a worst-case...
research
12/17/2009

Composite Binary Losses

We study losses for binary classification and class probability estimati...
research
08/08/2016

Boosting as a kernel-based method

Boosting combines weak (biased) learners to obtain effective learning al...
research
09/01/2022

The Geometry and Calculus of Losses

Statistical decision problems are the foundation of statistical machine ...
research
03/04/2021

Lower-bounded proper losses for weakly supervised classification

This paper discusses the problem of weakly supervised learning of classi...
research
01/27/2023

LegendreTron: Uprising Proper Multiclass Loss Learning

Loss functions serve as the foundation of supervised learning and are of...
research
12/12/2012

Advances in Boosting (Invited Talk)

Boosting is a general method of generating many simple classification ru...

Please sign up or login with your details

Forgot password? Click here to reset