Boosted CVaR Classification

10/26/2021
by   Runtian Zhai, et al.
0

Many modern machine learning tasks require models with high tail performance, i.e. high performance over the worst-off samples in the dataset. This problem has been widely studied in fields such as algorithmic fairness, class imbalance, and risk-sensitive decision making. A popular approach to maximize the model's tail performance is to minimize the CVaR (Conditional Value at Risk) loss, which computes the average risk over the tails of the loss. However, for classification tasks where models are evaluated by the zero-one loss, we show that if the classifiers are deterministic, then the minimizer of the average zero-one loss also minimizes the CVaR zero-one loss, suggesting that CVaR loss minimization is not helpful without additional assumptions. We circumvent this negative result by minimizing the CVaR loss over randomized classifiers, for which the minimizers of the average zero-one loss and the CVaR zero-one loss are no longer the same, so minimizing the latter can lead to better tail performance. To learn such randomized classifiers, we propose the Boosted CVaR Classification framework which is motivated by a direct relationship between CVaR and a classical boosting algorithm called LPBoost. Based on this framework, we design an algorithm called α-AdaLPBoost. We empirically evaluate our proposed algorithm on four benchmark datasets and show that it achieves higher tail performance than deterministic model training methods.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/08/2023

Safe Collaborative Filtering

Excellent tail performance is crucial for modern machine learning tasks,...
research
08/09/2023

Controlling Tail Risk in Online Ski-Rental

The classical ski-rental problem admits a textbook 2-competitive determi...
research
02/14/2020

Statistical Learning with Conditional Value at Risk

We propose a risk-averse statistical learning framework wherein the perf...
research
03/28/2022

Risk regularization through bidirectional dispersion

Many alternative notions of "risk" (e.g., CVaR, entropic risk, DRO risk)...
research
05/28/2020

Adversarial Classification via Distributional Robustness with Wasserstein Ambiguity

We study a model for adversarial classification based on distributionall...
research
10/15/2020

Minimax Classification with 0-1 Loss and Performance Guarantees

Supervised classification techniques use training samples to find classi...
research
11/01/2018

Minimizing Close-k Aggregate Loss Improves Classification

In classification, the de facto method for aggregating individual losses...

Please sign up or login with your details

Forgot password? Click here to reset