Robust and Efficient Boosting Method using the Conditional Risk

06/21/2018
by   Zhi Xiao, et al.
0

Well-known for its simplicity and effectiveness in classification, AdaBoost, however, suffers from overfitting when class-conditional distributions have significant overlap. Moreover, it is very sensitive to noise that appears in the labels. This article tackles the above limitations simultaneously via optimizing a modified loss function (i.e., the conditional risk). The proposed approach has the following two advantages. (1) It is able to directly take into account label uncertainty with an associated label confidence. (2) It introduces a "trustworthiness" measure on training samples via the Bayesian risk rule, and hence the resulting classifier tends to have finite sample performance that is superior to that of the original AdaBoost when there is a large overlap between class conditional distributions. Theoretical properties of the proposed method are investigated. Extensive experimental results using synthetic data and real-world data sets from UCI machine learning repository are provided. The empirical study shows the high competitiveness of the proposed method in predication accuracy and robustness when compared with the original AdaBoost and several existing robust AdaBoost algorithms.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/07/2021

Robustness and reliability when training with noisy labels

Labelling of data for supervised learning can be costly and time-consumi...
research
04/30/2012

A Conjugate Property between Loss Functions and Uncertainty Sets in Classification Problems

In binary classification problems, mainly two approaches have been propo...
research
03/01/2021

Adversarial Reciprocal Points Learning for Open Set Recognition

Open set recognition (OSR), aiming to simultaneously classify the seen c...
research
06/16/2022

Local overlap reduction procedure for dynamic ensemble selection

Class imbalance is a characteristic known for making learning more chall...
research
03/24/2022

Risk Consistent Multi-Class Learning from Label Proportions

This study addresses a multiclass learning from label proportions (MCLLP...
research
09/16/2011

A Characterization of the Combined Effects of Overlap and Imbalance on the SVM Classifier

In this paper we demonstrate that two common problems in Machine Learnin...
research
04/21/2020

Normalizing Flow Regression

In this letter we propose a convex approach to learning expressive scala...

Please sign up or login with your details

Forgot password? Click here to reset