Some Open Problems in Optimal AdaBoost and Decision Stumps

05/26/2015
by   Joshua Belanich, et al.
0

The significance of the study of the theoretical and practical properties of AdaBoost is unquestionable, given its simplicity, wide practical use, and effectiveness on real-world datasets. Here we present a few open problems regarding the behavior of "Optimal AdaBoost," a term coined by Rudin, Daubechies, and Schapire in 2004 to label the simple version of the standard AdaBoost algorithm in which the weak learner that AdaBoost uses always outputs the weak classifier with lowest weighted error among the respective hypothesis class of weak classifiers implicit in the weak learner. We concentrate on the standard, "vanilla" version of Optimal AdaBoost for binary classification that results from using an exponential-loss upper bound on the misclassification training error. We present two types of open problems. One deals with general weak hypotheses. The other deals with the particular case of decision stumps, as often and commonly used in practice. Answers to the open problems can have immediate significant impact to (1) cementing previously established results on asymptotic convergence properties of Optimal AdaBoost, for finite datasets, which in turn can be the start to any convergence-rate analysis; (2) understanding the weak-hypotheses class of effective decision stumps generated from data, which we have empirically observed to be significantly smaller than the typically obtained class, as well as the effect on the weak learner's running time and previously established improved bounds on the generalization performance of Optimal AdaBoost classifiers; and (3) shedding some light on the "self control" that AdaBoost tends to exhibit in practice.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/25/2021

Quantum Boosting using Domain-Partitioning Hypotheses

Boosting is an ensemble learning method that converts a weak learner int...
research
06/03/2022

Optimal Weak to Strong Learning

The classic algorithm AdaBoost allows to convert a weak learner, that is...
research
10/04/2018

Improved generalization bounds for robust learning

We consider a model of robust learning in an adversarial environment. Th...
research
01/31/2020

Boosting Simple Learners

We consider boosting algorithms under the restriction that the weak lear...
research
10/01/2022

Efficient Quantum Agnostic Improper Learning of Decision Trees

The agnostic setting is the hardest generalization of the PAC model sinc...
research
12/05/2012

On the Convergence Properties of Optimal AdaBoost

AdaBoost is one of the most popular machine-learning algorithms. It is s...
research
07/24/2013

When is the majority-vote classifier beneficial?

In his seminal work, Schapire (1990) proved that weak classifiers could ...

Please sign up or login with your details

Forgot password? Click here to reset