On the Insufficiency of the Large Margins Theory in Explaining the Performance of Ensemble Methods

06/10/2019
by   Waldyn Martinez, et al.
0

Boosting and other ensemble methods combine a large number of weak classifiers through weighted voting to produce stronger predictive models. To explain the successful performance of boosting algorithms, Schapire et al. (1998) showed that AdaBoost is especially effective at increasing the margins of the training data. Schapire et al. (1998) also developed an upper bound on the generalization error of any ensemble based on the margins of the training data, from which it was concluded that larger margins should lead to lower generalization error, everything else being equal (sometimes referred to as the “large margins theory”). Tighter bounds have been derived and have reinforced the large margins theory hypothesis. For instance, Wang et al. (2011) suggest that specific margin instances, such as the equilibrium margin, can better summarize the margins distribution. These results have led many researchers to consider direct optimization of the margins to improve ensemble generalization error with mixed results. We show that the large margins theory is not sufficient for explaining the performance of voting classifiers. We do this by illustrating how it is possible to improve upon the margin distribution of an ensemble solution, while keeping the complexity fixed, yet not improve the test set performance.

READ FULL TEXT
research
06/07/2019

On the Current State of Research in Explaining Ensemble Performance Using Margins

Empirical evidence shows that ensembles, such as bagging, boosting, rand...
research
11/10/2020

Margins are Insufficient for Explaining Gradient Boosting

Boosting is one of the most successful ideas in machine learning, achiev...
research
06/09/2022

On Margins and Generalisation for Voting Classifiers

We study the generalisation properties of majority voting on finite ense...
research
01/30/2019

Optimal Minimal Margin Maximization with Boosting

Boosting algorithms produce a classifier by iteratively combining base h...
research
04/23/2021

Selecting a number of voters for a voting ensemble

For a voting ensemble that selects an odd-sized subset of the ensemble c...
research
10/22/2020

In Search of Robust Measures of Generalization

One of the principal scientific challenges in deep learning is explainin...

Please sign up or login with your details

Forgot password? Click here to reset