Early stopping for L^2-boosting in high-dimensional linear models

10/14/2022
by   Bernhard Stankewitz, et al.
0

Increasingly high-dimensional data sets require that estimation methods do not only satisfy statistical guarantees but also remain computationally feasible. In this context, we consider L^2-boosting via orthogonal matching pursuit in a high-dimensional linear model and analyze a data-driven early stopping time τ of the algorithm, which is sequential in the sense that its computation is based on the first τ iterations only. This approach is much less costly than established model selection criteria, that require the computation of the full boosting path. We prove that sequential early stopping preserves statistical optimality in this setting in terms of a fully general oracle inequality for the empirical risk and recently established optimal convergence rates for the population risk. Finally, an extensive simulation study shows that at an immensely reduced computational cost, the performance of these type of methods is on par with other state of the art algorithms such as the cross-validated Lasso or model selection via a high dimensional Akaike criterion based on the full boosting path.

READ FULL TEXT

page 39

page 40

page 41

page 42

research
02/29/2016

High-Dimensional L_2Boosting: Rate of Convergence

Boosting is one of the most significant developments in machine learning...
research
12/13/2018

On the Differences between L2-Boosting and the Lasso

We prove that L2-Boosting lacks a theoretical property which is central ...
research
10/19/2015

NYTRO: When Subsampling Meets Early Stopping

Early stopping is a well known approach to reduce the time complexity fo...
research
11/15/2018

Minimax Posterior Convergence Rates and Model Selection Consistency in High-dimensional DAG Models based on Sparse Cholesky Factors

In this paper, we study the high-dimensional sparse directed acyclic gra...
research
06/09/2021

Bayesian Boosting for Linear Mixed Models

Boosting methods are widely used in statistical learning to deal with hi...
research
02/10/2017

L_2Boosting for Economic Applications

In the recent years more and more high-dimensional data sets, where the ...
research
04/16/2017

Boosting with Structural Sparsity: A Differential Inclusion Approach

Boosting as gradient descent algorithms is one popular method in machine...

Please sign up or login with your details

Forgot password? Click here to reset