Robust Boosting for Regression Problems

02/06/2020
by   Xiaomeng Ju, et al.
0

The gradient boosting algorithm constructs a regression estimator using a linear combination of simple "base learners". In order to obtain a robust non-parametric regression estimator that is scalable to high dimensional problems we propose a robust boosting algorithm based on a two-stage approach, similar to what is done for robust linear regression: we first minimize a robust residual scale estimator, and then improve its efficiency by optimizing a bounded loss function. Unlike previous proposals, our algorithm does not need to compute an ad-hoc residual scale estimator in each step. Since our loss functions are typically non-convex, we propose initializing our algorithm with an L_1 regression tree, which is fast to compute. We also introduce a robust variable importance metric for variable selection that is calculated via a permutation procedure. Through simulated and real data experiments, we compare our method against gradient boosting with squared loss and other robust boosting methods in the literature. With clean data, our method works equally well as gradient boosting with the squared loss. With symmetric and asymmetrically contaminated data, we show that our proposed method outperforms in terms of prediction error and variable selection accuracy.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/19/2021

Unified Robust Boosting

Boosting is a popular machine learning algorithm in regression and class...
research
02/03/2022

Deselection of Base-Learners for Statistical Boosting – with an Application to Distributional Regression

We present a new procedure for enhanced variable selection for component...
research
01/15/2021

Causal Gradient Boosting: Boosted Instrumental Variable Regression

Recent advances in the literature have demonstrated that standard superv...
research
02/27/2017

An update on statistical boosting in biomedicine

Statistical boosting algorithms have triggered a lot of research during ...
research
09/24/2019

The column measure and Gradient-Free Gradient Boosting

Sparse model selection by structural risk minimization leads to a set of...
research
09/29/2020

Selective Cascade of Residual ExtraTrees

We propose a novel tree-based ensemble method named Selective Cascade of...
research
06/20/2017

SPLBoost: An Improved Robust Boosting Algorithm Based on Self-paced Learning

It is known that Boosting can be interpreted as a gradient descent techn...

Please sign up or login with your details

Forgot password? Click here to reset