Measuring the Algorithmic Convergence of Randomized Ensembles: The Regression Setting

08/04/2019
by   Miles E. Lopes, et al.
0

When randomized ensemble methods such as bagging and random forests are implemented, a basic question arises: Is the ensemble large enough? In particular, the practitioner desires a rigorous guarantee that a given ensemble will perform nearly as well as an ideal infinite ensemble (trained on the same data). The purpose of the current paper is to develop a bootstrap method for solving this problem in the context of regression --- which complements our companion paper in the context of classification (Lopes 2019). In contrast to the classification setting, the current paper shows that theoretical guarantees for the proposed bootstrap can be established under much weaker assumptions. In addition, we illustrate the flexibility of the method by showing how it can be adapted to measure algorithmic convergence for variable selection. Lastly, we provide numerical results demonstrating that the method works well in a range of situations.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/20/2019

Estimating the Algorithmic Variance of Randomized Ensembles via the Bootstrap

Although the methods of bagging and random forests are some of the most ...
research
06/01/2015

Bootstrap Bias Corrections for Ensemble Methods

This paper examines the use of a residual bootstrap for bias correction ...
research
10/24/2017

Estimating the Operating Characteristics of Ensemble Methods

In this paper we present a technique for using the bootstrap to estimate...
research
04/15/2022

Towards a Unified Framework for Uncertainty-aware Nonlinear Variable Selection with Theoretical Guarantees

We develop a simple and unified framework for nonlinear variable selecti...
research
05/19/2022

Consistent Interpolating Ensembles via the Manifold-Hilbert Kernel

Recent research in the theory of overparametrized learning has sought to...
research
02/20/2023

Progressive Knowledge Distillation: Building Ensembles for Efficient Inference

We study the problem of progressive distillation: Given a large, pre-tra...

Please sign up or login with your details

Forgot password? Click here to reset