Faster Boosting with Smaller Memory

01/25/2019
by   Julaiti Alafate, et al.
22

The two state-of-the-art implementations of boosted trees: XGBoost and LightGBM, can process large training sets extremely fast. However, this performance requires that memory size is sufficient to hold a 2-3 multiple of the training set size. This paper presents an alternative approach to implementing boosted trees. which achieves a significant speedup over XGBoost and LightGBM, especially when memory size is small. This is achieved using a combination of two techniques: early stopping and stratified sampling, which are explained and analyzed in the paper. We describe our implementation and present experimental results to support our claims.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/15/2023

An Efficient B-tree Implementation for Memory-Constrained Embedded Systems

Embedded devices collect and process significant amounts of data in a va...
research
07/16/2019

Online Local Boosting: improving performance in online decision trees

As more data are produced each day, and faster, data stream mining is gr...
research
06/28/2021

Experimental Comparison of PC-Trees and PQ-Trees

PQ-trees and PC-trees are data structures that represent sets of linear ...
research
10/31/2017

TF Boosted Trees: A scalable TensorFlow based framework for gradient boosting

TF Boosted Trees (TFBT) is a new open-sourced frame-work for the distrib...
research
04/30/2018

Practical Low-Dimensional Halfspace Range Space Sampling

We develop, analyze, implement, and compare new algorithms for creating ...
research
07/28/2022

Implementing the Diverting Fast Radix Algorithm

This paper presents ThielSort, a practical implementation of the Diverti...
research
09/09/2020

Learning Slab Classes to Alleviate Memory Holes in Memcached

We consider the problem of memory holes in slab allocators, where an ite...

Please sign up or login with your details

Forgot password? Click here to reset