Sparse learning with CART

06/07/2020
by   Jason M. Klusowski, et al.
0

Decision trees with binary splits are popularly constructed using Classification and Regression Trees (CART) methodology. For regression models, this approach recursively divides the data into two near-homogenous daughter nodes according to a split point that maximizes the reduction in sum of squares error (the impurity) along a particular variable. This paper aims to study the statistical properties of regression trees constructed with CART. In doing so, we find that the training error is governed by the Pearson correlation between the optimal decision stump and response data in each node, which we bound by constructing a prior distribution on the split points and solving a quadratic program. We leverage this connection between the training error and Pearson correlation to show that CART with cost-complexity pruning achieves an optimal complexity/goodness-of-fit tradeoff when the depth scales with the logarithm of the sample size. Data dependent quantities, which adapt to the dimensionality and latent structure of the regression model, are seen to govern the rates of convergence of the prediction error.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/24/2019

Best Split Nodes for Regression Trees

Decision trees with binary splits are popularly constructed using Classi...
research
04/28/2021

Universal Consistency of Decision Trees in High Dimensions

This paper shows that decision trees constructed with Classification and...
research
10/26/2022

Convergence Rates of Oblique Regression Trees for Flexible Function Libraries

We develop a theoretical framework for the analysis of oblique decision ...
research
10/09/2020

Learning Binary Trees via Sparse Relaxation

One of the most classical problems in machine learning is how to learn b...
research
06/12/2023

Prediction Algorithms Achieving Bayesian Decision Theoretical Optimality Based on Decision Trees as Data Observation Processes

In the field of decision trees, most previous studies have difficulty en...
research
05/24/2019

LdSM: Logarithm-depth Streaming Multi-label Decision Trees

We consider multi-label classification where the goal is to annotate eac...
research
06/29/2022

conformalInference.multi and conformalInference.fd: Twin Packages for Conformal Prediction

Building on top of a regression model, Conformal Prediction methods prod...

Please sign up or login with your details

Forgot password? Click here to reset