Exploiting Categorical Structure Using Tree-Based Methods

04/15/2020
by   Brian Lucena, et al.
0

Standard methods of using categorical variables as predictors either endow them with an ordinal structure or assume they have no structure at all. However, categorical variables often possess structure that is more complicated than a linear ordering can capture. We develop a mathematical framework for representing the structure of categorical variables and show how to generalize decision trees to make use of this structure. This approach is applicable to methods such as Gradient Boosted Trees which use a decision tree as the underlying learner. We show results on weather data to demonstrate the improvement yielded by this approach.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/08/2020

StructureBoost: Efficient Gradient Boosting for Structured Categorical Variables

Gradient boosting methods based on Structured Categorical Decision Trees...
research
06/12/2017

Random Forests, Decision Trees, and Categorical Predictors: The "Absent Levels" Problem

One of the advantages that decision trees have over many other models is...
research
10/18/2022

Generalization Properties of Decision Trees on Real-valued and Categorical Features

We revisit binary decision trees from the perspective of partitions of t...
research
11/08/2022

A new BART prior for flexible modeling with categorical predictors

Default implementations of Bayesian Additive Regression Trees (BART) rep...
research
04/19/2020

Make E Smart Again

In this work in progress, we demonstrate a new use-case for the ENIGMA s...
research
11/06/2015

Finding structure in data using multivariate tree boosting

Technology and collaboration enable dramatic increases in the size of ps...
research
02/15/2021

Learning Accurate Decision Trees with Bandit Feedback via Quantized Gradient Descent

Decision trees provide a rich family of highly non-linear but efficient ...

Please sign up or login with your details

Forgot password? Click here to reset