Generalizing Gain Penalization for Feature Selection in Tree-based Models

06/12/2020
by   Bruna Wundervald, et al.
0

We develop a new approach for feature selection via gain penalization in tree-based models. First, we show that previous methods do not perform sufficient regularization and often exhibit sub-optimal out-of-sample performance, especially when correlated features are present. Instead, we develop a new gain penalization idea that exhibits a general local-global regularization for tree-based models. The new method allows for more flexibility in the choice of feature-specific importance weights. We validate our method on both simulated and real data and implement itas an extension of the popular R package ranger.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/07/2012

Feature Selection via Regularized Trees

We propose a tree regularization framework, which enables many tree mode...
research
09/02/2021

Inferring feature importance with uncertainties in high-dimensional data

Estimating feature importance is a significant aspect of explaining data...
research
11/08/2022

Individualized and Global Feature Attributions for Gradient Boosted Trees in the Presence of ℓ_2 Regularization

While ℓ_2 regularization is widely used in training gradient boosted tre...
research
10/13/2020

Neural Gaussian Mirror for Controlled Feature Selection in Neural Networks

Deep neural networks (DNNs) have become increasingly popular and achieve...
research
03/08/2022

Beam Search for Feature Selection

In this paper, we present and prove some consistency results about the p...
research
10/02/2020

Interactive Reinforcement Learning for Feature Selection with Decision Tree in the Loop

We study the problem of balancing effectiveness and efficiency in automa...
research
04/26/2020

Classification Trees for Imbalanced and Sparse Data: Surface-to-Volume Regularization

Classification algorithms face difficulties when one or more classes hav...

Please sign up or login with your details

Forgot password? Click here to reset