Gradient Boosted Feature Selection

01/13/2019
by   Zhixiang Eddie Xu, et al.
0

A feature selection algorithm should ideally satisfy four conditions: reliably extract relevant features; be able to identify non-linear feature interactions; scale linearly with the number of features and dimensions; allow the incorporation of known sparsity structure. In this work we propose a novel feature selection algorithm, Gradient Boosted Feature Selection (GBFS), which satisfies all four of these requirements. The algorithm is flexible, scalable, and surprisingly straight-forward to implement as it is based on a modification of Gradient Boosted Trees. We evaluate GBFS on several real world data sets and show that it matches or out-performs other state of the art feature selection algorithms. Yet it scales to larger data set sizes and naturally allows for domain-specific side information.

READ FULL TEXT
research
04/21/2017

Feature selection algorithm based on Catastrophe model to improve the performance of regression analysis

In this paper we introduce a new feature selection algorithm to remove t...
research
09/15/2019

Target-Focused Feature Selection Using a Bayesian Approach

In many real-world scenarios where data is high dimensional, test time a...
research
10/08/2013

Feature Selection Strategies for Classifying High Dimensional Astronomical Data Sets

The amount of collected data in many scientific fields is increasing, al...
research
10/22/2019

Orthogonal variance decomposition based feature selection

Existing feature selection methods fail to properly account for interact...
research
08/27/2019

Feature Gradients: Scalable Feature Selection via Discrete Relaxation

In this paper we introduce Feature Gradients, a gradient-based search al...
research
08/31/2016

A Mathematical Framework for Feature Selection from Real-World Data with Non-Linear Observations

In this paper, we study the challenge of feature selection based on a re...
research
04/30/2019

A scalable saliency-based Feature selection method with instance level information

Classic feature selection techniques remove those features that are eith...

Please sign up or login with your details

Forgot password? Click here to reset