SPSA-FSR: Simultaneous Perturbation Stochastic Approximation for Feature Selection and Ranking

04/16/2018
by   Zeren D. Yenice, et al.
0

This manuscript presents the following: (1) an improved version of the Binary Simultaneous Perturbation Stochastic Approximation (SPSA) Method for feature selection in machine learning (Aksakalli and Malekipirbazari, Pattern Recognition Letters, Vol. 75, 2016) based on non-monotone iteration gains computed via the Barzilai and Borwein (BB) method, (2) its adaptation for feature ranking, and (3) comparison against popular methods on public benchmark datasets. The improved method, which we call SPSA-FSR, dramatically reduces the number of iterations required for convergence without impacting solution quality. SPSA-FSR can be used for feature ranking and feature selection both for classification and regression problems. After a review of the current state-of-the-art, we discuss our improvements in detail and present three sets of computational experiments: (1) comparison of SPSA-FS as a (wrapper) feature selection method against sequential methods as well as genetic algorithms, (2) comparison of SPSA-FS as a feature ranking method in a classification setting against random forest importance, chi-squared, and information main methods, and (3) comparison of SPSA-FS as a feature ranking method in a regression setting against minimum redundancy maximum relevance (MRMR), RELIEF, and linear correlation methods. The number of features in the datasets we use range from a few dozens to a few thousands. Our results indicate that SPSA-FS converges to a good feature set in no more than 100 iterations and therefore it is quite fast for a wrapper method. SPSA-FS also outperforms popular feature selection as well as feature ranking methods in majority of test cases, sometimes by a large margin, and it stands as a promising new feature selection and ranking method.

READ FULL TEXT

page 27

page 28

page 29

page 30

page 31

page 32

page 33

page 35

research
08/30/2015

Feature Selection via Binary Simultaneous Perturbation Stochastic Approximation

Feature selection (FS) has become an indispensable task in dealing with ...
research
06/13/2019

Iterative subtraction method for Feature Ranking

Training features used to analyse physical processes are often highly co...
research
04/18/2017

Ranking to Learn: Feature Ranking and Selection via Eigenvector Centrality

In an era where accumulating data is easy and storing it inexpensive, fe...
research
06/15/2021

Canonical-Correlation-Based Fast Feature Selection

This paper proposes a canonical-correlation-based filter method for feat...
research
08/01/2023

Copula for Instance-wise Feature Selection and Ranking

Instance-wise feature selection and ranking methods can achieve a good s...
research
09/28/2016

Towards the effectiveness of Deep Convolutional Neural Network based Fast Random Forest Classifier

Deep Learning is considered to be a quite young in the area of machine l...
research
02/26/2019

A Feature Selection Based on Perturbation Theory

Consider a supervised dataset D=[A|b], where b is the outcome column, ro...

Please sign up or login with your details

Forgot password? Click here to reset