A new robust feature selection method using variance-based sensitivity analysis

by   Saman Sadeghyan, et al.

Excluding irrelevant features in a pattern recognition task plays an important role in maintaining a simpler machine learning model and optimizing the computational efficiency. Nowadays with the rise of large scale datasets, feature selection is in great demand as it becomes a central issue when facing high-dimensional datasets. The present study provides a new measure of saliency for features by employing a Sensitivity Analysis (SA) technique called the extended Fourier amplitude sensitivity test, and a well-trained Feedforward Neural Network (FNN) model, which ultimately leads to the selection of a promising optimal feature subset. Ideas of the paper are mainly demonstrated based on adopting FNN model for feature selection in classification problems. But in the end, a generalization framework is discussed in order to give insights into the usage in regression problems as well as expressing how other function approximate models can be deployed. Effectiveness of the proposed method is verified by result analysis and data visualization for a series of experiments over several well-known datasets drawn from UCI machine learning repository.



There are no comments yet.


page 1

page 2

page 3

page 4


On Feature Selection Using Anisotropic General Regression Neural Network

The presence of irrelevant features in the input dataset tends to reduce...

Deep Feature Selection using a Teacher-Student Network

High-dimensional data in many machine learning applications leads to com...

Relevant based structure learning for feature selection

Feature selection is an important task in many problems occurring in pat...

MS-BACO: A new Model Selection algorithm using Binary Ant Colony Optimization for neural complexity and error reduction

Stabilizing the complexity of Feedforward Neural Networks (FNNs) for the...

IVFS: Simple and Efficient Feature Selection for High Dimensional Topology Preservation

Feature selection is an important tool to deal with high dimensional dat...

A Feature Selection Based on Perturbation Theory

Consider a supervised dataset D=[A|b], where b is the outcome column, ro...

MISSION: Ultra Large-Scale Feature Selection using Count-Sketches

Feature selection is an important challenge in machine learning. It play...
This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.