FSMJ: Feature Selection with Maximum Jensen-Shannon Divergence for Text Categorization

06/20/2016
by   Bo Tang, et al.
0

In this paper, we present a new wrapper feature selection approach based on Jensen-Shannon (JS) divergence, termed feature selection with maximum JS-divergence (FSMJ), for text categorization. Unlike most existing feature selection approaches, the proposed FSMJ approach is based on real-valued features which provide more information for discrimination than binary-valued features used in conventional approaches. We show that the FSMJ is a greedy approach and the JS-divergence monotonically increases when more features are selected. We conduct several experiments on real-life data sets, compared with the state-of-the-art feature selection approaches for text categorization. The superior performance of the proposed FSMJ approach demonstrates its effectiveness and further indicates its wide potential applications on data mining.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/09/2016

Toward Optimal Feature Selection in Naive Bayes for Text Categorization

Automated feature selection is important for text categorization to redu...
research
05/11/2016

EEF: Exponentially Embedded Families with Class-Specific Features for Classification

In this letter, we present a novel exponentially embedded families (EEF)...
research
01/14/2021

A Nature-Inspired Feature Selection Approach based on Hypercomplex Information

Feature selection for a given model can be transformed into an optimizat...
research
05/03/2013

Feature Selection Based on Term Frequency and T-Test for Text Categorization

Much work has been done on feature selection. Existing methods are based...
research
02/04/2018

Heuristic Feature Selection for Clickbait Detection

We study feature selection as a means to optimize the baseline clickbait...
research
11/30/2018

Unsupervised learning with GLRM feature selection reveals novel traumatic brain injury phenotypes

Baseline injury categorization is important to traumatic brain injury (T...
research
10/09/2018

Deep supervised feature selection using Stochastic Gates

In this study, we propose a novel non-parametric embedded feature select...

Please sign up or login with your details

Forgot password? Click here to reset