Naive Feature Selection: Sparsity in Naive Bayes

05/23/2019
by   Armin Askari, et al.
0

Due to its linear complexity, naive Bayes classification remains an attractive supervised learning method, especially in very large-scale settings. We propose a sparse version of naive Bayes, which can be used for feature selection. This leads to a combinatorial maximum-likelihood problem, for which we provide an exact solution in the case of binary data, or a bound in the multinomial case. We prove that our bound becomes tight as the marginal contribution of additional features decreases. Both binary and multinomial sparse models are solvable in time almost linear in problem size, representing a very small extra relative cost compared to the classical naive Bayes. Numerical experiments on text data show that the naive Bayes feature selection method is as statistically effective as state-of-the-art feature selection methods such as recursive feature elimination, l_1-penalized logistic regression and LASSO, while being orders of magnitude faster. For a large data set, having more than with 1.6 million training points and about 12 million features, and with a non-optimized CPU implementation, our sparse naive Bayes model can be trained in less than 15 seconds.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/12/2015

Search Strategies for Binary Feature Selection for a Naive Bayes Classifier

We compare in this paper several feature selection methods for the Naive...
research
03/04/2020

Multiclass classification by sparse multinomial logistic regression

In this paper we consider high-dimensional multiclass classification by ...
research
02/05/2023

Revisiting Discriminative vs. Generative Classifiers: Theory and Implications

A large-scale deep model pre-trained on massive labeled or unlabeled dat...
research
07/11/2012

Pre-Selection of Independent Binary Features: An Application to Diagnosing Scrapie in Sheep

Suppose that the only available information in a multi-class problem are...
research
12/31/2013

Feature Augmentation via Nonparametrics and Selection (FANS) in High Dimensional Classification

We propose a high dimensional classification method that involves nonpar...
research
04/26/2020

Classification of Cuisines from Sequentially Structured Recipes

Cultures across the world are distinguished by the idiosyncratic pattern...
research
10/02/2018

Feature Selection Approach with Missing Values Conducted for Statistical Learning: A Case Study of Entrepreneurship Survival Dataset

In this article, we investigate the features which enhanced discriminate...

Please sign up or login with your details

Forgot password? Click here to reset