Scalable Greedy Feature Selection via Weak Submodularity

03/08/2017
by   Rajiv Khanna, et al.
0

Greedy algorithms are widely used for problems in machine learning such as feature selection and set function optimization. Unfortunately, for large datasets, the running time of even greedy algorithms can be quite high. This is because for each greedy step we need to refit a model or calculate a function using the previously selected choices and the new candidate. Two algorithms that are faster approximations to the greedy forward selection were introduced recently ([Mirzasoleiman et al. 2013, 2015]). They achieve better performance by exploiting distributed computation and stochastic evaluation respectively. Both algorithms have provable performance guarantees for submodular functions. In this paper we show that divergent from previously held opinion, submodularity is not required to obtain approximation guarantees for these two algorithms. Specifically, we show that a generalized concept of weak submodularity suffices to give multiplicative approximation guarantees. Our result extends the applicability of these algorithms to a larger class of functions. Furthermore, we show that a bounded submodularity ratio can be used to provide data dependent bounds that can sometimes be tighter also for submodular functions. We empirically validate our work by showing superior performance of fast greedy approximations versus several established baselines on artificial and real datasets.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/19/2011

Submodular meets Spectral: Greedy Algorithms for Subset Selection, Sparse Approximation and Dictionary Selection

We study the problem of selecting a subset of k random variables from a ...
research
02/28/2022

Fast Feature Selection with Fairness Constraints

We study the fundamental problem of selecting optimal features for model...
research
11/21/2020

Near-Optimal Data Source Selection for Bayesian Learning

We study a fundamental problem in Bayesian learning, where the goal is t...
research
03/06/2019

Fast Parallel Algorithms for Feature Selection

In this paper, we analyze a fast parallel algorithm to efficiently selec...
research
04/30/2020

Weakly Submodular Function Maximization Using Local Submodularity Ratio

Weak submodularity is a natural relaxation of the diminishing return pro...
research
04/24/2019

Beyond Adaptive Submodularity: Approximation Guarantees of Greedy Policy with Adaptive Submodularity Ratio

We propose a new concept named adaptive submodularity ratio to study the...
research
07/04/2020

Lazy Greedy Hypervolume Subset Selection from Large Candidate Solution Sets

Subset selection is a popular topic in recent years and a number of subs...

Please sign up or login with your details

Forgot password? Click here to reset