ADAGES: adaptive aggregation with stability for distributed feature selection

07/21/2020
by   Yu Gui, et al.
0

In this era of "big" data, not only the large amount of data keeps motivating distributed computing, but concerns on data privacy also put forward the emphasis on distributed learning. To conduct feature selection and to control the false discovery rate in a distributed pattern with multi-machines or multi-institutions, an efficient aggregation method is necessary. In this paper, we propose an adaptive aggregation method called ADAGES which can be flexibly applied to any machine-wise feature selection method. We will show that our method is capable of controlling the overall FDR with a theoretical foundation while maintaining power as good as the Union aggregation rule in practice.

READ FULL TEXT
research
04/28/2022

Controlling the False Discovery Rate via knockoffs: is the +1 needed?

Barber and Candès (2015) control of the FDR in feature selection relies ...
research
08/05/2022

Feature Selection for Machine Learning Algorithms that Bounds False Positive Rate

The problem of selecting a handful of truly relevant variables in superv...
research
02/21/2020

Aggregation of Multiple Knockoffs

We develop an extension of the Knockoff Inference procedure, introduced ...
research
10/17/2019

Dropping forward-backward algorithms for feature selection

In this era of big data, feature selection techniques, which have long b...
research
01/31/2019

Distributed Correlation-Based Feature Selection in Spark

CFS (Correlation-Based Feature Selection) is an FS algorithm that has be...
research
10/24/2014

Median Selection Subset Aggregation for Parallel Inference

For massive data sets, efficient computation commonly relies on distribu...
research
08/17/2020

Privacy-preserving feature selection: A survey and proposing a new set of protocols

Feature selection is the process of sieving features, in which informati...

Please sign up or login with your details

Forgot password? Click here to reset