Harnessing The Collective Wisdom: Fusion Learning Using Decision Sequences From Diverse Sources

08/21/2023
by   Trambak Banerjee, et al.
0

Learning from the collective wisdom of crowds enhances the transparency of scientific findings by incorporating diverse perspectives into the decision-making process. Synthesizing such collective wisdom is related to the statistical notion of fusion learning from multiple data sources or studies. However, fusing inferences from diverse sources is challenging since cross-source heterogeneity and potential data-sharing complicate statistical inference. Moreover, studies may rely on disparate designs, employ widely different modeling techniques for inferences, and prevailing data privacy norms may forbid sharing even summary statistics across the studies for an overall analysis. In this paper, we propose an Integrative Ranking and Thresholding (IRT) framework for fusion learning in multiple testing. IRT operates under the setting where from each study a triplet is available: the vector of binary accept-reject decisions on the tested hypotheses, the study-specific False Discovery Rate (FDR) level and the hypotheses tested by the study. Under this setting, IRT constructs an aggregated, nonparametric, and discriminatory measure of evidence against each null hypotheses, which facilitates ranking the hypotheses in the order of their likelihood of being rejected. We show that IRT guarantees an overall FDR control under arbitrary dependence between the evidence measures as long as the studies control their respective FDR at the desired levels. Furthermore, IRT synthesizes inferences from diverse studies irrespective of the underlying multiple testing algorithms employed by them. While the proofs of our theoretical statements are elementary, IRT is extremely flexible, and a comprehensive numerical study demonstrates that it is a powerful framework for pooling inferences.

READ FULL TEXT
research
11/13/2020

Nonparametric fusion learning: synthesize inferences from diverse sources using depth confidence distribution

Fusion learning refers to synthesizing inferences from multiple sources ...
research
06/24/2021

Multiple Testing for Composite Null with FDR Control Guarantee

False discovery rate (FDR) controlling procedures provide important stat...
research
10/19/2022

Combining Data from Surveys and Related Sources

To improve the precision of inferences and reduce costs there is conside...
research
05/19/2021

Testing partial conjunction hypotheses under dependency, with applications to meta-analysis

In many statistical problems the hypotheses are naturally divided into g...
research
03/18/2017

A unified treatment of multiple testing with prior knowledge using the p-filter

A significant literature studies ways of employing prior knowledge to im...
research
10/05/2020

Quantifying Statistical Significance of Neural Network Representation-Driven Hypotheses by Selective Inference

In the past few years, various approaches have been developed to explain...
research
07/27/2018

On the expected runtime of multiple testing algorithms with bounded error

Consider the testing of multiple hypotheses in the setting where the p-v...

Please sign up or login with your details

Forgot password? Click here to reset