On Estimating Many Means, Selection Bias, and the Bootstrap

11/15/2013
by   Noah Simon, et al.
0

With recent advances in high throughput technology, researchers often find themselves running a large number of hypothesis tests (thousands+) and esti- mating a large number of effect-sizes. Generally there is particular interest in those effects estimated to be most extreme. Unfortunately naive estimates of these effect-sizes (even after potentially accounting for multiplicity in a testing procedure) can be severely biased. In this manuscript we explore this bias from a frequentist perspective: we give a formal definition, and show that an oracle estimator using this bias dominates the naive maximum likelihood estimate. We give a resampling estimator to approximate this oracle, and show that it works well on simulated data. We also connect this to ideas in empirical Bayes.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/16/2014

Selection Bias Correction and Effect Size Estimation under Dependence

We consider large-scale studies in which it is of interest to test a ver...
research
01/09/2023

Testing for an ignorable sampling bias under random double truncation

In clinical and epidemiological research doubly truncated data often app...
research
07/27/2019

Estimating the Random Effect in Big Data Mixed Models

We consider three problems in high-dimensional Gaussian linear mixed mod...
research
12/20/2021

Adapting the Hill estimator to distributed inference: dealing with the bias

The distributed Hill estimator is a divide-and-conquer algorithm for est...
research
10/29/2019

Irrational Exuberance: Correcting Bias in Probability Estimates

We consider the common setting where one observes probability estimates ...

Please sign up or login with your details

Forgot password? Click here to reset