A Rademacher Complexity Based Method fo rControlling Power and Confidence Level in Adaptive Statistical Analysis

10/04/2019
by   Lorenzo De Stefani, et al.
0

While standard statistical inference techniques and machine learning generalization bounds assume that tests are run on data selected independently of the hypotheses, practical data analysis and machine learning are usually iterative and adaptive processes where the same holdout data is often used for testing a sequence of hypotheses (or models), which may each depend on the outcome of the previous tests on the same data. In this work, we present RadaBound a rigorous, efficient and practical procedure for controlling the generalization error when using a holdout sample for multiple adaptive testing. Our solution is based on a new application of the Rademacher Complexity generalization bounds, adapted to dependent tests. We demonstrate the statistical power and practicality of our method through extensive simulations and comparisons to alternative approaches.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/27/2018

The Structure of Optimal Private Tests for Simple Hypotheses

Hypothesis testing plays a central role in statistical inference, and is...
research
10/19/2021

Simulating the Power of Statistical Tests: A Collection of R Examples

This paper illustrates how to calculate the power of a statistical test ...
research
02/10/2021

Once is Never Enough: Foundations for Sound Statistical Inference in Tor Network Experimentation

Tor is a popular low-latency anonymous communication system that focuses...
research
07/27/2023

Rapid and Scalable Bayesian AB Testing

AB testing aids business operators with their decision making, and is co...
research
04/24/2017

Data-adaptive statistics for multiple hypothesis testing in high-dimensional settings

Current statistical inference problems in areas like astronomy, genomics...
research
10/31/2022

The role of prior information and computational power in Machine Learning

Science consists on conceiving hypotheses, confronting them with empiric...
research
11/01/2021

Statistical quantification of confounding bias in predictive modelling

The lack of non-parametric statistical tests for confounding bias signif...

Please sign up or login with your details

Forgot password? Click here to reset