Testing Hypotheses by Regularized Maximum Mean Discrepancy

05/02/2013
by   Somayeh Danafar, et al.
0

Do two data samples come from different distributions? Recent studies of this fundamental problem focused on embedding probability distributions into sufficiently rich characteristic Reproducing Kernel Hilbert Spaces (RKHSs), to compare distributions by the distance between their embeddings. We show that Regularized Maximum Mean Discrepancy (RMMD), our novel measure for kernel-based hypothesis testing, yields substantial improvements even when sample sizes are small, and excels at hypothesis tests involving multiple comparisons with power control. We derive asymptotic distributions under the null and alternative hypotheses, and assess power control. Outstanding results are obtained on: challenging EEG data, MNIST, the Berkley Covertype, and the Flare-Solar dataset.

READ FULL TEXT
research
11/22/2018

k-Sample problem based on generalized maximum mean discrepancy

In this paper we deal with the problem of testing for the quality of k p...
research
03/23/2022

Kernel Robust Hypothesis Testing

The problem of robust hypothesis testing is studied, where under the nul...
research
10/15/2012

The Perturbed Variation

We introduce a new discrepancy score between two distributions that give...
research
03/22/2017

Testing and Learning on Distributions with Symmetric Noise Invariance

Kernel embeddings of distributions and the Maximum Mean Discrepancy (MMD...
research
07/07/2022

Neural Stein critics with staged L^2-regularization

Learning to differentiate model distributions from observed data is a fu...
research
01/31/2022

Nyström Kernel Mean Embeddings

Kernel mean embeddings are a powerful tool to represent probability dist...
research
04/09/2017

Strictly Proper Kernel Scoring Rules and Divergences with an Application to Kernel Two-Sample Hypothesis Testing

We study strictly proper scoring rules in the Reproducing Kernel Hilbert...

Please sign up or login with your details

Forgot password? Click here to reset