Closed-form Expressions for Maximum Mean Discrepancy with Applications to Wasserstein Auto-Encoders

01/10/2019
by   Raif M. Rustamov, et al.
0

The Maximum Mean Discrepancy (MMD) has found numerous applications in statistics and machine learning, most recently as a penalty in the Wasserstein Auto-Encoder (WAE). In this paper we compute closed-form expressions for estimating the Gaussian kernel based MMD between a given distribution and the standard multivariate normal distribution. We introduce the standardized version of MMD as a penalty for the WAE training objective, allowing for a better interpretability of MMD values and more compatibility across different hyperparameter settings. Next, we propose using a version of batch normalization at the code layer; this has the benefits of making the kernel width selection easier, reducing the training effort, and preventing outliers in the aggregate code distribution. Finally, we discuss the appropriate null distributions and provide thresholds for multivariate normality testing with the standardized MMD, leading to a number of easy rules of thumb for monitoring the progress of WAE training. Curiously, our MMD formula reveals a connection to the Baringhaus-Henze-Epps-Pulley (BHEP) statistic of the Henze-Zirkler test and provides further insights about the MMD. Our experiments on synthetic and real data show that the analytic formulation improves over the commonly used stochastic approximation of the MMD, and demonstrate that code normalization provides significant benefits when training WAEs.

READ FULL TEXT
research
05/20/2020

Tessellated Wasserstein Auto-Encoders

Non-adversarial generative models such as variational auto-encoder (VAE)...
research
05/21/2023

Kernel Stein Discrepancy on Lie Groups: Theory and Applications

Distributional approximation is a fundamental problem in machine learnin...
research
06/06/2021

Neural Tangent Kernel Maximum Mean Discrepancy

We present a novel neural network Maximum Mean Discrepancy (MMD) statist...
research
02/08/2019

Asymptotics and practical aspects of testing normality with kernel methods

This paper is concerned with testing normality in a Hilbert space based ...
research
12/21/2019

Closed Form Variances for Variational Auto-Encoders

We propose a reformulation of Variational Auto-Encoders eliminating half...
research
06/11/2019

Maximum Mean Discrepancy Gradient Flow

We construct a Wasserstein gradient flow of the maximum mean discrepancy...

Please sign up or login with your details

Forgot password? Click here to reset