Central limit theorems for Sinkhorn divergence between probability distributions on finite spaces and statistical applications

11/24/2017
by   Jérémie Bigot, et al.
0

The notion of Sinkhorn divergence has recently gained popularity in machine learning and statistics, as it makes feasible the use of smoothed optimal transportation distances for data analysis. The Sinkhorn divergence allows the fast computation of an entropically regularized Wasserstein distance between two probability distributions supported on a finite metric space of (possibly) high-dimension. For data sampled from one or two unknown probability distributions, we derive central limit theorems for empirical Sinkhorn divergences. We also propose a bootstrap procedure which allows to obtain new test statistics for measuring the discrepancies between multivariate probability distributions. The strategy of proof uses the notions of directional Hadamard differentiability and delta-method in this setting. It is inspired by the results in the work of Sommerfeld and Munk (2016) on the asymptotic distribution of empirical Wasserstein distance on finite space using un-regularized transportation costs. Simulated and real datasets are used to illustrate our approach. A comparison with existing methods to measure the discrepancy between multivariate distributions is also proposed.

READ FULL TEXT

page 12

page 15

page 16

page 17

page 18

page 20

research
02/11/2022

Inference for Projection-Based Wasserstein Distances on Finite Spaces

The Wasserstein distance is a distance between two probability distribut...
research
11/21/2022

Limit distribution theory for f-Divergences

f-divergences, which quantify discrepancy between probability distributi...
research
06/14/2018

Statistical Aspects of Wasserstein Distances

Wasserstein distances are metrics on probability distributions inspired ...
research
03/04/2020

The empirical Christoffel function with applications in data analysis

We illustrate the potential applications in machine learning of the Chri...
research
03/11/2021

Non-Asymptotic Performance Guarantees for Neural Estimation of 𝖿-Divergences

Statistical distances (SDs), which quantify the dissimilarity between pr...
research
07/07/2022

Neural Stein critics with staged L^2-regularization

Learning to differentiate model distributions from observed data is a fu...
research
12/21/2018

Asymptotic distribution and convergence rates of stochastic algorithms for entropic optimal transportation between probability measures

This paper is devoted to the stochastic approximation of entropically re...

Please sign up or login with your details

Forgot password? Click here to reset