Sample Complexity of Sinkhorn divergences

10/05/2018
by   Aude Genevay, et al.
0

Optimal transport (OT) and maximum mean discrepancies (MMD) are now routinely used in machine learning to compare probability measures. We focus in this paper on Sinkhorn divergences (SDs), a regularized variant of OT distances which can interpolate, depending on the regularization strength ε, between OT (ε=0) and MMD (ε=∞). Although the tradeoff induced by that regularization is now well understood computationally (OT, SDs and MMD require respectively O(n^3 n), O(n^2) and n^2 operations given a sample size n), much less is known in terms of their sample complexity, namely the gap between these quantities, when evaluated using finite samples vs. their respective densities. Indeed, while the sample complexity of OT and MMD stand at two extremes, 1/n^1/d for OT in dimension d and 1/√(n) for MMD, that for SDs has only been studied empirically. In this paper, we (i) derive a bound on the approximation error made with SDs when approximating OT as a function of the regularizer ε, (ii) prove that the optimizers of regularized OT are bounded in a Sobolev (RKHS) ball independent of the two measures and (iii) provide the first sample complexity bound for SDs, obtained,by reformulating SDs as a maximization problem in a RKHS. We thus obtain a scaling in 1/√(n) (as in MMD), with a constant that depends however on ε, making the bridge between OT and MMD complete.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/03/2023

Sample Complexity of Probability Divergences under Group Symmetry

We rigorously quantify the improvement in the sample complexity of varia...
research
05/28/2019

Statistical bounds for entropic optimal transport: sample complexity and the central limit theorem

We prove several fundamental statistical bounds for entropic OT with the...
research
07/07/2021

Sample complexity of hidden subgroup problem

The hidden subgroup problem (𝖧𝖲𝖯) has been attracting much attention in ...
research
02/26/2020

Decidability of Sample Complexity of PAC Learning in finite setting

In this short note we observe that the sample complexity of PAC machine ...
research
05/21/2020

Computationally efficient stochastic MPC: a probabilistic scaling approach

In recent years, the increasing interest in Stochastic model predictive ...
research
10/09/2021

A Faster Algorithm for Max Cut in Dense Graphs

We design an algorithm for approximating the size of Max Cut in dense gr...
research
02/24/2020

Learning the mapping x∑_i=1^d x_i^2: the cost of finding the needle in a haystack

The task of using machine learning to approximate the mapping x∑_i=1^d x...

Please sign up or login with your details

Forgot password? Click here to reset