Controlling Wasserstein distances by Kernel norms with application to Compressive Statistical Learning

12/01/2021
by   Titouan Vayer, et al.
0

Comparing probability distributions is at the crux of many machine learning algorithms. Maximum Mean Discrepancies (MMD) and Optimal Transport distances (OT) are two classes of distances between probability measures that have attracted abundant attention in past years. This paper establishes some conditions under which the Wasserstein distance can be controlled by MMD norms. Our work is motivated by the compressive statistical learning (CSL) theory, a general framework for resource-efficient large scale learning in which the training data is summarized in a single vector (called sketch) that captures the information relevant to the considered learning task. Inspired by existing results in CSL, we introduce the Hölder Lower Restricted Isometric Property (Hölder LRIP) and show that this property comes with interesting guarantees for compressive statistical learning. Based on the relations between the MMD and the Wasserstein distance, we provide guarantees for compressive statistical learning by introducing and studying the concept of Wasserstein learnability of the learning task, that is when some task-specific metric between probability distributions can be bounded by a Wasserstein distance.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/17/2020

Statistical Learning Guarantees for Compressive Clustering and Compressive Mixture Modeling

We provide statistical learning guarantees for two unsupervised learning...
research
06/22/2017

Compressive Statistical Learning with Random Feature Moments

We describe a general framework --compressive statistical learning-- for...
research
01/11/2021

From Smooth Wasserstein Distance to Dual Sobolev Norm: Empirical Approximation and Statistical Applications

Statistical distances, i.e., discrepancy measures between probability di...
research
10/15/2021

Compressive Independent Component Analysis: Theory and Algorithms

Compressive learning forms the exciting intersection between compressed ...
research
06/14/2022

A theory of learning with constrained weight-distribution

A central question in computational neuroscience is how structure determ...
research
06/07/2020

Information Mandala: Statistical Distance Matrix with Its Clustering

In machine learning, observation features are measured in a metric space...
research
06/14/2018

Statistical Aspects of Wasserstein Distances

Wasserstein distances are metrics on probability distributions inspired ...

Please sign up or login with your details

Forgot password? Click here to reset