Learning Costs for Structured Monge Displacements

06/20/2023
by   Michal Klein, et al.
0

Optimal transport theory has provided machine learning with several tools to infer a push-forward map between densities from samples. While this theory has recently seen tremendous methodological developments in machine learning, its practical implementation remains notoriously difficult, because it is plagued by both computational and statistical challenges. Because of such difficulties, existing approaches rarely depart from the default choice of estimating such maps with the simple squared-Euclidean distance as the ground cost, c(x,y)=x-y^2_2. We follow a different path in this work, with the motivation of learning a suitable cost structure to encourage maps to transport points along engineered features. We extend the recently proposed Monge-Bregman-Occam pipeline <cit.>, that rests on an alternative cost formulation that is also cost-invariant c(x,y)=h(x-y), but which adopts a more general form as h=12 ℓ_2^2+τ, where τ is an appropriately chosen regularizer. We first propose a method that builds upon proximal gradient descent to generate ground truth transports for such structured costs, using the notion of h-transforms and h-concave potentials. We show more generally that such a method can be extended to compute h-transforms for entropic potentials. We study a regularizer that promotes transport displacements in low-dimensional spaces, and propose to learn such a basis change using Riemannian gradient descent on the Stiefel manifold. We show that these changes lead to estimators that are more robust and easier to interpret.

READ FULL TEXT
research
02/09/2023

The Monge Gap: A Regularizer to Learn All Transport Maps

Optimal transport (OT) theory has been been used in machine learning to ...
research
09/18/2020

SISTA: learning optimal transport costs under sparsity constraints

In this paper, we describe a novel iterative procedure called SISTA to l...
research
02/08/2023

Monge, Bregman and Occam: Interpretable Optimal Transport in High-Dimensions with Feature-Sparse Maps

Optimal transport (OT) theory focuses, among all maps T:ℝ^d→ℝ^d that can...
research
05/26/2019

Regularity as Regularization: Smooth and Strongly Convex Brenier Potentials in Optimal Transport

The problem of estimating Wasserstein distances in high-dimensional spac...
research
03/09/2019

Stochastic Incremental Algorithms for Optimal Transport with SON Regularizer

We introduce a new regularizer for optimal transport (OT) which is tailo...
research
06/12/2020

Linear Time Sinkhorn Divergences using Positive Features

Although Sinkhorn divergences are now routinely used in data sciences to...

Please sign up or login with your details

Forgot password? Click here to reset