Learning Latent Permutations with Gumbel-Sinkhorn Networks

02/23/2018
by   Gonzalo Mena, et al.
0

Permutations and matchings are core building blocks in a variety of latent variable models, as they allow us to align, canonicalize, and sort data. Learning in such models is difficult, however, because exact marginalization over these combinatorial objects is intractable. In response, this paper introduces a collection of new methods for end-to-end learning in such models that approximate discrete maximum-weight matching using the continuous Sinkhorn operator. Sinkhorn iteration is attractive because it functions as a simple, easy-to-implement analog of the softmax operator. With this, we can define the Gumbel-Sinkhorn method, an extension of the Gumbel-Softmax method (Jang et al. 2016, Maddison2016 et al. 2016) to distributions over latent matchings. We demonstrate the effectiveness of our method by outperforming competitive baselines on a range of qualitatively different tasks: sorting numbers, solving jigsaw puzzles, and identifying neural signals in worms.

READ FULL TEXT

page 4

page 8

page 21

research
07/01/2019

A Kernel Stein Test for Comparing Latent Variable Models

We propose a nonparametric, kernel-based test to assess the relative goo...
research
02/14/2018

DVAE++: Discrete Variational Autoencoders with Overlapping Transformations

Training of discrete latent variable models remains challenging because ...
research
06/15/2020

Gradient Estimation with Stochastic Softmax Tricks

The Gumbel-Max trick is the basis of many relaxed gradient estimators. T...
research
10/23/2015

Fast Latent Variable Models for Inference and Visualization on Mobile Devices

In this project we outline Vedalia, a high performance distributed netwo...
research
12/05/2018

Relative Entropy Regularized Policy Iteration

We present an off-policy actor-critic algorithm for Reinforcement Learni...
research
06/29/2020

SoftSort: A Continuous Relaxation for the argsort Operator

While sorting is an important procedure in computer science, the argsort...
research
12/16/2016

An Alternative Softmax Operator for Reinforcement Learning

A softmax operator applied to a set of values acts somewhat like the max...

Please sign up or login with your details

Forgot password? Click here to reset