Weakly Supervised Representation Learning with Sparse Perturbations

06/02/2022
by   Kartik Ahuja, et al.
30

The theory of representation learning aims to build methods that provably invert the data generating process with minimal domain knowledge or any source of supervision. Most prior approaches require strong distributional assumptions on the latent variables and weak supervision (auxiliary information such as timestamps) to provide provable identification guarantees. In this work, we show that if one has weak supervision from observations generated by sparse perturbations of the latent variables–e.g. images in a reinforcement learning environment where actions move individual sprites–identification is achievable under unknown continuous latent distributions. We show that if the perturbations are applied only on mutually exclusive blocks of latents, we identify the latents up to those blocks. We also show that if these perturbation blocks overlap, we identify latents up to the smallest blocks shared across perturbations. Consequently, if there are blocks that intersect in one latent variable only, then such latents are identified up to permutation and scaling. We propose a natural estimation procedure based on this theory and illustrate it on low-dimensional synthetic and image-based experiments.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/05/2023

Additive Decoders for Latent Variables Identification and Cartesian-Product Extrapolation

We tackle the problems of latent variables identification and "out-of-su...
research
10/29/2021

Properties from Mechanisms: An Equivariance Perspective on Identifiable Representation Learning

A key goal of unsupervised representation learning is "inverting" a data...
research
05/24/2022

RevUp: Revise and Update Information Bottleneck for Event Representation

In machine learning, latent variables play a key role to capture the und...
research
07/21/2023

Sparse plus low-rank identification for dynamical latent-variable graphical AR models

This paper focuses on the identification of graphical autoregressive mod...
research
02/27/2020

Fast and Three-rious: Speeding Up Weak Supervision with Triplet Methods

Weak supervision is a popular method for building machine learning model...
research
05/23/2018

Learning latent variable structured prediction models with Gaussian perturbations

The standard margin-based structured prediction commonly uses a maximum ...
research
11/22/2020

Non-Identifiability in Network Autoregressions

We study identification in autoregressions defined on a general network....

Please sign up or login with your details

Forgot password? Click here to reset