Prediction, Consistency, Curvature: Representation Learning for Locally-Linear Control

09/04/2019
by   Nir Levine, et al.
25

Many real-world sequential decision-making problems can be formulated as optimal control with high-dimensional observations and unknown dynamics. A promising approach is to embed the high-dimensional observations into a lower-dimensional latent representation space, estimate the latent dynamics model, then utilize this model for control in the latent space. An important open question is how to learn a representation that is amenable to existing control algorithms? In this paper, we focus on learning representations for locally-linear control algorithms, such as iterative LQR (iLQR). By formulating and analyzing the representation learning problem from an optimal control perspective, we establish three underlying principles that the learned representation should comprise: 1) accurate prediction in the observation space, 2) consistency between latent and observation space dynamics, and 3) low curvature in the latent space transitions. These principles naturally correspond to a loss function that consists of three terms: prediction, consistency, and curvature (PCC). Crucially, to make PCC tractable, we derive an amortized variational bound for the PCC loss function. Extensive experiments on benchmark domains demonstrate that the new variational-PCC learning algorithm benefits from significantly more stable and reproducible training, and leads to superior control performance. Further ablation studies give support to the importance of all three PCC components for learning a good latent space for control.

READ FULL TEXT
research
03/02/2020

Predictive Coding for Locally-Linear Control

High-dimensional observations and unknown dynamics are major challenges ...
research
06/24/2020

Control-Aware Representations for Model-based Reinforcement Learning

A major challenge in modern reinforcement learning (RL) is efficient con...
research
06/24/2015

Embed to Control: A Locally Linear Latent Dynamics Model for Control from Raw Images

We introduce Embed to Control (E2C), a method for model learning and con...
research
09/02/2022

Multi-Step Prediction in Linearized Latent State Spaces for Representation Learning

In this paper, we derive a novel method as a generalization over LCEs su...
research
08/18/2020

Heteroscedastic Uncertainty for Robust Generative Latent Dynamics

Learning or identifying dynamics from a sequence of high-dimensional obs...
research
03/29/2021

Bayesian Attention Networks for Data Compression

The lossless data compression algorithm based on Bayesian Attention Netw...
research
11/02/2019

Beta DVBF: Learning State-Space Models for Control from High Dimensional Observations

Learning a model of dynamics from high-dimensional images can be a core ...

Please sign up or login with your details

Forgot password? Click here to reset