Planning from Pixels using Inverse Dynamics Models

12/04/2020
by   Keiran Paster, et al.
0

Learning task-agnostic dynamics models in high-dimensional observation spaces can be challenging for model-based RL agents. We propose a novel way to learn latent world models by learning to predict sequences of future actions conditioned on task completion. These task-conditioned models adaptively focus modeling capacity on task-relevant dynamics, while simultaneously serving as an effective heuristic for planning with sparse rewards. We evaluate our method on challenging visual goal completion tasks and show a substantial increase in performance compared to prior model-free approaches.

READ FULL TEXT

page 2

page 17

page 18

research
11/12/2018

Learning Latent Dynamics for Planning from Pixels

Planning has been very successful for control tasks with known environme...
research
07/02/2019

Dynamics-Aware Unsupervised Discovery of Skills

Conventionally, model-based reinforcement learning (MBRL) aims to learn ...
research
07/14/2020

Goal-Aware Prediction: Learning to Model What Matters

Learned dynamics models combined with both planning and policy learning ...
research
12/02/2021

Robust Robotic Control from Pixels using Contrastive Recurrent State-Space Models

Modeling the world can benefit robot learning by providing a rich traini...
research
10/23/2020

CLOUD: Contrastive Learning of Unsupervised Dynamics

Developing agents that can perform complex control tasks from high dimen...
research
12/08/2020

Models, Pixels, and Rewards: Evaluating Design Trade-offs in Visual Model-Based Reinforcement Learning

Model-based reinforcement learning (MBRL) methods have shown strong samp...
research
06/13/2020

Hindsight Expectation Maximization for Goal-conditioned Reinforcement Learning

We propose a graphical model framework for goal-conditioned RL, with an ...

Please sign up or login with your details

Forgot password? Click here to reset