DreamerPro: Reconstruction-Free Model-Based Reinforcement Learning with Prototypical Representations

10/27/2021
by   Fei Deng, et al.
0

Top-performing Model-Based Reinforcement Learning (MBRL) agents, such as Dreamer, learn the world model by reconstructing the image observations. Hence, they often fail to discard task-irrelevant details and struggle to handle visual distractions. To address this issue, previous work has proposed to contrastively learn the world model, but the performance tends to be inferior in the absence of distractions. In this paper, we seek to enhance robustness to distractions for MBRL agents. Specifically, we consider incorporating prototypical representations, which have yielded more accurate and robust results than contrastive approaches in computer vision. However, it remains elusive how prototypical representations can benefit temporal dynamics learning in MBRL, since they treat each image independently without capturing temporal structures. To this end, we propose to learn the prototypes from the recurrent states of the world model, thereby distilling temporal structures from past observations and actions into the prototypes. The resulting model, DreamerPro, successfully combines Dreamer with prototypes, making large performance gains on the DeepMind Control suite both in the standard setting and when there are complex background distractions. Code available at https://github.com/fdeng18/dreamer-pro .

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/28/2023

Curious Replay for Model-based Adaptation

Agents must be able to adapt quickly as an environment changes. We find ...
research
04/08/2020

CURL: Contrastive Unsupervised Representations for Reinforcement Learning

We present CURL: Contrastive Unsupervised Representations for Reinforcem...
research
08/06/2020

Contrastive Variational Model-Based Reinforcement Learning for Complex Observations

Deep model-based reinforcement learning (MBRL) has achieved great sample...
research
03/27/2023

Model-Based Reinforcement Learning with Isolated Imaginations

World models learn the consequences of actions in vision-based interacti...
research
08/02/2018

Learning Actionable Representations from Visual Observations

In this work we explore a new approach for robots to teach themselves ab...
research
12/02/2021

Robust Robotic Control from Pixels using Contrastive Recurrent State-Space Models

Modeling the world can benefit robot learning by providing a rich traini...
research
11/15/2021

Learning Representations for Pixel-based Control: What Matters and Why?

Learning representations for pixel-based control has garnered significan...

Please sign up or login with your details

Forgot password? Click here to reset