Unsupervised State Representation Learning in Atari

06/19/2019
by   Ankesh Anand, et al.
5

State representation learning, or the ability to capture latent generative factors of an environment, is crucial for building intelligent agents that can perform a wide variety of tasks. Learning such representations without supervision from rewards is a challenging open problem. We introduce a method that learns state representations by maximizing mutual information across spatially and temporally distinct features of a neural encoder of the observations. We also introduce a new benchmark based on Atari 2600 games where we evaluate representations based on how well they capture the ground truth state variables. We believe this new framework for evaluating representation learning models will be crucial for future representation learning research. Finally, we compare our technique with other state-of-the-art generative and contrastive representation learning methods.

READ FULL TEXT

page 2

page 3

page 15

research
03/13/2023

Unsupervised Representation Learning in Partially Observable Atari Games

State representation learning aims to capture latent factors of an envir...
research
06/04/2018

Holographic Neural Architectures

Representation learning is at the heart of what makes deep learning effe...
research
07/13/2021

On Designing Good Representation Learning Models

The goal of representation learning is different from the ultimate objec...
research
02/09/2021

Benchmarks, Algorithms, and Metrics for Hierarchical Disentanglement

In representation learning, there has been recent interest in developing...
research
02/25/2022

Raman Spectrum Matching with Contrastive Representation Learning

Raman spectroscopy is an effective, low-cost, non-intrusive technique of...
research
09/03/2020

Action and Perception as Divergence Minimization

We introduce a unified objective for action and perception of intelligen...
research
05/02/2023

Long-Tailed Recognition by Mutual Information Maximization between Latent Features and Ground-Truth Labels

Although contrastive learning methods have shown prevailing performance ...

Please sign up or login with your details

Forgot password? Click here to reset