Stabilizing Contrastive RL: Techniques for Offline Goal Reaching

06/06/2023
by   Chongyi Zheng, et al.
0

In the same way that the computer vision (CV) and natural language processing (NLP) communities have developed self-supervised methods, reinforcement learning (RL) can be cast as a self-supervised problem: learning to reach any goal, without requiring human-specified rewards or labels. However, actually building a self-supervised foundation for RL faces some important challenges. Building on prior contrastive approaches to this RL problem, we conduct careful ablation experiments and discover that a shallow and wide architecture, combined with careful weight initialization and data augmentation, can significantly boost the performance of these contrastive RL approaches on challenging simulated benchmarks. Additionally, we demonstrate that, with these design decisions, contrastive approaches can solve real-world robotic manipulation tasks, with tasks being specified by a single goal image provided after training.

READ FULL TEXT

page 6

page 7

page 8

page 20

page 21

page 22

research
06/15/2022

Contrastive Learning as Goal-Conditioned Reinforcement Learning

In reinforcement learning (RL), it is easier to solve a task if given a ...
research
02/25/2021

A Primer on Contrastive Pretraining in Language Processing: Methods, Lessons Learned and Perspectives

Modern natural language processing (NLP) methods employ self-supervised ...
research
08/08/2023

BarlowRL: Barlow Twins for Data-Efficient Reinforcement Learning

This paper introduces BarlowRL, a data-efficient reinforcement learning ...
research
07/29/2022

Contrastive UCB: Provably Efficient Contrastive Self-Supervised Learning in Online Reinforcement Learning

In view of its power in extracting feature representation, contrastive s...
research
06/08/2023

A brief review of contrastive learning applied to astrophysics

Reliable tools to extract patterns from high-dimensionality spaces are b...
research
09/15/2021

SupCL-Seq: Supervised Contrastive Learning for Downstream Optimized Sequence Representations

While contrastive learning is proven to be an effective training strateg...
research
05/02/2020

Contrastive Self-Supervised Learning for Commonsense Reasoning

We propose a self-supervised method to solve Pronoun Disambiguation and ...

Please sign up or login with your details

Forgot password? Click here to reset