Action-Conditioned Contrastive Policy Pretraining

04/05/2022
by   Qihang Zhang, et al.
2

Deep visuomotor policy learning achieves promising results in control tasks such as robotic manipulation and autonomous driving, where the action is generated from the visual input by the neural policy. However, it requires a huge number of online interactions with the training environment, which limits its real-world application. Compared to the popular unsupervised feature learning for visual recognition, feature pretraining for visuomotor control tasks is much less explored. In this work, we aim to pretrain policy representations for driving tasks using hours-long uncurated YouTube videos. A new contrastive policy pretraining method is developed to learn action-conditioned features from video frames with action pseudo labels. Experiments show that the resulting action-conditioned features bring substantial improvements to the downstream reinforcement learning and imitation learning tasks, outperforming the weights pretrained from previous unsupervised learning methods. Code and models will be made publicly available.

READ FULL TEXT

page 6

page 7

page 13

page 14

page 15

page 16

page 17

page 18

research
07/16/2019

Improved Reinforcement Learning through Imitation Learning Pretraining Towards Image-based Autonomous Driving

We present a training pipeline for the autonomous driving task given the...
research
03/03/2019

End-to-end Driving Deploying through Uncertainty-Aware Imitation Learning and Stochastic Visual Domain Adaptation

End-to-end visual-based imitation learning has been widely applied in au...
research
06/06/2023

Vid2Act: Activate Offline Videos for Visual RL

Pretraining RL models on offline video datasets is a promising way to im...
research
02/11/2021

Representation Matters: Offline Pretraining for Sequential Decision Making

The recent success of supervised learning methods on ever larger offline...
research
06/23/2022

Video PreTraining (VPT): Learning to Act by Watching Unlabeled Online Videos

Pretraining on noisy, internet-scale datasets has been heavily studied a...
research
10/11/2021

On a Benefit of Mask Language Modeling: Robustness to Simplicity Bias

Despite the success of pretrained masked language models (MLM), why MLM ...
research
03/20/2021

Unsupervised Feature Learning for Manipulation with Contrastive Domain Randomization

Robotic tasks such as manipulation with visual inputs require image feat...

Please sign up or login with your details

Forgot password? Click here to reset