Policy-Induced Self-Supervision Improves Representation Finetuning in Visual RL

02/12/2023
by   Sébastien M. R. Arnold, et al.
0

We study how to transfer representations pretrained on source tasks to target tasks in visual percept based RL. We analyze two popular approaches: freezing or finetuning the pretrained representations. Empirical studies on a set of popular tasks reveal several properties of pretrained representations. First, finetuning is required even when pretrained representations perfectly capture the information required to solve the target task. Second, finetuned representations improve learnability and are more robust to noise. Third, pretrained bottom layers are task-agnostic and readily transferable to new tasks, while top layers encode task-specific information and require adaptation. Building on these insights, we propose a self-supervised objective that clusters representations according to the policy they induce, as opposed to traditional representation similarity measures which are policy-agnostic (e.g. Euclidean norm, cosine similarity). Together with freezing the bottom layers, this objective results in significantly better representation than frozen, finetuned, and self-supervised alternatives on a wide range of benchmarks.

READ FULL TEXT

page 3

page 4

research
04/13/2023

Lossless Adaptation of Pretrained Vision Models For Robotic Manipulation

Recent works have shown that large models pretrained on common visual le...
research
10/22/2020

Similarity Analysis of Self-Supervised Speech Representations

Self-supervised speech representation learning has recently been a prosp...
research
02/07/2021

Self-supervised driven consistency training for annotation efficient histopathology image analysis

Training a neural network with a large labeled dataset is still a domina...
research
09/19/2021

A Study of the Generalizability of Self-Supervised Representations

Recent advancements in self-supervised learning (SSL) made it possible t...
research
10/06/2020

Are "Undocumented Workers" the Same as "Illegal Aliens"? Disentangling Denotation and Connotation in Vector Spaces

In politics, neologisms are frequently invented for partisan objectives....
research
11/29/2022

MoDA: Map style transfer for self-supervised Domain Adaptation of embodied agents

We propose a domain adaptation method, MoDA, which adapts a pretrained e...
research
02/15/2022

Don't stop the training: continuously-updating self-supervised algorithms best account for auditory responses in the cortex

Over the last decade, numerous studies have shown that deep neural netwo...

Please sign up or login with your details

Forgot password? Click here to reset