SIRL: Similarity-based Implicit Representation Learning

01/02/2023
by   Andreea Bobu, et al.
13

When robots learn reward functions using high capacity models that take raw state directly as input, they need to both learn a representation for what matters in the task – the task “features" – as well as how to combine these features into a single objective. If they try to do both at once from input designed to teach the full reward function, it is easy to end up with a representation that contains spurious correlations in the data, which fails to generalize to new settings. Instead, our ultimate goal is to enable robots to identify and isolate the causal features that people actually care about and use when they represent states and behavior. Our idea is that we can tune into this representation by asking users what behaviors they consider similar: behaviors will be similar if the features that matter are similar, even if low-level behavior is different; conversely, behaviors will be different if even one of the features that matter differs. This, in turn, is what enables the robot to disambiguate between what needs to go into the representation versus what is spurious, as well as what aspects of behavior can be compressed together versus not. The notion of learning representations based on similarity has a nice parallel in contrastive learning, a self-supervised representation learning technique that maps visually similar data points to similar embeddings, where similarity is defined by a designer through data augmentation heuristics. By contrast, in order to learn the representations that people use, so we can learn their preferences and objectives, we use their definition of similarity. In simulation as well as in a user study, we show that learning through such similarity queries leads to representations that, while far from perfect, are indeed more generalizable than self-supervised and task-input alternatives.

READ FULL TEXT

page 1

page 4

page 8

research
01/18/2022

Inducing Structure in Reward Learning by Learning Features

Reward learning enables robots to learn adaptable behaviors from human i...
research
10/20/2021

Learning Universal User Representations via Self-Supervised Lifelong Behaviors Modeling

Universal user representation is an important research topic in industry...
research
11/11/2020

I Know What You Meant: Learning Human Objectives by (Under)estimating Their Choice Set

Assistive robots have the potential to help people perform everyday task...
research
09/13/2021

Cross Domain Robot Imitation with Invariant Representation

Animals are able to imitate each others' behavior, despite their differe...
research
12/07/2022

Teaching Matters: Investigating the Role of Supervision in Vision Transformers

Vision Transformers (ViTs) have gained significant popularity in recent ...
research
04/11/2023

Diagnosing and Augmenting Feature Representations in Correctional Inverse Reinforcement Learning

Robots have been increasingly better at doing tasks for humans by learni...
research
03/10/2023

Ignorance is Bliss: Robust Control via Information Gating

Informational parsimony – i.e., using the minimal information required f...

Please sign up or login with your details

Forgot password? Click here to reset