Improving Multimodal Interactive Agents with Reinforcement Learning from Human Feedback

11/21/2022
by   Josh Abramson, et al.
0

An important goal in artificial intelligence is to create agents that can both interact naturally with humans and learn from their feedback. Here we demonstrate how to use reinforcement learning from human feedback (RLHF) to improve upon simulated, embodied agents trained to a base level of competency with imitation learning. First, we collected data of humans interacting with agents in a simulated 3D world. We then asked annotators to record moments where they believed that agents either progressed toward or regressed from their human-instructed goal. Using this annotation data we leveraged a novel method - which we call "Inter-temporal Bradley-Terry" (IBT) modelling - to build a reward model that captures human judgments. Agents trained to optimise rewards delivered from IBT reward models improved with respect to all of our metrics, including subsequent human judgment during live interactions with agents. Altogether our results demonstrate how one can successfully leverage human judgments to improve agent behaviour, allowing us to use reinforcement learning in complex, embodied domains without programmatic reward functions. Videos of agent behaviour may be found at https://youtu.be/v_Z9F2_eKk4.

READ FULL TEXT

page 1

page 4

page 7

page 11

page 13

page 19

page 22

page 23

research
12/10/2020

Imitating Interactive Intelligence

A common vision from science fiction is that robots will one day inhabit...
research
09/30/2020

Learning Rewards from Linguistic Feedback

We explore unconstrained natural language feedback as a learning signal ...
research
10/02/2019

Unsupervised Doodling and Painting with Improved SPIRAL

We investigate using reinforcement learning agents as generative models ...
research
12/07/2021

Creating Multimodal Interactive Agents with Imitation and Self-Supervised Learning

A common vision from science fiction is that robots will one day inhabit...
research
01/20/2022

Safe Deep RL in 3D Environments using Human Feedback

Agents should avoid unsafe behaviour during both training and deployment...
research
05/18/2015

A Definition of Happiness for Reinforcement Learning Agents

What is happiness for reinforcement learning agents? We seek a formal de...
research
03/13/2023

Vision-Language Models as Success Detectors

Detecting successful behaviour is crucial for training intelligent agent...

Please sign up or login with your details

Forgot password? Click here to reset