Reward learning from human preferences and demonstrations in Atari

11/15/2018
by   Borja Ibarz, et al.
6

To solve complex real-world problems with reinforcement learning, we cannot rely on manually specified reward functions. Instead, we can have humans communicate an objective to the agent directly. In this work, we combine two approaches to learning from human feedback: expert demonstrations and trajectory preferences. We train a deep neural network to model the reward function and use its predicted reward to train an DQN-based deep reinforcement learning agent on 9 Atari games. Our approach beats the imitation learning baseline in 7 games and achieves strictly superhuman performance on 2 games without using game rewards. Additionally, we investigate the goodness of fit of the reward model, present some reward hacking problems, and study the effects of noise in the human labels.

READ FULL TEXT
research
06/12/2017

Deep reinforcement learning from human preferences

For sophisticated reinforcement learning (RL) systems to interact useful...
research
11/08/2021

Batch Reinforcement Learning from Crowds

A shortcoming of batch reinforcement learning is its requirement for rew...
research
06/06/2019

An Extensible Interactive Interface for Agent Design

In artificial intelligence, we often specify tasks through a reward func...
research
02/16/2016

Reinforcement Learning approach for Real Time Strategy Games Battle city and S3

In this paper we proposed reinforcement learning algorithms with the gen...
research
04/29/2021

Adapting to Reward Progressivity via Spectral Reinforcement Learning

In this paper we consider reinforcement learning tasks with progressive ...
research
11/28/2018

Trajectory-based Learning for Ball-in-Maze Games

Deep Reinforcement Learning has shown tremendous success in solving seve...

Please sign up or login with your details

Forgot password? Click here to reset