Perceptual Values from Observation

05/20/2019
by   Ashley D. Edwards, et al.
0

Imitation by observation is an approach for learning from expert demonstrations that lack action information, such as videos. Recent approaches to this problem can be placed into two broad categories: training dynamics models that aim to predict the actions taken between states, and learning rewards or features for computing them for Reinforcement Learning (RL). In this paper, we introduce a novel approach that learns values, rather than rewards, directly from observations. We show that by using values, we can significantly speed up RL by removing the need to bootstrap action-values, as compared to sparse-reward specifications.

READ FULL TEXT

page 3

page 4

research
07/25/2021

Reinforced Imitation Learning by Free Energy Principle

Reinforcement Learning (RL) requires a large amount of exploration espec...
research
10/14/2020

Self-Imitation Learning in Sparse Reward Settings

The application of reinforcement learning (RL) in real-world is still li...
research
12/05/2019

Reinforcement Learning Upside Down: Don't Predict Rewards – Just Map Them to Actions

We transform reinforcement learning (RL) into a form of supervised learn...
research
06/24/2023

Learning from Pixels with Expert Observations

In reinforcement learning (RL), sparse rewards can present a significant...
research
04/03/2017

Multi-Advisor Reinforcement Learning

We consider tackling a single-agent RL problem by distributing it to n l...
research
07/02/2019

On Conflicting and Conflicting Values

Values are things that are important to us. Actions activate values - th...
research
07/02/2019

On Conforming and Conflicting Values

Values are things that are important to us. Actions activate values - th...

Please sign up or login with your details

Forgot password? Click here to reset