Planning with RL and episodic-memory behavioral priors

07/05/2022
by   Shivansh Beohar, et al.
0

The practical application of learning agents requires sample efficient and interpretable algorithms. Learning from behavioral priors is a promising way to bootstrap agents with a better-than-random exploration policy or a safe-guard against the pitfalls of early learning. Existing solutions for imitation learning require a large number of expert demonstrations and rely on hard-to-interpret learning methods like Deep Q-learning. In this work we present a planning-based approach that can use these behavioral priors for effective exploration and learning in a reinforcement learning environment, and we demonstrate that curated exploration policies in the form of behavioral priors can help an agent learn faster.

READ FULL TEXT

page 1

page 4

research
03/26/2023

Inverse Reinforcement Learning without Reinforcement Learning

Inverse Reinforcement Learning (IRL) is a powerful set of techniques for...
research
06/23/2022

Video PreTraining (VPT): Learning to Act by Watching Unlabeled Online Videos

Pretraining on noisy, internet-scale datasets has been heavily studied a...
research
03/06/2022

MIRROR: Differentiable Deep Social Projection for Assistive Human-Robot Communication

Communication is a hallmark of intelligence. In this work, we present MI...
research
09/24/2021

Go-Blend behavior and affect

This paper proposes a paradigm shift for affective computing by viewing ...
research
08/13/2020

Imitating Unknown Policies via Exploration

Behavioral cloning is an imitation learning technique that teaches an ag...
research
03/02/2022

Improving the Diversity of Bootstrapped DQN via Noisy Priors

Q-learning is one of the most well-known Reinforcement Learning algorith...
research
08/26/2022

Play with Emotion: Affect-Driven Reinforcement Learning

This paper introduces a paradigm shift by viewing the task of affect mod...

Please sign up or login with your details

Forgot password? Click here to reset