Memory Augmented Self-Play

05/28/2018
by   Shagun Sodhani, et al.
0

Self-play is an unsupervised training procedure which enables the reinforcement learning agents to explore the environment without requiring any external rewards. We augment the self-play setting by providing an external memory where the agent can store experience from the previous tasks. This enables the agent to come up with more diverse self-play tasks resulting in faster exploration of the environment. The agent pretrained in the memory augmented self-play setting easily outperforms the agent pretrained in no-memory self-play setting.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/10/2020

Provable Self-Play Algorithms for Competitive Reinforcement Learning

Self-play, where the algorithm learns by playing against itself without ...
research
06/08/2020

A Comparison of Self-Play Algorithms Under a Generalized Framework

Throughout scientific history, overarching theoretical frameworks have a...
research
06/06/2021

Unbiased Self-Play

We present a general optimization framework for emergent belief-state re...
research
10/10/2020

Self-play for Data Efficient Language Acquisition

When communicating, people behave consistently across conversational rol...
research
02/06/2023

Intrinsic Rewards from Self-Organizing Feature Maps for Exploration in Reinforcement Learning

We introduce an exploration bonus for deep reinforcement learning method...
research
09/20/2021

Two Approaches to Building Collaborative, Task-Oriented Dialog Agents through Self-Play

Task-oriented dialog systems are often trained on human/human dialogs, s...
research
02/04/2023

Diversity Induced Environment Design via Self-Play

Recent work on designing an appropriate distribution of environments has...

Please sign up or login with your details

Forgot password? Click here to reset