Combining Experience Replay with Exploration by Random Network Distillation

05/18/2019
by   Francesco Sovrano, et al.
0

Our work is a simple extension of the paper "Exploration by Random Network Distillation". More in detail, we show how to efficiently combine Intrinsic Rewards with Experience Replay in order to achieve more efficient and robust exploration (with respect to PPO/RND) and consequently better results in terms of agent performances and sample efficiency. We are able to do it by using a new technique named Prioritized Oversampled Experience Replay (POER), that has been built upon the definition of what is the important experience useful to replay. Finally, we evaluate our technique on the famous Atari game Montezuma's Revenge and some other hard exploration Atari games.

READ FULL TEXT
research
05/15/2018

Advances in Experience Replay

This project combines recent advances in experience replay techniques, n...
research
10/19/2019

Reverse Experience Replay

This paper describes an improvement in Deep Q-learning called Reverse Ex...
research
10/30/2018

Exploration by Random Network Distillation

We introduce an exploration bonus for deep reinforcement learning method...
research
08/10/2020

Improving Intelligence of Evolutionary Algorithms Using Experience Share and Replay

We propose PESA, a novel approach combining Particle Swarm Optimisation ...
research
02/18/2021

Understanding algorithmic collusion with experience replay

In an infinitely repeated pricing game, pricing algorithms based on arti...
research
07/28/2020

A Novel Token-Based Replay Technique to Speed Up Conformance Checking and Process Enhancement

Token-based replay used to be the standard way to conduct conformance ch...
research
01/03/2018

ViZDoom: DRQN with Prioritized Experience Replay, Double-Q Learning, & Snapshot Ensembling

ViZDoom is a robust, first-person shooter reinforcement learning environ...

Please sign up or login with your details

Forgot password? Click here to reset