Noisy Networks for Exploration

06/30/2017
by   Meire Fortunato, et al.
0

We introduce NoisyNet, a deep reinforcement learning agent with parametric noise added to its weights, and show that the induced stochasticity of the agent's policy can be used to aid efficient exploration. The parameters of the noise are learned with gradient descent along with the remaining network weights. NoisyNet is straightforward to implement and adds little computational overhead. We find that replacing the conventional exploration heuristics for A3C, DQN and dueling agents (entropy reward and ϵ-greedy respectively) with NoisyNet yields substantially higher scores for a wide range of Atari games, in some cases advancing the agent from sub to super-human performance.

READ FULL TEXT

page 15

page 16

research
09/28/2021

Exploring More When It Needs in Deep Reinforcement Learning

We propose a exploration mechanism of policy in Deep Reinforcement Learn...
research
06/19/2020

NROWAN-DQN: A Stable Noisy Network with Noise Reduction and Online Weight Adjustment for Exploration

Deep reinforcement learning has been applied more and more widely nowada...
research
08/06/2019

Benchmarking Bonus-Based Exploration Methods on the Arcade Learning Environment

This paper provides an empirical evaluation of recently developed explor...
research
11/10/2020

Perturbation-based exploration methods in deep reinforcement learning

Recent research on structured exploration placed emphasis on identifying...
research
08/19/2022

Entropy Augmented Reinforcement Learning

Deep reinforcement learning has gained a lot of success with the presenc...
research
09/15/2022

Human-level Atari 200x faster

The task of building general agents that perform well over a wide range ...
research
02/07/2021

State-Aware Variational Thompson Sampling for Deep Q-Networks

Thompson sampling is a well-known approach for balancing exploration and...

Please sign up or login with your details

Forgot password? Click here to reset