Combine PPO with NES to Improve Exploration

05/23/2019
by   Lianjiang Li, et al.
0

We introduce two approaches for combining neural evolution strategy (NES) and proximal policy optimization (PPO): parameter transfer and parameter space noise. Parameter transfer is a PPO agent with parameters transferred from a NES agent. Parameter space noise is to directly add noise to the PPO agent`s parameters. We demonstrate that PPO could benefit from both methods through experimental comparison on discrete action environments as well as continuous control tasks

READ FULL TEXT

page 10

page 11

page 14

page 15

research
06/06/2017

Parameter Space Noise for Exploration

Deep reinforcement learning (RL) methods generally engage in exploratory...
research
09/18/2018

Switching Isotropic and Directional Exploration with Parameter Space Noise in Deep Reinforcement Learning

This paper proposes an exploration method for deep reinforcement learnin...
research
02/18/2019

Optimized data exploration applied to the simulation of a chemical process

In complex simulation environments, certain parameter space regions may ...
research
05/11/2021

ORCEA: Object Recognition by Continuous Evidence Assimilation

ORCEA is a novel object recognition method applicable for objects descri...
research
01/31/2019

Contrasting Exploration in Parameter and Action Space: A Zeroth-Order Optimization Perspective

Black-box optimizers that explore in parameter space have often been sho...
research
05/17/2013

Evolutionary optimization of an experimental apparatus

In recent decades, cold atom experiments have become increasingly comple...
research
08/25/2022

ExpoCloud: a Framework for Time and Budget-Effective Parameter Space Explorations Using a Cloud Compute Engine

Large parameter space explorations are among the most time consuming yet...

Please sign up or login with your details

Forgot password? Click here to reset