Online Hyper-parameter Tuning in Off-policy Learning via Evolutionary Strategies

06/13/2020
by   Yunhao Tang, et al.
1

Off-policy learning algorithms have been known to be sensitive to the choice of hyper-parameters. However, unlike near on-policy algorithms for which hyper-parameters could be optimized via e.g. meta-gradients, similar techniques could not be straightforwardly applied to off-policy learning. In this work, we propose a framework which entails the application of Evolutionary Strategies to online hyper-parameter tuning in off-policy learning. Our formulation draws close connections to meta-gradients and leverages the strengths of black-box optimization with relatively low-dimensional search spaces. We show that our method outperforms state-of-the-art off-policy learning baselines with static hyper-parameters and recent prior work over a wide range of continuous control benchmarks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/04/2020

On Hyper-parameter Tuning for Stochastic Optimization Algorithms

This paper proposes the first-ever algorithmic framework for tuning hype...
research
03/02/2020

Adaptive Structural Hyper-Parameter Configuration by Q-Learning

Tuning hyper-parameters for evolutionary algorithms is an important issu...
research
01/14/2021

A Pragmatic Approach for Hyper-Parameter Tuning in Search-based Test Case Generation

Search-based test case generation, which is the application of meta-heur...
research
12/11/2019

Efficacy of Modern Neuro-Evolutionary Strategies for Continuous Control Optimization

We analyze the efficacy of modern neuro-evolutionary strategies for cont...
research
06/05/2019

Revisiting Hyper-Parameter Tuning for Search-based Test Data Generation

Search-based software testing (SBST) has been studied a lot in the liter...
research
02/01/2022

Surrogate Gradients Design

Surrogate gradient (SG) training provides the possibility to quickly tra...

Please sign up or login with your details

Forgot password? Click here to reset