A Scalable Finite Difference Method for Deep Reinforcement Learning

10/14/2022
by   Matthew Allen, et al.
0

Several low-bandwidth distributable black-box optimization algorithms have recently been shown to perform nearly as well as more refined modern methods in some Deep Reinforcement Learning domains. In this work we investigate a core problem with the use of distributed workers in such systems. Further, we investigate the dramatic differences in performance between the popular Adam gradient descent algorithm and the simplest form of stochastic gradient descent. These investigations produce a stable, low-bandwidth learning algorithm that achieves 100% usage of all connected CPUs under typical conditions.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/23/2017

Stability and Generalization of Learning Algorithms that Converge to Global Optima

We establish novel generalization bounds for learning algorithms that co...
research
11/22/2019

Low-variance Black-box Gradient Estimates for the Plackett-Luce Distribution

Learning models with discrete latent variables using stochastic gradient...
research
05/25/2022

An Experimental Comparison Between Temporal Difference and Residual Gradient with Neural Network Approximation

Gradient descent or its variants are popular in training neural networks...
research
07/08/2020

Minimax Efficient Finite-Difference Stochastic Gradient Estimators Using Black-Box Function Evaluations

We consider stochastic gradient estimation using noisy black-box functio...
research
12/06/2018

Deep Reinforcement Learning and the Deadly Triad

We know from reinforcement learning theory that temporal difference lear...
research
01/03/2019

Machine Learning at the Wireless Edge: Distributed Stochastic Gradient Descent Over-the-Air

We study collaborative machine learning at the wireless edge, where powe...
research
12/18/2017

ES Is More Than Just a Traditional Finite-Difference Approximator

An evolution strategy (ES) variant recently attracted significant attent...

Please sign up or login with your details

Forgot password? Click here to reset