Efficient Wasserstein Natural Gradients for Reinforcement Learning

10/12/2020
by   Ted Moskovitz, et al.
0

A novel optimization approach is proposed for application to policy gradient methods and evolution strategies for reinforcement learning (RL). The procedure uses a computationally efficient Wasserstein natural gradient (WNG) descent that takes advantage of the geometry induced by a Wasserstein penalty to speed optimization. This method follows the recent theme in RL of including a divergence penalty in the objective to establish a trust region. Experiments on challenging tasks demonstrate improvements in both computational cost and performance over advanced baselines.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset