Natural Gradient Deep Q-learning

03/20/2018
by   Ethan Knight, et al.
0

This paper presents findings for training a Q-learning reinforcement learning agent using natural gradient techniques. We compare the original deep Q-network (DQN) algorithm to its natural gradient counterpart (NGDQN), measuring NGDQN and DQN performance on classic controls environments without target networks. We find that NGDQN performs favorably relative to DQN, converging to significantly better policies faster and more frequently. These results indicate that natural gradient could be used for value function optimization in reinforcement learning to accelerate and stabilize training.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/06/2020

Uniform State Abstraction For Reinforcement Learning

Potential Based Reward Shaping combined with a potential function based ...
research
06/02/2021

Smooth Q-learning: Accelerate Convergence of Q-learning Using Similarity

An improvement of Q-learning is proposed in this paper. It is different ...
research
12/10/2021

Deep Q-Network with Proximal Iteration

We employ Proximal Iteration for value-function optimization in reinforc...
research
12/22/2020

QVMix and QVMix-Max: Extending the Deep Quality-Value Family of Algorithms to Cooperative Multi-Agent Reinforcement Learning

This paper introduces four new algorithms that can be used for tackling ...
research
08/18/2015

Distributed Deep Q-Learning

We propose a distributed deep learning model to successfully learn contr...
research
05/25/2020

Gradient Monitored Reinforcement Learning

This paper presents a novel neural network training approach for faster ...
research
06/29/2021

A Convergent and Efficient Deep Q Network Algorithm

Despite the empirical success of the deep Q network (DQN) reinforcement ...

Please sign up or login with your details

Forgot password? Click here to reset