Smooth Q-learning: Accelerate Convergence of Q-learning Using Similarity

06/02/2021
by   Wei Liao, et al.
0

An improvement of Q-learning is proposed in this paper. It is different from classic Q-learning in that the similarity between different states and actions is considered in the proposed method. During the training, a new updating mechanism is used, in which the Q value of the similar state-action pairs are updated synchronously. The proposed method can be used in combination with both tabular Q-learning function and deep Q-learning. And the results of numerical examples illustrate that compared to the classic Q-learning, the proposed method has a significantly better performance.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/20/2018

Natural Gradient Deep Q-learning

This paper presents findings for training a Q-learning reinforcement lea...
research
07/23/2023

Development of pericardial fat count images using a combination of three different deep-learning models

Rationale and Objectives: Pericardial fat (PF), the thoracic visceral fa...
research
04/30/2015

Detecting and ordering adjectival scalemates

This paper presents a pattern-based method that can be used to infer adj...
research
03/18/2022

Learning Stabilizable Deep Dynamics Models

When neural networks are used to model dynamics, properties such as stab...
research
05/20/2021

A continuation multiple shooting method for Wasserstein geodesic equation

In this paper, we propose a numerical method to solve the classic L^2-op...
research
11/21/2019

Memory-Efficient Episodic Control Reinforcement Learning with Dynamic Online k-means

Recently, neuro-inspired episodic control (EC) methods have been develop...

Please sign up or login with your details

Forgot password? Click here to reset