Efficient Model-Free Reinforcement Learning Using Gaussian Process

12/11/2018
by   Ying Fan, et al.
6

Efficient Reinforcement Learning usually takes advantage of demonstration or good exploration strategy. By applying posterior sampling in model-free RL under the hypothesis of GP, we propose Gaussian Process Posterior Sampling Reinforcement Learning(GPPSTD) algorithm in continuous state space, giving theoretical justifications and empirical results. We also provide theoretical and empirical results that various demonstration could lower expected uncertainty and benefit posterior sampling exploration. In this way, we combined the demonstration and exploration process together to achieve a more efficient reinforcement learning.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/08/2022

An Empirical Evaluation of Posterior Sampling for Constrained Reinforcement Learning

We study a posterior sampling approach to efficient exploration in const...
research
12/06/2018

Active Deep Q-learning with Demonstration

Recent research has shown that although Reinforcement Learning (RL) can ...
research
07/14/2021

Model-free Reinforcement Learning for Robust Locomotion Using Trajectory Optimization for Exploration

In this work we present a general, two-stage reinforcement learning appr...
research
05/22/2023

Policy Representation via Diffusion Probability Model for Reinforcement Learning

Popular reinforcement learning (RL) algorithms tend to produce a unimoda...
research
11/17/2018

Recursive Sparse Pseudo-input Gaussian Process SARSA

The class of Gaussian Process (GP) methods for Temporal Difference learn...
research
09/06/2019

Gradient Q(σ, λ): A Unified Algorithm with Function Approximation for Reinforcement Learning

Full-sampling (e.g., Q-learning) and pure-expectation (e.g., Expected Sa...
research
02/05/2018

Coordinated Exploration in Concurrent Reinforcement Learning

We consider a team of reinforcement learning agents that concurrently le...

Please sign up or login with your details

Forgot password? Click here to reset