An Empirical Evaluation of Posterior Sampling for Constrained Reinforcement Learning

09/08/2022
by   Danil Provodin, et al.
6

We study a posterior sampling approach to efficient exploration in constrained reinforcement learning. Alternatively to existing algorithms, we propose two simple algorithms that are more efficient statistically, simpler to implement and computationally cheaper. The first algorithm is based on a linear formulation of CMDP, and the second algorithm leverages the saddle-point formulation of CMDP. Our empirical results demonstrate that, despite its simplicity, posterior sampling achieves state-of-the-art performance and, in some cases, significantly outperforms optimistic algorithms.

READ FULL TEXT
research
06/04/2013

(More) Efficient Reinforcement Learning via Posterior Sampling

Most provably-efficient learning algorithms introduce optimism about poo...
research
12/11/2018

Efficient Model-Free Reinforcement Learning Using Gaussian Process

Efficient Reinforcement Learning usually takes advantage of demonstratio...
research
09/06/2019

Gradient Q(σ, λ): A Unified Algorithm with Function Approximation for Reinforcement Learning

Full-sampling (e.g., Q-learning) and pure-expectation (e.g., Expected Sa...
research
08/18/2011

Feature Reinforcement Learning In Practice

Following a recent surge in using history-based methods for resolving pe...
research
04/30/2023

Posterior Sampling for Deep Reinforcement Learning

Despite remarkable successes, deep reinforcement learning algorithms rem...
research
02/05/2018

Coordinated Exploration in Concurrent Reinforcement Learning

We consider a team of reinforcement learning agents that concurrently le...
research
06/09/2020

Constrained episodic reinforcement learning in concave-convex and knapsack settings

We propose an algorithm for tabular episodic reinforcement learning with...

Please sign up or login with your details

Forgot password? Click here to reset