Monte-Carlo utility estimates for Bayesian reinforcement learning

03/11/2013
by   Christos Dimitrakakis, et al.
0

This paper introduces a set of algorithms for Monte-Carlo Bayesian reinforcement learning. Firstly, Monte-Carlo estimation of upper bounds on the Bayes-optimal value function is employed to construct an optimistic policy. Secondly, gradient-based algorithms for approximate upper and lower bounds are introduced. Finally, we introduce a new class of gradient algorithms for Bayesian Bellman error minimisation. We theoretically show that the gradient methods are sound. Experimentally, we demonstrate the superiority of the upper bound method in terms of reward obtained. However, we also show that the Bayesian Bellman error method is a close second, despite its significant computational simplicity.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/23/2019

Monte-Carlo Tree Search for Policy Optimization

Gradient-based methods are often used for policy optimization in deep re...
research
01/17/2018

rlsm: R package for least squares Monte Carlo

This short paper briefly describes the implementation of the least squar...
research
03/27/2013

ABC Reinforcement Learning

This paper introduces a simple, general framework for likelihood-free Ba...
research
01/15/2021

Estimation of future discretionary benefits in traditional life insurance

In the context of traditional life insurance, the future discretionary b...
research
07/02/2022

Reinforcement Learning Approaches for the Orienteering Problem with Stochastic and Dynamic Release Dates

In this paper, we study a sequential decision making problem faced by e-...
research
08/03/2023

Randomized approximation of summable sequences – adaptive and non-adaptive

We prove lower bounds for the randomized approximation of the embedding ...
research
04/03/2018

Renewal Monte Carlo: Renewal theory based reinforcement learning

In this paper, we present an online reinforcement learning algorithm, ca...

Please sign up or login with your details

Forgot password? Click here to reset