
Nearoptimal Reinforcement Learning in Factored MDPs
Any reinforcement learning algorithm that applies to all Markov decision...
read it

Active Reinforcement Learning with MonteCarlo Tree Search
Active Reinforcement Learning (ARL) is a twist on RL where the agent obs...
read it

RiskAverse BayesAdaptive Reinforcement Learning
In this work, we address riskaverse Bayesadaptive reinforcement learnin...
read it

Efficient BayesAdaptive Reinforcement Learning using SampleBased Search
Bayesian modelbased reinforcement learning is a formally elegant approa...
read it

Factoring Exogenous State for ModelFree Monte Carlo
Policy analysts wish to visualize a range of policies for large simulato...
read it

Bayesian Reinforcement Learning in Factored POMDPs
Bayesian approaches provide a principled solution to the explorationexp...
read it

Generalized Mean Estimation in MonteCarlo Tree Search
We consider MonteCarlo Tree Search (MCTS) applied to Markov Decision Pr...
read it
Learning is planning: near Bayesoptimal reinforcement learning via MonteCarlo tree search
Bayesoptimal behavior, while welldefined, is often difficult to achieve. Recent advances in the use of MonteCarlo tree search (MCTS) have shown that it is possible to act nearoptimally in Markov Decision Processes (MDPs) with very large or infinite state spaces. Bayesoptimal behavior in an unknown MDP is equivalent to optimal behavior in the known beliefspace MDP, although the size of this beliefspace MDP grows exponentially with the amount of history retained, and is potentially infinite. We show how an agent can use one particular MCTS algorithm, Forward Search Sparse Sampling (FSSS), in an efficient way to act nearly Bayesoptimally for all but a polynomial number of steps, assuming that FSSS can be used to act efficiently in any possible underlying MDP.
READ FULL TEXT
Comments
There are no comments yet.