PAC-Bayesian Policy Evaluation for Reinforcement Learning

02/14/2012
by   Mahdi Milani Fard, et al.
0

Bayesian priors offer a compact yet general means of incorporating domain knowledge into many learning tasks. The correctness of the Bayesian analysis and inference, however, largely depends on accuracy and correctness of these priors. PAC-Bayesian methods overcome this problem by providing bounds that hold regardless of the correctness of the prior distribution. This paper introduces the first PAC-Bayesian bound for the batch reinforcement learning problem with function approximation. We show how this bound can be used to perform model-selection in a transfer learning scenario. Our empirical results confirm that PAC-Bayesian policy evaluation is able to leverage prior distributions when they are informative and, unlike standard Bayesian RL approaches, ignore them when they are misleading.

READ FULL TEXT
research
10/23/2016

Simpler PAC-Bayesian Bounds for Hostile Data

PAC-Bayesian learning bounds are of the utmost interest to the learning ...
research
05/31/2022

Online PAC-Bayes Learning

Most PAC-Bayesian bounds hold in the batch learning setting where data i...
research
09/14/2016

Bayesian Reinforcement Learning: A Survey

Bayesian methods for machine learning have been widely investigated, yie...
research
12/11/2007

PAC-Bayesian Bounds for Randomized Empirical Risk Minimizers

The aim of this paper is to generalize the PAC-Bayesian theorems proved ...
research
10/07/2014

PAC-Bayesian AUC classification and scoring

We develop a scoring and classification procedure based on the PAC-Bayes...
research
11/12/2013

A PAC-Bayesian bound for Lifelong Learning

Transfer learning has received a lot of attention in the machine learnin...
research
10/24/2022

PAC-Bayesian Offline Contextual Bandits With Guarantees

This paper introduces a new principled approach for offline policy optim...

Please sign up or login with your details

Forgot password? Click here to reset