Approximate Kalman Filter Q-Learning for Continuous State-Space MDPs

09/26/2013
by   Charles Tripp, et al.
0

We seek to learn an effective policy for a Markov Decision Process (MDP) with continuous states via Q-Learning. Given a set of basis functions over state action pairs we search for a corresponding set of linear weights that minimizes the mean Bellman residual. Our algorithm uses a Kalman filter model to estimate those weights and we have developed a simpler approximate Kalman filter model that outperforms the current state of the art projected TD-Learning methods on several standard benchmark problems.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/23/2016

The discriminative Kalman filter for nonlinear and non-Gaussian sequential Bayesian filtering

The Kalman filter (KF) is used in a variety of applications for computin...
research
03/07/2017

Deep Robust Kalman Filter

A Robust Markov Decision Process (RMDP) is a sequential decision making ...
research
03/15/2012

A Scalable Method for Solving High-Dimensional Continuous POMDPs Using Local Approximation

Partially-Observable Markov Decision Processes (POMDPs) are typically so...
research
09/27/2022

Design of experiments for the calibration of history-dependent models via deep reinforcement learning and an enhanced Kalman filter

Experimental data is costly to obtain, which makes it difficult to calib...
research
05/26/2000

A Bayesian Reflection on Surfaces

The topic of this paper is a novel Bayesian continuous-basis field repre...
research
08/23/2023

Extended Linear Regression: A Kalman Filter Approach for Minimizing Loss via Area Under the Curve

This research enhances linear regression models by integrating a Kalman ...
research
05/01/2020

A continuous-time state-space model for rapid quality-control of Argos locations from animal-borne tags

State-space models are important tools for quality control of error-pron...

Please sign up or login with your details

Forgot password? Click here to reset