Does the Markov Decision Process Fit the Data: Testing for the Markov Property in Sequential Decision Making

02/05/2020
by   Chengchun Shi, et al.
0

The Markov assumption (MA) is fundamental to the empirical validity of reinforcement learning. In this paper, we propose a novel Forward-Backward Learning procedure to test MA in sequential decision making. The proposed test does not assume any parametric form on the joint distribution of the observed data and plays an important role for identifying the optimal policy in high-order Markov decision processes and partially observable MDPs. We apply our test to both synthetic datasets and a real data example from mobile health studies to illustrate its usefulness.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/27/2023

Decision Making for Autonomous Vehicles

This paper is on decision making of autonomous vehicles for handling rou...
research
08/21/2021

Sequential Stochastic Optimization in Separable Learning Environments

We consider a class of sequential decision-making problems under uncerta...
research
06/05/2012

A Mixed Observability Markov Decision Process Model for Musical Pitch

Partially observable Markov decision processes have been widely used to ...
research
01/13/2020

POPCORN: Partially Observed Prediction COnstrained ReiNforcement Learning

Many medical decision-making settings can be framed as partially observe...
research
10/19/2019

Optimal Immunization Policy Using Dynamic Programming

Decisions in public health are almost always made in the context of unce...
research
10/12/2019

The Search for Truth through Data: NP Decision Processes, ROC Functions, P-Functionals, Knowledge Updating and Sequential Learning

This paper re-visits the problem of deciding between two simple hypothes...
research
12/27/2019

Quantum Logic Gate Synthesis as a Markov Decision Process

Reinforcement learning has witnessed recent applications to a variety of...

Please sign up or login with your details

Forgot password? Click here to reset