Hitting time for Markov decision process

05/06/2022
by   Ruichao Jiang, et al.
0

We define the hitting time for a Markov decision process (MDP). We do not use the hitting time of the Markov process induced by the MDP because the induced chain may not have a stationary distribution. Even it has a stationary distribution, the stationary distribution may not coincide with the (normalized) occupancy measure of the MDP. We observe a relationship between the MDP and the PageRank. Using this observation, we construct an MP whose stationary distribution coincides with the normalized occupancy measure of the MDP and we define the hitting time of the MDP as the hitting time of the associated MP.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/22/2019

Efficient Policy Learning for Non-Stationary MDPs under Adversarial Manipulation

A Markov Decision Process (MDP) is a popular model for reinforcement lea...
research
04/01/2019

Elaboration Tolerant Representation of Markov Decision Process via Decision-Theoretic Extension of Probabilistic Action Language pBC+

We extend probabilistic action language pBC+ with the notion of utility ...
research
07/10/2019

Markov Decision Process for MOOC users behavioral inference

Studies on massive open online courses (MOOCs) users discuss the existen...
research
04/27/2023

Level Assembly as a Markov Decision Process

Many games feature a progression of levels that doesn't adapt to the pla...
research
04/30/2021

Optimal control policies for resource allocation in the Cloud: comparison between Markov decision process and heuristic approaches

We consider an auto-scaling technique in a cloud system where virtual ma...
research
06/10/2022

Conformal Prediction Intervals for Markov Decision Process Trajectories

Before delegating a task to an autonomous system, a human operator may w...
research
02/23/2018

Novel Approaches to Accelerating the Convergence Rate of Markov Decision Process for Search Result Diversification

Recently, some studies have utilized the Markov Decision Process for div...

Please sign up or login with your details

Forgot password? Click here to reset