Fast Value Iteration for Goal-Directed Markov Decision Processes

02/06/2013
by   Nevin Lianwen Zhang, et al.
0

Planning problems where effects of actions are non-deterministic can be modeled as Markov decision processes. Planning problems are usually goal-directed. This paper proposes several techniques for exploiting the goal-directedness to accelerate value iteration, a standard algorithm for solving Markov decision processes. Empirical studies have shown that the techniques can bring about significant speedups.

READ FULL TEXT

page 1

page 2

page 3

page 4

page 6

research
12/30/2015

A Notation for Markov Decision Processes

This paper specifies a notation for Markov decision processes....
research
06/14/2018

Configurable Markov Decision Processes

In many real-world problems, there is the possibility to configure, to a...
research
04/07/2016

Planning with Information-Processing Constraints and Model Uncertainty in Markov Decision Processes

Information-theoretic principles for learning and acting have been propo...
research
11/02/2022

Interval Markov Decision Processes with Continuous Action-Spaces

Interval Markov Decision Processes (IMDPs) are uncertain Markov models, ...
research
12/04/2019

A Variational Perturbative Approach to Planning in Graph-based Markov Decision Processes

Coordinating multiple interacting agents to achieve a common goal is a d...
research
04/14/2020

Extrapolation in Gridworld Markov-Decision Processes

Extrapolation in reinforcement learning is the ability to generalize at ...
research
09/26/2013

Approximation of Lorenz-Optimal Solutions in Multiobjective Markov Decision Processes

This paper is devoted to fair optimization in Multiobjective Markov Deci...

Please sign up or login with your details

Forgot password? Click here to reset