Integrating Planning and Execution in Stochastic Domains

02/27/2013
by   Richard Dearden, et al.
0

We investigate planning in time-critical domains represented as Markov Decision Processes, showing that search based techniques can be a very powerful method for finding close to optimal plans. To reduce the computational cost of planning in these domains, we execute actions as we construct the plan, and sacrifice optimality by searching to a fixed depth and using a heuristic function to estimate the value of states. Although this paper concentrates on the search algorithm, we also discuss ways of constructing heuristic functions suitable for this approach. Our results show that by interleaving search and execution, close to optimal policies can be found without the computational requirements of other approaches.

READ FULL TEXT

page 1

page 3

page 4

page 6

page 7

page 8

research
01/15/2014

Online Planning Algorithms for POMDPs

Partially Observable Markov Decision Processes (POMDPs) provide a rich f...
research
08/24/2017

Learning Generalized Reactive Policies using Deep Neural Networks

We consider the problem of learning for planning, where knowledge acquir...
research
09/12/2022

A Differentiable Loss Function for Learning Heuristics in A*

Optimization of heuristic functions for the A* algorithm, realized by de...
research
09/29/2011

FluCaP: A Heuristic Search Planner for First-Order MDPs

We present a heuristic search algorithm for solving first-order Markov D...
research
03/23/2023

Planning as Theorem Proving with Heuristics

Planning as theorem proving in situation calculus was abandoned 50 years...
research
04/08/2022

Iterative Depth-First Search for Fully Observable Non-Deterministic Planning

Fully Observable Non-Deterministic (FOND) planning models uncertainty th...
research
05/22/2019

Minimizing the Negative Side Effects of Planning with Reduced Models

Reduced models of large Markov decision processes accelerate planning by...

Please sign up or login with your details

Forgot password? Click here to reset