An On-Line POMDP Solver for Continuous Observation Spaces

11/04/2020
by   Marcus Hoerger, et al.
0

Planning under partial obervability is essential for autonomous robots. A principled way to address such planning problems is the Partially Observable Markov Decision Process (POMDP). Although solving POMDPs is computationally intractable, substantial advancements have been achieved in developing approximate POMDP solvers in the past two decades. However, computing robust solutions for problems with continuous observation spaces remains challenging. Most on-line solvers rely on discretising the observation space or artificially limiting the number of observations that are considered during planning to compute tractable policies. In this paper we propose a new on-line POMDP solver, called Lazy Belief Extraction for Continuous POMDPs (LABECOP), that combines methods from Monte-Carlo-Tree-Search and particle filtering to construct a policy reprentation which doesn't require discretised observation spaces and avoids limiting the number of observations considered during planning. Experiments on three different problems involving continuous observation spaces indicate that LABECOP performs similar or better than state-of-the-art POMDP solvers.

READ FULL TEXT
research
07/23/2019

Multilevel Monte-Carlo for Solving POMDPs Online

Planning under partial obervability is essential for autonomous robots. ...
research
09/18/2017

POMCPOW: An online algorithm for POMDPs with continuous state, action, and observation spaces

Online solvers for partially observable Markov decision processes have b...
research
05/14/2023

A Surprisingly Simple Continuous-Action POMDP Solver: Lazy Cross-Entropy Search Over Policy Trees

The Partially Observable Markov Decision Process (POMDP) provides a prin...
research
11/24/2020

Stochastic Motion Planning under Partial Observability for Mobile Robots with Continuous Range Measurements

In this paper, we address the problem of stochastic motion planning unde...
research
09/28/2019

Dual Sequential Monte Carlo: Tunneling Filtering and Planning in Continuous POMDPs

We present the DualSMC network that solves continuous POMDPs by learning...
research
09/13/2022

Adaptive Discretization using Voronoi Trees for Continuous-Action POMDPs

Solving Partially Observable Markov Decision Processes (POMDPs) with con...
research
09/19/2023

Asymptotically Optimal Belief Space Planning in Discrete Partially-Observable Domains

Robots often have to operate in discrete partially observable worlds, wh...

Please sign up or login with your details

Forgot password? Click here to reset