Sensor Synthesis for POMDPs with Reachability Objectives

09/29/2017
by   Krishnendu Chatterjee, et al.
0

Partially observable Markov decision processes (POMDPs) are widely used in probabilistic planning problems in which an agent interacts with an environment using noisy and imprecise sensors. We study a setting in which the sensors are only partially defined and the goal is to synthesize "weakest" additional sensors, such that in the resulting POMDP, there is a small-memory policy for the agent that almost-surely (with probability 1) satisfies a reachability objective. We show that the problem is NP-complete, and present a symbolic algorithm by encoding the problem into SAT instances. We illustrate trade-offs between the amount of memory of the policy and the number of additional sensors on a simple example. We have implemented our approach and consider three classical POMDP examples from the literature, and show that in all the examples the number of sensors can be significantly decreased (as compared to the existing solutions in the literature) without increasing the complexity of the policies.

READ FULL TEXT
research
11/26/2015

A Symbolic SAT-based Algorithm for Almost-sure Reachability with Small Strategies in POMDPs

POMDPs are standard models for probabilistic planning problems, where an...
research
01/29/2018

Bounded Policy Synthesis for POMDPs with Safe-Reachability Objectives

Planning robust executions under uncertainty is a fundamental challenge ...
research
07/16/2020

Strengthening Deterministic Policies for POMDPs

The synthesis problem for partially observable Markov decision processes...
research
06/30/2020

Enforcing Almost-Sure Reachability in POMDPs

Partially-Observable Markov Decision Processes (POMDPs) are a well-known...
research
01/26/2023

Robust Almost-Sure Reachability in Multi-Environment MDPs

Multiple-environment MDPs (MEMDPs) capture finite sets of MDPs that shar...
research
03/21/2020

Adaptive Informative Path Planning with Multimodal Sensing

Adaptive Informative Path Planning (AIPP) problems model an agent tasked...
research
04/27/2018

Expectation Optimization with Probabilistic Guarantees in POMDPs with Discounted-sum Objectives

Partially-observable Markov decision processes (POMDPs) with discounted-...

Please sign up or login with your details

Forgot password? Click here to reset