Enforcing Almost-Sure Reachability in POMDPs

06/30/2020
by   Sebastian Junges, et al.
0

Partially-Observable Markov Decision Processes (POMDPs) are a well-known formal model for planning scenarios where agents operate under limited information about their environment. In safety-critical domains, the agent must adhere to a policy satisfying certain behavioral constraints. We study the problem of synthesizing policies that almost-surely reach some goal state while a set of bad states is never visited. In particular, we present an iterative symbolic approach that computes a winning region, that is, a set of system configurations such that all policies that stay within this set are guaranteed to satisfy the constraints. The approach generalizes and improves previous work in terms of scalability and efficacy, as demonstrated in the empirical evaluation. Additionally, we show the applicability to safe exploration by restricting agent behavior to these winning regions.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/19/2023

Safe POMDP Online Planning via Shielding

Partially observable Markov decision processes (POMDPs) have been widely...
research
06/11/2014

Quantum POMDPs

We present quantum observable Markov decision processes (QOMDPs), the qu...
research
05/24/2023

Optimal Control of Logically Constrained Partially Observable and Multi-Agent Markov Decision Processes

Autonomous systems often have logical constraints arising, for example, ...
research
01/29/2018

Bounded Policy Synthesis for POMDPs with Safe-Reachability Objectives

Planning robust executions under uncertainty is a fundamental challenge ...
research
09/29/2017

Sensor Synthesis for POMDPs with Reachability Objectives

Partially observable Markov decision processes (POMDPs) are widely used ...
research
11/26/2016

Optimizing Expectation with Guarantees in POMDPs (Technical Report)

A standard objective in partially-observable Markov decision processes (...
research
06/01/2011

Asimovian Adaptive Agents

The goal of this research is to develop agents that are adaptive and pre...

Please sign up or login with your details

Forgot password? Click here to reset