Verification of indefinite-horizon POMDPs

06/30/2020
by   Alexander Bork, et al.
0

The verification problem in MDPs asks whether, for any policy resolving the nondeterminism, the probability that something bad happens is bounded by some given threshold. This verification problem is often overly pessimistic, as the policies it considers may depend on the complete system state. This paper considers the verification problem for partially observable MDPs, in which the policies make their decisions based on (the history of) the observations emitted by the system. We present an abstraction-refinement framework extending previous instantiations of the Lovejoy-approach. Our experiments show that this framework significantly improves the scalability of the approach.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/26/2023

Robust Almost-Sure Reachability in Multi-Environment MDPs

Multiple-environment MDPs (MEMDPs) capture finite sets of MDPs that shar...
research
05/15/2022

Reductive MDPs: A Perspective Beyond Temporal Horizons

Solving general Markov decision processes (MDPs) is a computationally ha...
research
06/30/2017

Tableaux for Policy Synthesis for MDPs with PCTL* Constraints

Markov decision processes (MDPs) are the standard formalism for modellin...
research
06/24/2011

On Polynomial Sized MDP Succinct Policies

Policies of Markov Decision Processes (MDPs) determine the next action t...
research
03/22/2023

Wasserstein Auto-encoded MDPs: Formal Verification of Efficiently Distilled RL Policies with Many-sided Guarantees

Although deep reinforcement learning (DRL) has many success stories, the...
research
12/12/2012

Inductive Policy Selection for First-Order MDPs

We select policies for large Markov Decision Processes (MDPs) with compa...
research
01/06/2022

An Abstraction-Refinement Approach to Verifying Convolutional Neural Networks

Convolutional neural networks have gained vast popularity due to their e...

Please sign up or login with your details

Forgot password? Click here to reset