Risk-Averse Planning Under Uncertainty

09/27/2019
by   Mohamadreza Ahmadi, et al.
11

We consider the problem of designing policies for partially observable Markov decision processes (POMDPs) with dynamic coherent risk objectives. Synthesizing risk-averse optimal policies for POMDPs requires infinite memory and thus undecidable. To overcome this difficulty, we propose a method based on bounded policy iteration for designing stochastic but finite state (memory) controllers, which takes advantage of standard convex optimization methods. Given a memory budget and optimality criterion, the proposed method modifies the stochastic finite state controller leading to sub-optimal solutions with lower coherent risk.

READ FULL TEXT

page 1

page 8

research
09/09/2021

Risk-Averse Decision Making Under Uncertainty

A large class of decision making under uncertainty problems can be descr...
research
01/21/2020

Stochastic Finite State Control of POMDPs with LTL Specifications

Partially observable Markov decision processes (POMDPs) provide a modeli...
research
03/15/2012

Characterizing the Set of Coherent Lower Previsions with a Finite Number of Constraints or Vertices

The standard coherence criterion for lower previsions is expressed using...
research
01/23/2013

Learning Finite-State Controllers for Partially Observable Environments

Reactive (memoryless) policies are sufficient in completely observable M...
research
03/26/2021

Risk-Averse Stochastic Shortest Path Planning

We consider the stochastic shortest path planning problem in MDPs, i.e.,...
research
09/24/2020

Robust Finite-State Controllers for Uncertain POMDPs

Uncertain partially observable Markov decision processes (uPOMDPs) allow...
research
04/27/2018

Expectation Optimization with Probabilistic Guarantees in POMDPs with Discounted-sum Objectives

Partially-observable Markov decision processes (POMDPs) with discounted-...

Please sign up or login with your details

Forgot password? Click here to reset