Markov Decision Processes under Ambiguity

07/04/2019
by   Nicole Bäuerle, et al.
0

We consider statistical Markov Decision Processes where the decision maker is risk averse against model ambiguity. The latter is given by an unknown parameter which influences the transition law and the cost functions. Risk aversion is either measured by the entropic risk measure or by the Average Value at Risk. We show how to solve these kind of problems using a general minimax theorem. Under some continuity and compactness assumptions we prove the existence of an optimal (deterministic) policy and discuss its computation. We illustrate our results using an example from statistical decision theory.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/30/2015

A Notation for Markov Decision Processes

This paper specifies a notation for Markov decision processes....
research
12/26/2022

Statistical minimax theorems via nonstandard analysis

For statistical decision problems with finite parameter space, it is wel...
research
02/28/2018

Verification of Markov Decision Processes with Risk-Sensitive Measures

We develop a method for computing policies in Markov decision processes ...
research
09/01/2023

Learning Risk Preferences in Markov Decision Processes: an Application to the Fourth Down Decision in Football

For decades, National Football League (NFL) coaches' observed fourth dow...
research
07/13/2023

Entropic Risk for Turn-Based Stochastic Games

Entropic risk (ERisk) is an established risk measure in finance, quantif...
research
12/01/2021

Comparing discounted and average-cost Markov Decision Processes: a statistical significance perspective

Optimal Markov Decision Process policies for problems with finite state ...
research
01/03/2023

Risk-Averse MDPs under Reward Ambiguity

We propose a distributionally robust return-risk model for Markov decisi...

Please sign up or login with your details

Forgot password? Click here to reset