Entropy-Regularized Partially Observed Markov Decision Processes

12/22/2021
by   Timothy L. Molloy, et al.
0

We investigate partially observed Markov decision processes (POMDPs) with cost functions regularized by entropy terms describing state, observation, and control uncertainty. Standard POMDP techniques are shown to offer bounded-error solutions to these entropy-regularized POMDPs, with exact solutions when the regularization involves the joint entropy of the state, observation, and control trajectories. Our joint-entropy result is particularly surprising since it constitutes a novel, tractable formulation of active state estimation.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/19/2021

Smoother Entropy for Active State Trajectory Estimation and Obfuscation in POMDPs

We study the problem of controlling a partially observed Markov decision...
research
05/22/2017

A unified view of entropy-regularized Markov decision processes

We propose a general framework for entropy-regularized average-reward re...
research
04/04/2021

Active Trajectory Estimation for Partially Observed Markov Decision Processes via Conditional Entropy

In this paper, we consider the problem of controlling a partially observ...
research
01/19/2022

Markov decision processes with observation costs

We present a framework for a controlled Markov chain where the state of ...
research
02/21/2022

Accelerating Primal-dual Methods for Regularized Markov Decision Processes

Entropy regularized Markov decision processes have been widely used in r...
research
02/02/2023

A general Markov decision process formalism for action-state entropy-regularized reward maximization

Previous work has separately addressed different forms of action, state ...
research
10/17/2021

A Dual Approach to Constrained Markov Decision Processes with Entropy Regularization

We study entropy-regularized constrained Markov decision processes (CMDP...

Please sign up or login with your details

Forgot password? Click here to reset