POMDP-lite for Robust Robot Planning under Uncertainty

02/16/2016
by   Min Chen, et al.
0

The partially observable Markov decision process (POMDP) provides a principled general model for planning under uncertainty. However, solving a general POMDP is computationally intractable in the worst case. This paper introduces POMDP-lite, a subclass of POMDPs in which the hidden state variables are constant or only change deterministically. We show that a POMDP-lite is equivalent to a set of fully observable Markov decision processes indexed by a hidden parameter and is useful for modeling a variety of interesting robotic tasks. We develop a simple model-based Bayesian reinforcement learning algorithm to solve POMDP-lite models. The algorithm performs well on large-scale POMDP-lite models with up to 10^20 states and outperforms the state-of-the-art general-purpose POMDP algorithms. We further show that the algorithm is near-Bayesian-optimal under suitable conditions.

READ FULL TEXT
research
01/30/2013

Planning with Partially Observable Markov Decision Processes: Advances in Exact Solution Method

There is much interest in using partially observable Markov decision pro...
research
09/17/2018

Hidden Markov Model Estimation-Based Q-learning for Partially Observable Markov Decision Process

The objective is to study an on-line Hidden Markov model (HMM) estimatio...
research
10/30/2022

Planning to the Information Horizon of BAMDPs via Epistemic State Abstraction

The Bayes-Adaptive Markov Decision Process (BAMDP) formalism pursues the...
research
06/07/2022

Learning in Observable POMDPs, without Computationally Intractable Oracles

Much of reinforcement learning theory is built on top of oracles that ar...
research
09/21/2022

Partially Observable Markov Decision Processes in Robotics: A Survey

Noisy sensing, imperfect control, and environment changes are defining c...
research
01/10/2013

Planning and Acting under Uncertainty: A New Model for Spoken Dialogue Systems

Uncertainty plays a central role in spoken dialogue systems. Some stocha...
research
09/12/2016

DESPOT: Online POMDP Planning with Regularization

The partially observable Markov decision process (POMDP) provides a prin...

Please sign up or login with your details

Forgot password? Click here to reset