Value Iteration with Options and State Aggregation

01/16/2015
by   Kamil Ciosek, et al.
0

This paper presents a way of solving Markov Decision Processes that combines state abstraction and temporal abstraction. Specifically, we combine state aggregation with the options framework and demonstrate that they work well together and indeed it is only after one combines the two that the full benefit of each is realized. We introduce a hierarchical value iteration algorithm where we first coarsely solve subgoals and then use these approximate solutions to exactly solve the MDP. This algorithm solved several problems faster than vanilla value iteration.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/29/2016

Algorithms for Batch Hierarchical Reinforcement Learning

Hierarchical Reinforcement Learning (HRL) exploits temporal abstraction ...
research
07/23/2021

An Adaptive State Aggregation Algorithm for Markov Decision Processes

Value iteration is a well-known method of solving Markov Decision Proces...
research
02/10/2016

Iterative Hierarchical Optimization for Misspecified Problems (IHOMP)

For complex, high-dimensional Markov Decision Processes (MDPs), it may b...
research
09/17/2013

Models and algorithms for skip-free Markov decision processes on trees

We introduce a class of models for multidimensional control problems whi...
research
09/20/2018

Logically-Constrained Neural Fitted Q-Iteration

This paper proposes a method for efficient training of the Q-function fo...
research
05/28/2019

Planning with State Abstractions for Non-Markovian Task Specifications

Often times, we specify tasks for a robot using temporal language that c...
research
09/29/2011

FluCaP: A Heuristic Search Planner for First-Order MDPs

We present a heuristic search algorithm for solving first-order Markov D...

Please sign up or login with your details

Forgot password? Click here to reset