Planning with State Abstractions for Non-Markovian Task Specifications

05/28/2019
by   Yoonseon Oh, et al.
0

Often times, we specify tasks for a robot using temporal language that can also span different levels of abstraction. The example command "go to the kitchen before going to the second floor" contains spatial abstraction, given that "floor" consists of individual rooms that can also be referred to in isolation ("kitchen", for example). There is also a temporal ordering of events, defined by the word "before". Previous works have used Linear Temporal Logic (LTL) to interpret temporal language (such as "before"), and Abstract Markov Decision Processes (AMDPs) to interpret hierarchical abstractions (such as "kitchen" and "second floor"), separately. To handle both types of commands at once, we introduce the Abstract Product Markov Decision Process (AP-MDP), a novel approach capable of representing non-Markovian reward functions at different levels of abstractions. The AP-MDP framework translates LTL into its corresponding automata, creates a product Markov Decision Process (MDP) of the LTL specification and the environment MDP, and decomposes the problem into subproblems to enable efficient planning with abstractions. AP-MDP performs faster than a non-hierarchical method of solving LTL problems in over 95 tasks, and this number only increases as the size of the environment domain increases. We also present a neural sequence-to-sequence model trained to translate language commands into LTL expression, and a new corpus of non-Markovian language commands spanning different levels of abstraction. We test our framework with the collected language commands on a drone, demonstrating that our approach enables a robot to efficiently solve temporal commands at different levels of abstraction.

READ FULL TEXT

page 1

page 8

research
11/30/2018

Online abstraction with MDP homomorphisms for Deep Learning

Abstraction of Markov Decision Processes is a useful tool for solving co...
research
10/19/2012

Implementation and Comparison of Solution Methods for Decision Processes with Non-Markovian Rewards

This paper examines a number of solution methods for decision processes ...
research
07/18/2018

Representational efficiency outweighs action efficiency in human program induction

The importance of hierarchically structured representations for tractabl...
research
02/18/2022

SMC4PEP: Stochastic Model Checking of Product Engineering Processes

Product Engineering Processes (PEPs) are used for describing complex pro...
research
11/10/2022

Switching Attention in Time-Varying Environments via Bayesian Inference of Abstractions

Motivated by the goal of endowing robots with a means for focusing atten...
research
02/28/2023

Exploiting Multiple Abstractions in Episodic RL via Reward Shaping

One major limitation to the applicability of Reinforcement Learning (RL)...
research
01/16/2015

Value Iteration with Options and State Aggregation

This paper presents a way of solving Markov Decision Processes that comb...

Please sign up or login with your details

Forgot password? Click here to reset