Interpretable Apprenticeship Learning with Temporal Logic Specifications

10/28/2017
by   Daniel Kasenberg, et al.
0

Recent work has addressed using formulas in linear temporal logic (LTL) as specifications for agents planning in Markov Decision Processes (MDPs). We consider the inverse problem: inferring an LTL specification from demonstrated behavior trajectories in MDPs. We formulate this as a multiobjective optimization problem, and describe state-based ("what actually happened") and action-based ("what the agent expected to happen") objective functions based on a notion of "violation cost". We demonstrate the efficacy of the approach by employing genetic programming to solve this problem in two simple domains.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/05/2018

Compositional planning in Markov decision processes: Temporal abstraction meets generalized logic composition

In hierarchical planning for Markov decision processes (MDPs), temporal ...
research
12/03/2020

Verifiable Planning in Expected Reward Multichain MDPs

The planning domain has experienced increased interest in the formal syn...
research
04/23/2023

Probabilistic Planning with Prioritized Preferences over Temporal Logic Objectives

This paper studies temporal planning in probabilistic environments, mode...
research
04/14/2017

Environment-Independent Task Specifications via GLTL

We propose a new task-specification language for Markov decision process...
research
03/10/2021

Inverse Reinforcement Learning of Autonomous Behaviors Encoded as Weighted Finite Automata

This paper presents a method for learning logical task specifications an...
research
10/26/2021

Average-Reward Learning and Planning with Options

We extend the options framework for temporal abstraction in reinforcemen...
research
03/02/2020

Learning and Solving Regular Decision Processes

Regular Decision Processes (RDPs) are a recently introduced model that e...

Please sign up or login with your details

Forgot password? Click here to reset