Policy Optimization with Linear Temporal Logic Constraints

06/20/2022
by   Cameron Voloshin, et al.
12

We study the problem of policy optimization (PO) with linear temporal logic (LTL) constraints. The language of LTL allows flexible description of tasks that may be unnatural to encode as a scalar cost function. We consider LTL-constrained PO as a systematic framework, decoupling task specification from policy selection, and an alternative to the standard of cost shaping. With access to a generative model, we develop a model-based approach that enjoys a sample complexity analysis for guaranteeing both task satisfaction and cost optimality (through a reduction to a reachability problem). Empirically, our algorithm can achieve strong performance even in low sample regimes.

READ FULL TEXT

page 6

page 7

page 9

page 10

page 11

page 12

page 31

page 34

research
08/01/2020

Learning with Safety Constraints: Sample Complexity of Reinforcement Learning for Constrained MDPs

Many physical systems have underlying safety considerations that require...
research
10/12/2020

Is Plug-in Solver Sample-Efficient for Feature-based Reinforcement Learning?

It is believed that a model-based approach for reinforcement learning (R...
research
11/30/2022

Reinforcement Learning for Signal Temporal Logic using Funnel-Based Approach

Signal Temporal Logic (STL) is a powerful framework for describing the c...
research
09/27/2016

Task Specific Adversarial Cost Function

The cost function used to train a generative model should fit the purpos...
research
03/03/2023

Eventual Discounting Temporal Logic Counterfactual Experience Replay

Linear temporal logic (LTL) offers a simplified way of specifying tasks ...
research
10/11/2020

Safe Reinforcement Learning with Natural Language Constraints

In this paper, we tackle the problem of learning control policies for ta...

Please sign up or login with your details

Forgot password? Click here to reset