Safety Considerations in Deep Control Policies with Probabilistic Safety Barrier Certificates

01/22/2020
by   Tom Hirshberg, et al.
0

Recent advances in Deep Machine Learning have shown promise in solving complex perception and control loops via methods such as reinforcement and imitation learning. However, guaranteeing safety for such learned deep policies has been a challenge due to issues such as partial observability and difficulties in characterizing the behavior of the neural networks. While a lot of emphasis in safe learning has been placed during training, it is non-trivial to guarantee safety at deployment or test time. This paper extends the work on Safety Barrier Certificates to guarantee safety with deep control policies despite uncertainty arising due to perception and other latent variables. In particular, the proposed framework wraps around the existing deep control policy and generates safe actions by dynamically evaluating and modifying the policy from the embedded network. Our framework utilizes control barrier functions to create spaces of control actions that are probabilistically safe, and when the original actions are found to be in violation of the safety constraint, uses quadratic programming to minimally modify the original actions to ensure they lie in the safe set. Representations of the environment are built through Euclidean signed distance fields that are then used to infer the safety of actions and to guarantee forward invariance. We implement this method in simulation in a drone-racing environment and show that our method results in safer actions compared to a baseline that only relies on imitation learning to generate control actions.

READ FULL TEXT

page 4

page 5

page 7

page 8

research
03/07/2023

ConBaT: Control Barrier Transformer for Safe Policy Learning

Large-scale self-supervised models have recently revolutionized our abil...
research
12/01/2022

Safe Reinforcement Learning with Probabilistic Control Barrier Functions for Ramp Merging

Prior work has looked at applying reinforcement learning and imitation l...
research
07/22/2018

EnsembleDAgger: A Bayesian Approach to Safe Imitation Learning

While imitation learning is often used in robotics, this approach often ...
research
03/03/2022

Fail-Safe Generative Adversarial Imitation Learning

For flexible yet safe imitation learning (IL), we propose a modular appr...
research
09/14/2021

Reactive and Safe Road User Simulations using Neural Barrier Certificates

Reactive and safe agent modelings are important for nowadays traffic sim...
research
09/18/2017

DropoutDAgger: A Bayesian Approach to Safe Imitation Learning

While imitation learning is becoming common practice in robotics, this a...

Please sign up or login with your details

Forgot password? Click here to reset