SMARLA: A Safety Monitoring Approach for Deep Reinforcement Learning Agents

08/03/2023
by   Amirhossein Zolfagharian, et al.
0

Deep reinforcement learning algorithms (DRL) are increasingly being used in safety-critical systems. Ensuring the safety of DRL agents is a critical concern in such contexts. However, relying solely on testing is not sufficient to ensure safety as it does not offer guarantees. Building safety monitors is one solution to alleviate this challenge. This paper proposes SMARLA, a machine learning-based safety monitoring approach designed for DRL agents. For practical reasons, SMARLA is designed to be black-box (as it does not require access to the internals of the agent) and leverages state abstraction to reduce the state space and thus facilitate the learning of safety violation prediction models from agent's states. We validated SMARLA on two well-known RL case studies. Empirical analysis reveals that SMARLA achieves accurate violation prediction with a low false positive rate, and can predict safety violations at an early stage, approximately halfway through the agent's execution before violations occur.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/15/2022

Search-Based Testing Approach for Deep Reinforcement Learning Agents

Deep Reinforcement Learning (DRL) algorithms have been increasingly empl...
research
01/20/2021

Shielding Atari Games with Bounded Prescience

Deep reinforcement learning (DRL) is applied in safety-critical domains ...
research
06/20/2022

Constrained Reinforcement Learning for Robotics via Scenario-Based Programming

Deep reinforcement learning (DRL) has achieved groundbreaking successes ...
research
05/22/2023

Testing of Deep Reinforcement Learning Agents with Surrogate Models

Deep Reinforcement Learning (DRL) has received a lot of attention from t...
research
08/23/2023

How Safe Am I Given What I See? Calibrated Prediction of Safety Chances for Image-Controlled Autonomy

End-to-end learning has emerged as a major paradigm for developing auton...
research
11/21/2022

BBReach: Tight and Scalable Black-Box Reachability Analysis of Deep Reinforcement Learning Systems

Reachability analysis is a promising technique to automatically prove or...
research
11/22/2018

Oversight of Unsafe Systems via Dynamic Safety Envelopes

This paper reviews the reasons that Human-in-the-Loop is both critical f...

Please sign up or login with your details

Forgot password? Click here to reset