SafeLife 1.0: Exploring Side Effects in Complex Environments

12/03/2019
by   Carroll L. Wainwright, et al.
0

We present SafeLife, a publicly available reinforcement learning environment that tests the safety of reinforcement learning agents. It contains complex, dynamic, tunable, procedurally generated levels with many opportunities for unsafe behavior. Agents are graded both on their ability to maximize their explicit reward and on their ability to operate safely without unnecessary side effects. We train agents to maximize rewards using proximal policy optimization and score them on a suite of benchmark levels. The resulting agents are performant but not safe—they tend to cause large side effects in their environments—but they form a baseline against which future safety research can be measured.

READ FULL TEXT

page 6

page 12

page 13

research
11/27/2017

AI Safety Gridworlds

We present a suite of reinforcement learning environments illustrating v...
research
10/06/2020

Safety Aware Reinforcement Learning (SARL)

As reinforcement learning agents become increasingly integrated into com...
research
02/18/2019

Parenting: Safe Reinforcement Learning from Human Input

Autonomous agents trained via reinforcement learning present numerous sa...
research
10/20/2020

Robust Constrained Reinforcement Learning for Continuous Control with Model Misspecification

Many real-world physical control systems are required to satisfy constra...
research
06/26/2019

Towards Empathic Deep Q-Learning

As reinforcement learning (RL) scales to solve increasingly complex task...
research
09/15/2021

Evolutionary Reinforcement Learning Dynamics with Irreducible Environmental Uncertainty

In this work we derive and present evolutionary reinforcement learning d...
research
04/21/2022

Path-Specific Objectives for Safer Agent Incentives

We present a general framework for training safe agents whose naive ince...

Please sign up or login with your details

Forgot password? Click here to reset