Reinforcement Learning Under Moral Uncertainty

06/08/2020
by   Adrien Ecoffet, et al.
11

An ambitious goal for artificial intelligence is to create agents that behave ethically: The capacity to abide by human moral norms would greatly expand the context in which autonomous agents could be practically and safely deployed. While ethical agents could be trained through reinforcement, by rewarding correct behavior under a specific moral theory (e.g. utilitarianism), there remains widespread disagreement (both societally and among moral philosophers) about the nature of morality and what ethical theory (if any) is objectively correct. Acknowledging such disagreement, recent work in moral philosophy proposes that ethical behavior requires acting under moral uncertainty, i.e. to take into account when acting that one's credence is split across several plausible ethical theories. Inspired by such work, this paper proposes a formalism that translates such insights to the field of reinforcement learning. Demonstrating the formalism's potential, we then train agents in simple environments to act under moral uncertainty, highlighting how such uncertainty can help curb extreme behavior from commitment to single theories. The overall aim is to draw productive connections from the fields of moral philosophy and machine ethics to that of machine learning, to inspire further research by highlighting a spectrum of machine learning research questions relevant to training ethically capable reinforcement learning agents.

READ FULL TEXT

page 25

page 26

page 27

page 28

page 29

page 31

page 32

page 33

research
08/30/2022

Towards Artificial Virtuous Agents: Games, Dilemmas and Machine Learning

Machine ethics has received increasing attention over the past few years...
research
12/12/2017

A Low-Cost Ethics Shaping Approach for Designing Reinforcement Learning Agents

This paper proposes a low-cost, easily realizable strategy to equip a re...
research
09/08/2023

SHAPE: A Framework for Evaluating the Ethicality of Influence

Agents often exert influence when interacting with humans and non-human ...
research
01/20/2023

Modeling Moral Choices in Social Dilemmas with Multi-Agent Reinforcement Learning

Practical uses of Artificial Intelligence (AI) in the real world have de...
research
12/07/2018

Toward the Engineering of Virtuous Machines

While various traditions under the 'virtue ethics' umbrella have been st...
research
03/30/2022

Reinforcement Learning Guided by Provable Normative Compliance

Reinforcement learning (RL) has shown promise as a tool for engineering ...
research
08/26/2020

Ethical behavior in humans and machines – Evaluating training data quality for beneficial machine learning

Machine behavior that is based on learning algorithms can be significant...

Please sign up or login with your details

Forgot password? Click here to reset