Legible Normativity for AI Alignment: The Value of Silly Rules

11/03/2018
by   Dylan Hadfield-Menell, et al.
0

It has become commonplace to assert that autonomous agents will have to be built to follow human rules of behavior--social norms and laws. But human laws and norms are complex and culturally varied systems, in many cases agents will have to learn the rules. This requires autonomous agents to have models of how human rule systems work so that they can make reliable predictions about rules. In this paper we contribute to the building of such models by analyzing an overlooked distinction between important rules and what we call silly rules--rules with no discernible direct impact on welfare. We show that silly rules render a normative system both more robust and more adaptable in response to shocks to perceived stability. They make normativity more legible for humans, and can increase legibility for AI systems as well. For AI systems to integrate into human normative systems, we suggest, it may be important for them to have models that include representations of silly rules.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/26/2021

How Should AI Interpret Rules? A Defense of Minimally Defeasible Interpretive Argumentation

Can artificially intelligent systems follow rules? The answer might seem...
research
07/07/2023

Specification, Validation and Verification of Social, Legal, Ethical, Empathetic and Cultural Requirements for Autonomous Agents

Autonomous agents are increasingly being proposed for use in healthcare,...
research
03/20/2023

Heterogeneity of AI-Induced Societal Harms and the Failure of Omnibus AI Laws

AI-induced societal harms mirror existing problems in domains where AI r...
research
10/13/2021

Truthful AI: Developing and governing AI that does not lie

In many contexts, lying – the use of verbal falsehoods to deceive – is h...
research
03/08/2023

Computational-level Analysis of Constraint Compliance for General Intelligence

Human behavior is conditioned by codes and norms that constrain action. ...
research
02/01/2014

Godseed: Benevolent or Malevolent?

It is hypothesized by some thinkers that benign looking AI objectives ma...
research
02/21/2022

Learning Behavioral Soft Constraints from Demonstrations

Many real-life scenarios require humans to make difficult trade-offs: do...

Please sign up or login with your details

Forgot password? Click here to reset