Playing the Blame Game with Robots

02/08/2021
by   Markus Kneer, et al.
0

Recent research shows – somewhat astonishingly – that people are willing to ascribe moral blame to AI-driven systems when they cause harm [1]-[4]. In this paper, we explore the moral-psychological underpinnings of these findings. Our hypothesis was that the reason why people ascribe moral blame to AI systems is that they consider them capable of entertaining inculpating mental states (what is called mens rea in the law). To explore this hypothesis, we created a scenario in which an AI system runs a risk of poisoning people by using a novel type of fertilizer. Manipulating the computational (or quasi-cognitive) abilities of the AI system in a between-subjects design, we tested whether people's willingness to ascribe knowledge of a substantial risk of harm (i.e., recklessness) and blame to the AI system. Furthermore, we investigated whether the ascription of recklessness and blame to the AI system would influence the perceived blameworthiness of the system's user (or owner). In an experiment with 347 participants, we found (i) that people are willing to ascribe blame to AI systems in contexts of recklessness, (ii) that blame ascriptions depend strongly on the willingness to attribute recklessness and (iii) that the latter, in turn, depends on the perceived "cognitive" capacities of the system. Furthermore, our results suggest (iv) that the higher the computational sophistication of the AI system, the more blame is shifted from the human user to the AI system.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/15/2021

Can Artificial Intelligence Make Art?

In two experiments (total N=693) we explored whether people are willing ...
research
01/06/2023

Improving Human-AI Collaboration With Descriptions of AI Behavior

People work with AI systems to improve their decision making, but often ...
research
02/19/2021

To Trust or to Think: Cognitive Forcing Functions Can Reduce Overreliance on AI in AI-assisted Decision-making

People supported by AI-powered decision support tools frequently overrel...
research
05/15/2023

Capturing Humans' Mental Models of AI: An Item Response Theory Approach

Improving our understanding of how humans perceive AI teammates is an im...
research
03/15/2021

Crossing the Tepper Line: An Emerging Ontology for Describing the Dynamic Sociality of Embodied AI

Artificial intelligences (AI) are increasingly being embodied and embedd...
research
01/18/2019

Friend, Collaborator, Student, Manager: How Design of an AI-Driven Game Level Editor Affects Creators

Machine learning advances have afforded an increase in algorithms capabl...
research
03/15/2023

Robot Navigation in Risky, Crowded Environments: Understanding Human Preferences

Risky and crowded environments (RCE) contain abstract sources of risk an...

Please sign up or login with your details

Forgot password? Click here to reset