In conversation with Artificial Intelligence: aligning language models with human values

09/01/2022
by   Atoosa Kasirzadeh, et al.
0

Large-scale language technologies are increasingly used in various forms of communication with humans across different contexts. One particular use case for these technologies is conversational agents, which output natural language text in response to prompts and queries. This mode of engagement raises a number of social and ethical questions. For example, what does it mean to align conversational agents with human norms or values? Which norms or values should they be aligned with? And how can this be accomplished? In this paper, we propose a number of steps that help answer these questions. We start by developing a philosophical analysis of the building blocks of linguistic communication between conversational agents and human interlocutors. We then use this analysis to identify and formulate ideal norms of conversation that can govern successful linguistic communication between humans and conversational agents. Furthermore, we explore how these norms can be used to align conversational agents with human values across a range of different discursive domains. We conclude by discussing the practical implications of our proposal for the design of conversational agents that are aligned with these norms and values.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/13/2022

Mediators: Conversational Agents Explaining NLP Model Behavior

The human-centric explainable artificial intelligence (HCXAI) community ...
research
01/20/2022

Learning Norms via Natural Language Teachings

To interact with humans, artificial intelligence (AI) systems must under...
research
02/17/2023

Value Engineering for Autonomous Agents

Machine Ethics (ME) is concerned with the design of Artificial Moral Age...
research
11/02/2020

An ontology-based chatbot for crises management: use case coronavirus

Today is the era of intelligence in machines. With the advances in Artif...
research
07/04/2023

Insert-expansions for Tool-enabled Conversational Agents

This paper delves into an advanced implementation of Chain-of-Thought-Pr...
research
04/02/2020

Improving Confidence in the Estimation of Values and Norms

Autonomous agents (AA) will increasingly be interacting with us in our d...
research
12/07/2019

Learning Norms from Stories: A Prior for Value Aligned Agents

Value alignment is a property of an intelligent agent indicating that it...

Please sign up or login with your details

Forgot password? Click here to reset