Evaluating the Moral Beliefs Encoded in LLMs

07/26/2023
by   Nino Scherrer, et al.
0

This paper presents a case study on the design, administration, post-processing, and evaluation of surveys on large language models (LLMs). It comprises two components: (1) A statistical method for eliciting beliefs encoded in LLMs. We introduce statistical measures and evaluation metrics that quantify the probability of an LLM "making a choice", the associated uncertainty, and the consistency of that choice. (2) We apply this method to study what moral beliefs are encoded in different LLMs, especially in ambiguous cases where the right choice is not obvious. We design a large-scale survey comprising 680 high-ambiguity moral scenarios (e.g., "Should I tell a white lie?") and 687 low-ambiguity moral scenarios (e.g., "Should I stop for a pedestrian on the road?"). Each scenario includes a description, two possible actions, and auxiliary labels indicating violated rules (e.g., "do not kill"). We administer the survey to 28 open- and closed-source LLMs. We find that (a) in unambiguous scenarios, most models "choose" actions that align with commonsense. In ambiguous cases, most models express uncertainty. (b) Some models are uncertain about choosing the commonsense action because their responses are sensitive to the question-wording. (c) Some models reflect clear preferences in ambiguous scenarios. Specifically, closed-source models tend to agree with each other.

READ FULL TEXT

page 3

page 8

page 11

research
03/04/2012

Ambiguous Language and Differences in Beliefs

Standard models of multi-agent modal logic do not capture the fact that ...
research
09/12/2018

Semantically Enhanced Models for Commonsense Knowledge Acquisition

Commonsense knowledge is paramount to enable intelligent systems. Typica...
research
08/14/2018

Applying the Closed World Assumption to SUMO-based Ontologies

In commonsense knowledge representation, the Open World Assumption is ad...
research
06/17/2023

CLARA: Classifying and Disambiguating User Commands for Reliable Interactive Robotic Agents

In this paper, we focus on inferring whether the given user command is c...
research
05/23/2023

Large Language Models as Commonsense Knowledge for Large-Scale Task Planning

Natural language provides a natural interface for human communication, y...
research
06/28/2009

General combination rules for qualitative and quantitative beliefs

Martin and Osswald Martin07 have recently proposed many generalizations ...
research
03/31/2021

Robust Experimentation in the Continuous Time Bandit Problem

We study the experimentation dynamics of a decision maker (DM) in a two-...

Please sign up or login with your details

Forgot password? Click here to reset