Truthful AI: Developing and governing AI that does not lie

10/13/2021
by   Owain Evans, et al.
0

In many contexts, lying – the use of verbal falsehoods to deceive – is harmful. While lying has traditionally been a human affair, AI systems that make sophisticated verbal statements are becoming increasingly prevalent. This raises the question of how we should limit the harm caused by AI "lies" (i.e. falsehoods that are actively selected for). Human truthfulness is governed by social norms and by laws (against defamation, perjury, and fraud). Differences between AI and humans present an opportunity to have more precise standards of truthfulness for AI, and to have these standards rise over time. This could provide significant benefits to public epistemics and the economy, and mitigate risks of worst-case AI futures. Establishing norms or laws of AI truthfulness will require significant work to: (1) identify clear truthfulness standards; (2) create institutions that can judge adherence to those standards; and (3) develop AI systems that are robustly truthful. Our initial proposals for these areas include: (1) a standard of avoiding "negligent falsehoods" (a generalisation of lies that is easier to assess); (2) institutions to evaluate AI systems before and after real-world deployment; and (3) explicitly training AI systems to be truthful via curated datasets and human interaction. A concerning possibility is that evaluation mechanisms for eventual truthfulness standards could be captured by political interests, leading to harmful censorship and propaganda. Avoiding this might take careful attention. And since the scale of AI speech acts might grow dramatically over the coming decades, early truthfulness standards might be particularly important because of the precedents they set.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/26/2020

Could regulating the creators deliver trustworthy AI?

Is a new regulated profession, such as Artificial Intelligence (AI) Arch...
research
11/03/2018

Legible Normativity for AI Alignment: The Value of Silly Rules

It has become commonplace to assert that autonomous agents will have to ...
research
07/06/2023

Frontier AI Regulation: Managing Emerging Risks to Public Safety

Advanced AI models hold the promise of tremendous benefits for humanity,...
research
07/10/2023

International Institutions for Advanced AI

International institutions may have an important role to play in ensurin...
research
09/21/2023

On the relationship between Benchmarking, Standards and Certification in Robotics and AI

Benchmarking, standards and certification are closely related processes....
research
05/11/2023

Towards best practices in AGI safety and governance: A survey of expert opinion

A number of leading AI companies, including OpenAI, Google DeepMind, and...
research
07/19/2023

Absolutist AI

This paper argues that training AI systems with absolute constraints – w...

Please sign up or login with your details

Forgot password? Click here to reset