Towards best practices in AGI safety and governance: A survey of expert opinion

05/11/2023
by   Jonas Schuett, et al.
0

A number of leading AI companies, including OpenAI, Google DeepMind, and Anthropic, have the stated goal of building artificial general intelligence (AGI) - AI systems that achieve or exceed human performance across a wide range of cognitive tasks. In pursuing this goal, they may develop and deploy AI systems that pose particularly significant risks. While they have already taken some measures to mitigate these risks, best practices have not yet emerged. To support the identification of best practices, we sent a survey to 92 leading experts from AGI labs, academia, and civil society and received 51 responses. Participants were asked how much they agreed with 50 statements about what AGI labs should do. Our main finding is that participants, on average, agreed with all of them. Many statements received extremely high levels of agreement. For example, 98 conduct pre-deployment risk assessments, dangerous capabilities evaluations, third-party model audits, safety restrictions on model usage, and red teaming. Ultimately, our list of statements may serve as a helpful foundation for efforts to develop best practices, standards, and regulations for AGI labs.

READ FULL TEXT

page 5

page 9

research
07/17/2023

Risk assessment at AGI companies: A review of popular risk assessment techniques from other safety-critical industries

Companies like OpenAI, Google DeepMind, and Anthropic have the stated go...
research
01/25/2023

Requirements Practices and Gaps When Engineering Human-Centered Artificial Intelligence Systems

[Context] Engineering Artificial Intelligence (AI) software is a relativ...
research
08/28/2023

The Promise and Peril of Artificial Intelligence – Violet Teaming Offers a Balanced Path Forward

Artificial intelligence (AI) promises immense benefits across sectors, y...
research
07/06/2023

Frontier AI Regulation: Managing Emerging Risks to Public Safety

Advanced AI models hold the promise of tremendous benefits for humanity,...
research
05/26/2023

AGI labs need an internal audit function

The paper argues that organizations that have the stated goal of buildin...
research
10/13/2021

Truthful AI: Developing and governing AI that does not lie

In many contexts, lying – the use of verbal falsehoods to deceive – is h...
research
12/01/2015

Recommendations for the Technical Infrastructure for Standardized International Rights Statements

This white paper is the product of a joint Digital Public Library of Ame...

Please sign up or login with your details

Forgot password? Click here to reset