Provably safe systems: the only path to controllable AGI

09/05/2023
by   Max Tegmark, et al.
0

We describe a path to humanity safely thriving with powerful Artificial General Intelligences (AGIs) by building them to provably satisfy human-specified requirements. We argue that this will soon be technically feasible using advanced AI for formal verification and mechanistic interpretability. We further argue that it is the only path which guarantees safe controlled AGI. We end with a list of challenge problems whose solution would contribute to this positive outcome and invite readers to join in this work.

READ FULL TEXT
research
08/20/2021

Safe Transformative AI via a Windfall Clause

Society could soon see transformative artificial intelligence (TAI). Mod...
research
06/01/2020

Revisiting Bounded-Suboptimal Safe Interval Path Planning

Safe-interval path planning (SIPP) is a powerful algorithm for finding a...
research
09/25/2022

Generating Formal Safety Assurances for High-Dimensional Reachability

Providing formal safety and performance guarantees for autonomous system...
research
08/01/2022

Safe Policy Improvement Approaches and their Limitations

Safe Policy Improvement (SPI) is an important technique for offline rein...
research
09/26/2020

Neurosymbolic Reinforcement Learning with Formally Verified Exploration

We present Revel, a partially neural reinforcement learning (RL) framewo...
research
02/01/2023

Safe Interval Path Planning With Kinodynamic Constraints

Safe Interval Path Planning (SIPP) is a powerful algorithm for solving s...
research
04/27/2023

Appropriateness is all you need!

The strive to make AI applications "safe" has led to the development of ...

Please sign up or login with your details

Forgot password? Click here to reset