works
Richard Ngo AGI safety from first principles: control online The effects of a misaligned artificial general intelligence (AGI) are a major concern in the field of AI safety. Two scenarios are possible: either a single AGI gains power through technological breakthroughs and seizes control of the world, or several misaligned AGIs gain influence and eventually become more powerful than both humans and aligned AIs. Whether one of these scenarios will occur depends on the speed of AI development, the transparency of AI systems, constrained deployment strategies, and human political and economic coordination. – AI-generated abstract.

AGI safety from first principles: control

Richard Ngo

AI Alignment Forum, September 28, 2020

Abstract

The effects of a misaligned artificial general intelligence (AGI) are a major concern in the field of AI safety. Two scenarios are possible: either a single AGI gains power through technological breakthroughs and seizes control of the world, or several misaligned AGIs gain influence and eventually become more powerful than both humans and aligned AIs. Whether one of these scenarios will occur depends on the speed of AI development, the transparency of AI systems, constrained deployment strategies, and human political and economic coordination. – AI-generated abstract.

PDF

First page of PDF