[ExI] Why “Everyone Dies” Gets AGI All Wrong by Ben Goertzel

BillK pharos at gmail.com
Wed Oct 1 09:49:48 UTC 2025


Why “Everyone Dies” Gets AGI All Wrong
A Response to Yudkowsky and Soares from the Front Lines of AGI Development
Ben Goertzel    Oct 01, 2025

Being: A reaction to Eliezer Yudkowsky and Nate Soares’s book “If
anybody builds it everyone dies” which is getting a bit of media
attention.

<https://bengoertzel.substack.com/p/why-everyone-dies-gets-agi-all-wrong>
Quote:
An intelligence capable of recursive self-improvement and transcending
from AGI to ASI would naturally tend toward complexity, nuance, and
relational adaptability rather than monomaniacal optimization.
------------------

A good description of why we should survive the arrival of AGI. (probably!).
BillK



More information about the extropy-chat mailing list