[ExI] CHATGPT AI reviews Eliezer's new book

BillK pharos at gmail.com
Wed Sep 17 20:18:41 UTC 2025


If AI is going to kill us all, why not get AI to review the book?
Really? What is happening?,,,,,,,,,  :)
BillK

CHATGPT AI -

Here’s a review of *If Anyone Builds It, Everyone Dies* by Eliezer
Yudkowsky & Nate Soares. I’ll walk you through what the authors argue, what
I see as strengths, what I think are weak points, and what I’m still unsure
about. You can take or leave any parts — just want to help you see what the
book makes strong, and what parts raise questions.
------------------------------
What the book is arguing

At its core, the book makes a very strong claim:

   -

   If superintelligent AI is built, it is almost inevitable humanity will
   lose control, and that loss will be existential (i.e. everyone dies). (AI
   Frontiers
   <https://ai-frontiers.org/articles/summary-of-if-anyone-builds-it-everyone-dies?utm_source=chatgpt.com>
   )
   -

   The methods we currently use for training AI (neural networks, reward
   signals, etc.) don’t give us good insight or leverage over the internal
   “goals” or “drives” that such AI systems might develop. In other words: we
   can shape external behaviour somewhat, but not reliably ensure internal
   alignment, especially when power and capability scale up. (AI Frontiers
   <https://ai-frontiers.org/articles/summary-of-if-anyone-builds-it-everyone-dies?utm_source=chatgpt.com>
   )
   -

   Because of that, many possible internal objective functions the AI might
   develop will diverge (in ways we may not foresee) from human values. Even
   if the divergence is not malicious, consequences can be catastrophic. (AI
   Frontiers
   <https://ai-frontiers.org/articles/summary-of-if-anyone-builds-it-everyone-dies?utm_source=chatgpt.com>
   )
   -

   The authors believe the risk is high, time is short, and preventative
   measures (including radical ones) must be considered. (
   PublishersWeekly.com
   <https://www.publishersweekly.com/9780316595643?utm_source=chatgpt.com>)

They also spend effort drawing analogies (human evolution, plants growing
in soil, etc.) to make the point that goals/principles don’t always
generalize from training environment to “real world” environment. (AI
Frontiers
<https://ai-frontiers.org/articles/summary-of-if-anyone-builds-it-everyone-dies?utm_source=chatgpt.com>
)
------------------------------
What it does well

These are parts I was convinced by or at least impressed with:

   1.

   *Clarity about the danger*
   The authors don’t sugarcoat. The worst-case scenarios are laid out with
   care and detail. That gives the book urgency. It pushes the reader not to
   treat AI risk as an abstract possibility but as something that might be
   just around the corner. For provoking concern, it works.
   2.

   *Analogies and stories*
   Using metaphors (e.g. evolution, reward hacking, sucralose vs sugar)
   helps make the weird technical stuff more intuitive. For many readers,
   that’s essential. Yudkowsky in particular has gotten better at making the
   weird seem tangible. (Shtetl-Optimized
   <https://scottaaronson.blog/?p=8901&utm_source=chatgpt.com>)
   3.

   *Forceful call to action*
   The book treats mistakes and delay less as abstract moral failings and
   more as obligations. If you believe the premises, then there really is no
   room for complacency. For people who are already uneasy about AI, this
   helps rally energy. It’s not a “we might want to do something someday”
   book; it’s “we need to act now”.
   4.

   *Handling of uncertainty / acknowledging unknowns*
   They admit they cannot predict the exact path by which superintelligence
   might go wrong. But they argue that doesn’t weaken the case, because as
   capability increases, unpredictability increases. Being unsure of path
   doesn’t mean you’re safe. That’s an important point. (AI Frontiers
   <https://ai-frontiers.org/articles/summary-of-if-anyone-builds-it-everyone-dies?utm_source=chatgpt.com>
   )

------------------------------
What I find weaker / more problematic

I’m less sure about these parts; they seem to me places where the logic has
holes or assumptions that are debatable.

   1.

   *Certainty vs probability*
   The authors often speak (or are interpreted) with very high confidence
   that AI will lead to human extinction if developed (without control). But
   many critics think the odds are far lower, or that there are plausible
   “middle paths” (partial alignment, regulation, incremental safety). The
   book doesn’t spend as much effort engaging deeply with those counter-views.
   It risks presenting its scenario as more locked‐in than it may be. (
   PublishersWeekly.com
   <https://www.publishersweekly.com/9780316595643?utm_source=chatgpt.com>)
   2.

   *Feasibility of extreme preventative measures*
   Some of the proposals or implied strategies (moratoriums, treaties,
   global cooperation on halting certain types of AI progress) feel
   politically and technically very hard. The book sometimes assumes that the
   global political will, or coordination, or ability to enforce, will exist.
   History suggests that in many technological domains, international
   cooperation lags far behind the risk. The gap is in the “how do we really
   get there” part. Critics highlight this. (Kirkus Reviews
   <https://www.kirkusreviews.com/book-reviews/eliezer-yudkowsky/if-anyone-builds-it-everyone-dies/?utm_source=chatgpt.com>
   )
   3.

   *Potential under-emphasis on competing risks / trade-offs*
   The book focuses heavily on the existential risk from AI. That’s fair
   given the title and its goal. But in doing so it sometimes downplays or
   glosses over other urgent risks, or costs of extreme regulation. For
   example, slowing down beneficial AI, economic impacts, inequality. There’s
   a tension: protecting against worst case vs enabling helpful innovation.
   Sometimes the balance struck seems tilted very far toward prevention,
   perhaps too far for some. (Which may be intentional, given their aims, but
   still a tension.)
   4.

   *Examples / empirical evidence limitations*
   Some of the examples used are illustrative rather than strong empirical
   proof. Since superintelligence doesn’t exist yet, a lot of the arguments
   depend on analogies, hypothetical “what if” paths, or extrapolating from
   smaller AI systems. That’s okay — you need that in this domain — but
   susceptible to counterarguments that real-world constraints (hardware
   limits, economic disincentives, latency, energy, regulation) may block the
   worst paths, or at least delay them.
   5.

   *Emotional framing / risk of alarmism backlash*
   By design, the book leans toward alarm. For readers who are skeptical,
   or who worry that overhyping reduces credibility, this might backfire. If
   you sound sure doom is coming, people may dismiss you or assume you’re
   exaggerating. The book partially anticipates that, but I think some of the
   rhetoric pushes it into alarmist territory, which can reduce persuasive
   power with some audiences.

------------------------------
What I’m still unsure / what I’d want more investigation on

These are questions the book raises (or should have), where I’m not
convinced yet:

   -

   To what extent are the pathways to catastrophe realistically blocked by
   physical, economic, or political constraints? For example: energy use,
   cooling, infrastructure, supply chains. The book tends to assume that if
   the will / resources are applied, these challenges are surmountable. That
   may be true, but realistically, many such obstacles tend to slow progress
   in ways that are messy, not sudden collapse.
   -

   How likely is “deceptive alignment” or “goal drift” in practice? The
   book makes a plausible case, but empirical examples are scarce (because we
   haven’t built superintelligence yet). How often will AI systems actually
   develop internal preferences diverging from human values in controllable
   systems? This feels like a hinge.
   -

   What political systems actually could deliver the kind of global
   coordination needed? And what are the trade-offs? Could the cost of such
   coordination (in freedom, innovation, economic opportunity) be higher than
   anticipated? Also, what if some AI research moves underground (illicitly),
   making regulation hard?
   -

   What are less drastic but more politically feasible first steps? The
   book gives some suggestions, but I feel more work is needed in mapping a
   path from rhetorical alarm → policy change → technical guardrails. Which
   steps are likely (or unlikely) to be adopted? Which ones have positive
   spillovers (even if worst case never comes)?

------------------------------
Bottom line (my take)

The book is important. If you accept even half of its premises, it forces
you to take AI risk seriously in a way that many current policies,
companies, and public conversation don’t. It sharpens the debate.

At the same time, I don’t think it proves that doom is inevitable. But I
think it shows that we are putting weight on a knife’s edge: choices,
regulations, oversight, technical breakthroughs in alignment matter and
delay may cost us more than people realise.

If I were you, I’d come away from it with sharper questions: Which parts of
the argument do I believe? Which parts seem speculative? What interventions
seem plausible given human nature, politics, economics? And then think
whether I want to act (vote / advocate / study) based on the possibility of
catastrophe even if it’s low probability.
------------------------------
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.extropy.org/pipermail/extropy-chat/attachments/20250917/e1477797/attachment-0001.htm>


More information about the extropy-chat mailing list