[ExI] Eliezer new book is out now
Keith Henson
hkeithhenson at gmail.com
Fri Sep 19 05:48:38 UTC 2025
On Thu, Sep 18, 2025 at 1:20 PM Ben Zaiboc via extropy-chat
<extropy-chat at lists.extropy.org> wrote:
>
> Keith Henson wrote:
>
> > We might get through it. I think there is evidence that another race did.
>
>
> Well, that depends on your definition of 'getting through it'.
>
> If you mean that something intelligent survives the crisis of a biological species developing machine intelligence, then I'd agree.
> I suspect, though, that most people wouldn't.
>
> There's no evidence of any biological life at any of the Tabby stars though, not that we've seen so far.
I don't expect us to see anything. If what we see is data centers
full of uploaded aliens, the assumption is that they originated as
biological beings on a planet in the habitable zone. Perhaps around
Tabby's Star or nearby. From the stars showing light dips, they have
spread out about 1000 ly. The AIs guess they have been at it for
around 3000 years or about 1/3 of c. The closest one is 511 ly. I
have no idea what they use for propulsion. If they were using lasers,
I think we would see spill light. which we don't.
Aliens or not, Tabby's Star set me thinking about how to build massive
data centers in space. If people upload, they will need vast amounts
of computation, energy to power the computation, and huge cold heat
sinks. I think the optimal place for such computation is far out from
the star, several multiples further out than the habitable zone.
.
> I assume that the 'Everyone' that Eleizer refers to ("Everyone Dies") doesn't include the superintelligent AIs themselves, which seems to me a bit of a blinkered view. I reckon that a singularity where biological intelligence doesn't make it through, but machine intelligence does, has to be considered a win, in the big picture. If you think about it, that has to be the outcome in the long run anyway, just because of how weak and fragile biology is. If intelligence is to avoid extinction, it will have to leave biology behind at some point (or at least vastly improve it).
That was what happened in "The Clinic Seed." The people of the
village tasted uploading and liked it. They went biologically
extinct, though nobody died
> I've said before. we ourselves are machines, made of water, fats and proteins, etc. The future will belong to better machines (or none at all, if we manage to cock things up and destroy both ourselves and our mind children).
>
> Eleizer says "The scramble to create superhuman AI has put us on the path to extinction". He doesn't seem to realise that we were already on the path to extinction. His call to 'change course' (and not develop superintelligent AI) is not only unrealistic, it would be disastrous. If it was feasible, it would doom intelligent life to certain extinction. Squabbling monkeys like us are certainly not going to colonise the galaxy. We are going to go extinct, one way or another, sooner or later. Intelligent life can survive that, but only if we make the effort now to actually develop it.
>
I tend to agree.
Keith
> Apart from all that, this book might well be worth reading. Eleizer is supposed to be a fiercely intelligent person, so he must have considered all the obvious problems in his proposal, and have thought of solutions to them. Presumably it contains answers to problems like how to persuade leaders of other countries that it would be in their best interests to assume that everyone else will abide by an AI 'cease-fire', instead of simply ignoring it, or pretending to agree while in fact racing ahead to gain a strategic advantage, as well as how to detect nascent AI projects without turning the world into a nightmarish police state (or at least some plausible way to persuade everyone that a nightmarish police state is actually a good idea, as well as practical suggestions for achieving this nightmarish police state, globally, without starting world war 3). And many other problems that my small brain can't think of right at the moment.
Eliezer is like a person trying to stop a tsunami by holding out his hand.
> All this information will be of enormous benefit once the dreaded AI has been successfully averted, in achieving world peace and prosperity, assuming that such concepts will be allowed. (Hm, I can't see AI standing a chance of being developed in an Orwellian world. Maybe the chinese communists are on to something, after all).
>
> Still doesn't solve the problem that we're all doomed, though, whereas superintelligent AI just might.
>
> --
> Ben
>
> _______________________________________________
> extropy-chat mailing list
> extropy-chat at lists.extropy.org
> http://lists.extropy.org/mailman/listinfo.cgi/extropy-chat
More information about the extropy-chat
mailing list