[ExI] AI and Eliezer
Dylan Distasio
interzone at gmail.com
Wed Mar 20 16:00:27 UTC 2024
You strike to the quick of it! He's willing to bomb non-compliant data
centers out of existence. After his Time interview, I can't take him
seriously any longer.
Climate change is another great example, and I would add a third playing
out in real time now in the US (and on this list), some have an irrational
fear that Trump is an infinite threat and are pretty much willing to do
ANYTHING to prevent his re-election.
On Wed, Mar 20, 2024 at 5:11 AM efc--- via extropy-chat <
extropy-chat at lists.extropy.org> wrote:
>
>
> On Wed, 20 Mar 2024, Samantha via extropy-chat wrote:
>
> > Oh my. Eliezer the luddite. I have been off the opinion for over two
> > decades that there is no way for humanity to continue at the current
> level,
> > much less transcend this level, without massively more effective
> intelligence
> > deployed than today whether human or artificial.
> >
> > If this is so then AGI is *essential* to human flourishing.
> >
> > Also "shut it down" means shutting down and controlling the entire
> internet
> > at this point. The work has gone well outside a few large tech
> companies.
> > It is vibrantly alive and being advanced in the open source world.
> Surely
> > Eliezer doesn't believe humanity could survive the level of tyranny it
> would
> > actually take to "shut it down"?
>
> The problem with Eliezer is that he deals with infinite threats. If you
> deal with infinite threats every action is excusable and preferable in
> order to avoid infinite evil.
>
> Just look at the more extreme climate change activists, they deal with
> infinite threats (everyone will die tomorrow) and therefore anything is
> allowed.
>
> Another classic is Pascals wager. Assign infinite good and infinite bad,
> and all calculation, nuance and rationality goes out the window.
>
> Best regards,
> Daniel
>
>
> > - samantha
> >
> > On 3/13/24 17:31, Keith Henson via extropy-chat wrote:
> >> I have been perusing X recently to see what Anders Sandberg and
> >> Eliezer have been saying
> >>
> >> Eliezer Yudkowsky
> >> @ESYudkowsky
> >> ·
> >> 4h
> >> I don't feel like I know. Does anyone on the ground in DC feel like
> >> making a case? To be clear, the correct action is "Shut it down
> >> worldwide" and other actions don't matter; I'm not interested in who
> >> tries to regulate it more strictly or has nicer goals.
> >>
> >> ·
> >> 16m
> >> Replying to
> >> @ESYudkowsky
> >> Given the economic, political, and international realities, do you
> >> think "shut it down" is possible? If not, are there any other options?
> >> You might remember a story I posted on sl4 where a very well aligned
> >> AI caused humans to go extinct (though nobody died).
> >>
> >>
> >> Keith
> >>
> >> _______________________________________________
> >> extropy-chat mailing list
> >> extropy-chat at lists.extropy.org
> >> http://lists.extropy.org/mailman/listinfo.cgi/extropy-chat
> >_______________________________________________
> extropy-chat mailing list
> extropy-chat at lists.extropy.org
> http://lists.extropy.org/mailman/listinfo.cgi/extropy-chat
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.extropy.org/pipermail/extropy-chat/attachments/20240320/ab2c0e21/attachment.htm>
More information about the extropy-chat
mailing list