[ExI] Elon Musk, Emad Mostaque, and other AI leaders sign open letter to 'Pause Giant AI Experiments'
    spike at rainier66.com 
    spike at rainier66.com
       
    Thu Mar 30 05:46:19 UTC 2023
    
    
  
 
 
From: extropy-chat <extropy-chat-bounces at lists.extropy.org> On Behalf Of Will Steinberg via extropy-chat
 
 
>….  We almost killed ourselves with nukes in the middle of the last century… Will
 
Doesn’t it seem like we should be able to retire an old existential risk when a new one shows up?  It feels to me like we are in as much danger of old-fashioned nuclear holocaust now as we have ever been in the past.  But now two new existential risks pop up: man-made weaponized viruses and ambiguously human level AI.  But we don’t get to retire the nuke risk.
 
Sheesh.  I liked it better when we only had one serious existential risk.
 
spike
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.extropy.org/pipermail/extropy-chat/attachments/20230329/8459876c/attachment.htm>
    
    
More information about the extropy-chat
mailing list