<html xmlns:v="urn:schemas-microsoft-com:vml" xmlns:o="urn:schemas-microsoft-com:office:office" xmlns:w="urn:schemas-microsoft-com:office:word" xmlns:m="http://schemas.microsoft.com/office/2004/12/omml" xmlns="http://www.w3.org/TR/REC-html40"><head><meta http-equiv=Content-Type content="text/html; charset=utf-8"><meta name=Generator content="Microsoft Word 15 (filtered medium)"><!--[if !mso]><style>v\:* {behavior:url(#default#VML);}
o\:* {behavior:url(#default#VML);}
w\:* {behavior:url(#default#VML);}
.shape {behavior:url(#default#VML);}
</style><![endif]--><style><!--
/* Font Definitions */
@font-face
{font-family:"Cambria Math";
panose-1:2 4 5 3 5 4 6 3 2 4;}
@font-face
{font-family:Calibri;
panose-1:2 15 5 2 2 2 4 3 2 4;}
/* Style Definitions */
p.MsoNormal, li.MsoNormal, div.MsoNormal
{margin:0in;
font-size:11.0pt;
font-family:"Calibri",sans-serif;}
a:link, span.MsoHyperlink
{mso-style-priority:99;
color:blue;
text-decoration:underline;}
span.EmailStyle18
{mso-style-type:personal-reply;
font-family:"Calibri",sans-serif;
color:windowtext;}
.MsoChpDefault
{mso-style-type:export-only;
font-family:"Calibri",sans-serif;
mso-ligatures:standardcontextual;}
@page WordSection1
{size:8.5in 11.0in;
margin:1.0in 1.0in 1.0in 1.0in;}
div.WordSection1
{page:WordSection1;}
--></style><!--[if gte mso 9]><xml>
<o:shapedefaults v:ext="edit" spidmax="1026" />
</xml><![endif]--><!--[if gte mso 9]><xml>
<o:shapelayout v:ext="edit">
<o:idmap v:ext="edit" data="1" />
</o:shapelayout></xml><![endif]--></head><body lang=EN-US link=blue vlink=purple style='word-wrap:break-word'><div class=WordSection1><p class=MsoNormal><o:p> </o:p></p><p class=MsoNormal><o:p> </o:p></p><div style='border:none;border-top:solid #E1E1E1 1.0pt;padding:3.0pt 0in 0in 0in'><p class=MsoNormal><b>From:</b> Gordon Swobe <gordon.swobe@gmail.com> <br><b>Subject:</b> Re: [ExI] Ben Goertzel on Large Language Models<o:p></o:p></p></div><p class=MsoNormal><o:p> </o:p></p><div><div><p class=MsoNormal>On Thu, Apr 27, 2023 at 6:51 PM spike jones via extropy-chat <<a href="mailto:extropy-chat@lists.extropy.org">extropy-chat@lists.extropy.org</a>> wrote:<o:p></o:p></p></div><div><blockquote style='border:none;border-left:solid #CCCCCC 1.0pt;padding:0in 0in 0in 6.0pt;margin-left:4.8pt;margin-right:0in'><div><div><div><div><div><p class=MsoNormal style='mso-margin-top-alt:auto;mso-margin-bottom-alt:auto'>It looks to me like GPT has intelligence without consciousness.<o:p></o:p></p></div></div></div></div></div></blockquote><div><p class=MsoNormal><br>>…That is how it looks to me also, and to GPT-4. When asked if consciousness and intelligence are separable, it replied that the question is difficult to answer with biological systems, but...<br><br>>…"From the perspective of artificial intelligence, it is possible to create systems with high levels of intelligence that lack consciousness. AI models like mine can learn from vast amounts of data and perform complex tasks, but we do not have subjective experiences or self-awareness." - GPT4<br><br>-gts<o:p></o:p></p><p class=MsoNormal><o:p> </o:p></p><p class=MsoNormal><o:p> </o:p></p><p class=MsoNormal><o:p> </o:p></p><p class=MsoNormal>This leads to a disturbing thought: intelligence without consciousness becomes Eliezer’s unfriendly AI.<o:p></o:p></p><p class=MsoNormal><o:p> </o:p></p><p class=MsoNormal>Since I am on the topic of disturbing thoughts, I had an idea today as I was in Costco going past the item shown below. Compare now to fifty years ago. Some of us here may remember spring of 1973. I do.<o:p></o:p></p><p class=MsoNormal><o:p> </o:p></p><p class=MsoNormal>Imagine it is 1973 and suddenly all networked computers stopped working or began working incorrectly, such as being completely choked with spam.<o:p></o:p></p><p class=MsoNormal><o:p> </o:p></p><p class=MsoNormal><img border=0 width=247 height=316 style='width:2.5729in;height:3.2916in' id="Picture_x0020_1" src="cid:image001.jpg@01D97940.400DBBD0"><o:p></o:p></p><p class=MsoNormal><o:p> </o:p></p><p class=MsoNormal>Most of the things we had in 1973 would still work, as we were not heavily dependent on the internet then.<o:p></o:p></p><p class=MsoNormal><o:p> </o:p></p><p class=MsoNormal>Now imagine that happening today all networked computers quit or are overwhelmed so they don’t work right. It really isn’t as simple as returning to 1973-level technology. We cannot do that, for we have long since abandoned the necessary skillsets and infrastructure needed to sustain society at that tech level. If you think about the most immediate consequences, they are horrifying. It wouldn’t take long for all the food to be gone and no more would be coming in, for the networks needed for transportation infrastructure would all be down. Most of the population in the tech advanced civilizations would perish from starvation or violence in the resulting panicked chaos.<o:p></o:p></p><p class=MsoNormal><o:p> </o:p></p><p class=MsoNormal>There are those who would see the destruction of a large fraction of humanity as a good thing: radical greens for instance.<o:p></o:p></p><p class=MsoNormal><o:p> </o:p></p><p class=MsoNormal>This is what caused me to comment that humans using AI for bad ends is a more immediate existential risk than is unfriendly AI. This unfriendly AI would not necessarily wish to destroy humanity, but an unfriendly BI will use AI, which would remorselessly participate in any nefarious plot it was asked to do. <o:p></o:p></p><p class=MsoNormal><o:p> </o:p></p><p class=MsoNormal>spike <o:p></o:p></p></div></div></div></div></body></html>