Feedback appreciated on this.<br><br>-Tyler<br><br>---------- Forwarded message ----------<br><span class="gmail_quote">From: <b class="gmail_sendername">The Singularity Institute Blog</b> <<a href="mailto:tyleremerson@gmail.com" target="_blank" onclick="return top.js.OpenExtLink(window,event,this)">
tyleremerson@gmail.com</a>><br>
Date: Sep 22, 2007 7:41 AM<br>Subject: The Singularity Institute Blog<br>To: <a href="mailto:tyleremerson@gmail.com" target="_blank" onclick="return top.js.OpenExtLink(window,event,this)">tyleremerson@gmail.com</a><br><br>
</span>
<div>
<div style="margin: 0pt 2em; font-family: Georgia,Helvetica,Arial,Sans-Serif; line-height: 140%; font-size: 13px; color: rgb(0, 0, 0);">
<table style="border: 0pt none ; margin: 0pt; padding: 0pt; width: 100%;">
<tbody><tr>
<td style="vertical-align: top;" width="99%">
<h1 style="margin: 0pt; padding-bottom: 6px;">
<a style="color: rgb(136, 136, 136); font-size: 22px; font-family: Arial,Helvetica,sans-serif; font-weight: normal; text-decoration: none;" href="http://www.singinst.org/blog" title="(http://www.singinst.org/blog)" target="_blank" onclick="return top.js.OpenExtLink(window,event,this)">
The Singularity Institute Blog</a>
</h1>
</td><td width="1%"><br></td>
</tr>
</tbody></table>
<table style="border-top: 1px solid rgb(153, 153, 153); clear: both; padding-top: 0.5em;">
<tbody><tr>
<td style="margin-bottom: 0pt; line-height: 1.4em;">
<p style="margin: 1em 0pt 3px; line-height: 115%;">
<a style="font-family: Arial,Helvetica,sans-serif; font-size: 18px;" href="http://www.singinst.org/blog/2007/09/21/summit-coverage-in-the-wall-street-journal-raises-questions/" target="_blank" onclick="return top.js.OpenExtLink(window,event,this)">
Summit Coverage in The Wall Street Journal raises questions</a>
</p>
<p style="margin: 9px 0pt 3px; color: rgb(85, 85, 85); font-family: Georgia,Helvetica,Arial,Sans-Serif; line-height: 140%; font-size: 13px;">
<span>Posted:</span> 21 Sep 2007 11:55 PM CDT</p>
<div style="margin: 0pt; font-family: Georgia,Helvetica,Arial,Sans-Serif; line-height: 140%; font-size: 13px; color: rgb(0, 0, 0);"><p>Earlier this week SIAI and the Singularity Summit got some major coverage in The Wall Street Journal. Lee Gomes, the Portals columnist for The Journal attended the Summit, and has some challenging thoughts about our movement and its perceived relevancy to the business community and the public at large.
</p>
<p>In his article, Gomes likens Singularitarians at times to 12-year-old sci-fi addicts, alien worshipers, and even gynephobics (don't tell my 3 daughters). While it is always fun to play "knock the nerds" in the popular press, I think Gomes raises key issues that point out why we sometimes struggle for credibility outside of our safety net in The Valley.
</p>
<p>As we start to organize our thoughts about next year's Singularity Summit, it is apparent that we need to focus more on bridging the knowledge and perception gaps between the scientific community, the business and investment community, and the public at large. Our success in crossing this chasm over the next couple of years will dictate how successfully the mission of the Singularity Institute will be embraced by broader segments of humanity.
</p>
<p>I'd like to open this discussion up to our community at large to get your ideas and feedback. How do we stay true to the vision of Singularity Institute, and at the same time create a partnership with the business community that creates an exciting and positive perspective on what we can accomplish? And how do we shake some of the more adverse associations to the lunatic fringe?
</p>
<p>I look forward to your thoughts. I've posted Lee's article below. Leave a comment to this post or contact me directly at <a href="mailto:lamis@singinst.org" target="_blank" onclick="return top.js.OpenExtLink(window,event,this)">
lamis@singinst.org</a>.</p>
<p><em>Reprinted from The Wall Street Journal</em></p>
<p>—————————————————————————</p>
<blockquote><p>The Singular Question Of Human vs. Machine Has a Spiritual Side</p>
<p>The Wall Street Journal<br>
PORTALS<br>
By LEE GOMES </p>
<p>September 19, 2007; Page B1</p>
<p>You can tell a lot about people from what they worry about. A few Saturdays ago, I spent the day in an auditorium full of fellow citizens concerned with "singularity." The word refers to the day when the intelligence of computers will exceed our own.
</p>
<p>The auditorium was filled with people who listed many things that might occur with singularity, such as a human-machine synthesis into a new, superintelligent life-form. The date has been projected as anytime from nine to 40 years hence.
</p>
<p>Singularity-believers say humanity urgently needs to begin preparing for this moment, if only to make sure that humans don't become kabobs at the first post-singularity tailgate party held by supersmart computers. There is even a Singularity Institute, bankrolled by Silicon Valley wealthoids.
</p>
<p>The weekend session featured speeches, panel discussions and informal chatting. About 800 people were on hand, more, frankly, than I would have expected. Who but 12-year-old sci-fi addicts still fret over malevolent, superintelligent machines? Most of us, living every day with computers, appreciate how even the world's most powerful one not only is incapable of an autonomous thought, it can't even distinguish spam from real email.
</p>
<p>To get to the singularity that we are supposed to be preparing for, we are going to need AGI, or Artificial General Intelligence, a topic the singularists go on about endlessly.</p>
<p>A computer with AGI thinks and reasons the same way a human being does, only much more quickly. But don't singularity people know that AI researchers have been trying to make such machines since the 1950s, without much success?
</p>
<p>It turns out, there is a schism between the AGI and the AI worlds. The AGI faction thinks AI researchers have sold out, abandoning their early dreams of "general" intelligence to concentrate on more attainable (and more lucrative) projects.
</p>
<p>They're right. The machines today that recognize speech or play chess are one-trick wonders. Of course, AI researchers defend that approach by saying their early dreams of general intelligence were naïve.</p>
<p>The singularists, though, don't seem bothered by those earlier AI failures; new approaches will bear fruit, they insist. They thus didn't think it a waste of either time or carbon offsets to be gathering at a conference to ask such questions as, "If you made a superintelligent robot, then forced it to work only for you, would that be slavery?"
</p>
<p>Robots are just computers with moving parts, of course, but the public is still confused about them, just like they used to be about computers themselves. The Great Metallic Hope of the robotics industry, for example, is currently a small, round vacuum cleaner that ambles across the floor by itself.
</p>
<p>A high-tech wonder? Actually, Consumer Reports said that even cheap vacuum cleaners did better than the first model. A little more of this, and no one will ever again worry about enslaving robots.</p>
<p>There is another way of thinking about the obsession with robots. John Huntington, professor of English, University of Illinois, has studied the genre and says sci-fi authors, especially the early ones who wrote about robots or aliens, were working out their own unacknowledged anxieties about closer-to-home topics.
</p>
<p>Most commonly, he said, these anxieties involved women, who were seen as becoming threatening as they gained social power. Racial and class tensions also were involved, he added.</p>
<p>I have a supplemental theory: that the discussion of singularity involves a sublimated spiritual yearning for some form of eternal life and an all-powerful being, but one articulated by way of technical, secular discourse.
</p>
<p>As it happens, there is considerable overlap between the singularity and the "life extension" communities. Ray Kurzweil, the best-known singularity writer, also co-wrote a lengthy guide to life extension. He once told me he expects literally to live forever — first by prolonging his life via a daily regimen that includes hundreds of pills and the nonstop consumption of green tea, then, once super-powerful computers arrive, by uploading his consciousness into one.
</p>
<p>Singularists also have an affinity for the Search for Extraterrestrial Intelligence, or SETI, program, which scans the skies looking for other civilizations. Isn't that a longing by some for an intergalactic messiah?
</p>
<p>Then, consider a poem read at the singularity conference that described an Aquarian Age scene in which humans and other mammals frolicked in a "cybernetic meadow … all watched over by machines of loving grace." Those computer protectors sound a lot like the guardian angels my grade-school nuns told us about.
</p>
<p>Years ago, a friend and I spent an evening with Arthur C. Clarke, the creator in "2001″ of HAL, the malevolent computer of every singularist's nightmare. He brought along slides, showing himself with some astronauts and with the authors of the musical "Hair."
</p>
<p>We talked about science and had our picture taken, which I still have. It proves that while I may have reached a different conclusion, at least I studied with the master.</p></blockquote>
</div>
</td>
</tr>
</tbody></table>
<table style="border-top: 1px solid rgb(153, 153, 153); padding-top: 4px; margin-top: 1.5em; width: 100%;">
<tbody><tr>
<td style="margin: 0pt 6px 1.2em 0pt; text-align: left; font-family: Helvetica,Arial,Sans-Serif; font-size: 11px; color: rgb(51, 51, 51);">You are subscribed to email updates from <a href="http://www.singinst.org/blog" target="_blank" onclick="return top.js.OpenExtLink(window,event,this)">
The Singularity Institute Blog</a>
<br>To stop receiving these emails, you may <a href="http://www.feedburner.com/fb/a/emailunsub?id=1830945&key=IPDZtRQeUq" target="_blank" onclick="return top.js.OpenExtLink(window,event,this)">unsubscribe now</a>.</td>
<td style="margin: 0pt 6px 1.2em 0pt; font-family: Helvetica,Arial,Sans-Serif; font-size: 11px; color: rgb(51, 51, 51); text-align: right; vertical-align: top;">Email Delivery powered by FeedBurner</td>
</tr>
<tr>
<td colspan="2" style="margin: 0pt 6px 1.2em 0pt; text-align: left; font-family: Helvetica,Arial,Sans-Serif; font-size: 11px; color: rgb(51, 51, 51);">Inbox too full? <a href="http://feeds.feedburner.com/siaiblog" target="_blank" onclick="return top.js.OpenExtLink(window,event,this)">
<img style="vertical-align: middle;" alt="(feed)"></a> <a href="http://feeds.feedburner.com/siaiblog" target="_blank" onclick="return top.js.OpenExtLink(window,event,this)">Subscribe</a> to the feed version of The Singularity Institute Blog in a feed reader.
</td>
</tr>
<tr>
<td style="margin: 0pt 6px 1.2em 0pt; text-align: left; font-family: Helvetica,Arial,Sans-Serif; font-size: 11px; color: rgb(51, 51, 51);" colspan="2">If you prefer to unsubscribe via postal mail, write to: The Singularity Institute Blog, c/o FeedBurner, 20 W Kinzie, 9th Floor, Chicago IL USA 60610
</td>
</tr>
</tbody></table>
</div>
</div>