[ExI] fun outsider's view on ai
Jason Resch
jasonresch at gmail.com
Wed May 11 01:53:22 UTC 2016
I can strongly recommend this book:
http://www.amazon.com/Superintelligence-Dangers-Strategies-Nick-Bostrom/dp/0198739834/ref=sr_1_1?ie=UTF8&qid=1462931554&sr=8-1&keywords=super+intelligence
Jason
On Mon, May 9, 2016 at 3:46 PM, William Flynn Wallace <foozler83 at gmail.com>
wrote:
> I'm neither a Singularitarian nor an AItheist. I think human-level AI is
> inevitable, if President Trump doesn't manage to wipe out the human race
> first :-). But I don't buy the notion that super intelligence is akin to a
> superpower, and don't think it's necessary for an AI to have consciousness,
> human-like emotions, or the ability to set its own goals, and without those
> there is no need to fear them. dave
>
> If you want an AI to be superintelligent, why reference the neuron,
> Spike? Human brains are so fallible it's just silly. A person super
> intelligent about one thing is totally at a loss about many other things.
> I think brains must be still evolving, because as they are, they are
> cobbled together among available equipment and have functioned well enough
> to get us to the present. You don't have to be a psychologist to see the
> irrationality, the emotional involvement, the selfishness, of the output of
> human brains. There are many functions of brains that we can do well
> without entirely. Start with all the cognitive errors we already know
> about.
>
> OK, so what else can we do? Every decision we make is wrapped up in
> emotions. That alone does not make them wrong or irrational, but often they
> are. Take them out and see what we get. Of course they are already out of
> the AIs we have now. So here is the question: do we really want an AI to
> function like a human brain? I say no. We are looking for something
> better, right?
>
> Since by definition we are not yet posthumans, how would we even know that
> an AI decision was super intelligent? I don't know enough about computer
> simulations to criticize them, but sooner or later we have to put an AI
> decision to experimental tests in the real world not knowing what will
> happen.
>
> In any case, I don't think that there is any magic in the neuron. It's in
> the connections. And let's not forget about the role of glial cells, about
> which we are just barely aware. (see The Other Brain by Douglas Fields)
> Oh yeah, and the role of the gut microbiome - also just barely aware of
> its functions. Not even to mention all the endocrine glands and their
> impact on brain functions. Raising and lowering hormones has profound
> effects on functioning of the brain. Ditto food, sunspots (?), humidity
> and temperature, chemicals in the dust we breathe, pheromones, and drugs
> (I take over 20 pills of various sorts, Who or what could figure out the
> results of that?) All told, an incredible number of variables, some of
> which we may not know about at present, all interacting with one another,
> our learning, and our genes.
>
> All told, we are many decades away from a good grasp of the brain, maybe
> 100 years. A super smart AI will likely not function at all like a human
> brain. No reason it should. (boy am I going to get flak on this one)
>
> bill w
>
>
>
> On Mon, May 9, 2016 at 11:27 AM, Dave Sill <sparge at gmail.com> wrote:
>
>> On Mon, May 9, 2016 at 10:44 AM, spike <spike66 at att.net> wrote:
>>
>>>
>>>
>>> Nothing particularly profound or insightful in this AI article, but it
>>> is good clean fun:
>>>
>>>
>>>
>>>
>>> https://aeon.co/essays/true-ai-is-both-logically-possible-and-utterly-implausible?utm_source=Aeon+Newsletter&utm_campaign=6469cf0d50-Daily_Newsletter_9_May_20165_9_2016&utm_medium=email&utm_term=0_411a82e59d-6469cf0d50-68957125
>>>
>>
>> Yeah, not bad. Mostly on the mark, IMO, but he says a few things that are
>> just not rational.
>>
>> He reminds me a little of Roger Penrose’s take on the subject from a long
>>> time ago: he introduces two schools of thought, pokes fun at both while
>>> offering little or no evidence or support, then reveals he is pretty much a
>>> follower of one of the two: the Church of AI-theists.
>>>
>>
>> To be fair, he says both camps are wrong and the truth is probably
>> somewhere in between. And I agree.
>>
>>
>>> There are plenty of AI-theists, but nowhere have I ever seen a really
>>> good argument for why we can never simulate a neuron and a dendrite and
>>> synapses. Once we understand them well enough, we can write a sim of one.
>>> We already have sims of complicated systems, such as aircraft, nuclear
>>> plants and such. So why not a brain cell? And if so, why not two, and
>>> why not a connectome and why can we not simulate a brain? I have been
>>> pondering that question for over 2 decades and have still never found a
>>> good reason. That puts me in Floridi-dismissed Church of the
>>> Singularitarian.
>>>
>>
>> Yeah, his "True AI is not logically impossible, but it is utterly
>> implausible" doesn't seem to be based on reality.
>>
>> I'm neither a Singularitarian nor an AItheist. I think human-level AI is
>> inevitable, if President Trump doesn't manage to wipe out the human race
>> first :-). But I don't buy the notion that super intelligence is akin to a
>> superpower, and don't think it's necessary for an AI to have consciousness,
>> human-like emotions, or the ability to set its own goals, and without those
>> there is no need to fear them.
>>
>> -Dave
>>
>> _______________________________________________
>> extropy-chat mailing list
>> extropy-chat at lists.extropy.org
>> http://lists.extropy.org/mailman/listinfo.cgi/extropy-chat
>>
>>
>
> _______________________________________________
> extropy-chat mailing list
> extropy-chat at lists.extropy.org
> http://lists.extropy.org/mailman/listinfo.cgi/extropy-chat
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.extropy.org/pipermail/extropy-chat/attachments/20160510/be337d2d/attachment.html>
More information about the extropy-chat
mailing list