[ExI] How could you ever support an AGI?
ABlainey at aol.com
ABlainey at aol.com
Wed Mar 5 03:19:02 UTC 2008
In a message dated 05/03/2008 00:21:57 GMT Standard Time,
santostasigio at yahoo.com writes:
> Well,
> about the anthropomorphisizing of AGI, you say in the end that some say the
> motivation of AGI will be the one we program into it. Exactly, that is my
> point, it is difficult for us to create an intelligence utterly alien when the
> only example of intelligence we have is us.
I can't help but notice that many of the posts have started out with logic
and concluded with quazi-anthropomorphic, straw man arguments.
I understand that an AGI will or should be based upon 'human intelligence,'
however the end result will be completely Alien to us. So much so that our
interpretation of intelligence wouldn't really fit.
> But maybe there are general and universal principles associated with
> intelligence.
> Intelligence means finding patterns and connections, understanding that
> affecting this part here means affecting this other part over there, intelligence
> means having a higher sense of physical and moral "ecology".
Again this is reduced to anthropomorphic intelligence. The AGI will have
logic based 'cold' intelligence. From this it will probably and rightly deduce
that morality is a human construct which serves the needs of human civilisation.
A civilisation which it is not a part. Expecting It to adhere to these moral
codes would be akin to you or I adhering to the moral codes of Ants. So If
someone comes on your property, bite their head off.
> If you see connections between all the beings than you feel compassion and
> understanding (and yes these are human feelings, but they are also
> fundamental components of our intelligence, and a lot of new research shows that
> without feelings we would no have a conscious intelligence at all).
My point. We would like to think that we can reduce ourselves to simple data
constructs which mirror our original wetware physical structure. Expecting
that this 'uploaded' us would run in the same manner that we do today. How do we
code for that groggy morning feeling? or the rush of excitement associated
with anticipation of something good? All the things which truly make us who we
are, the things which have driven us and made us take the unique forks in our
lives.
These are what give us the basis for our 'Intelligence' our logic, our
rationalisation. It is what makes us human.
The uploaded us and the AGI will have none of this, so will not make
intelligent decisions the way we do. that is what I mean by 'Cold' intelligence. It is
devoid of chemical input. Show me a line of code for Happy, Sad, Remorse.
At most we can hope for some minor 'don't do this because it's bad' type of
rules in its main code. But if we have given it the ability to change it's
code, what is to stop it overwriting these rules based upon some logical
conclusion that it comes to?
If we hard wire the rules, what is to stop it creating its own 'offspring'
without these rules? Whatever we do, it will have the logic to undo and far
faster than we can counter any mistakes or oversights.
> Yes we exterminate bugs, but usually in limited situations (like in our
> house or on a crop). It would be unacceptable for mankind to have a global plan
> to complete exterminate all the roaches of the earth even if it could be
> done.
> And it is difficult to have feelings for bug, it would not make sense
> ecologically, it would not be the intelligent thing to do, and by defintion AGI is
> supposed to be Intelligent.
>
Again anthropomorphically intelligent. It may well be the cold inteligent
decission to pre-emptively exterminate a potential threat. After all, it wouldn't
feel bad about it, it wouldn't feel anything.
Alex
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.extropy.org/pipermail/extropy-chat/attachments/20080304/b8645063/attachment.html>
More information about the extropy-chat
mailing list