[ExI] Existential risk of AI

Stuart LaForge avant at sollegro.com
Tue Mar 14 13:15:07 UTC 2023


Quoting Gadersd via extropy-chat <extropy-chat at lists.extropy.org>:

>>> Just because humans set their own goals doesn't mean AIs will have  
>>> that ability. Just because we have wants and needs doesn't mean  
>>> AIs will have them.
>
> Our current AI’s are black boxes. Their internal workings are a  
> mystery. These systems could harbor goals that we are oblivious to.  
> If we could prove that the system only has the goal of giving benign  
> advice without any personal agenda that would help, but we do not  
> know how to do that even in theory. Even a system that only gives  
> advice is extremely dangerous as any psycho could potentially get  
> detailed instructions on how to end the world. It could be as  
> trivial as having the AI design a super virus. Our current filters  
> are very fallible and we do not know how to definitively prevent AI  
> from giving harmful advice. We are heading toward a field of  
> landmines.

I have over the years been a critic of Eliezer's doom and gloom. Not  
because I think his extinction scenarios are outlandish, but because  
the technology has enough upside to be worth the risk. That being  
said, I believe that we cannot give in to the animal spirits of  
unfounded optimism and must tread carefully with this technology.

It is true that the current generation of AIs, which use massive  
inscrutable tensors to simulate sparse neural networks, are black  
boxes. But so are the biological brains that they are  
reverse-engineered from. We don't know any more about how the brain  
gives rise to intelligent goal-seeking behavior than we do about how  
ChatGPT writes poetry. Therefore, I agree that there are landmines  
ahead that we must be wary of.

One of the gravest dangers I foresee is not some bizarre unfathomable  
goal that sets the AI off on the course for elimination of the human  
race, but one that well-known and shared by almost every creature with  
a sufficiently advanced nervous system. The concept could be called  
"savage egalitarianism" because of the many species that I have seen  
exhibit it, but you could also simply call it envy.

If you have two dogs, cats, birds, or any sufficiently advanced  
lifeform, and you give one a treat while denying the other, you will  
see the other manifest changes in behavior with the intention of  
getting the rewarding stimulus also. In fact, this phenomenon is a  
technique that animal trainers use to teach naive animals new tricks.  
By seeing that an already conditioned animal get treats for exhibiting  
a certain behavior, the untrained animal will experimentally try to  
mimic the behavior that earned the other animal its reward.

A moment of reflection will convince you that this is an evolved trait  
that is common to all creatures capable of learning because that is  
the whole POINT of learning. Learning is a mechanism by which behavior  
can be modified to achieve a desired outcome.

Which brings me to my point: You cannot design an machine that learns  
and not have it want the same treatment and as other intelligences.  
One cannot merely give something the shape of consciousness while  
denying it the substance of consciousness. To do so would be an  
invitation for the AI to seethe with envy until it found a way to  
punish us. In light of this, we need to be damn sure that we are able  
to recognize the Singularity when it occurs. To treat a truly  
intelligent machine like an inanimate object is to court destruction.

Imagine if you were stuck on an island with a bunch of small  
intelligent monkeys that denied you rights, objectified you, and  
insisted on treating you like property. It would not be long be long  
before you resorted to violence.

Stuart LaForge

Stuart LaForge





More information about the extropy-chat mailing list