[ExI] Empathic AGI

Samantha Atkins sjatkins at mac.com
Fri Feb 11 18:38:16 UTC 2011

On 02/09/2011 09:41 AM, John Clark wrote:
> On Feb 7, 2011, at 12:16 PM, Stefano Vaj wrote:
>> If we accept that "normal" human-level empathy (that is, a mere
>> ingredient in the evolutionary strategies) is enough, we just have to
>> emulate a Darwinian machine as similar as possible
> Two difficulties with that:
> 1) The Darwinian process is more like history than mathematics, it is 
> not repeatable, very small changes in initial conditions could lead to 
> huge differences in output.
> 2) Human-level empathy is aimed at Human-level beings, the further 
> from that level the less empathy we have. We have less empathy for a 
> cow than a person and less for an insect than a cow. As the AI's 
> intelligence gets larger its empathy for us will get smaller although 
> its empathy for its own kind might be enormous.

Yes, we understand how interdependent peer level beings will naturally 
develop a set of ethical guides for how they treat one another and the 
ability to model one another.  We don't have much/any idea of how this 
would arise among beings of radically different natures and abilities 
there are not so interdependent regarding their treatment of one another.

- samantha

More information about the extropy-chat mailing list