[ExI] Future of Humanity Institute at Oxford University £1 million grant for AI

Anders Sandberg anders at aleph.se
Sun Jul 5 22:11:26 UTC 2015


Från:   Giulio Prisco <giulio at gmail.com> 

Anders, the paperclip maximizer is a good example indeed. For all we
know, it could reason that humans are the only users of paperclips in
the known universe, so that humans give a cosmic meaning to paper clip
production, so that humans must be revered and protected... My point
is that we don't know - just like Columbus didn't know what he would
find at the end of his journey.
But that is not my example. That would be a paperclip-meaning maximizer or something. I was talking about something that evaluates value as being proportional to the number of paperclips (defined according to some internal paperclip model). Humans using paperclips is negative by that standard (they occasionally break them, and by not being paperclips they represent untapped resources for more paperclips). 

It is this kind of slip of meaning that makes a lot of verbal AI arguments mistaken. Think of this as software code maximizing something, not as a human-like mind that cares and reveres. Yes, the *right* utility function might indeed producing caring behavior and maybe also revering mental states. But *most* utility functions that maximize the number of paperclips do not. Indeed, trying to write a caring utility function is way harder than a mere maximizer, so unless the programmer was strongly motivated to try from the start the most likely outcome is one of the other utility functions. 

Writing your ethical system as computer executable code (or code that reliably learns it) is a nontrivial challenge. Writing an explicit ethical system that doesn't have crazy side effects that render it contradictory to your intentions is even harder. 

Saying it is OK that we don't know what will happen is a bit like pouring out chemicals at random from your lab into the water supply: maybe they will make a cure for cancer. But very probably it is not a wise move.


Anders Sandberg, Future of Humanity Institute Philosophy Faculty of Oxford University
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.extropy.org/pipermail/extropy-chat/attachments/20150706/ec7c5e44/attachment.html>


More information about the extropy-chat mailing list