[ExI] The symbol grounding problem in strong AI
    Stathis Papaioannou 
    stathisp at gmail.com
       
    Sun Dec 20 05:45:44 UTC 2009
    
    
  
2009/12/20 Gordon Swobe <gts_2000 at yahoo.com>:
> --- On Sat, 12/19/09, John Clark <jonkc at bellsouth.net> wrote:
>
>> A mechanical punch card reader from half a century ago knows that
>> particular hole is a symbol that means "put this card in the third
>> column from the left". How do we know this, because the
>> machine put the card in the third column from the left.
>
> In other words you want to think that if something causes X to do Y then we can assume X actually knows how to do Y.
>
> That idea entails panpsychism; the theory that everything has mind. As I mentioned to someone here a week or so ago, panpsychism does refute my position here that only brains have minds, and it does so coherently. But most people find panpsychism implausible if not outrageous.
Not everything has a mind, just information-processing things. Mind is
not a binary quality: even in biology there is a gradation between
bacteria and humans. The richer and more complex the information
processing, the richer and more complex the mind.
-- 
Stathis Papaioannou
    
    
More information about the extropy-chat
mailing list