[Paleopsych] Reason: Are We Just Really Smart Robots?
Premise Checker
checker at panix.com
Mon Apr 11 22:06:09 UTC 2005
Are We Just Really Smart Robots?: Two books on the mind put the
human back into human beings.
http://www.reason.com/0504/cr.ks.are.shtml
5.4
by [6]Kenneth Silber
[7]On Intelligence, by Jeff Hawkins with Sandra Blakeslee, New York:
Times Books, 261 pages, $25
[8]Mind: A Brief Introduction, by John R. Searle, New York: Oxford
University Press, 326 pages, $26
Neurobiologys advances generate anxiety as well as joy and hope. On
the joyful and hopeful side, there are the prospect and reality of
improved treatments for brain diseases and debilities. But anxiety
arises over what the science tells us, or will tell us, about
ourselves. Thoughts and feelings may be reduced to brain structures
and processes. Consciousness and free will may be proven unimportant
or illusory. Much of what we value about ourselves, in short, may be
explainedor, worse, explained away.
The prevailing trends in the philosophy of mind reinforce such
concerns. The field is dominated by schools of materialism that
describe mental phenomena as types or side products of physical
phenomena. Mind-body dualism, which posits a separate existence for
the mind, has been effectively eclipsed (although it seems to receive
continued implicit acceptance from many nonexperts). Some forms of
materialism argue that the mental phenomena in question do not even
exist.
This turn toward the mechanistic could have baleful cultural and
political consequences. It threatens to undermine peoples sense of
responsibility and self-worth. There is the danger of what philosopher
Daniel Dennett calls creeping exculpation, as more and more human
behavior is attributed to material causes. Criminal violence, for
example, might be excused as a consequence of low levels of serotonin
or monoamine oxidase in the brain. Many philosophers, including
Dennett, argue that humans should be regarded as responsible agents
even if human behavior is fully determined. But the very fact that
such arguments need to be made shows how the deterministic premise has
altered the terms of debate.
If humans are mechanistic beings, it becomes harder to understand why
they should not be used as means to an end or why there should be much
concern with what they are thinking or feeling. At a political level,
such quandaries pose a threat to liberal democracy, which relies
heavily on the assumption that we are autonomous beings with the
capacity to make meaningful decisions. Mechanistic theories have
enjoyed an authoritarian cachet in the past. Stalins regime embraced
the work of Ivan Pavlov, famous for conditioning dogs to salivate at
the ringing of a bell. In Walden Two (1948), the American psychologist
B.F. Skinner described a society whose managers use operant
conditioning to suppress competitiveness and other undesired
behaviors.
Alongside the conception of human beings as biological machines looms
another specter: that human mental capacities will be equaled or
exceeded by machines of our own creation. An influential doctrine in
the philosophy of mind, congruent not only with neurobiology but with
cognitive psychology and computer science, is computer functionalism.
This view holds that the mind is fundamentally a computer program
implemented in the brains hardware one which could be replicated in a
different physical substrate. Notwithstanding the limited progress of
artificial intelligence (A.I.), many experts expect it to achieve vast
advances in coming decades. More important, the general public expects
this too. The prospect arouses considerable anxiety, as reflected in
the Terminators and Matrixes that populate science fiction.
The scientific and philosophical quest to understand human beings as
part of the natural world thus seems to come with a hefty price. It
forces us to regard ourselves as mere machinesindeed, as potentially
obsolescent machines, given advances in computing. Or does it?
Technologist Jeff Hawkins and philosopher John Searle both approach
matters of mind and brain from a naturalistic perspective, but their
arguments veer sharply from the grim picture sketched above. Both
provide valuable analysis and speculation about mental phenomena while
taking issue with much current scientific and philosophical thinking
about the subject.
In On Intelligence, Hawkins portrays human intelligence as more subtle
and flexible than anything computers do. His model suggests that while
future artificial systems may possess remarkable intelligence, they
will be neither human-like nor the malevolent superhuman entities of
science fiction. In Mind: A Brief Introduction, Searle provides an
iconoclastic overview of the philosophy of mind, arguing for a
position that accepts that the mind is materially based without
dismissing or downplaying mental phenomena. Searles discussion ranges
across such topics as the limitations of computers, the nature of the
unconscious, and free will as a possible feature of the brain.
Hawkins, who wrote On Intelligence with science journalist Sandra
Blakeslee, is a computer entrepreneur with a longstanding interest in
how the brain works. He is the inventor of the original Palm Pilot and
the founder of the Redwood Neuroscience Institute. In 1980, as an
Intel employee, he proposed a project to develop memory chips that
operate on brain-like principles. Intels chief scientist turned him
down, reasoning (correctly, Hawkins now believes) that such an effort
was premature. Hawkins then sought to do graduate work at the
Massachusetts Institute of Technology, to study brains as a means
toward developing intelligent machines. MIT, suffused with the idea
that A.I. had little need for brain research, rejected his
application. In the late 1980s, Hawkins viewed with interest but
growing skepticism the rise of neural networks, programs that bore a
resemblancebut only a very loose oneto brain operations. He found no
use for neural networks in developing the handwriting recognition
system later used in Palm Pilots.
Such experiences fed Hawkins convictions that intelligent machines
must be more genuinely brain-like, and that making them so requires a
new theory of how the brain operates. Neurobiology, he argues, has
amassed an impressive array of detail but lacks a compelling framework
for understanding intelligence and brain function. On Intelligence is
an attempt to provide such a framework.
Hawkins focuses mainly on the cortex, the most evolutionarily recent
part of the brain. The cortex, in his view, uses memory rather than
computation to solve problems. Consider the problem of catching a
ball. A robotic arm might be programmed for this task, but achieving
it is extremely difficult and involves reams of calculations. The
brain, by contrast, draws upon stored memories of how to catch a ball,
modifying those memories to suit the particular conditions each time a
ball is thrown.
The cortex also uses memories to make predictions. It is engaged in
constant, mostly unconscious prediction about everything we observe.
When something happens that varies from predictionif you detect an
unusual motion, say, or an odd textureit is passed up to a higher
level in the cortexs hierarchy of neurons. The new memories are then
parlayed into further predictions. Prediction, in Hawkins telling, is
the sine qua non of intelligence. To understand something is to be
able to make predictions about it.
A key concept in this memory-prediction model is that of invariant
representations. The cortex is presented with a flux of sensory data
but manages to perceive objects as stable. The magazine youre now
holding, or the computer screen youre looking at, sends constantly
changing inputs to your eye and optic nerve, but the subsequent
pattern of neurons firing in your visual cortex displays an underlying
stability. This capacity to pick out unchanging relationships gives
humans considerable cognitive flexibility. Imagine looking at a
picture of a face formed by dots (like those drawings in The Wall
Street Journal). Now imagine each dot is moved a few pixels to the
left. A human, unlike a conventional A.I. program or neural net,
easily will see it as the same face.
Hawkins buttresses his memory-prediction model with a fair amount of
neurobiological detail. Much of the model is speculative. There is,
for instance, considerable evidence of invariant representations in
the workings of the visual cortex, but it is not yet clear whether the
concept applies broadly to other sensory areas and to motor regions of
the cortex. Hawkins presents a list of neurobiological predictions to
test his models validity. He posits, for example, that certain layers
of the cortex contain neurons that become activated in anticipation of
a sensory input. Such anticipatory activity is in keeping with the
idea that perception involves prediction, as well as receipt, of
sensory inputs. When you glance around your living room, your brain
fills in some details based on what it has seen before.
As Hawkins notes, invariant representations can be viewed as a bug, as
well as a feature, in human cognition; negative stereotyping and
bigotry might have roots in such invariance. The strong element of
prediction involved in perception also has a downside: It could
underlie peoples tendency to see what they want to see. Overall,
though, Hawkins model underscores the considerable capabilities of
human intelligence. It provides a plausible explanation of how the
speed and agility of human thought can exceed the capacities of
computers, even though the latter have components that operate far
faster than neurons.
The model may also offer insight into creativity, which arguably
arises from the brains propensity to make predictions. In Hawkins
view, there is a continuum between everyday actions and perceptions
and the production of great novels or symphonies. The cortex during
normal waking moments combines its invariant memories with the details
of what is happening now; it is constantly predicting things that are
similar to, but at least slightly different from, what it has
experienced in the past. Our brains are geared to come up with
something new.
Hawkins ventures that memory and prediction will be crucial to an
understanding of consciousness, but he acknowledges that his model
does not probe deeply into how and why consciousness exists. He draws
a link between consciousness and memory through a thought experiment:
If your memories of yesterdays activities were erased, so would be
your sense that your behavior had been conscious. He speculates as to
why vision, hearing, and other senses are (normally) experienced as
qualitatively distinct, even though their inputs are all converted
into patterns in the cortex. The answer, he suggests, might involve
the diverse connections between the cortex and other parts of the
brain.
In his final chapter, Hawkins writes enthusiastically about the
prospects for intelligent machines. He expects rapid progress in the
development of brain-like systems in the next several decades, citing
speech recognition, vision, and smart cars as promising near-term
applications. He imagines super-intelligent systems that will predict
the weather, foresee political unrest, and understand
higher-dimensional spaces. Yet he emphasizes that intelligent machines
will not be similar to us. They will have something like a cortex and
senses, but not human-like bodies, emotions, or experiencesthings it
would be very difficult, and generally pointless, to give them. They
will not strive for power, wealth, status, or pleasure. They will not
be angry at being enslaved.
To illustrate the error of likening machines to human beings (and vice
versa), Hawkins draws on a well-known thought experiment: A man who
understands no Chinese is placed in a room with a wall slot through
which he receives questions written in Chinese. Following a rule book,
he replies to the questions with other Chinese symbols. To an outside
observer, he seems to understand Chinese. But in fact, he has no idea
what the questions or answers are about.
For Hawkins, the story of the Chinese Room points to limitations of
conventional A.I. and of the Turing Test, the standard that a computer
is intelligent if a human inquirer cannot distinguish its replies from
a persons. Hawkins adds, however, that the Chinese Room would display
intelligence if it contained a memory system that could make
predictions about the content of the Chinese messages passed through
the slot. This is an interesting wrinkle but a debatable point. One
can imagine the man in the room adeptly foreseeing which symbols will
follow which others but still not knowing what they mean.
The man who first asked us to imagine the Chinese Room was John
Searle, a Berkeley philosopher who has written influentially about
mind, language, and other subjects. His point was that a computer
manipulates symbols but attaches no meaning to them; it understands
nothing. Searle revisits the Chinese Room in Mind: A Brief
Introduction. He rebuts the common counterargument that it is the
overall systemman, room, rule bookthat understands Chinese. The point
is the same, he contends, even if the man is in an open field and has
memorized the rule book. Indeed, Searle believes his original argument
did not go far enough in debunking computer intelligence; something is
a computer, he elaborates, only if an intelligent observer interprets
it as such.
At the core of Mind: A Brief Introduction is Searles effort to situate
mental activity in the physical world. Consciousness, he argues, is a
biological phenomenon; it is a process of the brain, much as digestion
is a process of the stomach. He emphasizes, however, that
consciousness cannot be dismissed as an illusion or defined in terms
of lower-level neurobiological processes. Conscious states exist
insofar as someone experiences themthey have a first-person
ontologyand in this regard they are distinct from physical phenomena
that have a third-person ontology. The pain of banging into a coffee
table (unlike the table itself) is real only because you feel it.
Searle terms his position biological naturalism and contrasts it with
the conventional categories of materialism and dualism.
Searles picture leaves open the possibility of free will, defined here
in contradistinction to determinism. In this view, quantum mechanical
indeterminism at the micro-level may produce free will as a
higher-level feature of the brain. In making decisions, the brain
would draw upon the unpredictable behavior of its constituent
particles. But wouldnt such freedom consist of mere randomness? Searle
argues that this objection involves a fallacy of composition,
confusing the properties of a system with those of its parts. Our
pervasive experience of free will, he acknowledges, may be an
illusion. But if so, it is a strange illusion, one that requires vast
biological resources to maintain yet somehow survived evolutions
travails.
Searle ranges broadly across the subject of mental phenomena, poking
holes in much received philosophical and scientific wisdom. A key
feature of conscious experience, he notes, is its unified structure;
one normally encounters sights, sounds, and so on as part of ones
overall environment. Neurobiology, he ventures, will ultimately
benefit more from a unified-field approach to consciousness than from
the currently favored building-block emphasis. Searle also takes issue
with philosophical arguments that humans perceive not the real world
but merely sense data. Such claims, he contends, rely on slippery
language and dubious assumptions.
The concept of the unconscious, Searle argues, is indispensable for
explaining some forms of human behavior, but it is sometimes pushed
beyond its applicability. Unconscious mental states, in his telling,
are states that could in principle become conscious. It is possible,
for instance, to believe that George W. Bush is president even when
you are sound asleep. In Searles view, however, cognitive scientists
are incorrect to say, for example, that people see by performing
unconscious computations on visual stimuli. The brain processes
involved, much like the workings of the liver, are not the sort of
thing that could be conscious; hence they are nonconscious rather than
unconscious.
Searle closes with a discussion of the elusive concept of the self. A
longstanding philosophical tradition, initiated by David Hume, regards
the self as a bundle of perceptions; we have a series of experiences
but not an inner essence. Searle argues, to the contrary, that
consciousness, a capacity to initiate action, and an ability to act on
the basis of reasons do amount to a selfa non-Humean self that is more
than just a set of experiences. Having such a self provides continuity
between ones past, present, and future; it is what enables a person to
take responsibility and make plans.
Mind: A Brief Introduction and On Intelligence are thought-provoking
and, no less important, anxiety-reducing. By dispelling overstated
mechanistic claims arising from recent trends in neurobiology and
philosophy, these books serve to combat public fears and forestall a
possible backlash against science and technology. Humans can be part
of the natural world without being mere machines, and without being
outdone by our own machines.
These books cast light on how it is possible to have a rich mental
life while living in a physical universe. In so doing, they throw up
roadblocks against any push for political authoritarianism or social
engineering that might arise from increased knowledge of how brains
work. Far from advancing tyranny, neurobiology may be starting to
provide a deeper understanding of what human freedom is all about.
-------------------------------------
Kenneth Silber writes about science, technology, and economics for
Mental Floss and Tech Central Station, among other publications.
References
6. mailto:kensilber at yahoo.com
7. http://www.amazon.com/exec/obidos/ISBN=0805074562/ref=nosim/reasonmagazineA/
8. http://www.amazon.com/exec/obidos/ISBN=0195157338/ref=nosim/reasonmagazineA/
More information about the paleopsych
mailing list