Pit Schultz on Sat, 25 May 96 22:29 MDT


[Date Prev] [Date Next] [Thread Prev] [Thread Next] [Date Index] [Thread Index]

nettime: Reification, Schizophrenia, Artificial Intelligence - Phoebe Sengers



                        Fabricated Subjects:
         Reification, Schizophrenia, Artificial Intelligence

                          Phoebe Sengers 
           Computer Science --- Literary and Cultural Theory
                    Carnegie Mellon University
                        phoebe@cs.cmu.edu

Schizophrenia is the ego-crisis of the cyborg.  How could it be any
other way?  Cyborgs are the fabrications of a science invested in the
reproduction of subjects it takes to be real, a science whose first
mistake was the belief that cyborg subjects were autonomous agents,
that they existed outside any web of pre-existing significations.
Pre-structured by all comers, but taken to be pristine, the artificial
agent is caught in the quintessential double bind.  Fabricated by the
the techniques of mass production, the autonomous agent shares in the
modern malady of schizophrenia.  This paper tells the story of that
cyborg, of the ways it has come into being, how it has been
circumscribed and defined, how this circumscription has led to its
schizophrenia, and the ways in which it might one day be cured.

The Birth of the Cyborg:  Classical AI
--------------------------------------

The cyborg was born in the 1950's, the alter ego of the computer.  It
was launched into a world that had already defined it, a world whose
notions of subjectivity and mechanicity not only structured it but
provided the very grounds for its existence.  It was born from the
union of technical possibility with the attitudes, dreams, symbols,
concepts, prejudices of the men who had created it.  Viewed by its
creator as pure potentiality, it was, from the start, hamstrung by the
expectations and understandings which defined its existence.

Those expectations were, and are, almost unachievable.  The artificial
subject is one of the end points of science, the point at which the
knowledge of the subject will be so complete that its reproduction is
possible.  The twin births of Artificial Intelligence and Cognitive
Science represent two sides of the epistemological coin: the move to
reduce human existence to a set of algorithms and heuristics and the
desire to re-integrate those algorithms into a complete agent.  This
resulting agent carries all the burden of proof on its back; its
``correctness'' provides the objective foundation for a huge and
complicated system of knowledge whose centerpiece is rationality.
 
Make no mistake, rationality is the central organizing principle of
classical AI.  The artificial agent is fabricated in a world where
`intelligence,' not `existence,' is paramount, and `intelligence' is
identified with the problem-solving behavior of the scientist.  For
classical AI, the goal is to break intelligent behavior down into a
set of more-or-less well-defined puzzles, to solve each puzzle in a
rational, preferably provably correct, manner, and, one day, to
integrate all those puzzle-solvers to create an agent
indistinguishable (within a sufficiently limited framework) from a
human.

That limited framework had better not exceed reason.  Despite initial
dreams of agents as emotionally volatile as humans, the baggage of a
background in engineering quickly reduced agenthood to rationality.
For example, Allen Newell, one of the founders of AI, wrote an
influential paper which stated that the decision procedure of an agent
must necessarily follow the ``principle of rationality.''  Any agent
worthy of its name must have a set of goals it is pursuing, and any
action taken must, in its opinion, help to achieve one of its goals.
In the narrow constraints of this system, any agent that defies pure
rationality is explicitly stated to be completely incomprehensible,
and hence scientifically invalid.

Given these expectations, it was all too ironic when the artificial
agent began to show signs of schizophrenia.  Designing a rational
decision procedure to solve a clearly defined puzzle was
straight-forward; connecting these procedures together to function
wholistically in novel situations proved to be well-nigh impossible.
Bound in the straitjacket of pure rationality, the cyborg began to
show signs of disintegration: uttering words it did not understand
upon hearing, reasoning about events that didn't affect its actions,
suffering complete breakdown on coming across situations that did not
fit into its limited system of pre-programmed concepts.  Being
understood purely on its own terms and not with respect to any
environment, the agent lived in a fabricated world of its own making,
with only tenuous connections to shared physical and social
environments.  Autistic?  Schizophrenic?  In any case, deranged.


The Promise of Alternative AI
-----------------------------

It was time for therapy.  The shortcomings of the classical agent were
becoming more and more obvious: it could play chess like a master,
re-arrange blocks on command in its dream world, configure computer
boards, but it could not see, find her way around a room, or maintain
routine behavior in a changing world.  It was defined and fabricated
in an ideal, Platonic world, and could not function outside the
boundaries of neat definitions.  Faced with an uncertain, incompletely
knowable world, it ground to a halt.

Understanding that the cyborg was caught in a rational, disembodied
double bind, some AI researchers abandoned the terrain of classical
AI.  Alternative AI --- aka Artificial Life, behavior-based AI,
situated action --- sought to treat agents by the redefinition of the
grounds of their existence.  No longer limiting itself to the
Cartesian subject, the principle of situated action shattered notions
of atomic individualism by redefining an agent in terms of its
environment.  An agent is, and should be, understood as engaged in
interactions with its environment, and its `intelligence' can only be
gauged by understanding the patterns of these interactions.
`Intelligence' is not located in an agent but is the sum total of a
pattern of events occuring in the agent and in the world---the agent
no longer `solves problems,' but `behaves;' the goal is not
`intelligence' per se but `life.'

Redefining the conditions of existence of the agent breathed new life
into the field, if not into the agent itself.  Where once there had
been puzzle-solvers and theorem-provers as far as the eye could see,
there were now herds of walking robots, self-navigating cans-on-wheels
and other varieties of charming stupidity.  Alternative AI had given
the cyborg its body and had lifted some of the constraints on its
behavior.  No longer required to be rational, or even to use mental
representations, the artificial agent found new vistas open to itself.
It did not, however, escape schizophrenia.  Liberated from the
constraints of pure rationality, practitioners of alternative AI,
unwittingly following the latest rages in postmodernism, embrace
schizophrenia as a factor of living experience.  Rather than creating
schizophrenia as a side-effect, they explicitly engineer it in: the
more autonomous an agent's behaviors are, the fewer traces of
Cartesian ego left, the better.  May the most fractured win!

At the same time, that schizophrenia becomes a limit-point for
alternative AI, just as it has been for classical AI.  While
acknowledging that schizophrenia is not a fatal flaw, alternativists
have become frustrated at the extent to which schizophrenia hampers
them from building extensive agents.  Alternativists build agents by
creating behaviors; the integration of those behaviors into a larger
agent has been as much of a stumbling block in alternative AI as the
integration of problem-solvers is in classical AI.  Alternativists are
stuck with the major unsolved question of ``how to combine many
(e.g. more than a dozen) behavior generating modules in a way which
lets them be productive and cooperative.''  Despite their differences
in philosophy, neither alternativists nor classicists know how to keep
an agent's schizophrenia from becoming overwhelming.  What is it about
the engineering of subjectivities that has made such divergent
approaches ground on the same problem?

Fabricating Schizophrenias
--------------------------

There can be no doubt that alternative and classical AI have very
different stakes in their definitions of artificial subjectivity.
These different definitions lead to widely divergent possibilities for
the range of constructed subjects.  At the same time, these subjects
share a mode of breakdown; could it be that these agent-rearing
practices, at first blush so utterly opposed and motivated by
radically dissimilar politics, really have more in common than one
might suspect?

The agents' schizophrenia itself can point the way to a diagnosis of
the common problem.  Far from being autonomous and pristine objects,
artificial agents carry within themselves the fault lines, not only of
their physical environment, but also of the scientific and cultural
environment that created them.  The breakdowns of the agent reflect
the weak points of their construction.  It is not only the agents
themselves that are suffering from schizophrenia, but the very
methodology that is used to create them -- a methodology which, at its
most basic, both alternative and classical AI share.

In classical AI, the emphasis is on agent as problem-solver and
rational goal-seeker, and agents are built using functional
decomposition.  The agent is presumed to have a variety of modules
corresponding more or less to problem-solving methods in its mind.
Researchers work to `solve' each method, creating self-contained
modules for vision, speaking and understanding natural language,
reasoning, planning out behavior, learning, and so on.  They hope that
once they've built each module, they can with not too much effort glue
them back together again and, presto, a complete problem-solving agent
appears.  This is generally an untested hope, since integration, for
classicists, is at once undervalued and nonobvious.  Here,
schizophrenia appears as an inability to seamlessly integrate the 
various competences into a complete whole; the various parts have
conflicting presumptions and divergent belief systems, turning local
rationality into global irrationality.

For practitioners of alternative AI, the agent is thought of
behaviorally, and the preferred methodology is behavioral
decomposition.  Instead of dividing the agent into modules
corresponding to the various abstract abilities of the agent, the
agent is striated along the lines of the behaviors it engages in.  An
agent might typically be constructed by building modules that each
engage in a particular observable behavior: hunting, exploring,
sleeping, fighting.  Alternativists hope to avoid the form of
schizophrenia under which classicists suffer by integrating all the
agent's abilities from the start into specific behaviors in which the
agent is capable of seamlessly engaging.  The problem, again, comes
when those behaviors must be combined into a complete agent: the agent
knows what to do, but not when to do it or how to juggle its
separate-but-equal behaviors.  The agent sleeps instead of fighting,
or tries to do both at once.  Once again the agent is not a seamlessly
integrated whole but a jumble of ill-organized parts.

At its most fundamental, in both forms of AI, an artificial agent is
an engineered reproduction of a `natural' phenomenon and consists of a
semi-random collection of rational decision procedures. Both classical
and alternative AI use an analytic methodology, a methodology that was
described by Marx long before computationally engineering
subjectivities became possible: ``the process as a whole is examined
objectively, in itself, that is to say, without regard to the question
of its execution by human hands, it is analysed into its constituent
phases; and the problem, how to execute each detail process, and bind
them all into a whole, is solved by the aid of machines, chemistry,
&c'' (Marx 380).  In AI, one analyzes human behavior without reference
to cultural context, then attempts, by analysis, to determine and
reproduce the process that generates it.  The methodology of both
types of AI follows the straight, narrow, and ancient road of
objective analysis, with the following formula:

1. Identify a phenomenon in the world to reproduce.

2. Characterize that phenomenon by making a finite list of properties
that it has.

3. Reproduce each one of these properties in a rational decision
procedure.

4. Put the rational decision procedures together, perhaps under another
rational decision procedure, and presume that the original phenomenon
results.

The hallmarks of objectivity, reification, and exclusion of external
context are clear. Through their methodology, both alternative and
classical AI betray themselves as, not singularly novel sciences, but
only the latest step in the process of industrialization.

        In a sense, the mechanical intelligence provided by computers
        is the quintessential phenomenon of capitalism.  To replace
        human judgement with mechanical judgement - to record and
        codify the logic by which rational, profit-maximizing
        decisions are made - manifests the process that distinguishes
        capitalism: the rationalization and mechanization of
        productive processes in the pursuit of profit.... The modern
        world has reached the point where industrialisation is being
        directed squarely at the human intellect. (Kennedy 6)

This is no surprise, given that AI as an engineering discipline has
often been a cozy bedfellow of big business.  Engineering and capital
are co-articulated; fueled by money that encourages simple problem
statements, clear-cut answers, and quick profit unmitigated by social
or cultural concerns, it would in fact be a little surprising if
scientists had managed to develop a different outlook. Reificatory
methods seem almost inevitable.

But reification and industrialization lead to schizophrenia - the hard
lesson of Taylorism. And the methodology of AI seems almost a
replication of Taylorist techniques. Taylorists engaged in analyses of
workers' behavior that attempted to optimize the physical relation
between the worker and the machine. The worker was reduced to a set of
functions, each of which was optimized with complete disregard for the
psychological state of the worker. Workers were then given orders to
behave according to the generated optimal specifications; the result
was chaos. Workers' bodies fell apart under the strain of repetitive
motion. Workers' minds couldn't take the stress of mind-numbing
repetition. Taylorism fell prey to the limits of its own myopic
vision. 

Taylorism, like AI, demands that, not only the process of production,
but the subject itself become rationalized.  ``With the modern
`psychological' analysis of the work-process (in Taylorism) this
rational mechanisation extends right into the worker's `soul': even
his psychological attributes are separated from his total personality
and placed in opposition to it so as to facilitate their integration
into specialized rational systems and their reduction to statistically
viable concepts'' (88). This rationalization turns the subject into an
incoherent jumble of semi-rationalized processes, since ``not every
mental faculty is suppressed by mechanisation; only one faculty (or
complex of faculties) is detached from the whole personality and
placed in opposition to it, becoming a thing, a commodity'' (99).  At
this point, faced with the machine, the subject becomes schizophrenic.

And just the same thing happens in AI; a set of faculties is chosen as
representative of the desired behavior, is separately rationalized,
and is reunited in a parody of wholism. It is precisely the reduction
of subjectivity to reified faculties or behaviors and the naive
identification of the resultant system with subjectivity as a whole
that leads to schizophrenia in artificial agents.  When it comes to
the problem of schizophrenia, the analytic method is at fault.


Schizophrenization and Science
------------------------------

Where does this leave our cyborg?  Having traced its schizophrenia to
the root, it would seem that the antidote is straightforward: jettison
the analytic method, and our patient is cured.  However, just as there
are times when a patient cannot recover because his/her family needs
him/her to be sick, the cyborg cannot recover because its creators
cannot give up analysis.  The analytic method is not incidental to
present AI, something that could be thrown away and replaced with a