Postmodernism and truth
By Daniel Dennett
Here is a story you probably haven't heard, about how a team of
American researchers inadvertently introduced a virus into a third
world country they were studying.(1) They were experts in their
field, and they had the best intentions; they thought they were
helping the people they were studying, but in fact they had never
really seriously considered whether what they were doing might
have ill effects. It had not occurred to them that a side-effect of
their research might be damaging to the fragile ecology of the
country they were studying. The virus they introduced had some
dire effects indeed: it raised infant mortality rates, led to a general
decline in the health and wellbeing of women and children, and,
perhaps worst of all, indirectly undermined the only effective
political force for democracy in the country, strengthening the
hand of the traditional despot who ruled the nation. These
American researchers had something to answer for, surely, but
when confronted with the devastation they had wrought, their
response was frustrating, to say the least: they still thought that
what they were doing was, all things considered, in the interests of
the people, and declared that the standards by which this so-called
devastation was being measured were simply not appropriate.
Their critics, they contended, were trying to impose "Western"
standards in a cultural environment that had no use for such
standards. In this strange defense they were warmly supported by
the country's leaders--not surprisingly--and little was heard--not
surprisingly--from those who might have been said, by Western
standards, to have suffered as a result of their activities.
These researchers were not biologists intent on introducing new
strains of rice, nor were they agri-business chemists testing new
pesticides, or doctors trying out vaccines that couldn't legally be
tested in the U.S.A. They were postmodernist science critics and
other multiculturalists who were arguing, in the course of their
professional researches on the culture and traditional "science" of
this country, that Western science was just one among many
equally valid narratives, not to be "privileged" in its competition
with native traditions which other researchers--biologists,
chemists, doctors and others--were eager to supplant. The virus
they introduced was not a macromolecule but a meme (a
replicating idea): the idea that science was a "colonial"
imposition, not a worthy substitute for the practices and beliefs
that had carried the third-world country to its current condition.
And the reason you have not heard of this particular incident is
that I made it up, to dramatize the issue and to try to unsettle what
seems to be current orthodoxy among the literati about such
matters. But it is inspired by real incidents--that is to say, true
reports. Events of just this sort have occurred in India and
elsewhere, reported, movingly, by a number of writers, among
them:
Meera Nanda, "The Epistemic Charity of the Social Constructivist
Critics of Science and Why the Third World Should Refuse the
Offer," in N. Koertge, ed., A House Built on Sand: Exposing
Postmodernist Myths about Science, Oxford University Press,
1998, pp286-311
Reza Afshari, "An Essay on Islamic Cultural Relativism in the
Discourse of Human Rights," in Human Rights Quarterly, 16,
1994, pp.235-76.
Susan Okin, "Is Multiculturalism Bad for Women?" Boston
Review, October/November, 1997, pp 25-28.
Pervez Hoodbhoy, Islam and Science: Religious Orthodoxy and
the Battle for Rationality, London and New Jersey, Zed Books
Ltd. 1991.
My little fable is also inspired by a wonderful remark of E. O.
Wilson, in Atlantic Monthly a few months ago: "Scientists, being
held responsible for what they say, have not found postmodernism
useful." Actually, of course, we are all held responsible for what
we say. The laws of libel and slander, for instance, exempt none
of us, but most of us--including scientists in many or even most
fields--do not typically make assertions that, independently of
libel and slander considerations, might bring harm to others, even
indirectly. A handy measure of this fact is the evident
ridiculousness we discover in the idea of malpractice insurance
for . . . . literary critics, philosophers, mathematicians, historians,
cosmologists. What on earth could a mathematician or literary
critic do, in the course of executing her profession duties, that
might need the security blanket of malpractice insurance? She
might inadvertently trip a student in the corridor, or drop a book
on somebody's head, but aside from such outré side-effects, our
activities are paradigmatically innocuous. One would think. But in
those fields where the stakes are higher--and more direct--there is
a longstanding tradition of being especially cautious, and of taking
particular responsibility for ensuring that no harm results (as
explicitly honored in the Hippocratic Oath). Engineers, knowing
that thousands of people's safety may depend on the bridge they
design, engage in focussed exercises with specified constraints
designed to determine that, according to all current knowledge,
their designs are safe and sound. Even economists--often derided
for the risks they take with other people's livelihoods--when they
find themselves in positions to endorse specific economic
measures considered by government bodies or by their private
clients, are known to attempt to put a salutary strain on their
underlying assumptions, just to be safe. They are used to asking
themselves, and to being expected to ask themselves: "What if I'm
wrong?" We others seldom ask ourselves this question, since we
have spent our student and professional lives working on topics
that are, according both to tradition and common sense, incapable
of affecting any lives in ways worth worrying about. If my topic is
whether or not Vlastos had the best interpretation of Plato's
Parmenides or how the wool trade affected imagery in Tudor
poetry, or what the best version of string theory says about time,
or how to recast proofs in topology in some new formalism, if I
am wrong, dead wrong, in what I say, the only damage I am likely
to do is to my own scholarly reputation. But when we aspire to
have a greater impact on the "real" (as opposed to "academic")
world-- and many philosophers do aspire to this today--we need to
adopt the attitudes and habits of these more applied disciplines.
We need to hold ourselves responsible for what we say,
recognizing that our words, if believed, can have profound effects
for good or ill.
When I was a young untenured professor of philosophy, I once
received a visit from a colleague from the Comparative Literature
Department, an eminent and fashionable literary theorist, who
wanted some help from me. I was flattered to be asked, and did
my best to oblige, but the drift of his questions about various
philosophical topics was strangely perplexing to me. For quite a
while we were getting nowhere, until finally he managed to make
clear to me what he had come for. He wanted "an epistemology,"
he said. An epistemology. Every self-respecting literary theorist
had to sport an epistemology that season, it seems, and without
one he felt naked, so he had come to me for an epistemology to
wear--it was the very next fashion, he was sure, and he wanted the
dernier cri in epistemologies. It didn't matter to him that it be
sound, or defensible, or (as one might as well say) true; it just had
to be new and different and stylish. Accessorize, my good fellow,
or be overlooked at the party.
At that moment I perceived a gulf between us that I had only
dimly seen before. It struck me at first as simply the gulf between
being serious and being frivolous. But that initial surge of self-
righteousness on my part was, in fact, a naive reaction. My sense
of outrage, my sense that my time had been wasted by this man's
bizarre project, was in its own way as unsophisticated as the
reaction of the first-time theater-goer who leaps on the stage to
protect the heroine from the villain. "Don't you understand?" we
ask incredulously. "It's make believe. It's art. It isn't supposed to be
taken literally!" Put in that context, perhaps this man's quest was
not so disreputable after all. I would not have been offended,
would I, if a colleague in the Drama Department had come by and
asked if he could borrow a few yards of my books to put on the
shelves of the set for his production of Tom Stoppard's play,
Jumpers. What if anything would be wrong in outfitting this
fellow with a snazzy set of outrageous epistemological doctrines
with which he could titillate or confound his colleagues?
What would be wrong would be that since this man didn't
acknowledge the gulf, didn't even recognize that it existed, my
acquiescence in his shopping spree would have contributed to the
debasement of a precious commodity, the erosion of a valuable
distinction. Many people, including both onlookers and
participants, don't see this gulf, or actively deny its existence, and
therein lies the problem. The sad fact is that in some intellectual
circles, inhabited by some of our more advanced thinkers in the
arts and humanities, this attitude passes as a sophisticated
appreciation of the futility of proof and the relativity of all
knowledge claims. In fact this opinion, far from being
sophisticated, is the height of sheltered naiveté, made possible
only by flatfooted ignorance of the proven methods of scientific
truth-seeking and their power. Like many another naif, these
thinkers, reflecting on the manifest inability of their methods of
truth-seeking to achieve stable and valuable results, innocently
generalize from their own cases and conclude that nobody else
knows how to discover the truth either.
Among those who contribute to this problem, I am sorry to say, is,
my good friend Dick Rorty. Richard Rorty and I have been
constructively disagreeing with each other for over a quarter of a
century now. Each of us has taught the other a great deal, I
believe, in the reciprocal process of chipping away at our residual
points of disagreement. I can't name a living philosopher from
whom I have learned more. Rorty has opened up the horizons of
contemporary philosophy, shrewdly showing us philosophers
many things about how our own projects have grown out of the
philosophical projects of the distant and recent past, while boldly
describing and prescribing future paths for us to take. But there is
one point over which he and I do not agree at all--not yet--and that
concerns his attempt over the years to show that philosophers'
debates about Truth and Reality really do erase the gulf, really do
license a slide into some form of relativism. In the end, Rorty tells
us, it is all just "conversations," and there are only political or
historical or aesthetic grounds for taking one role or another in an
ongoing conversation.
Rorty has often tried to enlist me in his campaign, declaring that
he could find in my own work one explosive insight or another
that would help him with his project of destroying the illusory
edifice of objectivity. One of his favorite passages is the one with
which I ended my book Consciousness Explained (1991):
It's just a war of metaphors, you say--but metaphors are not "just"
metaphors; metaphors are the tools of thought. No one can think
about consciousness without them, so it is important to equip
yourself with the best set of tools available. Look what we have
built with our tools. Could you have imagined it without them?
[p.455]
"I wish," Rorty says, "he had taken one step further, and had added
that such tools are all that inquiry can ever provide, because
inquiry is never 'pure' in the sense of [Bernard] Williams' 'project
of pure inquiry.' It is always a matter of getting us something we
want." ("Holism, Intrinsicality, Transcendence," in Dahlbom, ed.,
Dennett and his Critics. 1993.) But I would never take that step,
for although metaphors are indeed irreplaceable tools of thought,
they are not the only such tools. Microscopes and mathematics
and MRI scanners are among the others. Yes, any inquiry is a
matter of getting us something we want: the truth about something
that matters to us, if all goes as it should.
When philosophers argue about truth, they are arguing about how
not to inflate the truth about truth into the Truth about Truth,
some absolutistic doctrine that makes indefensible demands on
our systems of thought. It is in this regard similar to debates
about, say, the reality of time, or the reality of the past. There are
some deep, sophisticated, worthy philosophical investigations into
whether, properly speaking, the past is real. Opinion is divided,
but you entirely misunderstand the point of these disagreements if
you suppose that they undercut claims such as the following:
Life first emerged on this planet more than three thousand million
years ago.
The Holocaust happened during World War II.
Jack Ruby shot and killed Lee Harvey Oswald at 11:21 am, Dallas
time, November 24, 1963.
These are truths about events that really happened. Their denials
are falsehoods. No sane philosopher has ever thought otherwise,
though in the heat of battle, they have sometimes made claims
that could be so interpreted.
Richard Rorty deserves his large and enthralled readership in the
arts and humanities, and in the "humanistic" social sciences, but
when his readers enthusiastically interpret him as encouraging
their postmodernist skepticism about truth, they trundle down
paths he himself has refrained from traveling. When I press him
on these points, he concedes that there is indeed a useful concept
of truth that survives intact after all the corrosive philosophical
objections have been duly entered. This serviceable, modest
concept of truth, Rorty acknowledges, has its uses: when we want
to compare two maps of the countryside for reliability, for
instance, or when the issue is whether the accused did or did not
commit the crime as charged.
Even Richard Rorty, then, acknowledges the gap, and the
importance of the gap, between appearance and reality, between
those theatrical exercises that may entertain us without pretence
of truth-telling, and those that aim for, and often hit, the truth. He
calls it a "vegetarian" concept of truth. Very well, then, let's all be
vegetarians about the truth. Scientists never wanted to go the
whole hog anyway.
So now, let's ask about the sources or foundations of this mild,
uncontroversial, vegetarian concept of truth.
Right now, as I speak, billions of organisms on this planet are
engaged in a game of hide and seek. It is not just a game for them.
It is a matter of life and death. Getting it right, not making
mistakes, has been of paramount importance to every living thing
on this planet for more than three billion years, and so these
organisms have evolved thousands of different ways of finding out
about the world they live in, discriminating friends from foes,
meals from mates, and ignoring the rest for the most part. It
matters to them that they not be misinformed about these matters-
-indeed nothing matters more--but they don't, as a rule, appreciate
this. They are the beneficiaries of equipment exquisitely designed
to get what matters right but when their equipment malfunctions
and gets matters wrong, they have no resources, as a rule, for
noticing this, let alone deploring it. They soldier on, unwittingly.
The difference between how things seem and how things really
are is just as fatal a gap for them as it can be for us, but they are
largely oblivious to it. The recognition of the difference between
appearance and reality is a human discovery. A few other species-
-some primates, some cetaceans, maybe even some birds--shows
signs of appreciating the phenomenon of "false belief"--getting it
wrong. They exhibit sensitivity to the errors of others, and perhaps
even some sensitivity to their own errors as errors, but they lack
the capacity for the reflection required to dwell on this possibility,
and so they cannot use this sensitivity in the deliberate design of
repairs or improvements of their own seeking gear or hiding gear.
That sort of bridging of the gap between appearance and reality is
a wrinkle that we human beings alone have mastered.
We are the species that discovered doubt. Is there enough food
laid by for winter? Have I miscalculated? Is my mate cheating on
me? Should we have moved south? Is it safe to enter this cave?
Other creatures are often visibly agitated by their own
uncertainties about just such questions, but because they cannot
actually ask themselves these questions, they cannot articulate
their predicaments for themselves or take steps to improve their
grip on the truth. They are stuck in a world of appearances,
making the best they can of how things seem and seldom if ever
worrying about whether how things seem is how they truly are.
We alone can be wracked with doubt, and we alone have been
provoked by that epistemic itch to seek a remedy: better truth-
seeking methods. Wanting to keep better track of our food
supplies, our territories, our families, our enemies, we discovered
the benefits of talking it over with others, asking questions,
passing on lore. We invented culture. Then we invented
measuring, and arithmetic, and maps, and writing. These
communicative and recording innovations come with a built-in
ideal: truth. The point of asking questions is to find true answers;
the point of measuring is to measure accurately; the point of
making maps is to find your way to your destination. There may
be an Island of the Colour-blind (allowing Oliver Sacks his usual
large dose of poetic license), but no Island of the People Who Do
Not Recognize Their Own Children. The Land of the Liars could
exist only in philosophers' puzzles; there are no traditions of False
Calendar Systems for mis-recording the passage of time. In short,
the goal of truth goes without saying, in every human culture.
We human beings use our communicative skills not just for truth-
telling, but also for promise-making, threatening, bargaining,
story-telling, entertaining, mystifying, inducing hypnotic trances,
and just plain kidding around, but prince of these activities is
truth-telling, and for this activity we have invented ever better
tools. Alongside our tools for agriculture, building, warfare, and
transportation, we have created a technology of truth: science. Try
to draw a straight line, or a circle, "freehand." Unless you have
considerable artistic talent, the result will not be impressive. With
a straight edge and a compass, on the other hand, you can
practically eliminate the sources of human variability and get a
nice clean, objective result, the same every time.
Is the line really straight? How straight is it? In response to these
questions, we develop ever finer tests, and then tests of the
accuracy of those tests, and so forth, bootstrapping our way to
ever greater accuracy and objectivity. Scientists are just as
vulnerable to wishful thinking, just as likely to be tempted by base
motives, just as venal and gullible and forgetful as the rest of
humankind. Scientists don't consider themselves to be saints; they
don't even pretend to be priests (who according to tradition are
supposed to do a better job than the rest of us at fighting off
human temptation and frailty). Scientists take themselves to be
just as weak and fallible as anybody else, but recognizing those
very sources of error in themselves and in the groups to which
they belong, they have devised elaborate systems to tie their own
hands, forcibly preventing their frailties and prejudices from
infecting their results.
It is not just the implements, the physical tools of the trade, that
are designed to be resistant to human error. The organization of
methods is also under severe selection pressure for improved
reliability and objectivity. The classic example is the double blind
experiment, in which, for instance, neither the human subjects nor
the experimenters themselves are permitted to know which
subjects get the test drug and which the placebo, so that nobody's
subliminal hankerings and hunches can influence the perception
of the results. The statistical design of both individual
experiments and suites of experiments, is then embedded in the
larger practice of routine attempts at replication by independent
investigators, which is further embedded in a tradition--flawed,
but recognized--of publication of both positive and negative
results.
What inspires faith in arithmetic is the fact that hundreds of
scribblers, working independently on the same problem, will all
arrive at the same answer (except for those negligible few whose
errors can be found and identified to the mutual satisfaction of
all). This unrivalled objectivity is also found in geometry and the
other branches of mathematics, which since antiquity have been
the very model of certain knowledge set against the world of flux
and controversy. In Plato's early dialogue, the Meno, Socrates and
the slave boy work out together a special case of the Pythagorean
theorem. Plato's example expresses the frank recognition of a
standard of truth to be aspired to by all truth-seekers, a standard
that has not only never been seriously challenged, but that has
been tacitly accepted--indeed heavily relied upon, even in matters
of life and death--by the most vigorous opponents of science. (Or
do you know a church that keeps track of its flock, and their
donations, without benefit of arithmetic?)
Yes, but science almost never looks as uncontroversial, as cut-
and-dried, as arithmetic. Indeed rival scientific factions often
engage in propaganda battles as ferocious as anything to be found
in politics, or even in religious conflict. The fury with which the
defenders of scientific orthodoxy often defend their doctrines
against the heretics is probably unmatched in other arenas of
human rhetorical combat. These competitions for allegiance--and,
of course, funding--are designed to capture attention, and being
well-designed, they typically succeed. This has the side effect that
the warfare on the cutting edge of any science draws attention
away from the huge uncontested background, the dull metal heft
of the axe that gives the cutting edge its power. What goes without
saying, during these heated disagreements, is an organized,
encyclopedic collection of agreed-upon, humdrum scientific fact.
Robert Proctor usefully draws our attention to a distinction
between neutrality and objectivity.(2) Geologists, he notes, know a
lot more about oil-bearing shales than about other rocks--for the
obvious economic and political reasons--but they do know
objectively about oil bearing shales. And much of what they learn
about oil-bearing shales can be generalized to other, less favored
rocks. We want science to be objective; we should not want
science to be neutral. Biologists know a lot more about the fruit-
fly, Drosophila, than they do about other insects--not because you
can get rich off fruit flies, but because you can get knowledge out
of fruit flies easier than you can get it out of most other species.
Biologists also know a lot more about mosquitoes than about
other insects, and here it is because mosquitoes are more harmful
to people than other species that might be much easier to study.
Many are the reasons for concentrating attention in science, and
they all conspire to making the paths of investigation far from
neutral; they do not, in general, make those paths any less
objective. Sometimes, to be sure, one bias or another leads to a
violation of the canons of scientific method. Studying the pattern
of a disease in men, for instance, while neglecting to gather the
data on the same disease in women, is not just not neutral; it is
bad science, as indefensible in scientific terms as it is in political
terms.
It is true that past scientific orthodoxies have themselves inspired
policies that hindsight reveals to be seriously flawed. One can
sympathize, for instance, with Ashis Nandy, editor of the
passionately anti-scientific anthology, Science, Hegemony and
Violence: A Requiem for Modernity, Delhi: Oxford Univ. Press,
1988. Having lived through Atoms for Peace, and the Green
Revolution, to name two of the most ballyhooed scientific
juggernauts that have seriously disrupted third world societies, he
sees how "the adaptation in India of decades-old western
technologies are advertised and purchased as great leaps forward
in science, even when such adaptations turn entire disciplines or
areas of knowledge into mere intellectual machines for the
adaptation, replication and testing of shop-worn western models
which have often been given up in the west itself as too dangerous
or as ecologically non-viable." (p8) But we should recognize this
as a political misuse of science, not as a fundamental flaw in
science itself.
The methods of science aren't foolproof, but they are indefinitely
perfectible. Just as important: there is a tradition of criticism that
enforces improvement whenever and wherever flaws are
discovered. The methods of science, like everything else under the
sun, are themselves objects of scientific scrutiny, as method
becomes methodology, the analysis of methods. Methodology in
turn falls under the gaze of epistemology, the investigation of
investigation itself--nothing is off limits to scientific questioning.
The irony is that these fruits of scientific reflection, showing us
the ineliminable smudges of imperfection, are sometimes used by
those who are suspicious of science as their grounds for denying it
a privileged status in the truth-seeking department--as if the
institutions and practices they see competing with it were no
worse off in these regards. But where are the examples of
religious orthodoxy being simply abandoned in the face of
irresistible evidence? Again and again in science, yesterday's
heresies have become today's new orthodoxies. No religion
exhibits that pattern in its history.
1. Portions of this paper are derived from "Faith in the Truth," my
Amnesty Lecture, Oxford, February 17, 1997.
2. Value-Free Science?, Harvard Univ. Press, 1991.
This is the final draft of a paper given at the 1998 World
Congress of Philosophy. Daniel Dennett's most recent book,
Freedom Evolves, has just been published by Viking Press.
---
To unsubscribe from the Virus list go to <
http://www.lucifer.com/cgi-bin/virus-l>