More Technology, Not Less
Alvin and Heidi Toffler are authors of such influential
books as Future Shock, The Third Wave, Powershift and Creating a New Civilization.
They are, respectively, chairman and vice chairman of Toffler Associates,
advisors to business and governments.
Los AngelesOne of Americas leading technologists has caused
a great stir by provocatively admitting that he sees "some merit"
in the Luddite reasoning of Theodore Kaczynski, the Unabomber, who once
terrorized scientists whose inventions he thought were enslaving the human
Bill Joy, chief scientist at Sun Microsystems and the chairman of the
presidential commission on the future of Internet technology research,
has gone so far as to call on the scientific community to "relinquish"
research that might lead to the domination of the human species by the
"destructive self-replication" of technologies made possible
by genetics, nano-science and robotics.
According to Joys essay in the April issue of Wired magazine titled
"Why the Future Doesnt Need Us," the alarm bell went off
when it recently became clear, thanks to new advances in molecular electronics,
that computer processing speeds would match the capacity of the human
brain by the year 2030. Theoretically that might make possible robots
as smart as humansand as processing capacities develop further,
smarter; indeed, smart enough to reproduce themselves.
Joy is not a Luddite; he is a serious man raising a responsible warning,
and certainly knows a lot about computers. He has spent 25 years on computer
networking where, as he himself writes, "the sending and receiving
of messages creates the opportunity for out-of-control replication."
All of us are now familiar with the damage that computer viruses can wreak.
But he too easily accepts the Unabombers argument that the human
race "might easily permit itself to drift into a position of such
dependence on machines that it would have no practical choice but to accept
all of the machines decisions." Or else, that the need for
control of such processes would lead to a takeover of society by an elite
that "domesticates" the masses like animals.
This reasoning is far too either-or. These are not the only alternatives.
The scenarios that both Kaczynski and Joy present are essentially mechanistic
and mono-causal. They ignore the rich complexity of the physical and social
environment, which is filled with thousands, if not millions of negative
feedback loops that, in fact, damp down most runaway processes before
they reach their ultimate limits. In fact, they assume that computer capacity
grows, while the human brain remains static.
But must it? The very technologies they regard as most dangerousrobotics,
genetics and nanotech may very well help us expand the human brains
capabilities and make it possible for us to use those technologies in
completely new ways. Recent advances in stem cell research, for example,
challenge the assumption that the brains capacity is fixed.
Like Kaczynski, Joy underestimates the ability of humans to mess things
up, to rebel, to create, whether by chance or not, counter-technologies,
and to step back from the brink of disaster.
Speaking of runaway processes, one of the most amazing facts about the
20th century is not the invention of the atomic bomb, but the fact that
after one demonstration of its power, the human race managed for more
than half a century never to use one again. That does not mean we wont,
and the dangers of proliferation are real. Nonetheless, the record so
far has been remarkable. We managed to chain the chain reaction.
We, too, worry about some of the effects of technology, whether self-replicating
or not, and raised warning flags long ago. In Future Shock (1970), we
forecast cloning of mammals and, eventually, humans; we warned about the
misuse of genetic engineering (even the possibility of race-selective
genetic weapons); we discussed the danger of eugenic manipulation and
a "biological Hiroshima." In War and Anti-War (1994), we wrote
about replication in the form of "self-reproducing war machines."
Unfortunately, Joys proposed remedy is potentially more frightening
than the disease. He recommends that we not only "relinquish"
certain technologies (stuffing genies back into bottles, as it were),
but that we limit the search for certain kinds of knowledge. Of course,
the search for knowledge is always limitedby funds, by cultural
blind spots, by political and other forces. Yet within this reality, the
ethos of science has always included a belief in free, unhindered curiosity
and research. Joys proposal strikes at the heart of that ethos,
which has, for the last three centuries, deepened our knowledge of the
universe we live in and made possible not merely atomic bombs and industrial
pollution, but longer life spans, a reduction of pain and hunger, and
the blessings of what limited democracy we have.
If, in fact, we set out to limit the range of scientific curiosity, the
obvious question is: Who decides? The Ayatollah Ali Khamenei? The Chinese
Communist Party (which regards even run-of-the-mill farm statistics as
state secrets)? Saddam Hussein? The limitations will be imposed by those
in power, and they may do exactly the reverse of what Joy proposes.
More fundamentally, trying to restrain the search for knowledge to stem
the "destructive self-replication of technology" is an attempt
to limit the self-replication of knowledge itself. That is no more possible
for an American computer scientist than for a Chinese apparatchik. Fortunately,
you just cant turn off 6 billion brains. If you could, you would
send the entire human race time-traveling back into the 12th century.
The solution? There isnt an easy one. But the answer is probably
more technology, not less. We will need new technologies that shut down
systems on their way out of control. We have such technologies now in
everything from jet planes to home space heaters. Pharmaceutical companies
are on their way to finding precisely how to use proteins in our DNA for
Joys fear of "destructive self-replication" is warranted.
His clarion call for responsible discussion of this issue is well taken.
But his extreme pessimism is not. Cancer is an example of a runaway, self-replicating
process. But do any of us anymore believe there is no cure for cancer?
Along with worrying about self-replication, it is worth thinking about
the following lines from Future Shock:
"The incipient worldwide movement for control of technology, however,
must not be permitted to fall into the hands of irresponsible technophobes,
nihilists and Rousseauian romantics....Reckless attempts to halt technology
will produce results quite as destructive as reckless attempts to advance
Kaczynski would never agree with that. But well bet Bill Joy does.
Back to Index