Source: Asia Times
(image by Asia Times)
The world was mired in Cyber Cold War mode; the talk was all about electronic counter-measures, with American C3s (command, control, communications) programmed to destroy Soviet C3s, and both the US and the USSR under MAD (mutually assured destruction) nuclear policies being able to destroy the earth 100 times over. Edward Snowden was not yet a three-year-old.
It was in this context that I set out to do a special report for a now defunct magazine about Artificial Intelligence (AI), roving from the Computer Museum in Boston to Apple in Cupertino and Pixar in San Rafael, and then to the campuses of Stanford, Berkeley and the MIT.
AI had been "inaugurated" in 1956 by Stanford's John McCarthy and MIT professor Marvin Minsky, then a student at Harvard. The basic idea, according to Minsky, was that any intelligence trait could be described so precisely that a machine could be created to simulate it.
My trip inevitably involved meeting a fabulous cast of characters. At the MIT's AI lab, there was Minsky and an inveterate iconoclast, Joseph Weizenbaum, who had coined the term "artificial intelligentsia" and believed computers could never "think" just like a human being.
At Stanford, there was Edward Feigenbaum, absolutely paranoid about Japanese scientific progress; he believed that if the Japanese a fifth-generation (5G) program, based on artificial intelligence, "the US will be able to bill itself as the first great post-industrial agrarian society."
And at Berkeley, still under the flame of hippie utopian populism, there was Robert Wilensky -- Brooklyn accent, Yale gloss, California overtones; and philosopher Robert Dreyfus, a tireless enemy of AI who got his kicks delivering lectures such as "Conventional AI as a Paradigm of Degenerated Research."
Meet Kim No-VAX
Soon I was deep into Minsky's "frames" -- a basic concept to organize every subsequent AI program -- and the "Chomsky paradigm"; the notion that language is at the root of knowledge, and that formal syntax is at the root of language. That was the Bible of cognitive science at the MIT.
Minsky was a serious AI enthusiast. One of his favorite themes was that people were afflicted with "carbon chauvinism"; "This is central to the AI phenomenon. Because it's possible that more sophisticated forms of intelligence are not incorporated in cellular form. If there are other forms of intelligent life, then we may speculate over other types of computer structure."
At the MIT cafeteria, Minsky delivered a futurist rap without in the least resembling Dr Emmet Brown in Back to the Future:
"I believe that in less than five centuries we will be producing machines very similar to us, representing our thoughts and point of view. If we can build a miniaturized human brain weighing, let's say, one gram, we can lodge it in a spaceship and make it travel at the speed of light. It would be very hard to build a spaceship to carry an astronaut and all his food for 10,000 years of travel ..."With Professor Feigenbaum, in Stanford's philosophical garden, the only space available was for the coming yellow apocalypse. But then one day I crossed Berkeley's post-hippie Rubicon and opened the door of the fourth floor of Evans Hall, where I met none other than Kim No-VAX.
No, that was not the Hitchcock blonde and Vertigo icon; it was an altered hardware computer (No-VAX because it had ceased to be a VAX), financed by the mellifluous Pentagon military agency DARPA, decorated with a photo of Kim Novak and humming with the sexy vibration of -- at the time immense -- 2,900 megabytes of electronic data spread over its body.
The US government's Defense Advanced Research Projects Agency -- or DARPA -- was all about computer science. In the mid-1980s, DARPA was immersed in a very ambitious program linking microelectronics, computer architecture and AI way beyond a mere military program. That was comparable to the Japanese 5G program. At the MIT, the overwhelming majority of scientists were huge DARPA cheerleaders, stressing how it was leading research. Yet Terry Winograd, a computer science professor at Stanford, warned that had DARPA been a civilian agency, "I believe we would have made much more progress."
It was up to Professor Dreyfus to provide the voice of reason amidst so much cyber-euphoria; "Computers cannot think like human beings because there's no way to represent all retrospective knowledge of an average human life -- that is, 'common sense' -- in a form that a computer may apprehend." Dreyfus's drift was that with the boom of computer science, philosophy was dead -- and he was a philosopher; "Heidegger said that philosophy ended because it reached its apex in technology. Philosophy in fact reached its limit with AI. They, the scientists, inherited our questions. What is the mind? Now they have to answer for it. Philosophy is over."
Yet Dreyfus was still teaching, as much as at the MIT Weizenbaum was condemning AI as a racket for "lunatics and psychopaths" but still continued to work at the AI lab.
NSA's wet web dream
In no time, helped by these brilliant minds, I figured out that the AI "secret" would be a military affair, and that meant the National Security Agency -- already in the mid-1980s vaguely known as "no such agency," with double the CIA's annual budget and snooping the whole planet. The mission back then was to penetrate and monitor the global electronic net -- that was years before all the hype over the "information highway" -- and at the same time reassure the Pentagon over the inviolability of its lines of communication. For those comrades -- remember, the Cold War, even with Gorbachev in power in the USSR, was still on -- AI was a gift from God (beating Pope Francis by almost three decades).
So what was the Pentagon/NSA up to, at the height of the star wars hype, and over a decade and a half before the Revolution in Military Affairs and the Full Spectrum Dominance doctrine?
1 | 2