NETFUTURE Technology and Human Responsibility -------------------------------------------------------------------------- Issue #138 November 7, 2002 -------------------------------------------------------------------------- A Publication of The Nature Institute Editor: Stephen L. Talbott (stevet@netfuture.org) On the Web: http://www.netfuture.org/ You may redistribute this newsletter for noncommercial purposes. Can we take responsibility for technology, or must we sleepwalk in submission to its inevitabilities? NetFuture is a voice for responsibility. It depends on the generosity of those who support its goals. To make a contribution, click here. CONTENTS --------- Quotes and Provocations High-Profile Doubts about Classroom Computers Ellen Ullman on Artificial Unintelligibility Mindlessness and the Brain (Stephen L. Talbott) Have you been nice to your two hemispheres today? DEPARTMENTS Correspondence Technology Does Not Make Us More Vulnerable (Michael Goldhaber) Response to Goldhaber (Langdon Winner) About this newsletter ========================================================================== QUOTES AND PROVOCATIONS High-Profile Doubts about Classroom Computers --------------------------------------------- That eminent news weekly, the Economist (Oct. 26, 2002), has now pronounced editorially -- and emphatically -- on the billions of dollars spent in order to "clutter classrooms with terminals and keyboards". These billions, the editors note, were "spent on a hunch", and the hunch turns out to have no apparent justification. In an accompanying story about a new study in Israel (which purports to show, among other things, "a consistently negative relation" between computer use and math test scores for fourth graders), the magazine writes: Back in 1922, Thomas Edison predicted that "the motion picture is destined to revolutionize our educational system and ... in a few years it will supplant largely, if not entirely, the use of textbooks". Well, we all make mistakes. But at least Edison did not squander vast quantities of public money on installing cinema screens in schools around the country. Actually, I wouldn't trust that Israeli study any further than I trust all the other studies supposedly elucidating the role of the computer in the classroom. Social science research, in general, remains primitive when it is undisciplined hackwork, and even more primitive when it tries to ape the "hard" sciences. And, in any case, the proponents of wired classrooms will hardly be fazed by the latest report. They will doubtless respond (with some reason) that, in order to get a really good education out of a computer, you can't just drop the machine into a conventional classroom. The computer imposes its own distinctive requirements upon the entire educational context, so that everything needs to change. This takes time. But if this is true, then what in the world were we thinking when we immediately decided that, given the availability of the networked PC, it was urgently required of us to re-found education upon program logic, web access, and email? Wouldn't the natural thing have been to look into that "total change" a little bit before committing huge resources to the computerization of education? Ellen Ullman on Artificial Unintelligibility -------------------------------------------- A few notes from my reading of Ellen Ullman's "Programming the Post-Human" in the October issue of Harper's Magazine: ** Struck by how radically computers have infected the natural sciences, Ullman quotes Lucia Jacobs, a Berkeley professor of psychology studying squirrel behavior: I am an ethologist and know virtually nothing about computers, simulations, programming, mathematical concepts, or logic. But the research is pulling me right into the middle of it. Jacobs is still interested in squirrels, but today she also works with robots. "It is now standard scientific practice", writes Ullman, "to study machine simulations as if they were indeed chipmunks, or squirrels .... Psychology and cognitive science -- and indeed biology -- are thus poised to become, in essence, branches of cybernetics". ** The faith buttressing this work is what Ullman refers to as "engineering empiricism". It is the conviction, long prevalent within artificial intelligence, that we should abandon "sterile philosophizing" and just get down to the practical business of building minds. "You don't have to understand thought to make a mind", as computer scientist Douglas Hofstadter puts it. And Rodney Brooks, director of MIT's artificial intelligence laboratory, told Ullman, "The definition of life is hard. You could spend five hundred years thinking about it or spend a few years doing it". Breathtaking statements. What we see here, in Ullman's apt phrase, is "anti-intellectualism in search of the intellect" -- a mindless search if ever there was one. It rarely seems to occur to researchers like Hofstadter and Brooks that our powers of making must be entirely distinct from any of the things we have made or could make, or that a little intelligent understanding might reveal gross misconceptions at the heart of the entire AI project. Moreover, the irony in their anti-intellectualism, as Ullman brings out, is that years of over-expectation and disappointment have now driven researchers back to the deeper questions -- for example, What is consciousness? Pick up many journals of cognitive science and you will find endless discourses on this question -- discourses, incidentally, whose philosophical naivete and conceptual inadequacy will make you yearn for the good old days of concise medieval hairsplitting. ** Ullman's primary concern in the essay is to combat the disembodied view of life and intelligence that takes hold wherever the computer model prevails. The logical extreme, she observes, is "pure software, unsullied by exigencies of carbon atoms, bodies, fuel, gravity, heat, or any other messy concern of either soft-tissued or metal-bodied creatures". But, she counters, we need more than abstraction. The organism's sentience is integral to the substrate from which it arose, not something that can be taken off and placed elsewhere. We drag along inside us the brains of reptiles, the tails of tadpoles, the DNA of fungi and mice; our cells are permuted paramecia; our salty blood is what's left of our birth in the sea. Genetically, we are barely more than roundworms. Evolution, that sloppy programmer, has seen fit to create us as a wild amalgam of everything that came before us: the whole history of life on Earth lives on, written in our bodies. And who is to say which piece of this history can be excised, separated, deemed "useless" as an essential part of our nature and being? It is surely important to recognize the importance of embodied existence for our mental life. But there is a problem here that Ullman does not address. The abstract, softwarish view of mentality is the conclusion and crown of a long development in science and technology. This development is now presenting us with a strange outcome: having focused single- mindedly upon what they think of as the "solid, physical world", scientists have found themselves sacrificing, not only mentality, but even materiality itself. That is, "hard" science has proven itself unable, in the last analysis, to reckon with anything but numbers, algorithms, and other airy abstractions. So it does Ullman little good to appeal to "the whole history of life on earth ... written in our bodies" -- not if, when we turn to examine these bodies, we find ourselves driven to reconceive them as manufactured from the disembodied equations of particle physics, or as machines governed computationally by genetic and evolutionary algorithms. We are right back with all those software abstractions. Ullman herself buys into this when she refers to evolution as a "programmer". If, as some physicists would have it, "its are bits", then there's not much use in saying (in the words Ullman borrows from linguists George Lakoff and Mark Johnson), "the very structure of reason itself comes from the details of our embodiment". This is merely to say that "these bits come from those bits". The body behind all this talk has disappeared, except as a kind of virtual machine capable of running software. Reified algorithms take the place of real bodies. ** When you buy into the general drift of modern thought, as Ullman seems to do, the results are not always pretty. For example, she quarrels with those who posit a disembodied essence of life, as if these messy experiences of alimentation and birth, these deepest biological imperatives -- stay alive, eat, create others who will stay alive -- were not the foundation, indeed the source, of intelligence; as if intelligence were not simply one of the many strategies that evolved to serve the striving for life. If sentience doesn't come from the body's desire to live ... where else would it come from? But what good does it do to explain intelligence as a strategy that evolved to serve the striving for life, when the idea of striving already includes intelligence? What good does it do to explain sentience as coming from "the body's desire to live", when desire is itself a form of sentience? This is simply to posit the thing you are trying to explain. You will find a similar explanatory strategy endemic in the literature of cognitive science, wherever the attempt is made to explain mind in terms of what is conceived as wholly other than mind. One way or another, the explanations smuggle in the very thing they are trying to explain, often in the guise of such terms as "information", "message", "code", "program", "algorithm", "tendency", and "pattern" -- none of which can be defined without assuming an unexplained thought-element. How could it be otherwise when you make it your task to get mind from something that is supposed to be wholly devoid of mind? ** It is her failure to overcome this confused state of affairs, I think, that leads Ullman to the disappointingly weak conclusion of her essay. In trying to find a positive basis for understanding the mind, she offers a few thoughts about how mammals form societies based on mutual recognition, which she takes to be the crucial thing. (She quotes Rodney Brooks as saying, "None of our robots can recognize their own kind".) And so: Uniqueness, individuality, specialness, is inherent to our strategy for living. It's not just a trick: there really is someone different in there. Yes, someone unique is "in there", and this is doubtless central. But I don't see how Ullman has made a case in any way that clearly separates her view from that of the AI researchers she is criticizing. In fact, she finally just throws in the towel: human sentience is "too complex to understand fully by rational means, something we observe, marvel at, fear. In the end, we give up and call it an 'act of God'". That's the conclusion, I think, of someone who is standing outside her own consciousness, trying to understand it as an objective thing rather than observe her own activity of understanding from within it. But despite its shortcomings on the positive side, Ellen Ullman's worthy article frames a vital set of issues, and is well worth checking out. (Thanks to Thomas Tommi for bringing Ullman's article to my attention.) SLT Goto table of contents ========================================================================== MINDLESSNESS AND THE BRAIN Stephen L. Talbott (stevet@netfuture.org) You have by now most likely read dozens of science news stories playing on the fact that researchers can watch areas of the brain "light up" as test subjects perform various activities. What the lighting up of a particular area means, stated more or less exhaustively, appears to be: "something's going on there". But in this field things are happening so fast that excited researchers can't afford to be slowed down by mere hopeless ignorance. One almost suspects a psychedelic element must reside in those glowing, multicolored, instrument-produced images of cerebral tissue, since most of the news reports carry the same howling absurdities. Here's a typical example. The New York Times (June 22, 1999) reported on research thought to confirm "a theory that the fear of pain is worse than the pain itself". The confirmation lay in the fact that particular areas of the brain lit up during the anticipation of pain, and these were mostly different from, though situated close to, areas that usually light up during the actual experience of pain. That's just about the entire substance of the story, which begins this way: It is a common reaction: fear of the dentist's drill. Now scientists say the feeling is not only real, but they can show just what happens in the brain to cause it. And to think that all this time I mistakenly thought my fear was caused by what the dentist's drill was about to do to me! I guess I should really have been fearing that ominous glow in my brain. At least I can take comfort in the researchers' conclusion that my fear is "real", although it's too bad they didn't give me, for comparison, an example of a fear that is not real. It would have been nice to know which fears of mine weren't really there. It's difficult to decipher what this article (like most others of the same ilk) is actually trying to tell us. But of one thing you can be sure: the chief scientist on the case hopes, as the article tells us, "to use this research to help people with chronic pain". According to the prevailing canons of journalism, every science story needs such a warm and fuzzy benediction, suggesting how the human estate may benefit from the work. This is decidedly not an equal opportunity affair, however; you rarely see such routine statements about the risks of such research. On Reforming Gray Matter ------------------------ For all I know, the brain investigations may indeed lead to chemical or other interventions that "work" in one way or another. But progress toward this end will not be aided by the acute conceptual confusions plaguing this kind of research. And the human pain resulting from these confusions may dwarf anything experienced in the dentist's chair. Given a vague grasp of the fact that "we are psychosomatic organisms", many people -- scientists among them -- seem content to flop blithely back and forth between a brain vocabulary and a mental vocabulary as if there were no distinction between the two. What makes this an inexcusable lack of discipline is the simple fact that, as these vocabularies now exist, no one has the slightest idea how to translate a single term of the one language into a term of the other. It's certainly true that we can correlate elements of brain physiology with elements of mentality. But this fact is fully consistent with opposite extremes of interpretation -- consistent, that is, with the idea that our thinking and other mental activities somehow "arise" as effects of brain matter, and also with the idea that thinking constructs and employs the brain for its own manifestation. However, which alternative we prefer is irrelevant to the point I'm making, which is that the sloppy shifting between different vocabularies results in the most shameless nonsense. What, after all, are we to make of references to the brain as if it were the stuff of mind? Should we try to "reform" those brain tissues that light up when we don't want them to, perhaps admonishing them or administering a slap to some recalcitrant gray matter? When researchers say they've found in the brain the "cause" of our fear of dentists, should we work to remove the cause by altering the physiological conditions responsible for the glow? To be a little more topical, are we to remedy the "cause" of our fear of terrorists by tweaking our brains, or do we need to look elsewhere? Locating Consciousness ---------------------- We are being swamped by this illuminated-brain craziness. A New York Times headline (Sep. 25, 2001) has researchers "Watching How the Brain Works as it Weighs a Moral Dilemma", while a science article in the Economist (May 25, 2002) talks about "how the brain actually makes decisions". Presumably, if the brain is really doing the decision-making and moral weighing, then the buck stops there and all moral education should be in the form of chemical or surgical "instruction". As for another article in the Times, "Looking for That Brain Wave Called Love" (Oct. 28, 2000), you might think it's pure jest. But, no, Rutgers anthropologist, Helen Fisher, has run madly infatuated test subjects through an MRI machine to record their brain activity. Which could be a perfectly interesting thing to do, except that she seems to think she is investigating the nature of love. In fact, she complains about the slowness of those who fail to see the value of her work: It's amazing how scientists don't regard depression or anxiety as a mystery but want to relegate romantic love to the realm of the supernatural. I'm not sure how the supernatural gets in there, except as a cheap way to declare her own point of view sane and rational. But if you want to consign love (or depression or anxiety) to a realm offering no hope of meaningful and non-mysterious understanding, I can't think of any better way than to equate it with physiology. Respond to your advice-seeking, lovesick friends by explaining how they should interact with or modify appropriate brain tissues, and I guarantee you'll produce a great deal of mystification. Another article in the Economist (Sep. 21, 2002) tells us that "neuroscientists think they may have pinned down the source of out-of-the- body experiences". The guilty party? None other than the right angular gyrus -- and if you want to fix it, I should warn you that it is not, after all, near the pineal gland, but rather located above and slightly behind the right ear. An odd place, perhaps, for out-of-the-body experiences to lurk, but if an experience has to hide out somewhere, I suppose that's as good a place as any. Harmonizing the Hemispheres --------------------------- As I was mulling over a pile of science stories like the ones mentioned above, I chanced upon a 1977 Owen Barfield essay. It concerned the 1976 Reith Lectures, entitled "Mechanics of the Mind", by neurophysiologist Colin Blakemore. Back then brain hemisphere research was becoming popular, and so Blakemore discussed it in his lectures. Picking up on this, Barfield began by noting that "if we know something about the physical structure of the brain, we can either make physical use of that knowledge (surgery, drugs, and so forth), or we can decide that another way of approaching our problem is more appropriate. Let us call it the 'consciousness' way". Take the two hemispheres, for instance. If a movement is set on foot for "liberating the right hemisphere", that is the imaginative, and relatively feminine, one (and according to the lecturer, there is such a movement), then the campaigners must mean by "liberation" one of two things -- either direct action on the brain itself, or indirect action by the ordinary means of agitation, argument, propaganda; by "the spread of ideas" in fact: in which case no difference whatever is made by calling it "liberation of the right hemisphere", instead of something like freeing the imagination, or liberation of women. Barfield goes on to "doubt whether the lecturer is capable of grasping such an uncomfortably disjunctive proposition". Which of the alternatives did Blakemore have in mind, for example, when he said, What we should be striving to achieve for ourselves and our brains is not the pampering of one hemisphere to the neglect of the other (whether right or left) or their independent development, but the marriage and harmony of the two. It's not that we should disavow either fork of Barfield's disjunctive proposition; both physical intervention and the attention of consciousness to its own contents have their place. It's just that we should not confuse the two or refuse to be clear regarding which one we are talking about. When Blakemore advocated the "marriage and harmony" of the two hemispheres, was he suggesting something like a surgical interweaving of tissues into a more artistically unified physical tapestry, or was he urging certain conscious disciplines? Or was he, through lamentable vagueness, implying the equivalence of the two approaches despite the fact that subjecting yourself to a scalpel doesn't seem to be quite the same activity as, say, participating in the discussions of a gender sensitivity group? In the actual case, Blakemore disavowed the surgical approach as impossible -- and also as crude compared to direct, cultural influences upon consciousness. And yet (as so often happens in these matters), he immediately flip-flopped. As Barfield puts it, "Dr. Blakemore was not going to let a tedious bit of logical consequence stand in the way of his march towards a peroration". So the scientist rose to the occasion by telling his audience: without a description of the brain, without an account of the forces that mould human behavior, there can never be a truly objective new ethic based on the needs and rights of man. So there we go again. If we want an understanding of our needs and rights and the influences upon our behavior -- things we might once have related to our families and workplaces, our social institutions and personal experiences -- now we see that what we really needed all along was a good description of the brain, presumably so we can whip those tissues into shape. No wonder Barfield gives way to near-despair: How much longer will it all go on? For how much longer will educated men go on being allured by the ignis fatuus of a "consciousness" accessible to physical experiment and investigation? How much longer will they go on spending untiring energy in pursuit of it? A Koan ------ Well, we've now gone on for another quarter century -- and things have only gotten worse. I will not bore you further with recitations from the many contemporary reports on brain research that are simultaneously passed off as reports on mentality. But I do wish to cite Barfield's wonderfully concise statement of the root of the confusion: Perceiving, and every other mode of consciousness, is categorically other than being perceptible, and therefore [is not] accessible to a merely physical investigation. That is, we cannot understand perceiving -- the inner reality of perceiving -- in terms of the kinds of outer things given through the act of perceiving, such as brain tissues. We cannot understand the act as the result of its own results. We cannot understand as just another object the activity that constitutes things as objects. I will leave you with those puzzling remarks, hoping they might serve as a koan of sorts, worthy of some perplexed reflection. I am fully aware that these statements will mostly provoke disbelieving resistance, if not outraged rebellion. They can carry little force for anyone who is still struggling to reconcile the impossible Cartesian notion of mindless matter with the impossible modern notion of mindless mind -- a struggle that yields, as we have seen above, just plain mindlessness. The real need, I'm convinced, is for us to overcome the entire Cartesian, mind-matter dualism, between the pincers of which our culture has been trapped for the past few centuries. "Overcome", I say, not "accept the original terms of the split and then claim to have overcome it by effectively denying just one of the two domains produced by the false cleavage" (which is the standard tack taken by those legions today who fervently disavow the "Cartesian dichotomy"). The idea of purely objective matter, uninformed by, and genetically unrelated to, the mind that perceives it is just as impossible as the idea of mind unrelated to the matter it informs. But we will make progress in all this only insofar as we begin to gain vivid experience, within consciousness, of our own activity in perceiving and thinking. The effect of this will be rather like turning much of modern thinking inside out. The exercise, however, is a long and difficult one. I do hope to write about it before long. Meanwhile, we are left with a view that leaves no room for the human being as anything other than a machine among machines. More alienation, pain, and suffering have flowed from this conviction than anyone could ever tally. If you want to meliorate this pain, you may watch the brain light up until the cows come home, and you can attempt to comfort, bathe in drugs, or otherwise manipulate the cerebral tissues to your heart's content, but you will never by these means touch the actual problem. It is a problem of consciousness, not a problem of the brain. Despite the confused rhetoric coming at us from all sides today, they are not the same thing. Goto table of contents ========================================================================== CORRESPONDENCE Technology Does Not Make Societies More Vulnerable -------------------------------------------------- Response to: "Technology, Trust and Terror" (NF #137) From: Michael Goldhaber (mgoldh@well.com) Langdon, I was frustrated by your piece "Complexity, Trust and Terror". Normally I like what you have to say. But this time, your main point that technological complexity leaves us particularly vulnerable -- say, to terrorism -- strikes me as cliched and mistaken. While it is no doubt the case that the complexities of our society, technological and otherwise, present a great many problems ranging from global warming to lack of active political involvement, the truth is we are far less vulnerable, even to these worries, than less sophisticated societies. Complex systems, among other things, tend to have great redundancy built in or simply lying around ready to be utilized if need be. History doesn't teach what you say. The example of the Goths' attack on Rome's aqueducts in 537 misses the context. After lasting for many centuries, and then undergoing centuries of decline, probably brought on by its inability to find a political form that could handle its size and diversity, the Western Roman Empire finally fell in 476. Sixty years later, it had been briefly reconquered by the remaining Eastern Roman (or Byzantine) Empire, which then failed to hold it. But by then Rome was far from being the technically sophisticated and advanced capital it had been centuries earlier. The aqueducts had survived, but not the engineers who had built them. The post-World War II Strategic Bombing Survey of Germany revealed that, contrary to its intent, the unprecedented level of allied bombing had not significantly reduced Germany's output of war machinery and materiel. Compare how quickly the less technologically complex Taliban fell when subject last year to American bombing at a much smaller scale. Or consider how Sierra Leone's society collapsed from an onslaught of ill- armed rebels a few years ago. As a more homely example, I offer my own experiences after the 1989 magnitude 7.1 Loma Prieta earthquake, which killed about a hundred people in and around the heavily populated San Francisco Bay Area. I was living in San Francisco at the time, and with the Bay Bridge out was concerned that food might not get into the city. My worries evaporated however, when I realized that the artisanal Acme Bread Bakery of Berkeley, which had only recently started supplying one store in my neighborhood, was sending its trucks around the Bay via San Jose to keep up deliveries. We were not only not going to starve, we would still have our luxuries. Contrast that to less technologically complex parts of Turkey, Iran, China, or Central or South America hit by similar magnitude quakes. The death toll is often in the thousands or tens of thousands; food and water supplies disappear; disease runs rampant. Your explanation, "modern, complex technologies succeed by wresting enormous stores of power from the natural realm, seeking to direct these powers in ways that are controllable and useful", is simply inapt. The Internet, complex as it is, ought to be subject to that definition, but just isn't. Even as a description of a skyscraper, the thought seems tortured, at best. The World Trade Center might well have been ugly and dehumanizing, but the reason it collapsed had to do less with its "wresting stores of power from the natural realm" than an inadequate fire- proofing system, inadequate concern for safety, etc. To some extent these problems and some others that you mention can be ascribed to trust. On the other hand, as you hint, our normal trust that mild levels of security are all that are needed for our safety from terrorists have mostly proved accurate. But the system just is not as vulnerable as you claim to the kind of attacks we witnessed. Saying otherwise feeds into the anti-terrorist hysteria. (I'm not saying no one will die, but even thousands of deaths, though horrible, are not the same as system breakdown.) If you were to argue that the complexity of modern life dumbfounds the electorate I think you would have something. Part of what you call trust is simply inattention resulting from the simple impossibility of taking seriously all the issues, etc., that seem to call for attention. (Some of the pieces, including those about terrorism, on my website http://www.well.com/user/mgoldh/ might be relevant.) Response to Goldhaber --------------------- From: Langdon Winner (winner@rpi.edu) Michael Goldhaber seems perfectly comfortable with the level of security that our complex technologies offer. Evidently, he sees no need for any fundamental change in the way we design, build and operate the complex systems that support our way of life. Even after a major earthquake, he notes, we can still rely on these helpful mechanisms of production and distribution to click into operation, allowing us to "have our luxuries." As he makes his case, Goldhaber engages in selective misreading of my piece. My comment about Rome, for example, explicitly avoids speculation on the ultimate causes of the fall of the empire and its timing. It highlighted the fact that the idea of attacking links of a crucial material infrastructure was not a new idea; Vitiges was aware of the value of this strategy and did, in fact, employ it successfully against the city. Perhaps I went a bit over the top by including Gray Brechin's colorful comment on the fate of the Roman caput mundi. Brechin's book is all about the strategies of imperial control and how they eventually come to grief. At a time in which techno-imperial schemes of military might and economic globalization are precisely the ones favored by American leaders, an echo of vulnerability from ancient times seemed worth noting. Goldhaber objects to my emphasis that "modern technologies succeed by wresting enormous powers from nature." He says that neither the Internet nor the skyscraper fit that definition. What is he thinking? Has he forgotten that equipment of computing and communication that comprise the Internet are produced from natural materials transformed through energy- and resource-intensive processes of chemical, electrical and industrial fabrication? What about the vast amounts of coal, gas, and oil burned, as well as atoms split, each day to transmit those trillions of bits of information? During my stay in southern California during the blackouts of February, 2001, my Internet connection certainly went down. As regards skyscrapers, is not the widely admired feat in their construction a mastery of nature expressed in the materials and ingeniously balanced structures that defy the force of gravity? In the wake of 9/11 architects and engineers are thoroughly reexamining the assumption and methods that were previously state of the art. My point was that arrangements of energy and resources that make large- scale systems apparently stable and useful also makes them precarious, a fact overlooked in the glory days of American modernism. As our dependency upon these technologies grows, protecting their structure and operation could well emerge as society's most urgent priority, even when that involves the sacrifice of key social and political ends. The alarm sounds repeatedly. A recent report by the Council on Foreign Relations, "America Still Unprepared -- America Still in Danger," warns, "The homeland infrastructure for refining and distributing energy to support our daily lives remains largely unprotected to sabotage." Stressing many of the areas I mentioned in my piece -- containerized cargo, transportation, food supply -- the report recommends an intense build up of police, surveillance, the National Guard, as well as the nation's public health and agricultural agencies. Spend more money! Mobilize all of society! Perhaps Michael would agree that we should reject this obsession and begin seeking other paths to security. My own view is that local, resilient, relatively small-scale, community-based technologies that rely upon renewable energy and resources provide the best solutions in the years ahead. In the short run we need to ask whether the sense of emergency requires dismantling central features of the Constitution and erecting fierce new institutions that endanger our freedoms far more than they threaten Al Qaeda. I appreciate Michael's response, but do not share his feeling that technological systems as we find them are probably good enough. In these dark and darkening times we need a much broader debate about what safety and security require. Goto table of contents ========================================================================== ABOUT THIS NEWSLETTER Copyright 2002 by The Nature Institute. You may redistribute this newsletter for noncommercial purposes. You may also redistribute individual articles in their entirety, provided the NetFuture url and this paragraph are attached. NetFuture is supported by freely given reader contributions, and could not survive without them. For details and special offers, see http://netfuture.org/support.html . Current and past issues of NetFuture are available on the Web: http://netfuture.org/ To subscribe or unsubscribe to NetFuture: http://netfuture.org/subscribe.html. Steve Talbott :: NetFuture #138 :: November 7, 2002 Goto table of contents