NETFUTURE Technology and Human Responsibility -------------------------------------------------------------------------- Issue #100 A Publication of The Nature Institute January 6, 2000 -------------------------------------------------------------------------- Editor: Stephen L. Talbott (stevet@netfuture.org) On the Web: http://www.netfuture.org/ You may redistribute this newsletter for noncommercial purposes. NetFuture is a reader-supported publication. CONTENTS --------- Editor's Note Quotes and Provocations Who Is on the Receiving End of the Genetic Engineer's Power? Of Gadgets and Real Needs The Trouble with Ubiquitous Technology Pushers (Part 1) (Stephen L. Talbott) or: Why We'd Be Better Off without the MIT Media Lab DEPARTMENTS Correspondence An Engineer's Lament (Gintas Jazbutis) What a Computer Can and Cannot Supply (Shyam Oberoi) Computing Should Relate to Life (Gail Campana) About this newsletter ========================================================================== EDITOR'S NOTE I continue to be hopelessly behind on email -- and I mistakenly deleted one batch. If you don't hear a reply to your inquiry in the next week or so, or are just wondering, feel free to re-send. Some developments relating to NetFuture: ** The Alliance for Childhood's draft statement on Technology Literacy ("Four Guiding Principles for Educators and Parents") published in NF #99 has stirred some things up. The New York Times Online ran a nice story on the statement, written by NetFuture reader Pamela Mendels. Also, the Christian Science Monitor put out a feature story entitled, "In World of High Tech, Everyone Is an Island". NetFuture reader Paul Van Slambrouck, the Monitor's Silicon Valley correspondent, authored the story, which cited both the Alliance and my own work at some length. For pointers to these articles, as well as to the statement on Technology Literacy and to responses from policy analyst Michael Goldhaber, columnist Gary Chapman and others, see the Alliance-related web pages I've set up: http://www.praxagora.com/~stevet/alliance. ** Also, NetFuture reader Dolores Brien expertly led me through an interview, which is now featured on the C. G. Jung website at http://www.cgjung.com/jpculture.html. [Added later: ural is now http://natureinstitute.org/txt/st/jung.htm.] I was very pleased with the result, which summarizes some of my primary concerns about our culture's abdication of consciousness in the presence of digital technologies. In the next issue I'll mention some websites that are making use of NetFuture content in one way or another. Let me hear from you if your site is reprinting, translating, or otherwise drawing upon NetFuture. SLT Goto table of contents ========================================================================== QUOTES AND PROVOCATIONS Who Is on the Receiving End of the Genetic Engineer's Power? ------------------------------------------------------------ C. S. Lewis wrote in The Abolition of Man several decades ago that, in the battle for mastery over nature, "there neither is nor can be any simple increase of power on Man's side. Each new power won by man is a power over man as well. Each advance leaves him weaker as well as stronger. In every victory, besides being the general who triumphs, he is also the prisoner who follows the triumphal car." This simple and disturbing truth has yet to find its proper place in the public dialog about genetic engineering, which has focused mostly on the (very real) likelihood of accident and miscalculation. But Lewis' concern is the more fundamental and inescapable one. In his own blunt terms, If any one age really attains, by eugenics and scientific education, the power to make its descendents what it pleases, all men who live after it are the patients of that power. They are weaker, not stronger. I don't think you can say, however, that one person's power over another is inherently evil. A mother has power over her child, an airline pilot has power over the passengers, and anyone who cares about community knows that we are all dependent upon each other. Some form of power is necessary if we are to act in the world at all. Everything depends, then, on the values with which we exercise our power. The traditional wisdom has it that the only healthy power is self- abnegating power, devoted to the service of others. This is power turned inside out and transformed into love. Those who are acted on by such power are not made weaker, but stronger. The cultural streams from which this wisdom has flowed are alien, not necessarily to the genetic engineers themselves, but certainly to the modern discipline of genetic engineering, founded as it is upon the habit of viewing the organism as a machine. The patients of the discipline can therefore expect to be treated like machines -- and more and more made into machines. Machines don't suffer, of course, and we do hear much talk of deliverance from suffering. But if you listen carefully you will notice that the suffering to be eradicated belongs more to the wielders of power than to its patients. The parent cannot bear the thought of a "deformed" child, nor can the engineer tolerate standing by helplessly. As most "deformed", "terminally ill", and "catastrophically suffering" individuals can tell you, it is above all the well-off who cannot face suffering. They are the ones who most readily forget that the truest aim of life on earth is not merely to be rid of suffering, but to redeem it, to bear its fruits, to escape it by achieving whatever it is that our own life is most forcefully asking of us. That there is, in any significant sense, a life to do this asking is, of course, a premise scarcely informing the apparatus of genetic engineering as we have it today. For the asking requires that there be an antecedent whole -- a being -- presupposed by all the "mechanisms" of our physical organism. Needless to say, this kind of language is anathema in the engineer's laboratory. I'm reminded of the world-famous artificial intelligence researcher at MIT who, taking sarcastic issue with my use of the word "human" in The Future Does Not Compute, wrote: Can I presume he's in favor of religion or something? Isn't that the most usual "human" solution to problems? I think it's fair to say that those who are most eager to take effective charge of human destiny are also those most reluctant to glimpse any coherent or respectable answer to the question, "What does it mean to be human?" So their enterprise unavoidably becomes arbitrary, which explains, among other things, their casual attitude toward gene transfers. It is not that we must refuse to change, develop, evolve. The power to transform ourselves is close to our essence. But there is all the difference in the world between mere arbitrary change and change that proceeds according to the inner necessities of our own being. And there is all the difference in the world between the diabolical power that imposes change upon others arbitrarily and from without (because it recognizes no inner being worth consulting), and the power of love that both recognizes and serves the other. Which sort of power drives the world's genetic engineering laboratories? Don't take my word for it. Why not ask the would-be engineers of humanity, before handing them the keys of power, "What does it mean to you to be human?" We should consider their answers carefully, for these answers will tell us, directly or indirectly, the significance of the engineering project. And if their answer is that it doesn't mean much of anything at all to be human, well then, what does it matter whether or not we suffer their messianic interventions? --------------------- Leon R. Kass cites the Lewis passage in his valuable article on "The Moral Meaning of Genetic Technology" in Commentary (September, 1999). Related articles: ** "What Does It Mean to Be a Sloth?" by Craig Holdrege in NF #97. How can one begin to think about the distinctive character, or being, of an organism? What is the antecendent unity that guarantees, rather than results from, the various "mechanisms" constituting the organism? Of course, the person who is determined to see nothing will see nothing. But this article can help those who are willing to see begin to do some disciplined looking. ** "Is Genetic Engineering Natural?" in NF #75. Are the genetic engineers doing nothing more than we've always done with our various breeding techniques? Also see the follow-up article, "Loosing Genetic Restraints", in NF #77. ** For other articles, see "Genetic engineering" in NetFuture's topical index. Of Gadgets and Real Needs ------------------------- In the ongoing discussion of ubiquitous computing (NF #94, #95, #98), Langdon Winner has noted "the utter disconnect between the supposed wants and needs studied in places like the MIT Media Lab and ones painfully evident in the human community at large" (NF #95). He invites us to compare the needs summarized in the yearly United Nations Human Development Report with the ones that galvanize those "high-tech wizards in Cambridge, Palo Alto and elsewhere" who worry that toilets, toasters, and doors aren't intelligent enough. Alan Wexelblat, who is one of the Cambridge high-tech wizards (he works in the Intelligent Agents Group of the MIT Media Lab) has rejected Winner's invitation because A response of the form "well xxx does nothing to address deep world problems such as hunger, overpopulation, etc." is unanswerable. It's true, but completely irrelevant to the discussion. (NF #98) Completely irrelevant? This is a stunning claim. Certainly it's true that no one subjects every daily activity to a minute "worthiness analysis", checking it against the great global challenges of the day. But surely it is just as true that, in those reflective moments when we do assess our own activities, we want them to prove worthy. We want them to weigh positively against the most pressing needs around us. We may healthily avoid becoming compulsive or too narrowly focused on questions of value, but Wexelblat seems to be saying that we should dismiss these questions even from our more serious reflection. And such reflection, of course, is what NetFuture is about. It's hard to believe that Wexelblat would hold consistently to his claim that "deep world problems" are irrelevant to the assessment of our technological investments. And, in fact, he goes on to say that Winner's argument invites people to put themselves in the position of claiming they don't care about such problems. Rather like folk in Congress making speeches about child pornography -- no one is going to stand up and defend the exploitation of children, but it misses the point that free speech rights are at issue. Here he implies that we should indeed care about world problems of the sort Winner mentions; it's just that other issues take priority. Unfortunately, he doesn't say what these other issues are. And if he should ever do so, wouldn't that be to identify them as the more pressing issues? But it was exactly the appeal to more pressing issues that Wexelblat disparaged in Winner: I would venture to guess that well over 90% of what goes on in the world at large fails to focus on deep significant problems. If that's what's on your mind, then you and I are having such vastly different conversations that it's not worth continuing. So I'm not sure what to make of Wexelblat's position. But I also feel considerable sympathy for his frustration. I honestly don't think that I or many other technology critics have formulated a fully adequate response to the enthusiasms and visions of progress that excite the legions of engineers working at places like the Media Lab. The questions are not easy ones. Did you notice that none of the letter writers responding to Wexelblat took up his comment about air bags having sensors to detect whether a child is sitting in the seat? And there is, I think, an almost unshakable conviction among many computer engineers that all criticisms of this or that silly device simply aren't relevant. The skills and tools we develop today will very likely find their full justification later; that's the nature of progress. Unfortunately, this is rarely a considered judgment. For all the reams of news copy the Media Lab generates each year, it's hard to find in this prolific output any coherent attempt to place the lab's inventive thrust in a context where its social value could be assessed. But, of course, this doesn't hinder the rest of us. In this issue I begin a series of brief essays that are my effort to flesh out a context for evaluating the various claims about ubiquitous computing. SLT Goto table of contents ========================================================================== THE TROUBLE WITH UBIQUITOUS TECHNOLOGY PUSHERS (PART 1) or: Why We'd Be Better Off without the MIT Media Lab Stephen L. Talbott (stevet@netfuture.org) The idea has seized our imaginations with all the force of a logical necessity. In fact, you could almost say that the idea is the idea of logical necessity -- the necessity of embedding little bits of silicon logic in everything around us. What was once the feverish dream of spooks and spies -- to plant a "bug" in every object -- has been enlarged and re-shaped into the millennial dream of ubiquitous computing. In this new dream, of course, the idea of a bug in every object carries various unpleasant overtones. But there are also overtones in the larger and better-promoted notion of ubiquitous computing, despite the fact that our ears are not yet attuned to them. Why Not Omnipotence? -------------------- I suppose Bill Gates' networked house is the reigning emblem of ubiquitous computing. When the door knows who is entering the room and communicates this information to the multimedia system, the background music and the images on the walls can be adjusted to suit the visitor's tastes. When the car and garage talk to each other, the garage door can open automatically whenever the car approaches. Once your mind gets to playing with such scenarios -- and there are plenty of people of good will at places like the MIT Media Lab and Xerox PARC who are playing very seriously with them -- the unlimited possibilities crowd in upon you, spawning visions of a future where all things stand ready to serve our omnipotence. Refrigerators that tell the grocery shopper what is in short supply, shopping carts that communicate with products on the shelves, toilets that assay their clients' health, clothes that network us, kitchen shelves that make omelets, smart cards that record all our medical data, cars that know where they're going -- clearly we can proceed down this road as far and fast as we wish. And why shouldn't we move quickly? Why shouldn't we welcome innovation and technical progress without hesitation? I have done enough computer programming to recognize the inwardly compelling force of the knowledge that I can give myself crisp new capabilities. It is hard to prefer not having a particular capability, whatever it might be, over having it. Moreover, I'm convinced that to say "we should not have technical capability X" is a dead-end argument. It's the kind of argument that makes the proponents of ubiquitous computing conclude, with some justification, that you are simply against progress. You can only finally assess a tool in its context of use, so that to pronounce the tool intrinsically undesirable would require an assessment of every currently possible or conceivable context. You just can't do it -- and if you try, you underestimate the fertile, unpredictable winds of human creativity. But this cuts both ways. You also cannot pronounce a tool desirable (or worth the investment of substantial resources) apart from a context of desirability. Things are desirable only insofar as a matrix of needs, capacities, yearnings, practical constraints, and wise judgments confirms them. This leads me to my first complaint against the ubiquitous technology pushers. Asking the Wrong Questions -------------------------- When we are asked to accept or reject a particular bit of technology -- and, more broadly, when we are asked to embrace or condemn ubiquitous computing as a defining feature of the coming century -- we should flatly refuse the invitation. Technologies as such are the wrong kinds of things to embrace or condemn. To focus our judgments on them is to mistake what is empty for something of value. Take, for example, the questions we face in the classroom. They are educational questions. They have to do, in the first place, with the nature, destiny, and capacities of the child. Such questions are always deeply contextual. They arise from a consideration of this child in this family in this community, against the backdrop of this culture and this physical environment. It's one thing if, deeply immersed in this educational context, pursuing the child's education, we come up against a gap, a shortfall, a felt need, and if, casting about for a solution, we conclude: The computer might offer the best way to fulfill this particular need. But it's quite another thing to begin by assuming that the computer is important for education and then to ask the backward and destructive question, "How can we use the computer in the classroom?" This is to deprive our inquiry of its educational focus and to invite the reduction of educational questions to merely technical ones -- a type of reduction that is the reigning temptation of our age. It leads us, for example, to reconceive learning as information transfer -- fact shoveling. Spurred by this backward thinking, we've felt compelled to spend billions of dollars wiring schools, retraining (or dismissing) teachers, hiring support staff, buying and updating software, rewriting job descriptions, and designing a new curriculum. Then Secretary of Education Richard Riley comes along after the fact and says, Oh, by the way, We have a great responsibility .... We must show that [all this expenditure] really makes a difference in the classroom. (Education Week on the Web, May 14, 1997, via Edupage) The same concerns arise in the workplace. Why do we work? Surely it is, in the first place, in order to discover and carry out our human vocations and to achieve something of value for society. What shape this productive effort might take -- and what tools might be embraced healthily -- can follow only from the most profound assessment of the needs and capacities of both the individual and society. Yet such assessment is increasingly forgotten as social "progress" and vocational decisions are handed to us by automatic, technology-driven processes. It is no accident that we see today a growing consensus among entrepreneurs that all considerations of human value should be jettisoned from the business enterprise as such. Seek first the Kingdom of Profitability, we are advised -- that is, seek what can be perfectly calculated by a machine -- and all else will somehow be added to you. Here again is the reduction of real questions to one-dimensional, abstract, decontextualized, technical ones. The availability of the precise, computational techniques of accounting have encouraged us toward a crazy reversal, whereby the healthy discipline of profitability no longer serves us in work that we independently choose as worthy and fulfilling, but rather we choose our work according to its profitability. It is always easier to make our choices according to rules that can be clearcut, precise, and automatic -- the kind of rules that can be embedded in ubiquitous silicon -- than to ask what sort of human beings we want to become. We can answer the latter question only through our own struggling and suffering -- that is, only by embedding ourselves in real-world contexts. Is Technology Really Context-free? ---------------------------------- So my first complaint is this: the most visible pronouncements in favor of ubiquitous computing take the form of huge investments in places like the MIT Media Lab where the whole aim is to pursue new technologies out of context, as if they were inherently desirable. This mistaking of mere technical capacity for what really matters is the one thing guaranteed to make the new inventions undesirable. The healthy way to proceed would be to concern ourselves with this or that activity in its fullest context -- and then, in the midst of the activity, ask ourselves how its meaning might be deepened, its purpose more satisfyingly fulfilled. Only in that meditation can we begin to sense which technologies might be introduced in appropriate ways and which would be harmful. If the researchers at the Media Lab pursue their work via such immersion in problem contexts -- that is, by exploring significant questions as a basis for seeking answers -- they've done a miserable job of communicating the fact to the rest of us. What we actually receive from them (via the news media) is a steady stream of exclamations about the wonders of this or that technical capability. Typical, so far as I can tell, is the fact, reported in the New York Times, that one of the Media Lab staffers most concerned to render kitchen appliances intelligent is "a bachelor who rarely uses his kitchen". Is it such people who will point us toward the realization of the kitchen's highest and most humane potentials? A technology-focused consciousness -- and you could fairly say that our society is becoming obsessively technology-focused -- is a consciousness always verging upon emptiness. It is a consciousness whose problems are purely formal or technical, with precisely definable solutions. They can be precisely defined because they lack context, they have no significance of their own. Now, it needs adding that no technology perfectly achieves this "ideal" of emptiness and self-containment. As I have pointed out before, a complex device like the computer evolves historically and has numerous tendencies of ours, numerous habits and contexts of use, built into it. This is why you can never say say that such devices are neutral in their implications for society. And, of course, its non-neutrality is what enables us to assess a technology: does it fit into and serve this particular context or not? So when I speak of "technology as such", I am to some degree falsifying things. But the point is that this is the very falsification the ubiquitous technology pushers are encouraging through their strongly decontextualized celebration of ... technology as such. This makes a certain self-deception easy, whereby new technical capacities are much too quickly assumed to represent the answers to problems. And it diverts massive social resources into the production of technologies that, because they will be injected into real contexts with alien force, will certainly prove socially destructive. --------------------- That, in fact, will be the argument of the next installment. More widely, the remaining parts of this essay will deal with issues such as these: When the ubiquitous technology pushers do claim social benefit for their inventions, how well-grounded are those claims? Why do time- and labor-saving devices leave us more pressured and harried? What are we really aiming for when we try to embed intelligence in all the objects around us? Related articles: ** Go to part 2 of this article. ** Go to part 3 of this article. ** "The Fascination with Ubiquitous Control" in NF #95. A kind of preparatory article for this current series. ** "How Technology Co-opted the Good (Parts 1 and 2)" in NF #64 and #65. This is a review of Albert Borgmann's crucially important book, Technology and the Character of Contemporary Life. Borgmann's idea of a "focal practice" captures beautifully what I have merely hinted at here: how meaningful contexts can guide our thinking about technology. The second part of the review is accessible from the first part. Goto table of contents ========================================================================== CORRESPONDENCE An Engineer's Lament -------------------- From: Gintas Jazbutis (gintas.jazbutis@sdrc.com) I often wonder how I could get a PhD in engineering from Georgia Tech while not once having to think about the history and philosophy of technology and its impact on society. There are lots of us out here just churning away on new technology without once thinking about its implications. We do because we can -- scary... -- Gintas Jazbutis Metaphase Deployment Competence Center Bellevue, WA 425-452-5838 What a Computer Can and Cannot Supply ------------------------------------- Response to: "We'll Get What We Choose" (NF-98) From: Shyam Oberoi (shyam.oberoi@thortech.com) It's no exaggeration, I think, to suggest that in order for the great dream of "ubiquitous computing" to be fully realized (that is, for it to be, in the true sense of the word, ubiquitous), one must first agree that every aspect of a human existence must be able to be reduced to a series of repeatable tasks and transactions. Certainly, Mr. Leavitt's wish-list would seem to suggest that his life is precisely that. Should anyone be surprised, then, when an existence is defined primarily through the purchase and consumption of goods, that that person would consider it essential to his well-being that those goods be more readily available and more speedily delivered? Curiously absent from this list is any sort of longing for those experiences that the computer cannot easily supply. Does Mr. Leavitt want to play the piano? Read a novel? Go to a museum? He will probably argue that he hasn't the time anymore (if he ever had) for these quaint pastimes, but he would do well to remember Goethe: "one ought, every day at least, to hear a little song, read a good poem, see a fine picture, and, if it were possible, to speak a few reasonable words." And, while we are certainly drawing close to that happy day when computers will compose symphonies and write sonnets indistinguishable (to most) from the former, old-fashioned productions of their human counterparts, we should not forget that the pleasure that some of us still take in Goethe's arts has nothing to do with the ease with which they can be transmitted or experienced. If anything, the very opposite: we will seek out difficult works of art for that obdurate difficulty, for that complexity that cannot be easily assimilated, for that ineffable experience in the very physical and mental act of reading a book or playing a piano that cannot (and probably will not ever) be replicated or facilitated by a computer. shyam Computing Should Relate to Life ------------------------------- Response to: "We'll Get What We Choose" (NF-98) From: Gail Campana (Gail.Campana@ht.msu.edu) Steve, In Thomas Leavitt's reply to you, I was struck by the fact that most of his support for computer technology revolves around buying things; with only a couple of exceptions, his examples were about spending money and "consuming." I don't see much in his reply that has anything to do with living one's life in a better or more humane way. I do not see anything which shows how technology will do anything to make us wiser, gentler, more peaceful or more happy. I do see that technology is about managing money in our vaunted "free-trade" neo-liberal world where the making of money at the expense of most of the world's population and environment has been raised to the status once held by worship of a supreme being. You were also taken to task by a respondent who said that most things do not have to do with world problems, such as reducing hunger, and as such, you were simply not discussing technology in any meaningful way. Hence, no discussion about technology with you was really possible. That is telling. Real world problems are, after all, the context in which computing happens. I do not want to underestimate the ways in which "technology" has made life easier and healthier than it was 500 years ago. For those of us in the US, particularly. But the lack of any coherent discussion about life, its meaning and what we should be doing as humans to fulfill that meaning, continues to be a problem for "advanced" technological societies which must grapple with wholesale emotional, psychological and spiritual emptiness. Computing, as far as I can tell, exacerbates this for many people.....and heightens the separations which isolate so many. Because as noted in the first paragraph, computing is increasingly about managing material superfluity. Gail Goto table of contents ========================================================================== ABOUT THIS NEWSLETTER Copyright 2000 by The Nature Institute. You may redistribute this newsletter for noncommercial purposes. You may also redistribute individual articles in their entirety, provided the NetFuture url and this paragraph are attached. NetFuture is supported by freely given reader contributions, and could not survive without them. For details and special offers, see http://netfuture.org/support.html . Current and past issues of NetFuture are available on the Web: http://netfuture.org/ To subscribe or unsubscribe to NetFuture: http://netfuture.org/subscribe.html. Steve Talbott :: NetFuture #100 :: January 6, 2000 Goto table of contents