Date: Fri 29 Apr 1988 21:13-PDT From: AIList Moderator Kenneth Laws Reply-To: AIList@KL.SRI.COM Us-Mail: SRI Int., 333 Ravenswood Ave., Menlo Park, CA 94025 Phone: (415) 859-6467 Subject: AIList V6 #86 - Philosophy To: AIList@KL.SRI.COM Status: R AIList Digest Saturday, 30 Apr 1988 Volume 6 : Issue 86 Today's Topics: Philosophy - Sociology & Free Will and Self-Awareness ---------------------------------------------------------------------- Date: 26 Apr 88 18:05:43 GMT From: uflorida!novavax!maddoxt@gatech.edu (Thomas Maddox) Subject: Re: The future of AI In article <978@crete.cs.glasgow.ac.uk> gilbert@crete.UUCP (Gilbert Cockton) writes: > >Sociologists study the present, not the future. >I presume the "Megatrends" books >cited is Toffler style futurology, and this sort of railway journey light >reading has no connection with rigorous sociology/contemporary anthrolopology. > >The only convincing statements about the future which competent sociologists >generally make are related to the likely effects of social policy. Such >statements are firmly rooted in a defendible analysis of the present. > >This ignorance of the proper practices of historians, anthropologists, >sociologists etc. reinforces my belief that as long as AI research is >conducted in philistine technical vacuums, the whole research area >will just chase one dead end after another. "Rigorous sociology/contemporary anthropology"? Ha ha ha ha ha ha ha ha, &c. While much work in AI from its inception has consisted of handwaving and wishful thinking, the field has produced and continues to produce ideas that are useful. And some of the most interesting investigations of topics once dominated by the humanities, such as theory of mind, are taking place in AI labs. By comparison, sociologists produce a great deal of nonsense, and indeed the social "sciences" in toto are afflicted by conceptual confusion at every level. Ideologues, special interest groups, purveyors of outworn dogma (Marxists, Freudians, et alia) continue to plague the social sciences in a way that would be almost unimaginable in the sciences, even in a field as slippery, ill-defined, and protean as AI. So talk about "philistine technical vacuums" if you wish, but remember that by and large people know which emperor has no clothes. Also, if you want to say "one dead end after another," you might adduce actual dead ends pursued by AI research and contrast them with non-dead ends so that the innocent who stumbles across your remark won't be utterly misled by your unsupported assertions. ------------------------------ Date: 21 Apr 88 03:58:53 GMT From: SPEECH2.CS.CMU.EDU!yamauchi@pt.cs.cmu.edu (Brian Yamauchi) Subject: Free Will & Self-Awareness In article <3200014@uiucdcsm>, channic@uiucdcsm.cs.uiuc.edu writes: > > I can't justify the proposition that scientific endeavors grouped > under the name "AI" SHOULD NOT IGNORE issues of free wil, mind-brain, > other minds, etc. If these issues are ignored, however, I would > strongly oppose the use of "intelligence" as being descriptive > of the work. Is it fair to claim work in that direction when > fundamental issues regarding such a goal are unresolved (if not > unresolvable)? If this is the name of the field, shouldn't the > field at least be able to define what it is working towards? > I personally cannot talk about intelligence without concepts such > as mind, thoughts, free will, consciousness, etc. If we, as AI > researchers make no progress whatsoever in clarifying these issues, > then we should at least be honest with ourselves and society, and find a > new title for our efforts. Actually the slight modification, > "Not Really Intelligence" would be more than suitable. > > > Tom Channic I agree that AI researchers should not ignore the questions of free will, consciousness, etc, but I think it is rather unreasonable to criticise AI people for not coming up with definitive answers (in a few decades) to questions that have stymied philosophers for millenia. How about the following as a working definition of free will? The interaction of an individual's values (as developed over the long term) and his/her/its immediate mental state (emotions, senses, etc.) to produce some sort of decision. I don't see any reason why this could not be incorporated into an AI program. My personal preference would be for a connectionist implementation because I believe this would be more likely to produce human-like behavior (it would be easy to make it unpredictable, just introduce a small amount of random noise into the connections). Another related issue is self-awareness. I would be interested in hearing about any research into having AI programs represent information about themselves and their "self-interest". Some special cases of this might include game-playing programs and autonomous robots / vehicles. By the way, I would highly recommend the book "Vehicles: Experiments in Synthetic Psychology" by Valentino Braitenburg to anyone who doesn't believe that machines could ever behave like living organisms. ______________________________________________________________________________ Brian Yamauchi INTERNET: yamauchi@speech2.cs.cmu.edu Carnegie-Mellon University Computer Science Department ______________________________________________________________________________ ------------------------------ Date: 26 Apr 88 11:41:52 GMT From: mcvax!ukc!strath-cs!glasgow!gilbert@uunet.uu.net (Gilbert Cockton) Subject: Re: Free Will & Self-Awareness In article <1484@pt.cs.cmu.edu> yamauchi@SPEECH2.CS.CMU.EDU (Brian Yamauchi) writes: >I agree that AI researchers should not ignore the questions of free will, >consciousness, etc, but I think it is rather unreasonable to criticise AI >people for not coming up with definitive answers (in a few decades) to >questions that have stymied philosophers for millenia. It is not the lack of answers that is criticised - it is the ignorance of candidate answers and their problems which leads to the charge of self-perpetuating incompetence. There are philosophers who would provide arguments in defence of AI, so the 'free-will' issue is not one where the materialists, logicians and mechanical/biological determinists will find themselves isolated without an intellectual tradition. > >I don't see any reason why this could not be incorporated into an AI program So what? This is standard silly AI, and implies that what is true has anything to do with the quality of your imagination. If people make personal statements like this, unfortunately the rebuttals can only be personal too, however much the rebutter would like to avoid this position. > >By the way, I would highly recommend the book "Vehicles: Experiments >in Synthetic Psychology" by Valentino Braitenburg to anyone who doesn't >believe that machines could ever behave like living organisms. There are few idealists or romantics who believe that NO part of an organism can be modelled as a mechanical process. Such a position would require that a heart-lung machine be at one with the patient's geist, soul or psyche! The logical fallacy beloved in AI is that if SOME aspects of an organism can be modelled mechanically, than ALL can. This extension is utterly flawed. It may be the case, but the case must be proven, and there are substantial arguments as to why this cannot be the case. For AI workers (not AI developers/exploiters who are just raiding the programming abstractions), the main problem they should recognise is that a rule-based or other mechanical account of cognition and decision making is at odds with the doctrine of free will which underpins most Western morality. It is in no way virtuous to ignore such a social force in the name of Science. Scientists who seek moral, ethical, epistemological or methodological vacuums are only marginalising themselves into positions where social forces will rightly constrain their work. ------------------------------ Date: 28 Apr 88 15:42:18 GMT From: glacier!jbn@labrea.stanford.edu (John B. Nagle) Subject: Re: Free Will & Self-Awareness Could the philosophical discussion be moved to "talk.philosophy"? Ken Laws is retiring as the editor of the Internet AILIST, and with him gone and no replacement on the horizon, the Internet AILIST (which shows on USENET as "comp.ai.digest") is to be merged with this one, unmoderated. If the combined list is to keep its present readership, which includes some of the major names in AI (both Minsky and McCarthy read AILIST), the content of this one must be improved a bit. John Nagle ------------------------------ Date: 29 Apr 88 13:04:43 GMT From: bwk@mitre-bedford.arpa (Barry W. Kort) Subject: Re: Free Will & Self-Awareness One of the problems with the English Language is that most of the words are already taken. Rather than argue over whether AI should or should not include investigations into consciousness, awareness, free will, etc, why not just make up a new label for this activity. I would like to learn how to imbue silicon with consciousness, awareness, free will, and a value system. Maybe this is not considered a legitimate branch of AI, and maybe it is a bit futuristic, but it does need a name that people can live with. So what can we call it? Artificial Consiousness? Artificial Awareness? Artificial Value Systems? Artificial Agency? Suppose I were able to inculcate a Value System into silicon. And in the event of a tie among competing choices, I use a random mechanism to force a decision. Would the behavior of my system be very much different from a sentient being with free will? --Barry Kort ------------------------------ Date: 29 Apr 88 01:26:02 GMT From: quintus!ok@unix.sri.com (Richard A. O'Keefe) Subject: Re: Free Will & Self-Awareness In article <1029@crete.cs.glasgow.ac.uk>, gilbert@cs.glasgow.ac.uk (Gilbert Cockton) writes: > For AI workers (not AI developers/exploiters who are just raiding the > programming abstractions), the main problem they should recognise is > that a rule-based or other mechanical account of cognition and decision > making is at odds with the doctrine of free will which underpins most Western > morality. What about compatibilism? There are a lot of arguments that free will is compatible with strong determinism. (The ones I've seen are riddled with logical errors, but most philosophical arguments I've seen are.) When I see how a decision I have made is consistent with my personality, so that someone else could have predicted what I'd do, I don't _feel_ that this means my choice wasn't free. ------------------------------ End of AIList Digest ********************