Date: Fri 27 May 1988 00:42-EDT From: AIList Moderator Nick Papadakis Reply-To: AIList@AI.AI.MIT.EDU Us-Mail: MIT Mail Stop 38-390, Cambridge MA 02139 Phone: (617) 253-2737 Subject: AIList Digest V7 #8 To: AIList@AI.AI.MIT.EDU Status: R AIList Digest Friday, 27 May 1988 Volume 7 : Issue 8 Today's Topics: Philosophy and Social science Re: Arguments against AI are arguments against human formalisms Re: Re: Re: Exciting work in AI -- (stats vs AI learning) non-AI theories about symbols the mind of society Alternative to Probability (was Re: this is philosophy ??!!?) Assumptions in dialog (was Re: Acting irrationally) ---------------------------------------------------------------------- Date: 4 May 88 14:28:25 GMT From: attcan!houdi!marty1@uunet.uu.net (M.BRILLIANT) Subject: Re: this is philosophy ??!!? In article <1588@pt.cs.cmu.edu>, Anurag Acharya writes: > Gilbert Cockton writes: > ... > > Your system should prevaricate, stall, duck the > >issue, deny there's a problem, pray, write to an agony aunt, ask its > >mum, wait a while, get its friends to ring it up and ask it out ... > > Whatever does all that stuff have to do with intelligence per se ? > .... Pardon me for abstracting out of context. Also for daring to comment when I am not an AI researcher, only an engineer waiting for a useful result. But I see that as an illuminating bit of dialogue. Cockton wants to emulate the real human decision maker, and I cannot say with certainty that he's wrong. Acharya wants to avoid the pitfalls of human fallibility, and I cannot say with certainty that he's wrong either. I wish we could see these arguments as a conflict between researchers who want to model the human mind, and researchers who want to make more useful computer programs. Then we could acknowledge that both schools belong in AI, and stop arguing over which should drive out the other. M. B. Brilliant Marty AT&T-BL HO 3D-520 (201)-949-1858 Holmdel, NJ 07733 ihnp4!houdi!marty1 Disclaimer: Opinions stated herein are mine unless and until my employer explicitly claims them; then I lose all rights to them. ------------------------------ Date: 9 May 88 14:12:39 GMT From: trwrb!aero!venera.isi.edu!smoliar@bloom-beacon.MIT.EDU (Stephen Smoliar) Subject: Re: this is philosophy ??!!? In article byerly@cars.rutgers.edu (Boyce Byerly ) writes: > >Perhaps the logical deduction of western philosophy needs to take a >back seat for a bit and let less sensitive, more probalistic >rationalities drive for a while. > I have a favoire paper which I always like to recommend when folks like Boyce propose putting probabilistic reasoning "in the driver's seat:" Alvan R. Feinstein Clinical biostatistics XXXIX. The haze of Bayes, the aerial palaces of decision analysis, and the computerized Ouija board. CLINICAL PHARMACOLOGY AND THERAPUTICS Vol. 21, No. 4 pp. 482-496 This is an excellent (as well as entertaining) exposition of many of the pitfalls of such reasoning written by a Professor of Medicine and Epidemiology at the Yale University School of Medicine. I do not wish this endorsement to be interpreted as a wholesale condemnation of the use of probabilities . . . just a warning that they can lead to just as much trouble as an attempt to reduce the entire world of first-order predicate calculus. We DEFINITELY need abstractions better than such logical constructs to deal with issues such as uncertainty and belief, but it is most unclear that probability theory is going to provide those abstractions. More likely, we should be investigating the shortcomings of natural deduction as a set of rules which represent the control of reasoning and consider, instead, possibilities of alternative rules, as well as the possibility that there is no one rule set which is used universally but that different sets of rules are engaged under different circumstances. ------------------------------ Date: 9 May 88 14:56:25 GMT From: attcan!lsuc!spectrix!yunexus!geac!geacrd!cbs@uunet.uu.net (Chris Syed) Subject: Re: Social science gibber [Was Re: Various Future of AI This is a comment upon parts of two recent submissions, one by Simon Brooke and another from Jeff Dalton. Brooke writes: > AI has two major concerns: the nature of knowledge, and the nature of > mind. These have been the central subject matter of philosophy since > Aristotle, at any rate. The methods used by AI workers to address these > problems include logic - again drawn from Philosophy. So to summarise: > AI addresses philosophical problems using (among other things) > philosophers tools. Or to put it differently, Philosophy plus hardware - > plus a little computer science - equals what we now know as AI. The fact > that some workers in the field don't know this is a shameful idictment on > the standards of teaching in AI departments. If anyone doubts these claims, s/he might try reading something on Horne clause logic. {I know, Horne probably dosen't have an 'e' on it}. And, as Brooke says, a dose of Thomas Kuhn seems called for. It is no accident that languages such as Prolog seem to appeal to philosophers. In fact, poking one's head into a Phil common room these days is much like trotting down to the Comp Sci dept. All them philosophers is talking like programmers these days. And no wonder - at last they can simulate minds. Meanwhile, try Minsky's _Community of Mind_ for a peek at the crossover from the other direction. By the by, it's relatively hard to find a Phil student, even at the graduate level, who can claim much knowledge of Aristotle these days (quod absit)! Nevertheless, dosen't some AI research have more mundane concerns than the study of mind? Like how do we zap all those incoming warheads whilst avoiding wasting time on the drones? Jeff Dalton writes: > Speaking of outworn dogmas, AI seems to be plagued by behaviorists, > or at least people who seem to think that having the right behavior > is all that is of interest: hence the popularity of the Turing Test. I'm not sure that the Turing Test is quite in fashion these days, though there is notion of a 'Total Turing Test' (Daniel C. Dennet, I think?). Behaviourism, I must admit, gives me a itch (positively reinforcing, I'm sure). But I wonder just what 'the right behaviour' _is_, anyway? It seems to me that children (from a Lockean 'tabula rasa' point of view), learn & react differently from adults (with all that emotional baggage they carry around). One aspect of _adult_ behaviour I'm not sure AI should try to mimic is our nasty propensity to fear admitting one's wrong. AI research offers Philosophy a way to strip out all the social and cultural surrounds and explore reasoning in a vaccuum... to experiment upon artificial children. But adult humans cannot observe, judge, nor act without all that claptrap. As an Irishman from MIT once observed, "a unique excellence is always a tragic flaw". Maybe it depends on what you're after? {uunet!mnetor,yunexus,utgpu} o ~ !geac!geacrd!cbs (Chris Syed) ~ \-----\---/ GEM: CHRIS:66 ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ "There can be no virtue in obeying the law of gravity." - J.E.McTaggart. ------------------------------ Date: 9 May 88 21:31:36 GMT From: mcvax!ukc!its63b!aiva!jeff@uunet.uu.net (Jeff Dalton) Subject: Re: Arguments against AI are arguments against human formalisms In article <1103@crete.cs.glasgow.ac.uk> Gilbert Cockton writes: > BTW, Robots aren't AI. Robots are robots. I'm reminded of the Lighthill report that caused a significant loss of AI funding in the UK about 10 years ago. The technique is to divide up AI, attack some parts as useless or worse, and then say the others are "not AI". The claim then is that "AI", properly used, turns out to encompass only things best not done at all. All of the so-called AI that's worth supporting (and funding) belongs to other disciplines and so should be done there. Another example of this approach can be found earlier in the message: > Note how scholars like John Anderson restrict themselves to proper > psycholgical data. I regard Anderson as a psychologist, not as an AI > worker. A problem with this attack is that it is not at all clear that AI should be defined so narrowly as to exclude, for example, *all* robotics. That robots are robots does not preclude some of them being programmed using AI techniques. Nor would an artificial intelligence embodied in a robot automatically fail to be AI. The attack seeks to set the terms of debate so that the defenders cannot win. Any respectable result cited will turn out to be "not AI". Any argument that AI is possible will be sat on by something like the following (from <1069@crete.cs.glasgow.ac.uk>): Before the 5th Generation scare, AI in the UK had been sat on for dodging too many methodological issues. Whilst, like the AI pioneers, they "could see no reasons WHY NOT [add list of major controversial positions", Lighthill could see no reasons WHY in their work. In short, the burden of proof would be such that it could not be met. The researcher who wanted to persue AI would have to show the research would succeed before undertaking it. Fortunately, there is no good reason to accept the narrow definition of AI, and anyone seeking to reject the normal use of the term should accept the burden of proof. AI is not confined to attempts at human- level intelligence, passing the Turing test, or other similar things now far beyond its reach. Moreover, the actual argument against human-level AI, once we strip away all the misdirection, makes claims that are at least questionable. > The argument against AI depends on being able to use written language > (physical symbol hypothesis) to represent the whole human and physical > universe. AI and any degree of literate-ignorance are incompatible. > Humans, by contrast, may be ignorant in a literate sense, but > knowlegeable in their activities. AI fails as this unformalised > knowledge is violated in formalisation, just as the Mona Lisa is > indescribable. The claim that AI requires zero "literate-ignorance", for example, is far from proven, as is the implication that humans can call on abilities of a kind completely inaccessable to machines. For some reasons to suppose that humans and machines are not on opposite sides of some uncrossable line, see (again) Dennett's Elbow Room. ------------------------------ Date: 9 May 88 21:46:31 GMT From: mcvax!ukc!its63b!aiva!jeff@uunet.uu.net (Jeff Dalton) Subject: Re: Sorry, no philosophy allowed here. In article <1069@crete.cs.glasgow.ac.uk> gilbert@cs.glasgow.ac.uk (Gilbert Cockton) says: > If you can't write it down, you cannot possibly program it. Not so. I can write programs that I could not write down on paper because I can use other programs to so some of the work. So I might write programs that are too long, or too complex, to write on paper. ------------------------------ Date: Mon, 9 May 88 20:11:02 PDT From: larry@VLSI.JPL.NASA.GOV Subject: Philosophy: Informatique & Marxism --The following bounced when I tried two different ways to send it directly. Gilbert Cockton: Even one reference to a critique of Systems Theory would be helpful if it includes a bibliography. If you can't find one without too much trouble, please send at least a few sentences explaining the flaw(s) you see. I would dearly love to be able to advance beyond it, but don't yet see an alternative. --The difficulty of formalising knowledge. Cockton makes a good point here. The situation is even worse than he indicates. Many, perhaps most or all decisions seem to be made subconsciously, apparently by an analog vector-sum operation rather than logical, step-by-step process. We then rationalize our decision, usually so quickly & easily that we are never aware of the real reasons. This makes knowledge and procedure capture so difficult that I suspect most AI researchers will try (ultimately unsuccessfully) to ignore it. --Marxism. Economics is interesting in that it produced a cybernetic explanation of production, value, & exchange (at least as early as the 17th century) long before there was cybernetics. Marx (and Engels) spent a lot of time studying & explaining this process, and much of their work is still useful. Other parts have not been supported by advances in history & knowledge. The workers (at least in America) are predominantly opposed to revolution and support capitalism, despite much discontent about the way bosses treat them. Part of this may be because our society holds out the hope that many of us can become bosses ourselves. Another reason is that many workers, either directly or through retirement plans, have become owners. Technology has also differentiated the kinds of workers from mostly physical laborer to skilled workers of many different types who sympathize with their own subclass rather than workers in general. Further, once workers feel they have reached an acceptable subsistence, oftentimes they develop other motivations for work having nothing to do with material concerns. People from a middle-class or higher background often stereotype the "lowest proletariat" as beer-drinking slobs whose only interest is food, football, and sex. Coming from a working class background (farmers and factory laborers), I know that "doing a good job" is a powerful motivator for many workers. The "higher proletariat" (who are further from the desperate concern for survival) show this characteristic even more strongly. Most engineers I know work for reasons having nothing to do with money; the same is true of many academics and artists. (This is NOT to say money is unimportant to them.) Just as the practice of economics has deviated further & further from the classical Marxist viewpoint, so has theory. Materialism, for instance, has changed drastically in a world where energy is at least as important as matter, which has itself become increasingly strange. Too, the science of "substance" has been joined by a young, confused, but increasingly vigorous, fertile and rigorous science of "form," variously called (or parts of it) computer science, cybernetics, communications theory, information science, informatique, etc. This has interesting implications for theories of monetary value and the definition of capital, implications that Marx did not see (& probably could not, trapped in his time as he was). Informatique has practical implications of which most of us on this list are well aware. One of the most interesting economically is the future of guardian-angel programs that help us work: potentially putting us out of a job, elevating our job beyond old limits, or (as any powerful tool can) harming us. And in one of the greatest ironies of all, AI researchers working in natural language and robotics have come to realize the enormous sophistication of "common labor" and the difficulties and expense of duplicating it mechanically. Larry @ jpl-vlsi ------------------------------ Date: Tue, 10 May 88 09:44 EDT From: Stephen Robbins Subject: AIList V6 #97 - Philosophy Date: 6 May 88 22:48:09 GMT From: paul.rutgers.edu!cars.rutgers.edu!byerly@rutgers.edu (Boyce Byerly ) Subject: Re: this is philosophy ??!!? 2) In representing human knowledge and discourse, it fails because it does not recognize or deal with contradiction. In a rigorously logical system, if P ==> Q ~Q P Then we have the tautology ~Q and Q. If you don't believe human beings can have the above deriveably contradictory structures in their logical environments, I suggest you spend a few hours listening to some of our great political leaders :-) There's also an issue of classification, with people. How do they even \know/ something's a Q or ~Q? One of the most fascinating (to me) moments in the programming class I teach is when I hand out a sheet of word problems for people to solve in LISP. If I call them "mini program specs," the class grabs them and gets right to work. If I call them "word problems," I start to get grown men and women telling me that "they can't \do/ word problems." Despite their belief that they \can/ program! It seems to be a classification issue. ------------------------------ Date: 6 May 88 23:05:54 GMT From: oliveb!tymix!calvin!baba@sun.com (Duane Hentrich) Subject: Re: Free Will & Self Awareness In article <31024@linus.UUCP> bwk@mbunix (Barry Kort) writes: >Why then, when a human engages in undesirable behavior, do we resort >to such unenlightened corrective measures as yelling, hitting, or >deprivation of life-affirming resources? Probably some vague ideas that negative reinforcement works well. Or role-modeling parents who did the same thing. For the same reason that the Enter/Carriage Return key on many keyboards is hit repeatedly and with great force, i.e. frustration with an inefficient/ineffective interface which doesn't produce the desired results. Yeah. I've noticed this: if something doesn't work, people do it longer, harder, and faster. "Force it!" seems to be the underlying attitude. But in my experience, slowing down and trying carefully works much better. "Force it!" hasn't even been a particularly good heuristic for me. Actually, I wonder if it's people-in-general, or primarily a Western phenomenon. -- Stephen ------------------------------ Date: 10 May 88 17:31:38 GMT From: dogie!mish@speedy.wisc.edu Subject: Re: Sorry, no philosophy allowed here. In article <414@aiva.ed.ac.uk>, jeff@aiva.ed.ac.uk (Jeff Dalton) writes... >In article <1069@crete.cs.glasgow.ac.uk> gilbert@cs.glasgow.ac.uk >(Gilbert Cockton) says: >> If you can't write it down, you cannot possibly program it. > >Not so. I can write programs that I could not write down on paper >because I can use other programs to so some of the work. So I might >write programs that are too long, or too complex, to write on paper. YACC Lives! I've written many a program that included code 'written' by some other program (namely YACC). The point is that the computer allows us to extend what we know. I may not have actually written the code, but I knew how to tell the computer to write the code. In doing so, I created a program that I never (well, almost never) could have written myself even though I knew how _____________________ ____________________________ ___________________ \ / \ / Tom \ / Bit: mish@wiscmacc \ / Univ. Of Wis. Mish X Arpa: mish@vms.macc.wisc.edu X Madison Jr. / \ Phone: (608) 262-8525 / \ MACC _____________________/ \____________________________/ \___________________ ------------------------------ Date: 10 May 1988 14:10 (Tuesday) From: munnari!nswitgould.oz.au!wray@uunet.UU.NET (Wray Buntine) Subject: Re: Re: Re: Exciting work in AI -- (stats vs AI learning) It is worthwhile me clarifying Stuart Crawford's (Advanced Decision Systems, stuart@ads.com, Re: Exciting work in AI) recent comments on my own discussion of Quinlan's work, because it brings out a distinction between purely statistical approaches and approaches from the AI area for learning prediction rules from noisy data. This prediction problem is OF COURSE an applied statistics one. (My original comment never presumed otherwise---the reason I posted the comment to AIList in the first place was to point this out) But, it is NOT ALWAYS a purely applied statistics problem (hence my comments about Quinlan's "improvements"). 1. In knowledge acquisition, we usually don't have a purely statistical problem, we often have a small amount of data , a knowledgeable but only moderately articulate expert . To apply a purely statistical approach to the data alone is clearly ignoring a very good source of information: the "expert". To expect the expert to sprout forth relevant information is naive. We have to produce a curious mix of applied statistics and cognitive psych to get good results. With comprehensibility of statistical results, prevelant in learning work labelled as AI, we can draw the expert into giving feedback on statistical results (potential rules). This is a devious but demonstrably successful means of capturing some of his additional information. There are other advantages of comprehensibility in the "knowledge acquisition" context that again arise for non-statistical reasons. 2. Suffice it to say, trees may be more comprehensible than rules sometimes (when they're small they certainly give a better picture of the overall result), but when they're large they aren't always. Transforming trees to rules is not simply a process of picking a branch and calling it a rule. A set of disjunctive rules can be logically equivalent to a minimal size tree that is LARGER BY AN ORDER OF MAGNITUDE. In a recent application (reported in CAIA-88) the expert flatly refused to go over trees, but on being shown rules found errors in the data preparation, errors in the problem formulation, and provided substantial extra information (the rules jogged his memory), .... merely because he could easily comprehend what he was looking at. Need I say, subsequent results were far superior. In summary, when learning prediction rules from noisy data, AI approaches complement straight statistical ones in knowledge acquisition contexts, for reasons outside the domain of statistics. In our experience, and the experience of many others, this can be necessary to produce results. Wray Buntine wray@nswitgould.oz School of Computing Science University of Technology, Sydney PO Box 123, Broadway Australia, 2007 ------------------------------ Date: 10 May 88 11:56 PDT From: hayes.pa@Xerox.COM Subject: Re: AIList V6 #98 - Philosophy Nancyk has made a valuable contribution to the debate about logic and AI, raising the discussion to a new level. Of course, once one sees that to believe that ( for example ) if P and Q are both true, then P&Q is true, is merely an arifact of concealed organismal Anthropology, a relic of bougeois ideology, the whole matter becomes much clearer. We who have thought that psychology might be relevant to AI have missed the point: of course, political science - specifically, Marxist political science - is the key to making progress. Let us all properly understand the difference between the Organismus-Umwelt Zusammenhang and the Mensch-Umwelt Zusammenhang, and let our science take sides with the working people, and we will be in a wholly new area. And I expect Gilbert will be happier. Thanks, Nancy. Pat Hayes ------------------------------ Date: 11 May 88 04:39:31 GMT From: glacier!jbn@labrea.stanford.edu (John B. Nagle) Subject: Re: Arguments against AI are arguments against human formalisms In article <1103@crete.cs.glasgow.ac.uk> Gilbert Cockton writes: > BTW, Robots aren't AI. Robots are robots. Rod Brooks has written "Robotics is a superset of AI". Robots have all the problems of stationary artificial intelligences, plus many more. Several of the big names in AI did work in robotics back in the early days of AI. McCarthy, Minsky, Winograd, and Shannon all did robotics work at one time. But they did it in a day when the difficulty of the problems to be faced was not recognized. There was great optimism in the early days, but even such seemingly simple problems such as grasping turned out to be very hard. Non-trivial problems such as general automatic assembly or automatic driving under any but the most benign conditions turned out to be totally out of reach with the techniques available. Progress has been made, but by inches. Nevertheless, I suspect that over the next few years, robotics will start to make a contribution to the more classic AI problems, as the techniques being developed for geometric reasoning and sensor fusion start to become the basis for new approaches to artificial intelligence. I consider robotics a very promising field at this point in time. But I must offer a caution. Working in robotics is risky. Failure is so obvious. This can be bad for your career. John Nagle ------------------------------ Date: Wed 11 May 88 14:18:27-PDT From: Conrad Bock Subject: non-AI theories about symbols For those interested in non-AI theories about symbols, the following is a very quick summary of Freud and Marx. In Freud, the prohibition on incest forces children to express their sexuality through symbolic means. Sexual desire is repressed in the unconscious, leaving the symbols to be the center of people's attention. People begin to be concerned with things for which there seems no justification. Marx observed that the act of exchanging objects in an economy forces us to abstract from our individual labors to a symbol of labor in general (money). The abstraction becomes embodied in capital, which people constantly try to accumulate, forgetting about the value of the products themselves. Both Marx and Freud use term `fetish' to refer to the process in which symbols (of sex and labor) begin to form systems that operate autonomously. In Freud's fetishism, someone may be obsessed with feet instead of actual love; in Marx, people are interested in money instead of actual work. In both cases, we lose control of something of our own creation (the symbol) and it dominates us. Conrad Bock ------------------------------ Date: 12 May 88 19:18:17 GMT From: centro.soar.cs.cmu.edu!acha@pt.cs.cmu.edu (Anurag Acharya) Subject: Re: this is philosophy ??!!? In article <86@edai.ed.ac.uk> rjc@edai.ed.ac.uk (Richard Caley) writes: >> Imagine Mr. Cockton, you are standing on the 36th floor of a building >> and you and your mates decide that you are Superman and can jump out >> without getting hurt. >Then there is something going wrong in the negotiations within the group!! Oh, yes! There definitely is! But it is still is a "negotiation" and it is "social"!. Since 'reality' and 'truth' are being defined as "negotiated outcomes of social processes", there are no constraints on what these outcomes may be. I can see no reason why a group couldn't conclude that ( esp. since physical world constraints are not necessarilly a part of these "negotiations"). >Saying that Y is the result of process X does not imply that any result >from X is a valid Y. In particular 'reality is the outcome >of social negotiation' does not imply that "real world" (whatever that is) >constraints do not have an effect. Do we have "valid" and "invalid" realities around ? >If we decided that I was Superman then presumably there is good evidence >for that assumption, since it is pretty hard to swallow. _In_such_a_case_ >I might jump. Being a careful soul I would probably try some smaller drops >first! Why would it be pretty hard to swallow ? And why do you need "good" evidence ? For that matter, what IS good evidence - that ten guys ( possibly deranged or malicious ) say so ? Have you thought why would you consider getting some real hard data by trying out smaller drops ? It is because Physical World just won't go away and the only real evidence that even you would accept are actual outcomes of physical events. Physical world is the final arbiter of "reality" and "truth" no matter what process you use to decide on your course of action. >To say you would not jump would be to say that you would not accept that >you were Superman no matter _how_ good the evidence. If you accept consensus of a group of people as "evidence", does the degree of goodness depend on the number of people, or what ? > Unless you say that the >concept of you being Superman is impossible ( say logically inconsistent with >your basic assumptions about the world ), which is ruled out by the >presuppositions of the example ( since if this was so you would never come >to the consensus that you were him ), then you _must_ accept that sufficient >evidence would cause you to believe and hence be prepared to jump. Ah, well.. if you reject logical consistency as a valid basis for argument then you could come to any conclusion/consensus in the world you please - you could conclude that you (simultaneously) were and were not Superman! Then, do you jump out or not ? ( or maybe teeter at the edge :-)) On the other hand, if you accept logical consistency as a valid basis for argument - you have no need for a crowd to back you up. Come on, does anyone really believe that if he and his pals reach a consensus on some aspect of the world - the world would change to suit them ? That is the conclusion I keep getting out of all these nebulous and hazy stuff about 'reality' being a function of 'social processes'. -- Anurag Acharya Arpanet: acharya@centro.soar.cs.cmu.edu "There's no sense in being precise when you don't even know what you're talking about" -- John von Neumann ------------------------------ Date: Fri, 13 May 88 06:00:14 EDT From: "Bruce E. Nevin" Subject: the mind of society A confession first about _The Society of Mind_: I have not yet read all of this remarkable linearized hypertext document. I do not think this matters for what I am saying here, but I certainly could be mistaken and would welcome being corrected. SOM gives a useful and even illuminating account of what might be going on in mental process. I think this account misses the mark in the following way: While SOM demonstrates in an intellectually convincing way that the experience of being a unitary self or an ego is illusory, it still assumes that individual persons are separate. It even appears to identify minds with brains in a one-one correspondence. But the logic of interacting intelligent agents applies in obvious ways to interacting persons. What then, for example, of the mind of society? (I bring to this training and experience in family therapy. Much of the subconscious mental processes and communications of family members serve to constitute and maintain the family as a homeostatic system--analogous to the traffic of SOM agents maintaining the ego, see below. Individuals--especially those out of communication with family members-- recreate relationships from their families in their relationships outside the family; and vice versa, as any parent knows. I concur with Gilbert Cockton's remarks about social context. If in alienation we pretend social context doesn't matter, we just extend the scope of that which we ignore, i.e. that which we relegate to subconsciousness.) Consider the cybernetic/constructivist view that mind is not transcendent but rather is immanent (an emergent property) in the cybernetic loop structure of what is going on. I believe SOM is consistent with this view as far as it goes, but that the SOM account could and should go farther--cf e.g. writings of Gregory Bateson, Paul Watzlawick, Maturana, Varela, and others; and in the AI world, some reaching in this direction in Winograd & Flores _Understanding Computers_. Minsky cites Freud ("or possibly Poincare'") as introducing serious consideration of subconscious thought. However, the Buddhists, for example, are pretty astute students of the mind, and have been investigating these matters quite systematically for a long time. What often happens when walking, jogging, meditating, laughing, just taking a deep sighing breath, etc is that there are temporarily no decisions to be made, no distinctions to be discriminated, and the reactive mind, the jostling crowd of interacting agents, quiets down a bit. It then becomes possible to observe the mental process more "objectively". The activity doesn't stop ("impermanence" or ceaseless change is the byword here). It may become apparent that this activity is and always was out of control. What brings an activity of an agent (using SOM terms) above the threshold between subconscious and conscious mental activity? What lets that continuing activity slip out of awareness? Not only are the activities out of control--most of them being most of the time below the surface of the ocean, so to speak, out of awareness--but even the constant process of ongoing mental and emotional states, images, and processes coming to the surface and disappearing again below the surface turns out also to be out of control. This temporary abeyance in the need to make decisions has an obvious relation to Minsky's speculation about "how we stop deciding" (in AIList V6 #98): MM> I claim that we feel free when we decide to not try further to MM> understand how we make the decisions: the sense of freedom comes from a MM> particular act - in which one part of the mind STOPs deciding, and MM> accepts what another part has done. I think the "mystery" of free will MM> is clarified only when we realize that it is not a form of decision MM> making at all - but another kind of action or attitude entirely, namely, MM> of how we stop deciding. The report here is that if you stop deciding voluntarily--hold the discrimination process in abeyance for the duration of sitting in meditation, for example, not an easy task--there is more to be discovered than the subjective feeling of freedom. Indeed, it can seem the antithesis of freedom and free will, at times! So the agents of SOM are continually churning away, and if they're predetermined it's not in any way that amounts to prediction and control as far as personal awareness is concerned. And material from this ongoing chatter continually rises into awareness and passes away out of awareness, utterly out of personal control. (If you don't believe me, look for yourself. It is a humbling experience for one wedded to intellectual rigor and all the rest, I can tell you.) Evidently, this fact of impermanence is due to there being no ego there to do any controlling. Thus, one comes experientially to the same conclusion reached by the intellectual argumentation in SOM: that there is no self or ego to control the mind. Unless perhaps it be an emergent property of the loop structures among the various agents of SOM. Cybernetics has to do with control, after all. (Are emergent properties illusory? Ilya Prigogine probably says no. But the Buddhists say the whole ball of wax is illusory, all mental process from top to bottom.) Here, the relation between "free will" and creativity becomes more accessible. Try substituting "creativity" for "free will" in all the discussion thus far on this topic and see what it sounds like. It may not be so easy to sustain the claim that "there is no creativity because everything is either determined or random." And although there is a profound relation between creativity and "reaching into the random" (cf Bateson's discussions of evolution and learning wrt double-bind theory), that relation may say more about randomness than it does about creativity. If the elementary unit of information and of mind is a difference that makes a difference (Bateson), then we characterize as random that in which we can find no differences that make a difference. Randomness is dependent on perspective. Changes in perspective and access to new perspectives can instantly convert the random to the non-random or structured. As we have seen in recent years, "chaos" is not random, since we can discern in its nonlinearity differences that make a difference. (Indeed, a cardinal feature of nonlinearity as I understand it is that small differences of input can make very large differences of output, the socalled "butterfly effect".) From the point of view of personal creativity, "reaching into the random" often means reaching into an irrelevant realm for analogy or metaphor, which has its own structure unrelated in any obvious or known way to the problem domain. (Cf. De Bono.) "Man's reach shall e'er exceed his grasp,/Else what's a meta for?" (Bateson, paraphrasing Browning.) It is interesting that the Buddhist experience of the Void--no ego, no self, no distinctions to be found (only those made by mind)--is logically equivalent to the view associated with Vedanta, Qabala, neoplatonism, and other traditions, that there is but one Self, and that the personal ego is an illusory reflection of That i.e. of God. ("No distinctions because there is no self vs other" is logically equivalent to "no distinctions because there is but one Self and no other", the singularity of zero.) SOM votes with the Buddhists, if it matters, once you drop the presumed one-one correspondence of minds with brains. On the neoplatonist view, there is one Will, and it is absolutely free. It has many centers of expression. You are a center of expression for that Will, as am I. Fully expressing your particular share of (or perspective on) that Will is the most rewarding and fulfilling thing you can do for yourself; it is in fact your heart's desire, that which you want to be more than anything else. It is also the most rewarding and beneficial thing you can possibly do for others; this follows directly from the premise that there is but one Will. (This is thus a perspective that is high in synergy, using Ruth Benedict's 1948 sense of that much buzzed term.) You as a person are of course free not to discover and not to do that which is your heart's desire, so the artifactual, illusory ego has free will too. It's "desire" is to continue to exist, that is, to convince you and everyone else that it is real and not an illusion. Whether you buy this or not, you can still appreciate and use the important distinction between cleverness (self acting to achieve desired arrangement of objectified other) and wisdom (acting out of the recognition of self and other as one whole). I would add my voice to others asking that we develop not just artificial cleverness, but artificial wisdom. Winograd & Flores again point in this direction. Bruce Nevin bn@cch.bbn.com ------------------------------ Date: 13 May 88 14:19:00 GMT From: vu0112@bingvaxu.cc.binghamton.edu (Cliff Joslyn) Subject: Alternative to Probability (was Re: this is philosophy ??!!?) In article <5459@venera.isi.edu> Stephen Smoliar writes: >We DEFINITELY need abstractions better than such logical constructs to >deal with issues such as uncertainty and belief, but it is most unclear >that probability theory is going to provide those abstractions. More >likely, we should be investigating the shortcomings of natural deduction >as a set of rules which represent the control of reasoning and consider, >instead, possibilities of alternative rules, as well as the possibility >that there is no one rule set which is used universally but that >different sets of rules are engaged under different circumstances. Absolutely right. Furthermore, such a theory exists: Fuzzy Systems Theory. Over the past fifteen years, through the work of Zadeh, Prade, Dubois, Shafer, Gaines, Baldwin, Klir, and many others, we now understand that probability measures in particular are a very special case of Fuzzy Measures in general. Belief, Plausibility, Possibility, Necessity, and Basic Probability Measures all provide alternative, and very powerful, formalisms for representing uncertainty and indeterminism. The traditional concept of 'information' itself is also recognized as a special case. Other formalisms include measures of Fuzziness, Uncertainty, Dissonance, Confusion, and Nonspecificity. These methods are having a very wide impact in AI, especially with regards to the representation of uncertainty in artificial reasoning. Primary references: Klir, George, _Fuzzy Sets, Uncertainty, and Information_, Prentice Hall, 1988, Prentice Hall. Dubois, D., Prade, H. _Fuzzy Sets and Systems: Theory and Applications_, 1980, Academic. Shafer, G., _A Mathematical Theory of Evidence_, 1976, Princeton. Zadeh, L.A. "The Role of Fuzzy Logic in the Management of Uncertainty in Expert Systems," in Gupta MM et. al., _Approximatte Reasoning in Expert Systems_, 1985, U Cal Berkley. Journals: _Fuzzy Sets and Systems_ _International J. of Approximate Reasoning_ _International J. of Man-Machine Studies_ _Information Science_ _International J. of General Systems_ -- O----------------------------------------------------------------------> | Cliff Joslyn, Cybernetician at Large | Systems Science, SUNY Binghamton, vu0112@bingvaxu.cc.binghamton.edu V All the world is biscuit shaped. . . ------------------------------ Date: 16 May 88 10:29 PDT From: hayes.pa@Xerox.COM Subject: Re: AIList Digest V7 #1 I was fascinated by the correspondence between Gabe Nault and Mott Given in vol7#1, concerning "an artifical intelligence language ... something more than lisp or xlisp." Can anyone suggest a list of features which a programming language must have which would qualify it as an "artificial intelligence language" ? Pat Hayes ------------------------------ Date: 16 May 88 18:26:11 GMT From: bwk@mitre-bedford.arpa (Barry W. Kort) Subject: Re: AIList V6 #86 - Philosophy David Sher has injected some new grist into the discussion of "responsibility" for machines and intelligent systems. I tend to delegate responsibility to machines known as "feedback control systems". I entrust them to maintain the temperature of my house, oven, and hot water. I entrust them to maintain my highway speed (cruise control). When these systems malfunction, things can go awry in a big way. I think we would have no trouble saying that such feedback control systems "fail", and their failure is the cause of undesirable consequences. The only interesting issue is our reaction. I say fix them (or improve their reliability) and get on with it. Blame and punishment are pointless. If a system is unable to respond, doesn't it make more sense to restore its ability than to merely label it "irresponsible"? --Barry Kort ------------------------------ Date: Mon, 16 May 88 19:00 MST From: DanPrice@HIS-PHOENIX-MULTICS.ARPA Subject: Sociology vs Science Debate In regard to the Sociology vs Science debate it seems to me that in business & politics the bigger & more important a decision; the less rational is the process used to arrive at that decision. Or to put it another way, emotion overrides logic every time! Question: What is the relationship between emotion and intelligence and how do we program emotion into our logical ai machines?? Do we have an ai system that can tell whether a respondant is behaving in an emotional or a logical way?? ------------------------------ Date: 26 May 88 16:27:03 GMT From: krulwich-bruce@yale-zoo.arpa (Bruce Krulwich) Subject: Assumptions in dialog (was Re: Acting irrationally) In article <180@proxftl.UUCP> tomh@proxftl.UUCP (Tom Holroyd) writes: >True communication can only occur when both parties understand what all >the symbols used to communicate mean. This doesn't mean you have to >explicitly define what you mean by "tree" every time you use the word >tree, but it's a good idea to define it once, especially if it's something >more complex than "tree" (with due respect to all sentient hardwood). This can't be true. True communications occurs whenever the two party's understanding of the words used overlap in the areas in which they are in fact being used. Every word that a speaker uses will have a large set of information associated with it in the speaker's mind, but only a small subset of that information will actually be needed to understand what the speaker is saying. The trick is for the listener to (1) have the necessary information as a subset of the information that he has about the word (which is what you are considering above), and (2) correctly choose that subset from the information he has (which is a form of the indexing problem). >>The problem comes in deciding WHAT needs to be explicitly articulated >>and what can be left in the "implicit background." That is a problem >>which we, as humans, seem to deal with rather poorly, which is why >>there is so much yeeling and hitting in the world. Au contraire, humans do this quite well. True, there are problems, but most everyday communication is quite successful. Computers at this point can't succeed at this (the indexing problem) anywhere near as well as people do (yet). >Here's a simple rule: explicitly articulate everything, at least once. >The problem, as I see it, is that there are a lot of people who, for >one reason or another, keep some information secret (perhaps the >information isn't known). You are vastly underestimating the amount of knowledge you (and everybody) have for every word, entity, and concept you know about. More likely is the idea that people have a pretty good idea of what other people know (see Wilks' recent work, for example). Again, this breaks down seemingly often, but 99% of the time people seem to be correct. Just because it doesn't take any effort to understand simple sentences like "John hit Mary" doesn't mean that there isn't alot of information selection and assumption making going on. Bruce Krulwich ------------------------------ End of AIList Digest ********************