Date: Thu 1 Dec 1988 22:33-EST From: AIList Moderator Nick Papadakis Reply-To: AIList@AI.AI.MIT.EDU Us-Mail: MIT LCS, 545 Tech Square, Rm# NE43-504, Cambridge MA 02139 Phone: (617) 253-6524 Subject: AIList Digest V8 #135 To: AIList@AI.AI.MIT.EDU Status: RO AIList Digest Friday, 2 Dec 1988 Volume 8 : Issue 135 Philosophy: Defining Machine Intelligence (6 messages) ---------------------------------------------------------------------- Date: 17 Nov 88 21:16:04 GMT From: uwslh!lishka@speedy.wisc.edu (Fish-Guts) Subject: The difference between machine and human intelligence (was: AI and Intelligence) In article <4216@homxc.UUCP> marty@homxc.UUCP (M.B.BRILLIANT) writes: > >Any definition of ``artificial intelligence'' must allow intelligence >to be characteristically human, but not exclusively so. A very good point (IMHO). I believe that artificial intelligence is possible, but that machine intelligence will probably *NOT* resemble human intelligence all that closely. My main reason for this is that unless you duplicate much of what a human is (i.e. the neural structure, all of the senses, etc.), you will not get the same result. I propose that a machine without human-like senses cannot "understand" many ideas and imagery the way a human does, simply because it will not be able to perceive its surroundings in the same way as a human. Any comments? .oO Chris Oo.-- Christopher Lishka ...!{rutgers|ucbvax|...}!uwvax!uwslh!lishka Wisconsin State Lab of Hygiene lishka%uwslh.uucp@cs.wisc.edu Immunology Section (608)262-1617 lishka@uwslh.uucp "I'm not aware of too many things... I know what I know if you know what I mean" -- Edie Brickell & the New Bohemians ------------------------------ Date: 19 Nov 88 17:06:03 GMT From: uwslh!lishka@speedy.wisc.edu (Fish-Guts) Subject: Re: Defining Machine Intelligence. In article <1111@dukeac.UUCP> sbigham@dukeac.UUCP (Scott Bigham) writes: >In article <401@uwslh.UUCP> lishka@uwslh.UUCP (Fish-Guts) writes: >>I believe that artificial intelligence >>is possible, but that machine intelligence will probably *NOT* >>resemble human intelligence... > >So how shall we define machine intelligence? More importantly, how will we >recognize it when (if?) we see it? > > sbigham A good question, to which I do not have a good answer. I *have* thought about it quite a bit, though ... however, I haven't come up with much that I am satisfied with. Here is what my current lines or thought are on this subject: Many (if not most) attempts at definitions of "machine intelligence" relate it to "human intelligence." However, I have yet to find a good definition of "human intelligence" that is less vague than a dictionary's definition. It would seem (to me at least) that AI scientists (as well as scientists in many other fields) have yet to come up with a good, working definition of "human intelligence" that most will accept. Rather, most AI people I have spoken with (including myself ;-) have a vague notion of what "human intelligence" is, or else have definitions of "human intelligence" that relies on many personal assumptions. I still do not think that the AI community has developed a definition of "human intelligence" that can be universally presented in an introductory course on AI. It is no wonder, then, that there is no commonly accepted definition of machine intelligence (which would seem to be a crucial definition in AI, IMHO). So how do we define machine intelligence? I propose that we define it apart from human intelligence at first, and try to relate it to human intelligence afterwards. In my opinion, machine intelligence does not have to be the same as human intelligence (and probably will not), for reasons I have mentioned in other articles. From what I have read here, I believe that at least a few other people in this group also feel this way. First, the necessary "features" of machine intelligence should be discussed and decided upon. It is important that this be done *without* considering current architectures and AI knowledge; the "features" should be for an ideal "machine intelligence," and not geared towards something that can be achieve in fifty years. Also, human intelligence should be *considered* at this point, but not used as a *basis* for defining machine intelligence; intelligence in other beings (mammals, birds, insects, rocks (;-), whatever) should also be considered. Second, after having figured out what we want machine intelligence to be, we should then try and come up with some good "indicators" that could be used to tell whether an AI system exhibits machine intelligence. These indicators can include specific tests, but I have a feeling that tests for any form of intelligence have never been very good indicators (note that I do not put that much value on IQ tests as measures of intelligence). Indicators of intelligence in humans and other beings should be considered here as well (i.e. what do we feel is a good sign that someone is intelligent?). After all that is done (and it may never get done ;-), then we can try and compare it to human intelligence. Chances are the two definitions of intelligence (for machines and humans) will be different. Of course, if, in looking at human intelligence, some important points of machine intelligence have been missed, then revisions are in order ... there is always time to revise the definition. I am sorry that I could not provide a concrete definition of what machine intelligence is. However, I hoped I have provided a small framework for discussions on how to go about defining machine intelligence. And of course all the above is only my view on the subject, and is subject to change; do with it what you will ... if you want to print it up and use it as bird-cage liner, well that is fine by me ;-) .oO Chris Oo.-- Christopher Lishka ...!{rutgers|ucbvax|...}!uwvax!uwslh!lishka Wisconsin State Lab of Hygiene lishka%uwslh.uucp@cs.wisc.edu Immunology Section (608)262-1617 lishka@uwslh.uucp "I'm not aware of too many things... I know what I know if you know what I mean" -- Edie Brickell & the New Bohemians ------------------------------ Date: 20 Nov 88 06:53:44 GMT From: quintus!ok@unix.sri.com (Richard A. O'Keefe) Subject: Re: Defining Machine Intelligence. In article <404@uwslh.UUCP> lishka@uwslh.UUCP (Fish-Guts) writes: > Many (if not most) attempts at definitions of "machine >intelligence" relate it to "human intelligence." However, I have yet >to find a good definition of "human intelligence" that is less vague >than a dictionary's definition. It would seem (to me at least) that >AI scientists (as well as scientists in many other fields) have yet to >come up with a good, working definition of "human intelligence" that >most will accept. Rather, most AI people I have spoken with >(including myself ;-) have a vague notion of what "human intelligence" >is, or else have definitions of "human intelligence" that relies on >many personal assumptions. I still do not think that the AI community >has developed a definition of "human intelligence" that can be >universally presented in an introductory course on AI. It is no >wonder, then, that there is no commonly accepted definition of machine >intelligence (which would seem to be a crucial definition in AI, IMHO). I think it is useful to bear in mind that "intelligence" is a _social_ construct. We can identify particular characters which are associated with it, and we may be able to measure those. (For example, one of the old intelligence tests identified knowing that Crisco (sp?) is a cooking oil as a component of intelligence.) It is _NOT_ the responsibility of AI people to define "human intelligence". It is the job of sociologists to determine how the notion of "intelligence" is deployed in various cultures, and of psychologists to study whatever aspects turn out to be based on mental characteristics of the individual. The field called "Machine Intelligence" or "Artificial Intelligence" is something which originated in a particular related group of cultures and took the "folk" notion of "intelligence" as its starting point. We wave our hands a bit, and say "you know how smart people are, and how dumb machines are, well, we want to make machines smarter." At some point we will declare victory, and whatever we have at that point, _that_ will be the definition of "machine intelligence". ("Intelligent" is already used to mean "able to perform the operations of a computer", so is "smart" in the phrase "smart card".) Let's face it, 13th century philosophers didn't have a definition of "mass", "potential field", "tensor", or even "hadron" when they started out trying to make sense of motion. They used the ordinary language they had. The definitions came _last_. There are at least two approaches to AI, which may be caricatured as (1) "Let's build a god" (2) "Let's build amplifiers for the mind" I belong to the second camp: I don't give a Continental whether we end up with "machine intelligences" or not, just so long as we end up with cognitive tools which are far more intelligible to humans than what we have now. For the first camp, the possibility of "inhuman" machine intelligences is of interest. It would definitely be a kind of success. For the second camp, something which is not close enough to the human style to be readily comprehended by an ordinary human would be an utter failure. We are still close enough to the beginnings of AI (whatever that is) that both camps can pursue their goals by similar means, and have useful things to say to each other, but don't confuse them! ------------------------------ Date: 23 Nov 88 11:49 EST From: SDEIBEL%ZEUS.decnet@ge-crd.arpa Subject: What the heck is intelligence and should we care? In Vol8 Issue 131 of the BITNET distribution of AILIST, Nick Taylor mentioned the problem of defining intelligence. This is indeed a problem: What really are we talking about when we set ourselves off from the "animals", etc? I'm not foolish enough to pretend I have any answers but did find some interesting ideas in Ray Jackendoff's book "Conciousness and the Computational Mind". Jackendoff suggests (in Chapter 2, I believe) that one fundamental characterestic of intelligence that seperates the actions of humans (and possibly, other animals) from non-intelligent systems/animals/etc is the way in which components of intelligent entities interact. The matter of interest in intelligent entities is the way in which independently acting sub-parts (e.g. neurons) interact and the way in which the states of these sub-parts combinatorily combine. On the other hand, the matter of interest in non-intelligent entities (e.g. a stomach) is the way in which the action of subparts (e.g. secreting cells) SUM into a coherent whole. While vague, this idea of intelligence as arising from complexity and the interaction of independent units seemed interesting to me in that it offers a nice and simplistic general description of intelligence. Oh, yes it could start to imply that computers are intelligent, etc, etc but one must not forget the complexity gap between the brian and the most complex computers in existence today! Rather that wrestle with the subtleties and complexities of words like "intelligence" (among others), it might be better to accept the fact that we may never be able to decide what intelligence is. How about "The sum total of human cognitive abilities" and forget about it to concentrate on deciding how humans might acheive some of their cognitive feats? Try deciding what we really mean when we say "bicycle" and you'll run into monumental problems. Why should we expect to be able to characterise "intelligence" any easier? Stephan Deibel (sdeibel%zeus.decnet@ge-crd.arpa) ------------------------------ Date: 25 Nov 88 08:55:39 GMT From: tramp!hassell@boulder.colorado.edu (Christopher Hassell) Subject: Re: Intelligent Displacement of Dirt (was: Re: Artificial Intelligence and Intelligence) In article <4561@phoenix.Princeton.EDU> eliot@phoenix.Princeton.EDU (Eliot Handelman) writes: >What I've come to admire as intelligence is the capacity to understand the >nature of one's limitations, and through that understanding to construct >alternative approaches to whichever goal one has undertaken to acheive. >Where real intelligence begins, I think, is the capacity to apply this idea >to itself, that is, the capacity to assess the machinery of discrimination >and criticism itself. I surmise that a finite level of recursion is sufficient >to justify intelligent behaviour. > >As an example, supposing that my goal is to displace an enormous pile of dirt >in the course of an afternoon. I may know that it takes me an afternoon to >displace a fraction of the total amount. The questions are, how would I know >this if I haven't tried, and how do I arrive at the idea of a shovel. I invite >discussion of this matter. On the whole subject, this does appear to be one of the better definitions of intelligence because it is self-propogating (it'll get smarter over time). I still believe that this analysis, though requiring agile thought, isn't even attempted by most of us 'intelligent' beings. We have our general inference mech's to say .. "well that possibility is still not tried" or "The outside world opened that option etc.." .. not too terribly difficult. Myself, I am a pragmatist and find sufficient evidence for `getting' intelligence from the outside world given a critical mass of inherently important initial syllogisms, (i.e. the original `how to learn' questions) I throw a verbose attempted `solution' to the world in this: One realizes the fact of a homogeneous material (dirt) needing `transport' from one place to another, and the motor recognition of gravity and its effect on the dirt (needing a bowl-like container to 'move' it) the inability of anything DIRECTLY equalling the task (no BIG auto-digging-and-moving-and-dumping-bowls to control) >From this comes the reduction that given the ability to 'integrate' over time the more human-sized act of moving 'some' dirt (homog materials) which requires the ability to break down this inhuman goal to a normal one. (this state can be more good than that original state .. so try it) (this does come from some recognition of being able to manipulate dirt at all) hands are the first suggestion but upon experimentation (remembrance too) one gets "bored" . upon this the extrapolations of 'what holds dirt' goes on towards other objects, this mixed with handyness would lead to a shovel and maybe even a wheelbarrow (a larger 'bowl' but one that can't be used directly to get dirt with) YES this is only the break-down-the-problem-to-managible-sub-parts but this is a general version for with X "resources" find a way Y fits into them upon a thing called an attempt. (Y being the problem) (yes also "resources" are nice and static too. Just change the problem to a set of responses that must propogate into themselves) I hope this gets some opinions (not all of them unfavorable?? /:-) -------------------------------------------------------------------------- In any situation the complete redefinition of the problem *IS* the answer itself, ... so let's get redefining. :-) {sunybcs, ncar, nbires}!boulder!tramp!hassell ## and oh so much of it ## #### C. H. #### ------------------------------ Date: 30 Nov 88 18:04:02 GMT From: uwslh!lishka@speedy.wisc.edu (Fish-Guts) Subject: Re: The difference between machine and human intelligence (was: AI and Intelligence) In article <960@dgbt.uucp> thom@dgbt.uucp (Thom Whalen) writes: >From article <401@uwslh.UUCP>, by lishka@uwslh.UUCP (Fish-Guts): >> I propose that a machine without human-like senses cannot "understand" >> many ideas and imagery the way a human does, simply because it will >> not be able to perceive its surroundings in the same way as a human. >> Any comments? > >Do you believe that Helen Keller "understood many ideas and imagery the >way a human does? She certainly lacked much of the sensory input that >we normally associate with intelligence. > >Thom Whalen I do not believe she *perceived* the world as most people with full senses do. I do believe she "understood many ideas and imagery" the way humans do because she had (1) touch, (2) taste, and (3) olfactory senses (she was not able to hear or see, if I remember correctly), as well as other internal sensations (i.e. sickness, pain, etc.). The way I remember it, she was taught to speak by having her "feel" the vibrations of her teacher's throat as words were said while associating the words with some sensation (i.e. the "feeling" or water as it ran over her hands). Also (and this is a highly personal judgement) I think the fact that she was a human, with a human nervous system and human reactions to other sensations (i.e. a sore stomach, human sicknesses, etc.), also added to her "human understanding." .oO Chris Oo.-- Christopher Lishka ...!{rutgers|ucbvax|...}!uwvax!uwslh!lishka Wisconsin State Lab of Hygiene lishka%uwslh.uucp@cs.wisc.edu Immunology Section (608)262-1617 lishka@uwslh.uucp "I'm not aware of too many things... I know what I know if you know what I mean" -- Edie Brickell & the New Bohemians ------------------------------ End of AIList Digest ********************