Date: Tue 28 Jun 1988 23:05-EDT From: AIList Moderator Nick Papadakis Reply-To: AIList@AI.AI.MIT.EDU Us-Mail: MIT Mail Stop 38-390, Cambridge MA 02139 Phone: (617) 253-2737 Subject: AIList Digest V7 #46 To: AIList@AI.AI.MIT.EDU Status: RO AIList Digest Wednesday, 29 Jun 1988 Volume 7 : Issue 46 Today's Topics: Philosophy: replicating the brain with a Turing machine Deep Thought. possible value of AI H. G. Wells Who else isn't a science? metaepistemology questions and answers about meta-epistemology ---------------------------------------------------------------------- Date: Sat, 25 Jun 88 21:16 O From: Subject: replicating the brain with a Turing machine Distribution-File: AILIST@AI.AI.MIT.EDU In AIList Digest V7 #29, agate!garnet!weemba@presto.ig.com (Obnoxious Math Grad Student) writes: >In article <517@dcl-csvax.comp.lancs.ac.uk>, simon@comp (Simon Brooke) writes: >>[...] >>If all this is so, then it is possible to exactly reproduce the workings >>of a human brain in a [Turing machine]. > >Your argument was pretty slipshod. I for one do not believe the above >is even possible in principle. Why? You must / at least should have a basis for the opinion. One possibility I can think of is the dualist position: we have a spirit but don't know how to make a machine with one. Any other Dualists out there? Andy Ylikoski ------------------------------ Date: Sun, 26 Jun 88 13:10:51 +0100 From: "Gordon Joly, Statistics, UCL" Subject: Deep Thought. Kurt Thearling quotes a quote in AIList Digest V7 #45. > An interesting quote from the article is "Fredkin believes that > the universe is very literally a computer and that it is being > used by someone, or something, to solve a problem." Excuse my ignorance, but I am not able to judge who is plagiarising whom. Douglas Adams invented a computer which does just this in "The Hitch Hikers Guide to the Galaxy". Most people regard this novel as a work of fiction. Gordon Joly. ------------------------------ Date: 26 Jun 88 15:41:38 GMT From: uwvax!uwslh!lishka@rutgers.edu (Fish-Guts) Reply-to: uwvax!uwslh!lishka@rutgers.edu (Fish-Guts) Subject: Re: possible value of AI In a previous article, DJS%UTRC@utrcgw.utc.COM writes: >Gilbert Cockton writes: > >"... Once again, what the hell can a computer program tell us about >ourselves? Secondly, what can it tell us that we couldn't find out by >studying people instead?" > >What do people use mirrors for? What the hell can a MIRROR tell us about >ourselves? Secondly, what can it tell us that we couldn't find out by >studying people instead? Personally, I think everything can tell us something about ourselves, be it mirror, computer, or rock. Maybe it depends on what one expects to find? > Isn't it possible that a computer program could have properties >which might facilitate detailed self analysis? I believe some people >have already seen the dim reflection of true intelligence in the primitive >attempts of AI research. Hopefully all that is needed is extensive >polishing and the development of new tools. Whether or not there has been a "dim reflection of true intelligence in the primitive attempts of AI research" is an opinion, varying greatly with who you talk to. I would think that there are "some people" who have already seen "true intelligence" (however bright or dim) in anything and everything. Again, their opinions. > David Sirag > UTRC The above are my opinions, although they might be those of my cockatiels as well. -Chris -- Christopher Lishka | lishka@uwslh.uucp Wisconsin State Lab of Hygiene | lishka%uwslh.uucp@cs.wisc.edu Immunology Section (608)262-1617 | ...!{rutgers|ucbvax|...}!uwvax!uwslh!lishka "...Just because someone is shy and gets straight A's does not mean they won't put wads of gum in your arm pits." - Lynda Barry, "Ernie Pook's Commeek: Gum of Mystery" ------------------------------ Date: Sun, 26 Jun 88 13:59:24 -0400 (EDT) From: David Greene Subject: Re: H. G. Wells >I was disappointed to see no reaction to John Cugini's quotation from >H. G. Wells. Is no one willing to admit that things haven't changed >since 1906? Of course things have changed... there are now far more than 4000 "scientists" in Washington and Herbert George Wells is dead. ------------------------------ Date: 26 Jun 88 18:17:37 GMT From: pasteur!agate!garnet!weemba@ames.arpa (Obnoxious Math Grad Student) Subject: Re: Who else isn't a science? In article , now expired here, ???? asked me for references. I find this request strange, since at least one of my references was in the very article being replied to, although not spelled out as such. Anyway, let me recommend the following works by neurophysiologists: G M Edelman _Neural Darwinism: The Theory of Neuronal Group Selection_ (Basic Books, 1987) C A Skarda and W J Freeman "How brains make chaos in order to make sense of the world", _Behavorial and Brain Sciences_, (1987) 10:2 pp 161-195. These researchers start by looking at *real* brains, *real* EEGs, they work with what is known about *real* biological systems, and derive very intriguing connectionist-like models. To me, *this* is science. GME rejects all the standard categories about the real world as the start- ing point for anything. He views brains as--yes, a Society of Mind--but in this case a *biological* society whose basic unit is the neuronal group, and that the brain develops by these neuronal groups evolving in classical Darwinian competition with each other, as stimulated by their environment. CAS & WJF have developed a rudimentary chaotic model based on the study of olfactory bulb EEGs in rabbits. They hooked together actual ODEs with actual parameters that describe actual rabbit brains, and get chaotic EEG like results. ------------------------------------------------------------------------ In article <34227@linus.UUCP>, marsh@mitre-bedford (Ralph J. Marshall) writes: > "The ability to learn or understand or to deal with new or > trying situations." >I'm not at all sure that this is really the focus of current AI work, >but I am reasonably convinced that it is a long-term goal that is worth >pursuing. Well, sure. So what? Everyone's in favor of apple pie. ------------------------------------------------------------------------ In article <2618@mit-amt.MEDIA.MIT.EDU>, bc@mit-amt (bill coderre) writes: >Oh boy. Just wonderful. We have people who have never done AI arguing >about whether or not it is a science [...] We've also got what I think a lot of people who've never studied the philosophy of science here too. Join the crowd. >May I also inform the above participants that a MAJORITY of AI >research is centered around some of the following: >[a list of topics] Which sure sounded like programming/engineering to me. > As it happens, I am doing simulations of animal >behavior using Society of Mind theories. So I do lots of learning and >knowledge acquisition. Well good for you! But are you doing SCIENCE? As in: If your simulations have only the slightest relevance to ethology, is your advisor going to tell you to chuck everything and try again? I doubt it. ucbvax!garnet!weemba Matthew P Wiener/Brahms Gang/Berkeley CA 94720 ------------------------------ Date: Mon, 27 Jun 88 09:18:11 EDT From: csrobe@icase.arpa (Charles S. Roberson) Subject: Re: metaepistemology Assume the "basic structure of the world is unknowable" [JMC@SAIL.Stanford.edu] and that we can only PERCEIVE our world, NOT KNOW that what we perceive is ACTUALLY how the world is. Now imagine that I have created an agent that interacts with *our* world and which builds models of the world as it PERCEIVES it (via sensors, nerves, or whatever). My question is this: Where does this agent stand, in relation to me, in its perception of reality? Does it share the same level of perception that I 'enjoy' or is it 'doomed' to be one level removed from my world (i.e. is its perception inextricably linked to my perception of the world, since I built it)? Assume now, that the agent is so doomed. Therefore, it may perceive things that are inconsistent with the world (though we may never know it) but are consistent with *my* perception of the world. Does this imply that "true intelligence" is possible if and only if an agent's perception is not nested in the perception of its creator? I don't think so. If it is true that we cannot know the "basic structure of the world" then our actions are based solely on our perceptions and are independent of the reality of the world. I believe we all accept perception as a vital part of an intelligent entity. (Please correct me if I am wrong.) However, a flawed perception does not make the entity any less intelligent (does it?). What does this say about the role of perception to intelligence? It has to be there but it doesn't have to function free of original bias? Perhaps, we have just created an agent that perceives freely but it can only perceive a sub-world that I defined based on my perceptions. Could it ever be possible to create an agent that perceives freely and that does not live in a sub-world? -chip +-------------------------------------------------------------------------+ |Charles S. Roberson ARPANET: csrobe@icase.arpa | |ICASE, MS 132C BITNET: $csrobe@wmmvs.bitnet | |NASA Langley Rsch. Ctr. UUCP: ...!uunet!pyrdc!gmu90x!wmcs!csrobe| |Hampton, VA 23665-5225 Phone: (804) 865-4090 +-------------------------------------------------------------------------+ ------------------------------ Date: Mon, 27 Jun 88 20:57 O From: Subject: questions and answers about meta-epistemology Distribution-File: AILIST@AI.AI.MIT.EDU Here come questions by csrobe@icase.arpa (Charles S. Roberson) and my answers to them: >Assume the "basic structure of the world is unknowable" >[JMC@SAIL.Stanford.edu] and that we can only PERCEIVE our >world, NOT KNOW that what we perceive is ACTUALLY how the >world is. > >Now imagine that I have created an agent that interacts >with *our* world and which builds models of the world >as it PERCEIVES it (via sensors, nerves, or whatever). > >My question is this: Where does this agent stand, in >relation to me, in its perception of reality? Does it >share the same level of perception that I 'enjoy' or is >it 'doomed' to be one level removed from my world (i.e. >is its perception inextricably linked to my perception >of the world, since I built it)? It has the perceptory and inferencing capabilities you designed and implemented, unless you gave it some kind of self-rebuilding or self-improving capability. Thus its perception is linked to your world. >Does this imply that "true intelligence" is possible >if and only if an agent's perception is not nested >in the perception of its creator? I don't think so. I also don't think so. The limitation of the perception of the robot being linked to the designer of the robot is unessential I think. >I believe we all accept perception as a vital part of an >intelligent entity. (Please correct me if I am wrong.) Perception is extremely essential. All observation of reality takes place by means of perception. >However, a flawed perception does not make the entity any >less intelligent (does it?). What does this say about >the role of perception to intelligence? It has to be >there but it doesn't have to function free of original >bias? A flawed perception can be lethal for example to an animal. Perception is a necessary requirement. It can be argued, though, that all human perception is biased (our education influeces how we interpret that which we perceive). >Perhaps, we have just created an agent that perceives >freely but it can only perceive a sub-world that I >defined based on my perceptions. Could it ever be >possible to create an agent that perceives freely and >that does not live in a sub-world? Yes, at least if the agent has the capability to extend itself for example by being able to redesign and rebuild itself. How much computational power in the Turing machine sense this capability requires is an interesting theoretical question which may already have been studied by the theoreticians out there. Andy Ylikoski ------------------------------ End of AIList Digest ********************