Date: Mon 18 Jul 1988 00:03-EDT From: AIList Moderator Nick Papadakis Reply-To: AIList@mc.lcs.mit.edu Us-Mail: MIT Mail Stop 38-390, Cambridge MA 02139 Phone: (617) 253-2737 Subject: AIList Digest V8 #9 To: AIList@mc.lcs.mit.edu Status: R AIList Digest Monday, 18 Jul 1988 Volume 8 : Issue 9 Today's Topics: Philosophy Reproducing the brain in low-power analog CMOS (LONG!) ASL notation Smart money AI (GOFAI) and cognitive psychology Metaepistemology & Phil. of Science Generality in Artificial Intelligence Critique of Systems Theory ---------------------------------------------------------------------- Date: 2 Jul 88 15:03:55 GMT From: uwvax!uwslh!lishka@rutgers.edu (Fish-Guts) Reply-to: uwvax!uwslh!lishka@rutgers.edu (Fish-Guts) Subject: Re: Reproducing the brain in low-power analog CMOS (LONG!) In a previous article,jbn@GLACIER.STANFORD.EDU (John B. Nagle) writes: >Date: Wed, 29 Jun 88 11:23 EDT >From: John B. Nagle >Subject: Reproducing the brain in low-power analog CMOS >To: AILIST@ai.ai.mit.edu > > > Forget Turing machines. The smart money is on reproducing the brain >with low-pwer analog CMOS VLSI. Carver Mead is down at Caltech, reverse >engineering the visual system of monkeys and building equivalent electronic >circuits. Progress seems to be rapid. Very possibly, traditional AI will >be bypassed by the VLSI people. > > John Nagle There is one catch: you do need to know what the Brain (be it human, monkey, or insect) is doing at a neuronal level. After two courses in Neurobiology, I am convinced that humans are quite far away from understanding even a fraction of what is going on. Although I do not know much about the research happening at Caltech, I would suspect that they are reverse engineering the visual system from the retina and working their way back to the visual cortex. From what I was taught, the first several stages in the visual system consist of "preprocessing circuits" (in a Professor's words), and serve to transform the signals from the rods and cones into higher-order constructs (i.e. lines, motion, etc.). If this is indeed what they are researching at Caltech, then it is a good choice, although a rather low level one. I would bet that these stages lend themselves well to being implemented in hardware, although I don't know how many "deep" issues about the Brain and Mind that implementation of these circuits will solve. One of my Professors pointed out that studying higher order visual functions is hard because of all of the preprocessing stages that occur before the visual cortex (where higher order functions are thought to occur). Since we do not yet know exactly what kinds of "data abstractions" are reaching the visual cortex, it is hard to come up with any hardcore theories of the visual cortex because noone really knows what the input to that area looks like. Sure, we know what enters the retina, and a fair bit about the workings of the rods and cones, but how the signals are transformed as they pass through the bipolar, horizontal, amacrine, and ganglion cells is not known to any great certainty. [NOTE: a fair bit seems to be known about how individual classes of neurons tranform the signals, but the overall *picture* is what is still missing]. Maybe this is what the folks at Caltech are trying to do; I am not sure. It would help a great deal in later studies of the visual system to know what kinds of inputs are reaching the visual cortex. However, there are other areas of research that do address higher order functions in cortex. The particular area that I am thinking of is the Olfactory System, specifically the Pyriform Cortex. This cortex is only one synapse away from the olfactory sensory apparatus, so the input into the Pyriform Cortex is *not* preprocessed to any great degree; much less so than the > 4 levels of preprocessing that occur in the visual system. Futhermore, the Pyriform Cortex seems to be similar in structure to several Neural Net models, including some proposed by Hopfield (who does much of his work in hardware). Therefore, it is much easier to figure out what sort of inputs are reaching the Pyriform Cortex, though even this has been elusive to a large degree. The current research indicates that certain Neural Net models effectively duplicate characteristics of the cortex to a good degree. I have read several interesting articles on the storage and retrieval of previous odors in the Pyriform Cortex; these papers use Content-Addressable Memory (also called Associative Memory) as the predominant model. There is even some direct modelling being done by a couple of Grad. students on a Sun workstatoin down in California. [If anyone wants these references, send me email; if there is enough interest, I will post them]. My point is: do not write off AI (especially Neural Net) theories so fast; there is much interesting work being done right now that has a much potential as being important in the long run as the work being done at Caltech. Just because they are implementing Brain *circuits* and *architecture* in hardware does not mean they will get any closer than AI researchers. I still believe that AI researchers are doing much more higher order research than anything that has been implemented in hardware. Furthermore, the future of AI does not lie only in Artificial Intelligence and Computer Science. You bring up a good point: look at other disciplines as well; your example was hardware. But look even further: Neurobiology is a fairly important area of reasearch in this area. So are the Computational sciences: Hopfield does quite a bit of research here. Furthermore, Hopfield is (or was, at least) a Physicist; in fact I saw him give a lecture here at the UW about his networks where he related his work to AI, Neurobiology, Physics, Computational Sciences, and more...and the talk took place (and was sponsored by) the Physics department. Research into AI and Brain related fields is being performed in many discplines, and each will influence all others in the end. Whew! Sorry to get up on my soapbox, but I had to let it out. Remember, these are just my humble opinions, so don't take them too seriously if you do not like them. I would be happy to discuss this further over email, and I can give you references to some interesting articles to read on ties between AI and Neurobiology (almost all dealing with the Olfactory System, as that looks the most promising from my point of view). -Chris-- Christopher Lishka | lishka@uwslh.uucp Wisconsin State Lab of Hygiene | lishka%uwslh.uucp@cs.wisc.edu Immunology Section (608)262-1617 | ...!{rutgers|ucbvax|...}!uwvax!uwslh!lishka "...Just because someone is shy and gets straight A's does not mean they won't put wads of gum in your arm pits." - Lynda Barry, "Ernie Pook's Commeek: Gum of Mystery" ------------------------------ Date: 3 Jul 88 18:47:04 GMT From: doug@feedme.UUCP (Doug Salot) Reply-to: doug@feedme.UUCP (Doug Salot) Subject: ASL notation hayes.pa@XEROX.COM writes: >On dance notation: >A quick suggestion for a similar but perhaps even thornier problem: a notation >for the movements involved in deaf sign language. I'm not sure how this would benefit the deaf community, as most signers are quite capable of reading written language, however work has been done to "quantize" ASL. Look to the work of Poizner, Bellugi, et al at the Salk Institute for suggested mappings between several movement attributes (planar locus, shape, iteration freq., etc) and language attributes (case, number, etc.). --- Doug Salot || doug@feedme.UUCP || ...{trwrb,hplabs}!felix!dhw68k!feedme!doug "Thinking: The Thinking Man's Sport" ------------------------------ Date: Thu, 7 Jul 88 09:09:00 pdt From: Ray Allis Subject: Smart money John Nagle says: > Forget Turing machines. The smart money is on reproducing the brain >with low-power analog CMOS VLSI. Carver Mead is down at Caltech, reverse >engineering the visual system of monkeys and building equivalent electronic >circuits. Progress seems to be rapid. Very possibly, traditional AI will >be bypassed by the VLSI people. Hear, hear! Smart money indeed! I don't know whether Carver prefers to work low-profile, but this has to be the most undervalued AI work of the century. This is the most likely area for the next breakthrough toward creating intelligent artifacts. And the key concept is that these devices are ANALOG. Ray Allis Boeing Computer Services-Aerospace Support CSNET: ray@boeing.com ------------------------------ Date: Tue, 12 Jul 88 12:48:25 EDT From: Ralf.Brown@b.gp.cs.cmu.edu To: comp-ai-digest@rutgers.edu Path: b.gp.cs.cmu.edu!Ralf.Brown@B.GP.CS.CMU.EDU From: Ralf.Brown@B.GP.CS.CMU.EDU Newsgroups: comp.ai.digest Subject: Re: Generality in Artificial Intelligence Message-ID: <22d9eea5@ralf> Date: 12 Jul 88 10:49:09 GMT Sender: ralf@b.gp.cs.cmu.edu Lines: 29 In-Reply-To: <19880712044954.9.NICK@HOWARD-JOHNSONS.LCS.MIT.EDU> In a previous article, YLIKOSKI@FINFUN.BITNET writes: }Thus, it would seem that in order for us to see true commonsense }knowledge exhibited by a program we need: } } * a vast amount of knowledge involving the world of a person } in virtual memory. The knowledge involves gardening, } Buddhism, the emotions of an ordinary person and so forth - } its amount might equal a good encyclopaedia. Actually, it would probably put the combined text of several good encyclopedias to shame. Even encyclopedias and dictionaries leave out a lot of "common-sense" information. The CYC project at MCC is a 10-year undertaking to build a large knowledge base of real world facts, heuristics, and methods for reasoning over the knowledge base.[1] The current phase of the project is to carefully represent 400 articles from a one-volume encyclopedia. They expect their system to contain about 10,000 frames once they've encoded the 400 articles, about half of them common-sense concepts. [1] CYC: Using Common Sense Knowledge to Overcome Brittleness and Knowledge Acquisition Bottlenecks, AI Magazine v6 #4. -- UUCP: {ucbvax,harvard}!cs.cmu.edu!ralf -=-=-=- Voice: (412) 268-3053 (school) ARPA: ralf@cs.cmu.edu BIT: ralf%cs.cmu.edu@CMUCCVMA FIDO: Ralf Brown 1:129/31 Disclaimer? I |Ducharm's Axiom: If you view your problem closely enough claimed something?| you will recognize yourself as part of the problem. ------------------------------ Date: 13 Jul 88 04:48:34 GMT From: pitt!cisunx!vangelde@cadre.dsl.pittsburgh.edu (Timothy J Van) Subject: AI (GOFAI) and cognitive psychology What with the connectionist bandwagon, everyone seems to be getting a lot clearer about just what AI is and what sort of a picture of cognition it embodies. The short story, of course, is that AI claims that thought in general and intelligence in particular is the rule governed manipulation of symbols. So AI is committed to symbolic representations with a combinatorial syntax and formal rules defined over them. The implemenation of those rules is computation. Supposedly, the standard or "classical" view in cognitive psychology is committed to exactly the same picture in the case of human cognition, and so goes around devising models and experiments based on these commitments. My question is - how much of cognitive psychology literally fits this kind of characterization? Some classics, for example the early Shepard and Metzler experiments on image rotation dont seem to fit the description very closely at all. Others, such as the SOAR system, often seem to remain pretty vague about exactly how much of their symbolic machinery they are really attributing to the human cognizer. So, to make my question a little more concrete - I'd be interested to know what people's favorite examples of systems that REALLY DO FIT THE DESCRIPTION are? (Or any other interesting comments, of course.) Tim van Gelder ------------------------------ Date: 14 Jul 88 09:55:40 EDT From: David Chess Subject: Metaepistemology & Phil. of Science >In this sense the reality is unknowable. We only have >descriptions of the actual world. This "only" seems to be the key to the force of the argument. If it were "we have descriptions of the actual world", it would sound considerably tamer. The "only" suggests that there is something else (besides "descriptions") that we *might* have, but that we do not. What might this something else be? What, besides "descriptions", could we have of the actual world? I certainly wouldn't want the actual world *itself* in my brain (wouldn't fit). Can anyone complete the sentence "The actual world is unknowable to us, because we have only descriptions/representations of it, and not..."? (I would tend to claim that "knowing" is just (roughly) "having the right kind of descriptions/representations of", and that there's no genuine "unknowability" here; but that's another debate...) Dave Chess Watson Research * Disclaimer: Who, me? ------------------------------ Date: Thu, 14 Jul 88 15:06:53 BST From: mcvax!doc.ic.ac.uk!sme@uunet.UU.NET Reply-to: sme@doc.ic.ac.uk (Steve M Easterbrook) Subject: Re: Generality in Artificial Intelligence In a previous article, YLIKOSKI@FINFUN.BITNET writes: >> "In my opinion, getting a language for expressing general >> commonsense knowledge for inclusion in a general database is the key >> problem of generality in AI." >... >Here follows an example where commonsense knowledge plays its part. A >human parses the sentence > >"Christine put the candle onto the wooden table, lit a match and lit >it." > >The difficulty which humans overcome with commonsense knowledge but >which is hard to a program is to determine whether the last word, the >pronoun "it" refers to the candle or to the table. After all, you can >burn a wooden table. > >Probably a human would reason, within less than a second, like this. > >"Assume Christine is sane. The event might have taken place at a >party or during her rendezvous with her boyfriend. People who do >things such as taking part in parties most often are sane. > >People who are sane are more likely to burn candles than tables. > >Therefore, Christine lit the candle, not the table." Aren't you overcomplicating it a wee bit? My brain would simply tell me that in my experience, candles are burnt much more often than tables. QED. This to me, is very revealing. The concept of commonsense knowledge that McCarthy talks of is simply a huge base of experience built up over a lifetime. If a computer program was switched on for long enough, with a set of sensors similar to those provided by the human body, and a basic abililty to go out and do things, to observe and experiment, and to interact with people, it would be able to gather a similar set of experiences to those possessed by humans. The question is then whether the program can store and index those experiences, in their totality, in some huge episodic memory, and whether it has the appropriate mechanisms to fire useful episodic recalls at useful moments, and apply those recalls to the present situation, whether by analogy or otherwise. >From this, it seems to me that the most important task that AI can address itself to at the present is the study of episodic memory: how it can be organised, how it can be accessed, and how analogies with past situations can be developed. This should lead to a theory of experience, ready for when robotics and memory capacities are advanced enough for the kind of exeriment I descibed above. With all due respect to McCarthy et al, attempts to hand code the wealth of experience of the real world that adult human beings have accumulated are going to prove futile. Human intelligence doesn't gather this commonsense by being explicitly programmed with rules (formal OR informal), and neither will artificial intelligences. >It seems to me that the inferences are not so demanding but the >inferencer utilizes a large amount of background knowledge and a good >associative access mechanism. Yes. Work on the associative access, and let the background knowledge evolve itself. >... >What kind of formalism should we use for expressing the commonsense >knowledge? Try asking what kind of formalism should we use for expressing episodic memories? Later on you suggest natural language. Is this suitable? Do people remember things by describing them in words to themselves? Or do they just create private "symbols", or brain activation patterns, which only need to be translated into words when being communicated to others? Note: I am not saying people don't think in natural language, only that they don't store memories as natural language accounts. I don't think this kind of experience can be expressed in any formalism, nor do I think it can be captured by natural language. It needs to evolve as a private set of informal symbols, of which the brain (human or computer) does not need to consciously realise are there. All it needs to do is associate the right thought with the symbols when they are retrieved, i.e. to interpret the memories. Again I think this kind of ability evolves with experience: at first, symbols (brain activity patterns) would be triggered which the brain would be unable to interpret. If this is beginning to sound like an advocation of neural nets/connectionist learning, then so be it. I feel that a conventional AI system coupled to a connectionist net for its episodic memory might be a very sensible achitecture. There are probably other ways of achieving the same behaviour, I don't know. One final note. Read the first chapter of the Dreyfus and Dreyfus book "Mind over Machine", for a thought-provoking account of how experts perform, using "intuition". Logic is only used in hindsight to support an intuitive choice. Hence "heuristics is compiled hindsight". Being arch-critics of AI, the Dreyfuses conclude that the intuition that experts develop is intrinsically human and can never be reproduced by machine. Being an AI enthusiast, I conclude that intuition is really the unconscious application of experience, and all that's needed to reproduce it is the necessary mechanisms for storing and retrieving episodic memories by association. >In my opinion, it is possible that an attempt to express commonsense >knowledge with a formalism is analogous to an attempt to fit a whale >into a tin sardine can. ... I agree. How did you generate this analogy? Is it because you had a vast amount of common sense knowledge about whales and sardines, and tins, (whether expressed in natural language (vast!), or some formal system) through which you had to wade to realise that sardines will fit in small tins but whales will not, and eventually synthesis this particular concept, or did you just try to recall (holistically) an image that matched the concept of forcing a huge thing into a rigid container? Steve Easterbrook. ------------------------------ Date: Fri, 15 Jul 88 13:55:04 PDT From: larry@VLSI.JPL.NASA.GOV Subject: Philosophy: Critique of Systems Theory -- Using Gilbert Cockton's references to works critical of systems theory, over the last month I've spent a few afternoons in the CalTech and UCLA libraries tracing down those and other criticisms. The works I was able to get and study are at the end of this message. I also examined a number of introduc- tory texts on psychology and sociology from the last 15 years or so. General Systems Theory was founded by biologist Ludwig von Bertallanfy in the late '40s. It drew heavily on biology, borrowed from many areas, and promised a grand unified theory of all the sciences. The ideas gained momentum till in the early '70s in the "humanics" or "soft sciences" it had reached fad proportions. Bertallanfy was made an honorary psychoanalyst, for instance, and a volume appeared containing articles by various prominent analysts discussing the effects of GST on their field. After that peak, interest died down. Indexes in social-sciences textbooks carried fewer references to smaller discussions. The GST journal became thinner and went from annual to biannual; the latest issue was in 1985. Interest still exists, however, and in various bibliographies of English publications for 1987 I found at total of seven new books. What seems to have happened is that the more optimistic promises of GST failed and lost it the support of most of its followers. Its more success- ful ideas were pre-empted by several fields. These include control theory in engineering, taxonomy of organizations in management, and the origins of psychosis in social psychology. For me the main benefit of GST has been a personally satisfactory resolution of the reduction paradox, which follows. Because of the limits of human intelligence, we simplify and view the universe as various levels of abstraction, each level fairly independent of the levels "above" and "below" it. This gives rises to arguments, however, about whether, for instance, physics or psychology is "truer." The simple answer is that both are only approximations of an ultimately unknowable reality and, since both views are too useful give up, their incompatibility is inevitable and we just have to live with it. This is what many physi- cists have done with the conflict between quantum and wave views of energy. The GST view is that each higher level of organization is built on a previous one via systems, which are new kinds of units made from binding together more elementary units. New kinds of systems exhibit synergy: attributes and abilities not observed in any of their constituent elements. But where do these attributes/abilities come from? I found no answer in the writings on system theory that I read, so I had to make my own answer. I finally settled on what interaction effects. Two atoms bound together chemically affect each other's electron shrouds, forming a new shroud around them both. This gives the resulting molecule a color, solubility, conduc- tivity, and so on that neither solitary atom has. Similarly, two people can cross a wall neither can alone by one standing on the other's shoulders to reach the top, then pulling the bottom partner up. A "living" machine can repair and reproduce itself. And consciousness can arise from elements that don't have it -- memories, processors, sensors, and effectors -- though no amount of study of individual elements will find life or consciousness. They are the result of interaction effects. _System Analysis in Public Policy: a Critique_, I. Hoos, 1972 _Central Problems in Social Theory_, Anthony Giddens, 1979 _System Theory, System Practice_, P. B. Checkland, 1981 _On Systems Analysis_, David Berlinski, 1976 _The Rise of Systems Theory_, R. Lilienfeld, 1978 The last two are reviewed in _Futures_, 10/2, p. 159 and 11/2, p. 165. They also contain criticisms of system theory which, they complain, Berlinski and Lilienfeld overlooked. ------------------------------ End of AIList Digest ********************