Artificial Intelligence

I study artificial intelligence, 'cause I have none of my own.

Artificial Intelligence (AI) comes in two varieties: Symbolic Processing, which is all about playing with set theoretic datatypes, and Neural Networks, which is all about acquiring stochastic datatypes. Regrettably, neither proves terribly intelligent except in the most limited domains, in which they're nothing more than fancy searching/sorting devices - really no more intelligent than a sieve. The trouble is that the moment you try to make 'em more than that, you hit a Combinatorial Explosion that kills you. So AI remains the stuff of marketing and pipe dreams.

At CMU, the Symbolic Processing people were called "Neats" and the Neural Networks types were the "Scruffies".

For a hopeful attempt at the Symbolic Processing variety, Semantic Network flavor, try Concepts Web.


What happened with Fuzzy Logic and Genetic Algorithms?

Is FuzzyLogic neat or scruffy? ''neither'' Is a GeneticAlgorithm neat or scruffy? ''neither''

Neither Fuzzy Logic nor Genetic Algorithms are inference systems. However, the use of Fuzzy Logic as part of an inference system, and the systematic use of genetic algorithms to develop inference heuristics, would probably appeal more to the scruffies.

What happened to them when? They're still in widespread use.

I've never seen a satisfactory explanation of fuzzy logic having descriptive power beyond (a subset of) what you can do with statistics. It may be a different way of looking at things, but doesn't seem to do anything new, really. It also doesn't seem to be a big win as a way of looking at things, either. So I've always been a bit puzzled by its popularity. It certainly isn't any gain in the direction of `artificial intelligence'.

Fuzzy Logic and Bayesian Logic (probability-logic) are orthogonal, and both can be used simultaneously in an inference system. That which can be described with Fuzzy Logic cannot be described with statistics. However, Fuzzy Logic is of questionable semantic value... saying that something (a bar stool) is '40% chair' is far less meaningful than saying 'it has 3 legs in a tripod formation supporting a single disk-shaped surface horizontally to the ground at a height slightly higher than has been recognized as suitable for a chair'. The latter allows a great deal more -correct- inference, whereas '40% chair' has semantic issues (what does 40% chair mean to you?) and doesn't lead to any sort of meaningful inference. I'll certainly agree with your statement that it is hardly a big win. Its primary examples are prototype-words (tall, short, warm, cool, fat, thin), but even these have semantic issues (60% warm???) and may be better handled by bayesian logics (I'm 90% confident that Bob would say it's 'warm', and I'm his thermostat, so I'll say it's warm.)


This should be called Artificial Human Intelligence. That is what the Turing Test is about as well; testing whether there is a perceptible difference between a human and a computer. This is an anthropocentric view of the world - not that this is necessarily a Bad Thing, but is is somewhat limiting. Would Bacteria classify the humans they live in as intelligent? Probably not. Would we classify the earth we live on or our Solar System as being intelligent? We do not now, but it may be intelligent in such a fundamentally different way that we will never understand it. -- Aalbert Torsius


I've often thought that Artificial Stupidity was much more interesting from a scientific perspective. Systems that are consistently wrong in interesting ways tell us more about the nature of intelligent behaviour than overly engineered Artificial Intelligence Big Science. In other words you would learn more about natural language by making a machine that could make the same mistakes that a two year old does while learning to speak than making one that perfectly recognizes a total of six words spoken by one speaker in clear well separated tones. -- Larry Price

When I was studying AI, I always explained to people that the real challenge was trying to achieve Artificial Common Sense.

While chatting in the office today we came to an interesting conclusion: AI is doomed to failure, because every apparent success is immediately written off as "conventional programming". To play chess as good as a human does was once an AI goal, who's impressed now? To prove theorems was once an AI goal, now theorem provers appear as the solutions to problems in programming textbooks. And so on. Poor old AI community, every time they get something right, it gets taken away from them.

The rest of the IT community demand miracles, that is, seemingly impossible things, from the AI folk (who do themselves no favours re managing expectations), and so anything that turns out to be definitely possible after all is reclassified as "not a miracle" and so doesn't count as "real" AI. The Ai community is, therefore, never going to be able to produce anything that other will look at and say "Yes! that is artificial intelligence". It'll always be: oh yeah, well that didn't turn out to be so hard after all, then.

Hmm, maybe there was always a bit of an incompatibility between the Turing Test as a goal for software within fifty years (I think Turing actually said less) and Do The Simplest Thing That Could Possibly Work. AI was always going to be hoist with its own petard. [What?] [ from www.m-w.com; hoist with one's own petard: victimized or hurt by one's own scheme ]

Turing said (in Computing Machinery And Intelligence, 1950):

[...] The original question, 'Can machines think?' I believe to be too meaningless to deserve discussion. Nevertheless I believe that at the end of the [20th] century the use of words and general educated opinion will have altered so much that one will be able to speak of machines thinking without expecting to be contradicted. [...]

I have always been of the opinion that AI systems were not going to be really intelligent until they started reaching conclusions whereby even the authors of the system could not understand the "thought process" used to reach the conclusions. We could thus be in awe of the unfathomable intelligence of the machine. Frankly. We are there. My word processor is stubborn and pig headed. It frequently decides to do things that I do not want it to do. :)

In order to understand what is going on one needs a lesson from history. 30 or forty years ago there was an argument about human intelligence. The chronology is enlightening.

[The lies you tell in order to make your point are enlightening. The point can be made without the falsehoods.]

Monkeys did not have human intelligence because they did not have language.

Monkeys (apes) were found to have language and construct novel sentences according to a grammar.

Monkeys were found to be capable of lying. This indicates that they had an awareness that the human at the other end of the conversation was capable of misunderstanding. This strongly suggests self awareness/sentience.

Move the goal posts. Monkeys do not have human intelligence because they do not use tools.

Monkeys were observed to use tools.

Monkeys do not have human intelligence because they do not make tools

Monkeys and a species of bird were observed to make tools.

silence.

monkeys were also observed to kidnap, exact revenge, have wars, and all sorts of other human traits.

I suspect monkeys are running the current "human" wars.

(I think you mean Ape, not monkey.)

My suggestion. Start believing in artificial intelligence before the computers declare war on you just to prove a point! :)

As an implementer of intelligent software, let me say that it's OK - I gave up worrying about where other people placed the goalposts years ago.

This is a Red Herring. The goalposts have never changed, just our perception of them. It's a trivial observation that apes and monkeys aren't like human beings. What exactly that means is ill-defined. The Map Is Not The Territory. Our understanding of the difference between human beings and mere animals can change without that difference ever changing.


Monkeys were found to have language and construct novel sentences according to a grammar.

Really? I'd be interested in a reference. So far as I know, the only apes that "construct novel sentences according to a grammar" are those of the species Homo sapiens. -- Gareth Mc Caughan

The following seems to contain relevant information: www.pigeon.psy.tufts.edu

Interesting stuff, but I wouldn't describe anything it reports as "constructing novel sentences according to a grammar". Would you?

Being able to talk at all is pretty much forming novel sentences according to a grammar, isn't it? In any case, the chimps were capable of forming new compounds - eg "cryhurtfood." I can't see what more you might need.

I think it has already been proven that apes/monkeys (Don't know the difference, sorry) can understand and use human language (Koko the gorilla anyone?). So I don't think it is much of a stretch to assume it is possible for apes to develop their own languages. I think it is rather arrogant to imply that Homo sapiens are the only species of sentient animal - the fact is we don't know to the extent most animals are intelligent, and quite a few have been shown to be self-aware. Heck, my dog can understand quite a few words (close to a hundred). Sorry if I sound angry. -- Timothy Seguine

Sorry, but it is ludicrous for apes and monkeys (apes are big, monkeys are small) to develop their own language. Don't you think that if it were at all possible for them to do so, they would have done it in the last tens of millions of years?

The fact is that apes and monkeys simply never create words, they don't expand its vocabulary. This is in sharp contrast to human beings, who will create entire new languages when young if the ones they're exposed to don't suit their needs. How did you think sign languages were invented? Well, they were invented multiple times by children who had been born deaf, without any exposure to any pre-existing sign language.

The only thing human beings require to learn a language (creating it if necessary) is exposure to sounds or signs. Not words, just sounds or signs. That's pretty freaking remarkable when you think about it. And it's infinitely more than those stupid apes have ever amounted to.

Homo sapiens are the only species of sentient animal on this planet. That's not arrogance, it's well-investigated, well-documented fact. I think it's pretty arrogant of you to devalue our species just because you don't like it. -- rk

''Seems like I have hit on a sore spot for most people here. Well first of all, I'd like to remark that the comment about apes creating their own languages was mostly a for-instance -- I never meant to imply that was an actual fact. As for sentience in animals, we can look to a standardized definition of what sentience is. I checked my handy-dandy dictionary and the primary definition read something like this: "Having sense perception; conscious." I then looked up the word conscious: "Having an awareness of one's environment and one's own existence, sensations, and thoughts." So sentience breaks down to being self-aware. It has been shown by scientists that a few species of primate(chimpanzees, bonobos, and orang-utans most notably) are in fact self-aware - I have read about it on multiple occasions and have seen it on television more than a few. The science books you grew up reading in high school and college and such probably would not agree with me, but most of those are outdated. Here are some references from a couple seconds searching(Please excuse my formatting errors): www.abc.net.au books.cambridge.org homepages.uc.edu I don't want to start a flame-war here but, if one claims something is well-documented fact, one should provide a reference to said documented fact. They do change after all. Your thoughts?'' -- Timothy Seguine

You chose to mention consciousness, so sorry, you lose. Because consciousness is very well defined above and beyond what you'll find in a stupid dictionary (see What Is Consciousness, under psychological consciousness) and it's very likely that human beings as little as 3000 years ago did not have consciousness (Origin Of Consciousness). So the chance of a lower animal having it? Yeah, right!

Oh, and forget self-awareness because that is not well-defined at all. I don't see why having a model of oneself in an environment matters so long as you're not aware that you are that model. And I just can't imagine that happening short of full-fledged consciousness, which we know that some humans don't have ....

And then there's philosophical consciousness which is undetectable .... -- rk

I can see that this is going to be difficult than I thought (I respect that). In reference to point 1, spatialization: sorry I don't have a link at the moment but there was a study with orang-utans involving a square room that was fully furnished and an exact model of that room. An object was chosen and hidden somewhere in the room. Then a miniature version af the object was shown to the ape and then they placed it in the same place in the model. The ape was then sent in alone to find it and was able to find it with the same accuracy as a 4-year-old child. You have provoked some great thought in me as to the nature of this issue. I don't have any knowledge regarding any scientific study of any of the other items on the list. But also to point 6, narratization: I am not sure if my memory serves me correctly, but I believe that Koko often told stories about things she did. I will look into this issue more deeply in the recent future. As a footnote, I would like to mention that my original post on this thread was about a previous fallacious post stating that apes cannot use language. This conversation has been interesting but I don't believe I have all of the necessary information to proceed. Thanks for making my work-day more interesting. -- Timothy Seguine

There's more than just Origin Of Consciousness. John Mc Crone has written extensively on consciousness and one of his key points is that it's a learned ability which humans don't possess naturally and can fail to develop if it isn't instilled by the age cutoff. His website (just google) seems to be down at the moment.

I saw a study once where apes demonstrated the ability to recollect. But I won't put much stock in it until it's compared with animal human abilities. I really need to read JM's books. Well, good luck. -- rk

The inner-voice is something I had not thought of as a hallmark of consciousness. It makes a heck of a lot of sense though. In most of my early memories, I distinctly remember my the voice in my head. I believe that the inner-voice would be at least partially separate from actual thought though. I have memory (maybe backfilled by my older brain, I dunno) of not being able to express what I was thinking tangibly and being frustrated by it. Not sure how this relates to the common experience or how much of my memory in this matter is purely imagined. I thought it might be of note to mention (Please Delete if too off topic), but perhaps this is the wrong place for this discussion Refactor Mercilessly. This is a topic of major interest to me because evaluating natural and artificial intelligences is one of my favorite past-times. -- Timothy Seguine

See Daniel Dennett for more about the model of consciousness as internalized speech. It's my favorite model right now. -- Eric Hodges


I've been doing a lot of thinking recently that perhaps people ought not be trying to invent artificial intelligence so much as reverse engineer natural intelligence. You figure most real systems of behavior have five or six simple rules of operation and all components are made up of collections of those rules. A perfect example is AND, OR, NOT, XOR (and even this set can be reduced to AND and NOT). {Actually just to one: NAND} So if someone can figure out the few basic rules of operation, you could effectively duplicate neural activity in silicon. Add a few million input lines and boom, you've got a human brain. This is my theory, which is mine, and it's my theory. :) -- Steve Sparks

One thing that's always stopped me cold about that level of approach is how, even if you got the underlying uber-kernel exactly right, likely would one then be to be able to hit on the precise instructional sequence giving rise to human-like intelligence? Seems about as likely or on the order of the complexity of coming up with the human genome to be fed into that little chemical assembler called DNA (or the double helix, or whatever the little engine itself is called - my biology is about as good as my Spanish: I had a couple years of it in school. :) -- Bill Kelly

Artificial Neural Networks are based on the idea of Reverse Engineering natural intelligence. Not at the level of making things exactly the same as human brains, though (I don't think we have the technology to do that, for a start). But the principle is to try building things that work in the same sort of way as real brains. Incidentally, Bill: DNA is the medium in which the genome is recorded. There are actually lots of "little engines"; they're called ribosomes.

Like everything else, it is not the fundamental units that really count, but how you put them all together. The thing about the brain is that we pretty much know the big picture and the small picture, but no way to understand the link between them. The big picture is the various areas of the brain that specialize in different things, and the small picture is the neurons.


I think that there are (at least) three fundamental questions about AI. They are all based on the assumption that we will eventually figure out how to make artifacts with an average human level of intelligence. For example, I think either of Rodney Brooks's grand projects of slow progress in AI via ontogeny (i.e., start with a computer with the intelligence of a human fetus, and gradually work through development one day at a time) or via phylogeny (start with a computer with the intelligence of a bacterium and work through our evolutionary branch) would/will work.

The first question is then: what short cuts are possible? Strong symbolic AI, the idea that human intelligence was largely a matter of logic-like conceptual processing that did not fundamentally depend on (for example) interaction with the external world, would be an example of a (potential, possibly already failed) short cut.

The second question is: will we ever build artifacts that greatly exceed human levels of intelligence? Arguably we have already done so, but of special interest is whether we will ever construct a robot that is better than humans at all of the things we think of as typically human: telling stories, raising families, cooking food, etc.

The final question is: to what extent will we understand the artificial intelligences that we create? In my line of work, we train and use statistical models all the time without in any real sense understanding what they do, which is not to say that such understanding is impossible: sometimes we do decide to go and try to figure out how such and such Neural Network is doing its job, and sometimes (depending on the size and complexity of the neural net) we even succeed. We almost certainly won't understand everything about the first AIs that we create, but the scientist in me hopes that asymptotic complete understanding is possible (and if we should be so lucky as to continually create ever more complex and intelligent intelligences, that the lag between our understanding and creation be bounded by a constant or sub-log function of time, amen.) -- Thomas Colthurst


AI could also stand for Automated Information Processing, where the inputs are all understood, the process not understood, and the output not too important.


Moved from SearchSpace

A search space is metaphor for many problems in Artificial Intelligence. For example, a program for Game Of Chess can calculate the combinations of different alternatives after each move. The new positions are nodes in the search space (also called a Game Tree). Game Trees can be computationally explored, starting from a root node. A search space can also be explored randomly or using a Genetic Algorithm which doesn't involve starting from a root node.


It would be interesting to try create an Artificial Neural Network with subsections for major parts of the brain (Artificial Baby?) to see what would result. Train a subnet to do simple language understanding, a subnet to generate the same (as text like an Eliza Program), a subnet to do basic problem solving, a subnet to simulate basic emotions, and a subnet with some basic knowledge about the world (which could be modified by the language subsystem), and interconnect them. The advantage would be you could incrementally increase the "knowledge" of the system by training from examples. Could even be like a Wiki where the topology, weights, training would be arbitrarily specified and modified by many users over internet. Even if each person added a few neurons/connections and test inputs/outputs once a week, or Nn Vandals scrambled it, most would want to see it grow it would become more and more intelligent over time. Perl has AI::Neural Net modules available with the standard distribution for free it would be relatively easy to set up a cgi interface to create this (use mysql to store the topology and weights). Similar to Open Mind (commonsense.media.mit.edu ) but that implementation does not let you query yet and is not based on Neural Networks as far as I know. Users can input random data over internet but the architecture is controlled centrally. Another that is NN based is "20 Questions" www.20q.net but seems very basic.

I thought of a similar idea recently that I wanted to fling into an open forum. I don't know the exact feasibility of this as I couldn't think of an acceptable algorithm. It involves seeding a relatively small net with values, then using quasi-realtime evaluation of inputs and neuron firing. Training would involve sending strings of characters (or other desired information) and a self-organizing learning mechanism. The learning algorithm would not only alter weights but connections and simulated cell growth(strategic addition of neurons). Decaying unused connections, strengthening commonly used ones, and sprouting new ones when localized activity seems to suggest it should happen. The web interface would be used to provide input from the masses. It seems to me that (only in theory) this is like the biological process in more ways than traditional forward and back propagation nets. I have know idea of various implementation details, even certain parts of the math. There is also an issue with continuous input being required to run the net. I don't know if it would even provide interesting results even because it relies on the ability of the input values to induce patterns in the net (of what variety I couldn't be sure). Haven't gotten very far in trying to implement anything of this nature as I completely break down when it comes to deciding on solid learning rules. -- Timothy Seguine


Are there systems out there that crawl the web and automatically create a Neural Network with the information that is on the web?

That's identical to asking if there are systems that crawl the web and create data structures based on statistical analysis. Probably, almost certainly, but the question is too vague to have much meaning. Neural nets are building blocks, not solutions. Look at them as bricks. You can make a building out of bricks, but there are an unlimited number of kinds of such buildings. Or non-buildings. You can make a nice pathway in your garden with bricks. You can use bricks as paperweights. There are many kinds of bricks, made out of many materials, but to most people bricks are not interesting, it's what you make out of them. Talking about "bricks" is just too vague to mean much. Same thing with neural nets.

Questions like this presumably spring out of the vast quantities of hype we've heard over the last twenty years. Neural nets are hugely over-hyped. There is nothing magic about them. They aren't even very powerful, in and of themselves, any more than a single transistor is very powerful.


If the new economy will treat Brains Asa Cheap Commodity, why bother with AI? It just takes cheaper bandwidth to connect to billions of cheap brains, and that is a more solvable problem then true AI it appears right now. The bandwidth needs seem to be about 5 to 10 years away, instead of the 25 or so for AI.


Integration Lacking

I think the biggest problem with AI is lack of integration between different intelligence techniques. Humans generally use multiple skills and combine the results to correct and hone in on the right answer. These include:

Physical modeling

Analogy application

Formal logic

Pattern recognition

Language parsing

Memory

Temporal Simulation

Others that I forgot

It takes connectivity and cordination between just about all of these. Lab AI has done pretty well at each of these alone, but has *not* found way to make them help each other.

There are things like Cyc, but Cyc has no physical modeling capability, for example. It couldn't figure out how to fix a toy beyond the trivial, for example. At best it would be like talking to somebody who has been blind and without a sense of touch all their life. Even Hellen Keller had a sense of touch to understand the physical dimensions of something. Cyc was designed to give something basic common sense about life, not about novel thinking or walking around a room. Thus, it lacks the integration of skills listed above. There are AI projects that deal more with physical models, but nobody knows how to link them to things like Cyc so they can reinforce each other. That is the bottleneck. We can build specialists, but just don't know how to connect them to make a generalist who can use a variety of clues (visual, logical, deductive, inductive, case history, simulation, etc.).


Is Hardware or Software the Bottleneck?

Related to the question of whether the bottleneck is hardware or software, the answer perhaps depends on what stage we are looking at. I suspect that the hardware technology currently exists to build an expensive demonstration *if* we knew how. We have the technology to have the same hardware power as a human brain: it just costs many millions. If such did happen in a big lab, it may still require too much hardware to be practical in the field. But for a proof-of-concept, I don't think expensive hardware is actually the bottleneck to building a smart machine.

However, another view is that experimentation on a wide scale is needed to explore ideas. And, this will probably only happen if the hardware is relatively cheap so that small universities and home tinkerers can also play around. In short, we need to distinquish between powerful hardware needed to make it practical, and powerful hardware needed to test the ideas in labs. The answer will probably depend on whether the needed breakthroughs are related to theory or to experimentation. The more they require experimentation, the more hardware will be a factor to at least make a prototype. At this point we don't know what the magic solution will look like.

Once it happens in the lab, companies will find a way to make it cheaper for commercial use in a relatively short time. It is too big an opportunity for investers to ignore. If they dump gazillions of bucks into pets.com, then real AI seems a no-brainer (bad pun). Thus, I don't think hardware is a big bottleneck to wide adoption of real AI.

As a person who studies AI, I can definitely say that if you give me a terabyte of memory, I'll use that entire terabyte and probably want a few more. However, the software needs to advance quite a bit, too. To have AI, we need to design algorithms that steadily reduce the computational cost to solve problems and answer question (why? how? what?). This is especially the search-cost when dealing with many possibilities. It has proven that true experts do not process multiple steps forward with a combinatorial explosion... humans are very bad at thinking about that sort of thing! No, instead we just think of the few correct solutions, and move forward... spending time thinking about the wrong solutions is entirely wasteful. This sort of reduction of search costs is definitely not a hardware problem... it is entirely'' a software problem.

I, personally, believe that we could build a reasonably intelligent machine on a modern desktop computer if we had a few terabytes of memory, a veritable ton of training material (also measured in terabytes), and software that will automatically develop and recognize useful patterns, learn associations between them, continuously optimizes and indexes, self-audits its learning, and 'forgets' those things that are too low-level to bother remembering. The ultimate task from this continuous process would be to go from low-level pixels and audio to high-level patterns (like recognizing apples vs. glasses of water vs. individual people, human vs. chair, etc.) with severe reduction of the wrong search paths. Something similar can be done with gaining higher concepts out of unprocessed text. Now, this process would be slow (possibly taking over a year), and the final intelligence would only be moderately snappy... but the depth of one's intelligence isn't really associated with the speed of one's intelligence; use of a massive HDD will do just fine for memory in the presence of the right sort of indices.

I'd go so far as saying that: if you don't know how to build a -deep- intelligence on a modern desktop machine, you won't be able to build any sort of deep intelligence given a thousand times the processing power and memory capacity. That isn't to say you should stick to the desktop machine if someone offers you the latter... but the more powerful machine will never substitute for building the right software.

The Cyc method of teaching via by-hand entry of phrases is absolutely the wrong way to go. To learn the concept of 'apple', an AI ought to know it through sensor-fusion just as we do: what does it look like, how does one sound when a person bites into it, etc. Teaching that the word 'apple' with its associated sound 'apple' just happens to be associated with this concept of 'apple' and can be used to communicate this concept is teaching vocabulary. Symbolics intelligences cannot teach concepts that are not, themselves, symbolic in nature... if you want an AI that can draw a 3D apple and put it on a 3D image of a teacher's desk that it synthesized from its own conceptions, it absolutely needs to learn from sensor fusion.

I have to disagree with your apple analogy. As a thought experiment only (hopefully), suppose we took an infant and severed their nerves such that they could not feel nor taste nor smell, and also blinded them. They would basically be an ear + brain + mouth. I beleive that such a child could grow up to still answer common-sense questions if given a decent education. In fact, I bet you could hold a phone conversation with them and never know there was any problem as long as certain senses (color, feel) weren't mentioned. Studies and evidence of injuries and birth defects show that the brain makes the best out of the information channels it has available to it. It does not depend on all the senses, or even most working in to function on an every-day level. In fact, I would bet that they only need just one of the following senses: Hearing, Sight, Touch (perhaps this is stretching it) -- top

That unfortunate child would be unable to draw an apple, remember the taste of an apple, know how an apple feels in one's hand, the crunch, the spray of juice, the scent of citrus. Further, she would be unable to even imagine these things; she'd have no frame of reference to learn things like red, shape, taste, etc. Placed in a situation where she must learn on her own (without educators providing artificial senses), the most she'd ever learn is that apples crunch between the teeth, and do so with far less effort and damage to one's teeth than do rocks. The concept of 'apple' as understood by other children includes a vast array of associated patterns including color, size, taste, stiffness, and more... all concepts that they understand through their senses. That is, they have learned these patterns by experiencing and remembering them, and they can synthesize -- imagine -- these patterns, reproducing a facsimilie of them in their own minds: imagined taste, imagined weight, imagined crunch, imagined spray of juice, imagined scent of citrus. Given the proper media, they might be able to draw an image, offer a verbal description, write down a description, shape clay as an apple - all are forms of expression; all are art. But, no matter what top's unfortunate child does on her own, she'll lack any true understanding of the apple. She'll be unable to learn these patterns. She'll be unable to imagine them. On her own, she'll be unable to ever gain that understanding or produce that art...

Thus she had best not operate 'on her own. Her understanding will depend on the understanding of others, who sense with their own eyes and hands, who can offer some summary of what they see. It may also depend on secondary tools, instruments, to measure signals that don't reach the brain directly and present them by some means through the remaining systems. By speaking into her ears, an educator can tell her of the taste of the apple, of the color, of the weight and shape. She will be unable to intuitively comprehend what weight and shape mean, but the educator could teach her that heavy things cannot be lifted by one man, that light things can be lifted easily, and that objects between one and ten stones can be lifted by men of varying strength. She'll gain a symbolic understanding of 'weight'. The educator can teach that apples grow on trees, then busy themselves for a few weeks teaching the concept of 'tree'. They can make a long list of 'features' for apples... and when later describing some subset of those features, the child may learn to infer (correctly) that they discuss an apple. Ultimately, these educators aren't inherently different from any other sort of sensor. What are eyes but instruments of high-bandwidth visual information that translate to higher-level patterns (people, places, objects) in one's mind? If someone speaks into your ear of the objects that are near, is not one providing a somewhat lower bandwidth, somewhat higher-level eye? Educators can speak even of things that are far beyond your own experience, bringing understanding and patterns and associations back to you would never fathom on your own. And about everything they teach, they can go into smaller detail or greater detail... perhaps at request.

On the other hand, perhaps you cannot gain details at request, and lacking anything but a symbolic input sensor, you cannot investigate on your own. Perhaps, as for Cyc, you must wait for some arse to enter the detail -by hand- along with millions of other things that the arses who crippled you forget to tell you about.

Suppose, for a moment, that through some miracle of modern medicine, we were able, at the age of twenty, to restore this child's eyes, sense of smell, sense of taste, and her somatosensory system (collectively 'touch'). Suddenly, she'd have all the senses of another human. However, she'd still be unable to recognize an apple, or draw an apple... she'd know a lot about apples after she is taught to see, of course; she has a lot of high-level knowledge on apples that won't just go away... but nothing in her education has prepared her for handling visual inputs. It will be a long road to learn what people commonly understand about 'apple' and 'tree'. The same goes for Cyc. You cannot just add mics and cameras to Cyc and expect -any- gain. Sight will need to be learned entirely and associated with higher-level concepts through pattern-recognition... invertible to allow for synthesis. (Not that Cyc is capable of automated pattern-recognition in any case. The limits based on dependency for higher-level symbolic information from other sources will hold true for any intelligence.)

It would have been much better to just have the senses to begin with, to have some perceptual understanding of color, weight, size, shape, etc. when first learning about apples. This is not harmful, at the very least... excepting that it adds to the software burden (e.g. translating visual input to meaningful data).

I don't disagree that an AI can learn the concept of 'apple' in the same manner as this crippled child... symbolically, with massive aid from educators. I only believe that it is absolutely the wrong way to go: that the AI ought to know it through sensor-fusion just as we do: what does it look like, how does one sound when a person bites into it, etc. Only in this manner can the AI even begin to really understand the apple as we do, or with sufficiently similar resolution (e.g. as an anosmic does). Anything else will result in a crippled understanding. For the full range of reasoning, one needs both higher-level symbolic information (origin, destination, 'physical object', etc. -- associations of patterns communicated by use of words) and lower-level information (all the way down to recognizing (or analyzing) from raw sensory inputs, or imagining (synthesizing) the appropriate inputs for projections... e.g. imagining the weight in one's hand before predicting how far one can lob an apple.)

One thing at a time. Passing a Turing Test and winning an art award are two very different goals. You didn't seem to refute my original goal as stated. -- top

I have no need to refute your 'original goal'. You disagreed with a statement about what ought to be by declaring what can be. I agree on the can, provided you give her educators that use what they have learned through their senses to teach her. However, it does not change my opinion on the ought. If you are unwilling to discuss the ought, then we're done on that subject because I have no intention to refute the can.

I do disagree with your more recent statement: "passing a Turing Test and winning an art award are two very different goals". The ability to pass a Turing Test requires the ability to synthesize dialogue that is believable as coming from another human. That is art. It might not win you an 'art award' for best script, but it does indicate that the two are definitely not "very different goals".

Couldn't an AI machine describe what an apple feels or tastes like based on a database of descriptions, say google and/or cyc? Plus, a person or machine is not necessarily going to fail a common sense test or Turing Test if they say, "Sorry, but I was born without arms, so I cannot really describe that". You are not going to ask Stevie Wonder what a pretty girl looks like. (What one feels like is probably for a mature audiance only.) I suppose this debate could not be settled unless such a deprivation experiment is actually performed. But it is not realistic (unless medical history could provide such data).

Does the ability to regurgitate information from a google database constitute real understanding of a perceptual concept? If Stevie Wonder memorized that "She has sharp grey eyes and long auburn tresses", does he have a real understanding of 'auburn' and 'grey'? I think not. However, perhaps I'm the only person here that thinks that understanding requires the ability to synthesize, to produce... to imagine. And proof of understanding requires going first to a high level of abstraction, then back down to a lower one... e.g. to prove you understood a book, you read it then you summarize it; summarizing a book requires building high-level concepts in your mind then synthesizing words to communicate those concepts. To me, that's fundamental to the difference between understanding and recording. However, you and top certainly seem to think otherwise.

If full honesty and understanding is required for intelligence, then most salesmen and most CEO's would flunk. Probably politicians also. Further, why couldn't Mr. S. Wonder answer, "I don't know, I am blind"; or the AI-box answering, "I don't know what an apple feels like; I was born without arms". That is an honest answer. What is an example dialog that you are envisioning?

I guess I am of the belief that "understanding" isn't necessary as long as a satisfactory answer is given. I don't want to get into a Laynes Law brawl over the meaning of "understanding", which is probably a tricky sticky philosophical problem. Thus, I choose to measure AI by the output rather than how processing is actually done. Cowardly, perhaps, but practical.

The "I don't know" above was about the test subject, not a description of AI technology. - top

If the test subject answers "I don't know", the test subject lacks understanding. If you are claiming that the test-subject may possess understanding regarding a subject and simultaneously answer "I don't know" about that subject, you've claimed a contradiction. This discussion began with a claim that a horribly sensory-deprived child can understand things as a normal human... that having more than just one sensor isn't necessary. Given secondary sensors such as educators, that claim is provably correct. However, by backing off to legitimize the answer "I don't know" to perfectly legitimate questions, you are suddenly offering your permission for understanding to be absent where you've (correctly) claimed it can exist... which simply isn't a relevant path for this discussion to take. Instead of answering "I don't know", you ought to be focusing on determining what sort of information would be necessary to "know", or to "understand" such things as color, even in the absence of sight.

If Stevie Wonder cannot describe the molted pattern of an apple skin, then leaving it at that does not make him non-intelligent. If we give humans such lattitude, then an AI box should get the same. --top

If you cannot describe 11th dimensional String Theory, then leaving it at that does not make you non-intelligent. If we give humans such latitude, then an AI box should get the same. We don't say that the human is incapable of understanding it... only that the human doesn't understand it. Unless you plan to argue that Stevie Wonder cannot learn about apple moltings if, for some odd reason, he chose to study it, then I don't see how your little quip here is at all relevant. I think your line of reasoning here needs some consideration... where are you trying to go with it?

:Measuring an AI by the output rather than how the processing is performed is okay. All one needs to do is define the correct sort of tests. The test I defined for 'proof of understanding' is the ability to take a lot of input data, (black-box process it), then produce and communicate a summary. A summary must have the following characteristics: (a) it is considerably smaller than the original data set, (b) it is at a higher abstraction than the original data-set, (c) it hits the most important features necessary to reproduce a facsimile of the data-set with a certain degree of fidelity.

:These are part and parcel to the common definition of 'summary'. One should note that 'features' are utilized opposite of abstractions... e.g. 'red' applied to 'apple' is a feature, whereas 'red' applied to a 'clump of pixels' is an abstraction. The production of a summary proves understanding to any intelligence that has both the original data-set and agrees that the reproduced facsimile is within the specified degree of fidelity (the summary may be translated, if necessary). It often is useful to prove understanding to oneself as a form of auditing one's knowledge. 'Higher abstraction' is defined in such a way that prevents many forms of lossless and lossy data-compression (including simple transformations of representation as per JPEG or LZW. Transformations and translations, in general, cannot constitute proof of understanding). Essentially, production of a 'summary' requires semantic compression.

:I've found nobody so far who does not agree that 'understanding' (in its common, English understanding) is not sufficiently proven by production of this summary. I've seen initial disagreements as to whether component is necessary, or to whether this is the only possible proof of understanding... but, in general, thought-experiments have revealed that alternative proofs are largely equivalent to this one, and removing requirements results in something less than understanding (... as it is commonly understood...). Anyhow, this provides you the opportunity to measure AI by the output. You can now measure understanding. And by that, I mean 'understanding' sufficient to qualify for the English use of the term.

Many humans are not able to create good or usable summaries, especially the poorly educated. I think that is too strict a test. If you wish to define that as *your* target AI goal, I have no problem with that as your working frame of reference; but, it is not a definition that I would use. I think it creates too big of a risk of a "summary savant". Or, at least I wouldn't put all my eggs in that one criteria. A multi-test score-card approach may be more effective. Summarizing would be one of many tests. If they flunk summaries but pass other tests, then perhaps they could be forgiven for such a flaw. This is generally how we evaluate human abilities also. Somebody who may be a great auto-mechanic may be crappy at descriptive summaries, for example. But they are still a valuable member of society even if the boss or coworkers have to help him/her sometimes when he talks to a customer. Some geeks are like this: unable to turn shop-talk into something non-techies can relate to. -- top

Many humans lack understanding of the topics about which they are tested by request for provision of a summary. This is especially true of the poorly educated. However, their ability to prove understanding is often hampered by their inability to communicate. Thus inability to prove understanding doesn't necessarily mean understanding doesn't exist (see my original discussion of how 'understanding' differs from 'proof of understanding', below). Every proof to some other intelligence requires communication, so there is no getting around that. However, there are alternative, often weaker proofs of understanding... e.g. (from stronger to weaker) via asking of open-ended trick questions (tests recognition and rejection, and synthesis... same as full summary), via asking of open-ended leading questions (tests recognition and synthesis), via asking of multiple-choice questions (tests recognition only), via multiple-choice questions with some 'none of the above' answers (tests recognition and rejection), request for examples (tests synthesis only). The open-ended trick questions test all the same things as a full summary... if you can provide a complete summary or definition, you can answer those questions, and vice versa. However, the others you might reasonably 'lesser forms' of understanding. I'd call them 'partial' understandings, seeing as none of them individually meet the proof requirements I set forward when I was initially formalizing my understanding of understanding.

Translation is a non-issue. I listed no constraints on the language in which the summary is provided, nor on the amount of translation between one system and another (so long as that translation has a well-defined lossiness... ideally zero). Obviously, the party receiving the summary must be able to read the summary (which may require that translation) because a summary only can only ever prove understanding to those intelligences that can both read and test it. In practice, that means the tester must have an equal or better education in the subject matter and jargon involved as does the testee. That doesn't change simply because we're talking about AIs.

:Supposing you are not writing a machine that plays the fool ("I don't summarize for no one"), understanding is a necessary component to pass any Turing Test. The proof for this is simple. One may ask at any time for a summary of the dialog thus far. However, that sort of understanding does not mean there is any underlying understanding... i.e. if one talks about apples, the system need not understand apples to pass the test. To understand apples would require that it know what you -mean- by 'apple' as more than just the word 'apple'.

:How could an AI that never sees, touches, tastes, smells, etc. an apple understand an apple? Well, through secondary sensors -- educators or instruments can tell it of the underlying characteristics common to apples ('some apples are red', 'some apples are green', 'some apples are sweet', 'some apples are tart'). So, yes, such an AI can come to understand apples. Stevie Wonder can come to understand 'blue' as 'the same color as the sky on a sunny day', even if he never sees the thing. Or he can understand it as 'the range of wavelengths 440-490 nm. After all, we don't need to see radio-waves to understand them. I don't disagree that this can be done.

:As far as sticky, philosophical issues go... understanding itself can be practically defined as those characteristics that are necessary to pass your proof of understanding. That's practical, not sticky, not philosophical. One could argue that you possess understanding so long as you have these characteristics... even if you never bother proving you have it, or lack the means to communicate a proof. The essence of understanding under the definition I provided is the ability to analyze data to a higher abstraction paired with the ability to synthesize an abstraction to facsimile data. When you can do both, you understand the abstraction. If you can look at the light reflecting off an apple and categorize (correctly) that the predominant color is 'red', and if you can later remember or predict about an apple that it is 'red' and you can synthesize a facsimile of low-level 'red' and attach it to the apple, then you understand 'red' as applied to objects (or at least apples). Understanding requires both synthesis and analysis; without both, semantic compression (low-level-data to 'red', 'red' to low-level-data) becomes impossible... you'd either be unable to produce or utilize a summary.

:I do care how one might go about implementing a system capable of understanding since I plan to implement systems capable of it, but I'm not concerned about which precise implementation is used insofar as it produces the measurable results. In that, I feel the same as you. On the other hand, if you can prove that the necessary characteristics of understanding do not exist, then you can prove that understanding does not exist. If you are forever failing to investigate what 'understanding' actually takes at the level of processing, you'll never make a step of progress towards any implementation.


I suppose this debate could not be settled unless such a deprivation experiment is actually performed.

Such a deprivation, at least in two sense vectors, has been "performed" and her name was Helen Keller. Obviously, the degree and nature of understanding is going to depend on the degree of sensory capability. An AI based purely on Cyc will exhibit (at least theoretically) only a symbolic and abstract "understanding". An AI with audition, vision, gustation, tactition, olfaction, thermoception, nociception, proprioception, and equilibrioception will be theoretically capable of human levels of understanding. An AI with all of these, plus electroception, broadband EM reception, magnetoception, and echolocation will theoretically be capable of "understanding" beyond human capability. That does not, however, diminish the value of AI systems built only upon Cyc; but they shouldn't be expected to be human in nature. -- Dave Voorhis

My main issue with Cyc isn't the lack of ultimate capability... it's the AI's utter dependence on human input, which is both erroneous and incomplete. A system like Cyc will be largely unable to audit or extend its own information. It also has little or no use for the knowledge it does possess, and lacks much ability to make conclusions based on patterns of data. Now, perhaps that is my personal bias for learning AIs speaking. I won't hide it.

As you say, an AI with more sensory capability than we possess would potentially be capable of obtaining a much deeper understanding of the world than we currently possess... and we'll feel at a loss when it tells us of how (for example) certain patterns of echolocation and broadband EM reception are meaningful to (some predictions here) just as Helen Keller felt at a loss dealing with colors and sounds. Again, it isn't to say we couldn't learn about these patterns (should the AI deign to explain it to us, or should we study it), but we'd have far less use for that knowledge and understanding than would the AI, and we'd never gain the knowledge intuitively.

Of course, a broader range of sensors goes beyond just accepting a broader range of physical signals. Merely having sensors placed differently, while still providing sensor fusion, can create a world of difference (or at least a paradigm of difference). Having three auditory channels allows you to triangulate sound. Having audio and/or cameras spread over a massive area (e.g. a city) can provide access to broad-area patterns that humans have troubles seeing due to scale (e.g. traffic patterns, crime patterns).

I have to still disagree that dealing with a variety of sensory input is the key. It may be helpful, but not a requirement. Hellen Keller could read braile books about all kinds of subjects and "understand" once she had a teacher to get her started. Nor did I claim that Cyc's rules alone could produce sufficient AI. I agree it probably requires a multi-pronged approach that "triangulates" the responses from different tools/skills, but not necessarily involving lots of sensory info. For example, it may require temporal and/or physical modeling in addition to cyc's rules. Or at least proximity models, such as graphs with distance links (stick-figure if you will). A lot of human activity deals with temporarl and physical proximity. --top

I'm left thinking you haven't read what I have written. Feel free to continue ignoring those locations where I more or less say, point blank, that a variety of sensory input is not a necessary requirement... that understanding can be gained without it, provided you have sufficient support from secondary sensors (including teachers and instruments).

There are some parts I have difficulty understanding of yours. Do you wish I bring those up? Please do. I'm quite voluminous, sometimes overly so, in my willingness to help another understand. It helps to be able to target the answers.

Then why do you keep talking about a variety of senses? If it is not a pivotal issue, then please drop mention of them. This must be among the things you missed. There is nothing to be lost and everything to be gained by utilizing sensor fusion. It is what ought to be done. You keep focusing on what can be done with just one crippled little sensor, and I don't disagree on that points. You keep saying you disagree that the variety is necessary, and I won't argue. If you have an argument with me, then you need to change to: this is why you shouldn't use sensor-fusion.. Now, I've answered why sensor-fusion ought to be done several times, above, but apparently the presence of those bits is just confusing to you, who is searching all the while for any direct argument against the position you've presented. I have the impression that you and I are alike in our tendency to first defend, then absorb. Anyhow, you won't find a direct counter to your position. If you wish, now is the time to absorb.

There is no inherent difference between understanding temporal vs. spatial patterns. There is a great deal of -intuitive- difference, mind you, and some hard-wired differences in humans... but data is data, and such things as space and time are just dimensions that exert themselves in reference to the ordering of sensory input. Useful patterns in 'space' are usually viewed as 'objects' (or 'nouns') whereas useful patterns in 'time' are usually viewed as actions (or 'verbs'), but that's just based on what happens to be useful. Anyhow, your new tactic has some problems (let's add a bunch of models atop Cyc! Yeah!). A model incorporates an understanding, a paradigm; it's quite possible for models to be wrong or inadequate just like anything else you teach Cyc. More importantly, though, a model should be something you develop -within- the AI if you wish to claim any understanding is present... otherwise the AI is doing the equivalent of pulling out a simulator to answer the questions, and it will never learn where it can abstract, cut corners, speed up the process with low loss of fidelity... or even improve on the process, gaining more fidelity. If you're going to add existing models to an AI, it'd be better to teach them the models so they are truly part of the AI's knowledge and understanding. And, yes, this is another ought... I won't argue that you cannot hack together a functioning AI by plugging in third-party models. I'll only argue that it is an ultimately redundant and harmful approach to the progress of AI technologies, and that you ought to know better.

We're not looking for infinite flexibility. I never ruled out the models from being expanded to accept new knowledge. But making entirely new learning frameworks or techniques on its own is building an unnecessarily high goalpost. We are going for average joe at this stage, not Dr. Codd. How about a sample dialog to illustrate your point. -- top

That can't possibly be the royal we. When I started searching for the an internally consistent and correct set of formal abstractions and definitions for data, pattern, information, knowledge, intelligence, wisdom, and behavior, it was definitely for flexibility purposes. My goal is to build an intelligence that can learn anything I can learn, and systematically grow its own understanding over time (which is why I needed a measurable definition of understanding in the first place). Interestingly, though, my studies have shown that it's actually more difficult to make models 'modules' integrate with a good knowledge system than it is to make the good knowledge system integrate the models directly.

I'm not sure which point you wish me to illustrate with a dialog, or whether a dialog is an appropriate mechanism to illustrate the point you desire.

You seem to be expecting the Wright brothers to build a jet the first time through. Walk before you run.

The Wright brothers didn't cobble something together and hoped it worked. They sat down with the pencil, the pad, and their brain... and worked on designs. They also knew exactly what they wanted: a machine that could fly and carry men while doing so. They also had access to plenty of literature revealing plans that failed in their past. They had a very decent chance at getting it... and, no, they didn't build machines that could 'walk' along the way. Their failures floundered about in the mud, just as yours will if you expect that cobbling together failing plans will get you Joe Average intelligence.

Testing is of course needed. The Wright brothers used wind tunnels (an apparent invention of theirs) and kite-like tests. They tested and tuned based on tests. Their first goal was flight, not necessarily far flight, fast flight, or even practical flight. They knew they had to get to stage 1 first in order to get patents and funding for stage 2. If they went for the whole kit-and-kaboodle, then the French etc. may have beaten them to the glory.

A fine example of Worse Is Better.

Of course, tuning is a useless waste of time before the architecture has proven itself capable 'with just a little tuning'. I don't disagree that there are stages on the path to Intelligence, with plenty of patents and such on the way. Among them are more generic data bases, pattern-analysis, information systems, and expert-system style correlation engines.

Yes. Get it going and when it has commercial success it will generate more research funds and activity. The fixed-wing plane techniques we now take for granted were tested with money and activity around flex-wing models. You seem to want to find the equivalent of fixed-wing technology first. The Write's used a test-based synthesis of existing technologies: gliders, bicycles, piston engines, and kites. Maybe there is a magic shortcut, but history is weighted against that being the most likely trigger.

When a system has commercial success, it also has a great deal more momentum. Better to observe what has already had commercial success and improve on those.

You should examine, for a moment, that earlier statement of yours: "We are going for average joe at this stage, not Dr. Codd." I'm quite dubious that Joe Average is somehow an 'easier target' than Dr. Codd. Joe Average can drive cars and taxis, can do dishes and laundry, can learn carpentry and architecture, can learn math and medicine. Dr. Codd doesn't have anything inherently different than Joe Average... just more patterns, more education, more knowledge... and, with time and effort, Joe Average could know everything Dr. Codd knows.

You need to do some serious and pervasive design work. YAGNI is a principle that keeps you on a plateau of solutions that can be reached with least expenditure of energy... and modern computer science has not yet lifted it to a stage where YAGNI will ever get you to human intelligence. You Are Gonna Need It. If you think that you can even get to Joe Average without significant design work, you'll fail. Or, worse, you'll succeed and haven't a darn clue why... which combination of cobbled-together pieces and third-party model-modules got you intelligence?

I think you focus too much on the Turing Test. If you wish to walk before you can run, go for the intelligence of a Mutt... which can get around in the world, survive, cannot learn Riemann Sums for its life, and absolutely cannot hold a conversation with a human. That will be the path I take.


Interesting that animals can do physical navigation pretty well, but do not communicate beyond the basics very effectively. A chimp may be able to make a nice banana split with training, but would fail a Turing Test.

Indeed, that is true. And pigeons can play piano. Do not underestimate the mutt.


A hypothetical AI Turing Test chat:

(Per = person, Bot = AI-box)

Per: Describe an apple

Bot searches dictionary

Bot: A round-shaped fruit

Per: What color is it?

Bot googles "apple color". "Red" appears most often, but green and yellow also appear. Using this rank, it uses its standard frequency-ranking response.

Bot: Apples are usually red, but can be green and yellow.

Per: Do apples bounce?

Bot googles for "apple bounce" (perhaps with variations such as "bounce apples"). Finds infrequent references. Uses standard low-frequency response.

Bot: I do not remember ever seeing apples bounce.

Per: Is that a yes or a no?

Bot has a canned response for forced choices.

Bot: I am only stating my experience. I shall not speak outside of it. Have you seen apples bounce?

Per: No, I guess not. Do apples fall?

Bot googles for falling apples. Finds many references to Isaac Newton's apple experience, but not much else. Bot bets on the Newton hits due to their high occurrence rate relative to other responses. Also checks standard biography database to make sure that Newton was a real person. Since current context is assumed not to be fiction, this check is standard. A Cyc rule says that people don't perform much activity when they are dead such that the event probably happened during Newton's life-time, of which the range can be found in the standard biography database. Thus, the event is guessed to have happened in the past.

Bot: An apple fell on Isaac Newton.

Per: Why?

Bot searches dictionary for "fell", which takes it to "fall". Finds definition "to descend due to force of gravity".

Bot: Probably due to gravity.

Person decides to try to knock bot off guard.

Per: But that was like 300 years ago. Do apples still fall?

The question "still?" causes bot to bring up time-related change research algorithm. Using pre-programmed synonyms for "change", it googles for items such as "gravity change", "gravity increase", "gravity decrease", "more gravity", "less gravity", etc. It finds a few physics-related articles that talk about speculation regarding changing of cosmological constants. They are too complicated for Bot to interpret. There is also a grade-school-targeted article on how earth can increase its gravity if it's mass changes. Bot can glean some basic facts about such, such as a change of mass can cause a change in gravity, but none that directly answer the question. Bot concludes that it is confused, and thus uses a standard "confusion" response for time-related change questions.

Bot: I don't know if gravity has changed.

Per: Have you seen it change?

Since bot is programmed to use google as its "experience record", it answers the same question, but from the perspective of personal observation since the question was phrased in terms of an observation request.

Bot: I have not seen it change.

Per: So why do you think it might have changed.

Bot inspects its past responses and concludes that Per's restatement does not appear to be a good match for anything Bot has said. Thus, it produces a canned response for probable misquotes.

Bot: I don't think I said that.

Per: If you say you don't know it changed, then you are entertaining the idea that gravity may have changed.

Based on bots English idiom dictionary, it knows that "entertaining the idea" is the same as "considering".Bot reviews its logic and Cyc heuristics and finds a rule about considering possibilities of the form: "If there is no evidence for lack of X, then one should consider X a possibility".

Bot: If I have no evidence of lack of change, I probably should consider change a possibility.

Per: That's dumb. I don't know of anybody who thought gravity might change in 300 years. At least not enough to notice.

Bot knows its view is being criticized. Bot's logic and evidence weighing heuristics realizes that it does not know why Per thinks that gravity has not changed. It is a statement made without any detectable evidence. It thus asks a canned question to try to get Per's reasoning.

Bot: Why are these people so certain of their viewpoint?

Per: I guess if nobody ever talks about it happening, then it didn't happen.

Bot's debate logic tells it to try to form a general rule based on Per's suggested IF-Then rule and ask for clarification.

Bot: So if nobody ever talks about something happening, then it probably didn't happen?

Per: There are probably exceptions, but I guess you can say that.

Bot's debate heuristics recognize that Per has admitted there may be exceptions.

Bot: Is it not possible that this case is an exception?

Per: Maybe. I am not a scientist. I just feel that somebody would have mentioned it.

Bot can conclude based usage of "feel" that intuitive or emotion-based evidence is being used and admitted to. It thus uses a canned debate-related response:

Bot: Well, emotions and gut feelings are sometimes wrong.

Per: Perhaps. I don't claim to always be right. If gravity changed, then something has to cause it.

Based on the previous article, it knows that a change in mass may affect gravity, but does not have information about whether mass has changed. Bot decides to float the idea.

Bot: I hear gravity can change if mass changes.

Per: Okay, but what masses changed?

Bot googles for articles related to mass changing using the technique already described. Finds lots of mention of stars losing mass from astronomy articles.

Bot: A star's mass can change.

Per: Do you mean the sun?

Basic classification and categorizing database is searched and finds that the sun is a member of "stars".

Bot: The sun is a star.

Per: Yip. So did it change mass within the last 300 years?

Bot performs the same kind of change study on mass of the sun. Finds an article that it can glean some basic information.

Bot: I've read that mass coronal ejections can cause a star to lose mass.

Per: Well, that is outside of my league. But I am not sure what happens on the Sun would change Earth's gravity much. We have tides and all, but I don't know much about the details. And, if the sun did barf stuff out, then wouldn't it have shrunk a lot over billions of years? If it can shrink noticeably in 300 years, then in a billion it might be gone.

Bot can find nothing wrong with the stated logic. Decides to change the subject a little bit by trying to clarify the apparent but not certain association with "barf" and coronal ejections because its database associates barf with living things.

Bot: You call mass coronal ejections "barfing stuff out"? I thought only humans and animals barf.

Per: I am just joking around with words. You are kind of a literal guy, aren't you?

Bot's designers encountered this criticism before in tests, so put in some canned responses.

Bot: I was born in another country, so I sometimes don't get some figures of speech.

Per: Understood. Not a problem. Your country does seem to have some odd notions of gravity, though. Good day, Sir.

Bot passes Turing Test, but it judged to be of merely "average intelligence" from a human perspective.


... And now the person proposing this just needs to make this hacked together approach work. As a note, none of the 'google experience records' will work. Here's what will really happen:

Per: Describe an apple.

Bot searches dictionary.

Bot: A usually round, edible fruit of a small tree, Malus sylvestris, of the rose family.

Per: What color is it?

Bot googles "apple color". Its primary references are apple.com with regards to color utilities and toner cartridges. It filters these responses for words that identify colors, and sorts by frequency of response.

Bot: Apples are usually black, but may be yellow, cyan, or magenta.

Per snickers, then asks another question.

Wikipedia often has "disambiguation" lists that can clue in a bot that there are different kinds of apples. Probability and heuristics can be used to pick the most fitting variation, or trigger the bot to ask for clarification if there is no clear-cut leader. See below.

Per: Do apples bounce?

Bot googles "apple bounce" and finds common responses involving Apple shares and commemoration key-chains. It makes what sense of this it can.

Bot: In 2000 there was a great deal of controversy as to whether apples could bounce back. In 2006, Wall Street Journal reported that apples do bounce. I am led to believe that apples bounce.

Per snickers again, then goes on to have fun stumping the Bot some more.

... Ambiguity and communications context must be handled explicitly, and must seamlessly integrate with any searches to independent sources of information like google. That seamless integration costs a great deal for the hacked-together system consisting of many component parts - a cost that increases at least linearly (and more likely with the square) of the number of component parts. It ends up being the humans that pay this cost, that program in, explicitly, which subsystem to use, when, and how... and it can be made to work for specific applications. However, the code will be rigid. It isn't nearly so flexible or long-term efficient as it is to create a system that learns when and which subsystems to use and how to best utilize them (i.e. learn which queries are sensible); if all you do is help out on the 'how' (via provision of a component communications service), the human cost is guaranteed to be linear and significantly smaller with the number of different component interfaces.

If it has Cyc-like rules, which was stated as an assumption, then it could detect that "apple" may mean different things (distant nodes in multiple taxonomies with the same handle) and use the most appropriate context, or at least ask to clearify. As far as making the machine self-learning versus human-fed, well that was not really a debate the example was meant to address.

Suppose, for the moment, that the Bot was tuned a little and now utilized its initial definition ('a round-shaped fruit' or 'A usually round, edible fruit of a small tree, Malus sylvestris, of the rose family.') and kept this in some sort of communications context for future conversation. Further, suppose that it was able to identify, in its Cyc-like ruleset, this particular form of the apple (e.g. by its attachments to 'fruit', 'tree', and 'Malus sylvestris'). That is, you've already integrated three components of the system: the dictionary, the Cyc-like ruleset, and a communications context management service. So far so good, right? Things are nicely integrated and working 'seamlessly'.

Then you decide to mess with this fragile, rigid balance by utilizing Google as the experience record. If you use statistics and hits as you originally suggested you'll STILL get the sorts of responses I describe above because you would have not integrated the Google service with the human-fed cyc-ruleset, the dictionary, and the communications context. What you must do is, at minimum, filter web page results by subject prior to obtaining statistical hits. (Better would be to make the 'Bot' read the pages and 'understand' them, too, but this will lead you directly into that rabbit hole you were attempting to avoid by use of statistical hits.) In any case, to distinguish the subject you must 'read' some other parts of the page (even if it's just to filter out all pages containing the sequence 'apple.com'). Now, as the programmer, you have the opportunity to decide how you are going to make the Bot 'read' the page and perform the filtering, and how this will integrate (if at all) with the communications context, dictionary, and cyc-like rules set. Tell me, how will you distinguish 'apple' (the fruit) and 'apple' (the company) in a systematic manner that extends to a great many subjects other than the apple?

Consider, also, the case where the rules-set isn't initially aware of 'apple' (the company) and must, therefore, self-learn to distinguish between them. Even if you take care of 'apple', this situation will show up again and again and again. Your apparent assumption is that humans have already hand-fed the machine all the answers it will need. I am extremely skeptical that such a thing would ever happen... that humans can hand-code anything like 'intelligence'. Even such 'simple' things as reliably recognizing and distinguishing faces is difficult, much less distinguishing behaviors (e.g. recognizing criminal activity, identifying error so one can suggest corrections), and pattern perception is only a quarter of intelligence: consequence prediction, behavior production, and planning are part of intelligence, too.

Wikipedia, which the bot can use, has "disambiguation pages". The bot can check those to see what context is the closest fit. And if there is not a good fit, the bot can simply ask. And even if there wasn't a disambiguation page. At the bottom of many Wikipedia pages is a classification or grouping. Millions of human fingers have already "educated the web" to some extent.

Example:

Per: Describe an Apple.

Bot: What kind of Apple? There is the fruit, Fiona Apple the singer-songwriter, the Apple Corporation, etc.

Per: I mean the fruit.

It may be annoying, but we are not measuring annoyance. (Hell, I'd probably fail an annoyance test too :-) As far as whether such is "intelligence", again that is a definition battle that I don't want to get into. I view it from a standpoint of utility, not a property that by itself we are trying to achieve (if even universally measurable). If something can "fake it" using combinations of google, wikipedia, and cyc to acheive its goal; so be it. On reviewing such a dialog, an AI engineer may add a new rule that says something like, "if the article 'an' is put before a word, it implies a physical object as apposed to people or organizations". (Cyc doesn't seem to do fuzzy associations very well, which is something I would probably fix or suppliment if I was building a knowledge base.)

This isn't relevant. Even if you can distinguish the two at the point of 'Describe an Apple', it isn't affecting the later questions: do apples fall, what color is it, etc. I.e. even if the Bot simply ASSUMES you mean 'fruit', you haven't sidestepped the problem. Did you misunderstand me, above?

Perhaps not, that is why I created a sample bot discussion so that problems can be shown instead of verbalized. Once the context of "apple" is established, that context stays locked until otherwise changed. You seem to be assuming that the bot forgets prior questions.

I am making no such assumption. My assumptions above state the opposite: that Bot has this communications context nicely integrated with its Cyc knowledge base, that it is tracking the topic of discussion. Despite this, the fact that the Bot has this context is completely irrelevant if what you are doing is measuring statistical hits on the Google web pages. You need Bot to read, distinguish, and filter between the 'apple' types returned in the search. And that is one of the places where the problem gets deep... it's a pattern recognition problem. There are some simple approaches for any specific domain (e.g. filter out pages with 'apple.com' when searching for colors of apples), but you'll need to handle it for more generic domains. You seem to be assuming this rabbit hole will be something you just casually step over... but if you try, it will swallow you whole.

I was not assuming we tuned the system just for apples.

I did not mean to imply you were making that assumption. What you're assuming, unconsciously perhaps, is that designing the Bot to distinguish relevant and irrelevant web-pages returned from a search will somehow be an 'easy' problem... or that it will be wholly solved for you by Google such that you can borrow off the intelligence already invested into the network organization. It is not an 'easy' problem, and the search will never reach that level of organization unless Google both solves this problem first and offers an unambiguous ontology for the searching (Web 3.0 quality stuff... it might happen). This problem is far deeper than you're giving it credit for... deep enough that accomplishing the task is, by itself, a very significant accomplishment. What I see you doing is either waving your hands to make the problem go away or (giving less credit to you) a complete ignorance of problem. I ask again: how will you distinguish web-pages about 'apple' (the fruit) and 'apple' (the company) in a systematic manner that extends to a great many subjects other than the apple? Think about it. Seriously.

The bot does not have to be a top trivial expert or professor to pass the Turing Test. I realize there are probably lots of ways to trip it up, but I suspect many of the same tricks work on an average person also. You have not presented any scenario that demonstrates a bot would do worse on average than an average person. The bot has high-speed access to the web, unlike a typical human (we'll assume here). Although the bot lacks common (human) sense, it does have the advantage of using the web and Cyc to "cheat", which would in general balance things out between humans and bots.

And I am not claiming it's an "easy" task. It would be a lot of work to join a Cyc-like rule base and google + wikipedia etc. It would probably require a lot of scenario testing to tune good enough to pass. And it may also fail badly sometimes. (The Turing Test doesn't say much about the success frequency needed to pass that I know of. Some humans I know would fail too sometimes.) But it is not beyond existing technology to get it to pass as an average Joe or Jane. I often suspect Sarah Palin would fail a Turing Test.

Re: "consequence prediction, behavior production, and planning are part of intelligence, too.

Many humans are not so good at those either. It may be possible to grab a plan from the web or knowledge base also. We need to strive for average joe, not Einstein at this point. We don't have to reinvent ideas that other people already documented. Once And Only Once. It seems you are looking for a "learning machine" more so than a task doer. And what is "behavior production"?

Humans are excellent at planning. We're so good that we reach Unconscious Competence and don't even think about how much planning we actually do. How do you plan to obtain dinner? to make yourself presentable for work? to drive home? to obtain your e-mail? to learn about the colors of apples? It is only when making plans that are well beyond our experience, beyond our ability to predict, into the realm of Conscious Incompetence that we experience frustration and weakness in our planning. Humans are also excellent at consequence prediction. What do you predict will happen if you throw a baseball at a window? or at your boss? what are the likely consequences of a car accident? Finally, humans are quite capable at behavior production on average, and outstanding for certain individuals. I.e. after planning to throw the ring onto the bottle, many of us will at least get close when we make the attempt... but some people will get it every time. We do lack the precision behavior available to machines, though.

There are plenty of web articles on how to dress for work.

I'm sad to hear you need a web page to make yourself presentable for work, and I suggest you push that skill to a level of at least Conscious Competence. (^_^ facetious). Anyhow, the point I'm making is that humans are very good at planning... we plan most of each day's activities without even really thinking about it. I plan the whole direction of my response to you with very little effort, the vast majority of it below the threshhold of conscious thought. When designing an Artificial Intelligence that can plan in most domains, however, things are not nearly so easy... they might, as you suggest, go find 'programs' to help them out in the form of 'How To' documents for dummies and robots, but even the effort to find such a guide is part of a plan (e.g. how might I dress for work? well, first I hit Google up with a request for pages on how to dress, then I read these pages, then I attempt to obtain clothing that is a close match to the suggested styles, etc...).

(On prediction, baseball at window): The answer can come from prior experience of others. If a person never saw a window before, they may have no idea what would happen. That comes from watching (TV or real). For example, it could bounce off, make a clean hole, or shatter the entire window, bend a dent, etc.

I agree. Much of our learning can come from the experiences of others as communicated to us. And if a person doesn't even know what a window is (or a baseball, for that matter) it wouldn't be expected that they could make predictions about one, and they'd need to ask questions about these until they had a clearer understanding of the question. I can't expect you to tell me about Bose-Einstein condensates or Relativity at a moment's notice... you'd need to put some effort into studying up, first... the ability to make predictions arises directly from one's knowledge base. It doesn't seem you're challenging any point I made. A human is likely to very rapidly distinguish consequences of throwing a baseball at an open window, a closed window, a glass window, a shuttered window... and even the social consequences of throwing the baseball at one's own window, the window of a neighbor, the window of one's landlord. These consequences include the physical (sound, holes, debris, etc.) and higher-level patterns (difficult conversations, apologies, guilt and shame or arrogant snobbery and lawsuits...).

A bot will have a boss? Physical similation is a separate issue. Stevie Wonder is probably bad at physical similation, but that doesn't make him non-intelligent.

Of course a bot will have a boss. ^_^ Or, at the very least, it will be a long time before we make bots 'free' entities that may pick and choose their own jobs of their own 'free will'. More seriously, physical 'simulation' (for prediction, planning, etc.) is NOT a separate issue. It's a separate domain. There's a difference. If you're going to handle prediction across generic domains, the physical domain is one of those domains. If you plan to handle every single effin' domain separately, you'll have quite the mess and it will take you a long time and a lot of money to build it. And Stevie Wonder was probably just fine at physical simulation, though he lacked the visual component. He might not be an expert at physics, mind you... but you don't need to be an expert at physics to perform a great deal of useful physical 'simulation' (synthesis, prediction of likely consequences). Even children can tell you a decent subset of what happens if someone bites your arm, jumping from a great height, the physical effects of lifting and slinging a baseball in the direction of a window, etc.

Children mostly learn these by experience, not physics simulation. Plus, we have not ruled out a physical simulation engine to suppliment the bot. It could triangulate info from all its potential sources of knowledge. Such a simulation engine would probably have to be initialized with common features from the human world, such as desks, dogs, cars, staplers, kick-balls, etc. A physical object encyclopedia, if you will. Yes, a self-learning bot would be nice, but at this point it seems stuffing it with pre-digested knowledge may be the better-understood route even if it does require some form of brute force. Brute force is sometimes the better solution (or at least more predictable) when you are struggling to get the brainier approach to work, which is what is currently happening. When they get self-learning to work and it overtakes knowledge-stuffing, fine. Until then, stuff it. It is similar to propeller planes versus jets. Jets had more potential (invented roughly around 1912), but they required much more R&D to get right (helped by Hitler's often wasteful appitite for "killer apps"). Propellers offered a more tame-able, visible, and incrimental path at the time. Same with transistors versus vacuum tubes. We are just not ready for the jet and transistor of AI yet. Knowledge stuffing almost works, while self-learning is pitiful right now.

I guess that comes down to how flexible your definition of 'simulation' is. To me, a 'simulation' doesn't require a formal mathematical approach. Any system that represents and predicts the behavior of another system qualifies for the term. To me, a computer that ran a 'simulation' by simply making educated guesses as to the properties of desks, dogs, cars, and staplers and their behaviors is still running a simulation... just as much as a computer that runs using rigid and predigested object models with properties hand-programmed by /humans/ that essentially made educated guesses. 'Experience' alone doesn't produce the ability to predict consequences. A child use his or her experience to to aide in imagining consequences, but it is the imagination that constitutes a simulation. When the domain is physics, it's a physics simulation.

I am not sure there is a real difference between "imagination" versus "simulation". When I am imagining something, I tend to have a little movie running in my head and explore difference scenarios that way. Thus, this may be another Laynes Law wall.

However, to avoid a Laynes Law debate, I'll consider for a moment what you mean by 'simulation': an 'engine' fully preprogrammed that handles such issues as velocities, forces, energies, collisions, etc. in accordance with some popular physics model (classic, quantum, relativity). I think such simulators are of great utility for details beyond human imagination, but for a computer that is more questionable... the details are being handled one way or another. Further, now that you have your physics-engine, you need to add components to the AI so it can use the simulator; by the time you've made it so the AI can figure out what to simulate, you're only be a few small steps from requesting that the AI predict the results of the simulation. By taking away the simulator, you can also reduce the set of rules constraining 'what' you decided to simulate and broaden the set of domains your AI's simulator can handle... to handle such things as social consequences and the behavior of the dog (which a physics engine cannot handle).

May I request an example? I'm not sure I am following you. Note that there are also things like time-line simulations which may involve little or no location-based simulating. Ideally we would have both course-level abstract simulations and finer-grain ones.

I'll provide an example, but I'm not sure whether you want an example of 'something to simulate' vs. an example of a situation where removing the simulator is an advantage (e.g. handling the behavior of the dog in response to the baseball) vs. an example of why just having the simulator doesn't build a bridge to using the simulator (which requires the AI build something of a 'scenario' and convert it into the simulator's language based on its own conception of the problem it wants to solve; building such a scenario is a constraints problem, and requires synthesis just like simulating would have in the first place).

As far as knowledge-stuffing vs. self-learning: it isn't an issue of 'versus'. To learn fast, you need both. To advance to human levels, an AI must be able to learn even from what it is taught.

I am not against research in both; however, we are not ready to integrate them yet to provide the triangulation effect sought above. Thus, the one further along is the best place to invest time right now because we need its success to fund the jet engine of self-learning.

Attaching more and more components won't be profitable. If you can limit the additional components a great deal, you can reduce the explosive complexity increase on meaningful integration. Focus on enhancing and using just a few components and you'll be able to move towards self-learning.

Behavior production (aka behavior synthesis) is the production of a sequence of signals to an agent's actuators... and while it includes such things as entirely random behavior, the 'quality' of behavior production would usually be measured by the disparity between intent and effect (including 'side'-effects, such as expenditure of energy).

You keep saying that you're going for Joe Average. I'll continue to insist that the difference between Joe Average and Einstein is negligible on the intellectual scale... like insisting you're attempting for $10billions instead of $30billions funding when you'll be working hard to reach $100thousands.

Different humans are better at different things. So why should a bot be good at everything?

The bot doesn't need to be good at everything. But if you're going for Joe Average, the bot needs to be mediocre at an enormously vast array of things, and 'good' at a merely 'vast' array of things. The width of domains a typical human's understanding, predictive ability, reasoning, planning, and behavior-set touches is, quite apparently, beyond your comprehension... beyond your sense of scale. To truly understand it, you need to think about how you'd program a Bot to understand or perform things you consider 'trivial', like folding laundry, summarizing a letter, making lunch, and predicting the likely immediate consequences of throwing a baseball at the president's dog. People like Einstein and Feynman don't really have anything special except lifetime curiosity, a habit of thinking deeply, and a mind at least briefly open to alternative paradigms; they might be more creative, have more talent for recognizing patterns, making connections... but those are just power-ups on components everyone has. Even Joe Average can understand Bose-Einstein condensates and Relativity if Joe Average can be motivated to learn about them; the potential of even 'average' humans is vast. Joe Average is not a significantly easier target than Einstein and Feynman.

That, we do not know, since it has never been done. I don't think anyone in the 1950s would have guessed that a chess bot would beat the master before a bot could do laundry. And it could have been the other way around.

I wouldn't be too surprised if we have 'Einstein' and 'Feynman' level bots of specific fields (capable of 'imagining' physics instead of using a rigid physics-engine) before we have the Joe Average robot capable of handling the thousands of fields.

{Are you talking about a Turing Test or doing the laundry?}



See original on c2.com

"Just because we can, doesn't mean we should." I have had a fear of the concept of artificial intelligence because I believe that we already have an over-reliance on "experts," technology and the external systems of care that we have created. My fear is in alignment with our friend, Derek Peterson's call to "de-professionalise caring."