John,

It is certainly clear that mental imagery plays a role in human thinking, but this role does appear to vary from person to person, both in extent and in nature. Take a look at Hadamard's old book "The Psychology of Mathematical Invention" for a fascinating discussion of the different sorts of mental imagery pursued by different people (visual, acoustic, verbal, etc.). I myself use a lot of visual and auditory imagery in my own thinking, but I know others who do not (at least not at the conscious level).

In the Novamente design this is dealt with via a currently unimplemented aspect of the design called the "internal simulation world." This is a very non-human-brain-like approach, but I think it's an interesting and ultimately very powerful one. What it means is that NM will actually have, internally, a private 3D world-simulation, complete with a simple physics engine. It can use this internal sim to experiment with hypothetical actions in hypothetical situations, but also to draw various abstract sketches and movies that don't correspond to any real-world phenomena. We haven't implemented this part yet due to the familiar lack of adequate human resources, but I think it will be a valuable addition to NM's cognitive arsenal. For the sim world, we would use the CrystalSpace engine that we are now using (in the AGISim project) to give NM a sim-world to use for embodiment and interaction with humans...

I don't really see mental imagery as a critical missing link btw the symbolic and the subsymbolic. In NM, there is interaction & translation between symbolic and subsymbolic knowledge without need for mental imagery. However, in some cases mental imagery can provide insights that would be hard to come by otherwise.

-- Ben G

John Scanlon wrote:
My philosophy of AI has never been logic-based or neural-based. I did explore neural nets during the neural-net mania of the nineties. I did a lot of reading, and experimented with some with feedforward nets I wrote using simulated annealing and backpropagation (which never did work very well). Neural nets seem to have potential as one tool among several types of incremental learning algorithms, including genetic algorithms and statistical methods, but in themselves, they are no more than that -- useful tools, but not the solution. Language, which includes logic, is a way of representing ideas simply and crudely. Good for communication and internal reasoning -- "if I do this then this will happen, unless state X is the case, which means that this other thing will happen," etc. My project uses an artificial language (Jinnteera) for both these things, and the language is integral to the whole thing. But it does not function as the core knowledge-representation scheme. So this brings us to what I've been calling the missing piece. Artificial neural nets (as they currently exist) can function as general-learning algorithms, but they don't represent knowledge of the real spatiotemporal world well. They are too low-level for handling what in human intelligence is thought of as mental imagery. Yes, in the brain, it is all neural based, but in a non-massively-parallel von Neuman computer system (even a PDP system), building a 100-billion-node neural net is computationally intractable (is that the right word?). It has to be done differently. The missing piece lies between low-level learning algorithms and highest-level logical-linguistic knowledge representation. When a human translator, at the U.N., for example, translates between Chinese and English, he (or she) does it infinitely more effectively than any translation software could do it, because there is an intermediate knowledge representation that is neither Chinese nor English, but that can be readily translated to or from either language by a fluent speaker. The intermediate knowledge representation is non-linguistic -- it consists of mental models constructed of sensorimotor patterns representing a 3-D temporal world. This sounds very vague and abstract, but I'm working on making it concrete, in my system (Gnoljinn) -- developing the data structures in code for implementing this knowledge-representation scheme. There's been some talk here recently about 3-D vision systems, and this points roughly in the direction I'm going in. Gnoljinn uses a single sensory modality right now -- vision -- and will be restricted to it for a good while, because, while it might be useful to have other sensory modalities, none of them are absolutely necessary for higher intelligence, and it's best to keep things as simple as possible starting out. I seriously wonder if I can do this project myself, or whether I need to try to find some collaborators. Yan King Yin wrote:

    John Scanlon wrote:
    > [...]
    > Logical deduction or inference is not thought.  It is mechanical
    symbol manipulation that can can be programmed into any scientific
    pocket calculator.
    > [...]
Hi John, I admire your attitude for attacking the core AI issues =) One is either neural-based or logic-based, using a crude
    dichotomy.  So your approach is closer to neural-based?  Mine is
    closer to the logic-based end of the spectrum.
You did not have a real argument against logical AI. What you
    said was just some sentiments about the ill-defined concept of
    "thought".  You may want to take some time to express an argument
    why logic-based AI is doomed.  In fact, both Ben's and my system
    have certain "neural" characteristics, eg being graphical, having
    numerical truth values, etc.
In the end we may all end up somewhere between logic and neural...
------------------------------------------------------------------------
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?list_id=303

-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?list_id=303

Reply via email to