RICHARD, I can't provide "concrete reasons" why Novamente and roughly similar approaches will work --- precisely because they are designed to operate in the realm Wolfram calls computationally irreducible --- meaning it cannot be modeled properly by anything substantially less complex than itself. Thus, whether or not it will work cannot be formally proven in advance.
I assume it would have been equally hard to provide "concrete reasons" why Hofstadter's Copycat and his similar systems would work before they were built. But they did work. Since computation irreducibility is something you make a big point about in one of your own papers you frequently cite --- you --- of all people --- should not deny the right to believe in approaches to AGI before they are supported by concrete proof. But I do have, what I consider, rational reasons for believing a Novamente-like systems will work. One of them is that you have failed to describe to me any major conceptual problem for the AGI community does not have what appear to be valid approaches. For more reasons why I believe a Novamente-like approach might work I copy the following portion of a former post of from about 5 months ago --- describing my understanding of how Novamente itself would probably work, based on reading material from Novamente and my own other reading and thinking. ======= Novamente starts with a focus on "General Intelligence", which it defines as "the ability to achieve complex goals in complex environments." It is focused on automatic, interactive learning, experiential grounding, self understanding, and both conscious (focus-of-attention) and unconscious (currently less attended) thought. It records experience, finds repeated patterns in it, makes generalizations and compositions out of such patterns -- all through multiple generalizational and compositional levels -- based on spatial, temporal, and learned-pattern-derived relationships. It uses a novel form of inference, firmly grounded in Bayesian mathematics, for deriving inferences from many millions of activated patterns at once. This provides probabilistic reasoning much more powerful and flexible than any prior Bayesian techniques. Patterns -- which can include behaviors (including mental behaviors) -- are formed, modified, generalized, and deleted all the time. They have to complete for their computational resources and continued existence. This results in a self-organizing network of similarity, generalizational, and compositional patterns and relationships, that all must continue to prove their worth in a survival-of-the-fittest, goal-oriented, experiential-knowledge ecology. Re-enforcement learning is used to weight patterns, both for general long term and context specific importance, based on the direct or indirect roles they have played in achieving the system's goals in the past. These indications of importance -- along with a deep memory for past similar experiences, goals, contexts, and similar inferencing and learning patterns -- significantly narrow and focus attention, avoiding the pitfalls of combinatorial explosion, and resulting in context-appropriate decision making. Genetic learning algorithms, made more efficient by the system's experience and probabilistic inferencing, give the system the ability to learn new behaviors, classifiers, and creative ideas. Taken together all these features, and many more, will allow the system to automatically learn, reason, plan, imagine, and create with a sophistication and power never before possible -- not even for the most intelligent of humans. Of course, it will take some time for the first such systems to learn the important aspects of world knowledge. Most of the valuable patterns in the minds of such machines will come from machine learning and not human programming. Such learning can be greatly speed if such machines are taught, at least partially, the way human children are. But once knowledge has been learned by such systems much of it can be quickly replicated into, or shared between, other machines. ======= -----Original Message----- From: Richard Loosemore [mailto:[EMAIL PROTECTED] Sent: Saturday, April 19, 2008 7:57 PM To: agi@v2.listbox.com Subject: Re: [agi] WHAT ARE THE MISSING CONCEPTUAL PIECES IN AGI? Ed Porter wrote: > WHAT ARE THE MISSING CONCEPTUAL PIECES IN AGI? > > With the work done by Goertzel et al, Pei, Joscha Bach > <http://www.micropsi.org/> , Sam Adams, and others who spoke at AGI 2008, I > feel we pretty much conceptually understand how build powerful AGI's. I'm > not necessarily saying we know all the pieces of the puzzle, but rather that > we know enough to start building impressive intelligences Ed, can you please specify *precisely* what, in the talks at AGI 2008, leads you to the conclusion that "we know enough to start building impressive intelligences"? Some might say that everything shown at AGI 2008 could be interpreted as lots of people working on the problem, but that in ten years time those same people will come back and report that they are still working on the problem, but with no substantial difference to what they showed this year. The reason that some people would say this is that if you went back ten years, you could find people achieving forms of AI that exhibited no *substantial* difference to anything shown at AGI 2008. So, I am looking for your concrete reason (not gut instinct, but concrete reason) to claim that "we know enough ...etc.". Richard Loosemore ------------------------------------------- agi Archives: http://www.listbox.com/member/archive/303/=now RSS Feed: http://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: http://www.listbox.com/member/?& Powered by Listbox: http://www.listbox.com ------------------------------------------- agi Archives: http://www.listbox.com/member/archive/303/=now RSS Feed: http://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: http://www.listbox.com/member/?member_id=8660244&id_secret=101455710-f059c4 Powered by Listbox: http://www.listbox.com