I think that there are two basic directions to better the Novamente architecture: the one Mark talks about more integration of MOSES with PLN and RL theory
On 11/13/07, Edward W. Porter <[EMAIL PROTECTED]> wrote: > Response to Mark Waser Mon 11/12/2007 2:42 PM post. > > > > >MARK>>>> Remember that the brain is *massively* parallel. Novamente and > any other linear (or minorly-parallel) system is *not* going to work in > the same fashion as the brain. Novamente can be parallelized to some > degree but *not* to anywhere near the same degree as the brain. I love > your speculation and agree with it -- but it doesn't match near-term > reality. We aren't going to have brain-equivalent parallelism anytime in > the near future. > > > > ED>>>> I think in five to ten years there could be computers capable of > providing every bit as much parallelism as the brain at prices that will > allow thousands or hundreds of thousands of them to be sold. > > > > But it is not going to happen overnight. Until then the lack of brain > level hardware is going to limit AGI. But there are still a lot of high > value system that could be built on say $100K to $10M of hardware. > > > > You claim we really need experience with computing and controlling > activation over large atom tables. I would argue that obtaining such > experience should be a top priority for government funders. > > > > >MARK>>>> The node/link architecture is very generic and can be used for > virtually anything. There is no rational way to attack it. It is, I > believe, going to be the foundation for any system since any system can > easily be translated into it. Attacking the node/link architecture is > like attacking assembly language or machine code. Now -- are you going to > write your AGI in assembly language? If you're still at the level of > arguing node/link, we're not communicating well. > > > > ED>>>> nodes and links are what patterns are made of, and each static > pattern can have an identifying node associated with it as well as the > nodes and links representing its sub-patterns, elements, the compositions > of which it is part, it associations, etc. The system automatically > organize patterns into a gen/comp hierarchy. So, I am not just dealing at > a node and link level, but they are the basic building blocks. > > > > > > >MARK>>>> ... I *AM* saying that the necessity of using probabilistic > reasoning for day-to-day decision-making is vastly over-rated and has been > a horrendous side-road for many/most projects because they are attempting > to do it in situations where it is NOT appropriate. The "increased, > almost ubiquitous adaptation of probabilistic methods" is the herd > mentality in action (not to mention the fact that it is directly > orthogonal to work thirty years older). Most of the time, most projects > are using probabilistic methods to calculate a tenth place decimal of a > truth value when their data isn't even sufficient for one. If you've got > a heavy-duty discovery system, probabilistic methods are ideal. If you're > trying to derive probabilities from a small number of English statements > (like "this raven is white" and "most ravens are black"), you're seriously > on the wrong track. If you go on and on about how humans don't understand > Bayesian reasoning, you're both correct and clueless in not recognizing > that your very statement points out how little Bayesian reasoning has to > do with most general intelligence. Note, however, that I *do* believe > that probabilistic methods *are* going to be critically important for > activation for attention, etc. > > > > ED>>>> I agree that many approaches accord too much importance to the > numerical accuracy and Bayesian purity of their approach, and not enough > importance on the justification for the Bayesian formulations they use. > I know of one case where I suggested using information that would almost > certainly have improved a perception process and the suggestion was > refused because it would not fit within the system's probabilistic > framework. At an AAAI conference in 1997 I talked to a programmer for a > big defense contractor who said he as a fan of fuzzy logic system; that > they were so much more simple to get up an running because you didn't have > to worry about probabilistic purity. He said his group that used fuzzy > logic was getting things out the door that worked faster than the more > probability limited competition. So obviously there is something to say > for not letting probabilistic purity get in the way of more reasonable > approaches. > > > > But I still think probabilities are darn important. Even your "this raven > is white" and "most ravens are black" example involves notions of > probability. We attribute probabilities to such statements based on > experience with the source of such statements or similar sources of > information, and the concept "most" is a probabilistic one. The reason we > humans are so good at reasoning from small data is based on our ability to > estimate rough probabilities from similar or generic patterns. > > > > >MARK>>>> ....The problem with probability-based conflict resolution is > that it is a hack to get around insufficient knowledge rather than an > attempt to figure out how to get more knowledge.... > > > > ED>>>> This agrees with what I said above about not putting enough > emphasis on selecting what probabilistic formulas are appropriate. But it > doesn't argue against the importance of probabilities It argues against > using them blindly. > > > > > >>ED>>>> So by "operating with small amounts of data" how small, very > roughly, are you talking about. And are you only talking about the active > goals or sources of activation, that will be small or are you saying that > all the computation in the system will only be dealing with a small amount > of data within, for example, one second of the processing of human-level > system operating at human-level speed? > > > > >MARK>>>> I mean like the way humans reason, there is only concentration > on a small number of objects -- which are only one link away from an > almost inconceivable number of related things -- and then the brain can > jump at least three of these links with lightning rapidity. > > > > ED>>>> So this implies you are not arguing against the idea that AGI will > be dealing with massive data, just that that use will be focused by a > concentration on a relatively small number of sources of activation at > once. > > > > > > >MARK>>>> Ask Ben how much actual work has been done on activation > control in very large, very sparse atom spaces in Novamente. He'll tell > you that it's a project for when he's further along. I'll insist (as will > Richard) that if it isn't baked in from the very beginning, you're > probably going to have to go back to the beginning to repair the lack. > > > > ED>>>> It is exactly such research I want to see funded. It strikes me > as one of the key things we must learn to do well to make powerful AGI. > But I think even with some fairly dumb activation control systems you > could get useful results. Such results would not be at all human-level in > may ways, but in other ways they could be much more powerful because such > systems could deal with many more explicit facts and could input and > output information at a much higher rate than humans. > > > > For example, what is the equivalent of the activation control (or search) > algorithm in Google sets. They operate over huge data. I bet the > algorithm for calculating their search or activation is relatively simple > (much, much, much less than a PhD theses) and look what they can do. So I > think one path is to come up with applications that can use and reason > with large data, having roughly world knowledge-like sparseness, (such as > NL data) and start with relatively simple activation algorithms and > develop then from the ground up. > > > > >MARK>>>> P.S. Oh yeah -- if you were public enemy number one, I > wouldn't bother answering you (and I probably should lay off of the > fan-boy crap :-). > > > > ED>>>> Thanks. > > > > I admit I am impressed with Novamente. Since it's the best AGI > architecture I currently know of; I am impressed with Ben; believe there > is a high probability all the gaps you address could be largely fixed > within five years with deep funding (which may never come); and since I > want to get such deep funding for just the type of large atom-base work > you say is so critical, I think it is important to focus on the potential > for greatness that Novamente and somewhat similar systems have, rather > than only think of its current gaps and potential problems. > > > > But of course, at the same time, we must look for and try to understand > its gaps and potential problems so that we can remove them. > > > > Ed Porter > > > > > -----Original Message----- > From: Mark Waser [mailto:[EMAIL PROTECTED] > Sent: Monday, November 12, 2007 2:42 PM > To: [email protected] > Subject: Re: [agi] What best evidence for fast AI? > > > >> It is NOT clear that Novamente documentation is NOT enabling, or could > not be made enabling, with, say, one man year of work. Strong argument > could be made both ways. > > I believe that Ben would argue that Novamente documentation is NOT > enabling even with one man-year of work. Ben? There is still way to much > *research* work to be done. > > >> But the standard for non-enablement is very arguably weaker than not > requiring a miracle. It would be more like "not requiring a leap of > creativity that is outside the normal skill of talented PhDs trained in > related fields". > > >> So although your position is reasonable, I hope you understand so is > that on the other side. > > > My meant-to-be-humorous miracle phrasing is clearly throwing you. The > phrase "not requiring a leap of creativity that is outside the normal > skill of talented PhDs trained in related fields" works for me. Novamente > is *definitely* not there yet. I'm rather sure that Ben would agree -- as > in, I'm not on the other side, *you* are on the other side from the > system's designer. Again, Ben please feel free to chime in. > > >> <much scaling stuff> > > Remember that the brain is *massively* parallel. Novamente and any > other linear (or minorly-parallel) system is *not* going to work in the > same fashion as the brain. Novamente can be parallelized to some degree > but *not* to anywhere near the same degree as the brain. I love your > speculation and agree with it -- but it doesn't match near-term reality. > We aren't going to have brain-equivalent parallelism anytime in the near > future. > > >> "with regard to serious review of memory design" I don't know what you > mean. Are you attacking the node, link architecture, or what? > > The node/link architecture is very generic and can be used for > virtually anything. There is no rational way to attack it. It is, I > believe, going to be the foundation for any system since any system can > easily be translated into it. Attacking the node/link architecture is > like attacking assembly language or machine code. Now -- are you going to > write your AGI in assembly language? If you're still at the level of > arguing node/link, we're not communicating well. > > >> I don't understand this. If there as been one major transformation in > AI since the mid-80's it is the increased, almost ubiquitous adaptation of > probabilistic methods. Are you claiming probabilistic reasoning is not > important?. > > It depends upon what you mean by probabilistic reasoning. I *AM* > saying that the necessity of using probabilistic reasoning for day-to-day > decision-making is vastly over-rated and has been a horrendous side-road > for many/most projects because they are attempting to do it in situations > where it is NOT appropriate. The "increased, almost ubiquitous adaptation > of probabilistic methods" is the herd mentality in action (not to mention > the fact that it is directly orthogonal to work thirty years older). Most > of the time, most projects are using probabilistic methods to calculate a > tenth place decimal of a truth value when their data isn't even sufficient > for one. If you've got a heavy-duty discovery system, probabilistic > methods are ideal. If you're trying to derive probabilities from a small > number of English statements (like "this raven is white" and "most ravens > are black"), you're seriously on the wrong track. If you go on and on > about how humans don't understand Bayesian reasoning, you're both correct > and clueless in not recognizing that your very statement points out how > little Bayesian reasoning has to do with most general intelligence. Note, > however, that I *do* believe that probabilistic methods *are* going to be > critically important for activation for attention, etc. > > >> With regard to knowledge-conflict-resolution, Novamente's probabilistic > reasoning is designed to deal with it. Most of the other system I know of > that deal with knowledge-conflict-resolution, such as constraint > relaxation techniques, are probability based. > > This is where I believe that probabilistic reasoning is most often > improperly used though I don't believe that "most" constraint-relaxation > systems are probability-based (except, occasionally as an add-on to just > why a given constraint was relaxed rather than another). The problem with > probability-based conflict resolution is that it is a hack to get around > insufficient knowledge rather than an attempt to figure out how to get > more knowledge. It works because you always take the highest probability > choice -- except when the system tells you that the sauna is hot because > it doesn't know about the ice frozen over the top. In data-rich > constrained environments, probabilistic reasoning works (and neural > networks are very successful). In every day life . . . . it still works > because all your probabilities are near 100% . . . . except when they > suddenly aren't. > > >> So by "operating with small amounts of data" how small, very roughly, > are you talking about. And are you only talking about the active goals or > sources of activation, that will be small or are you saying that all the > computation in the system will only be dealing with a small amount of data > within, for example, one second of the processing of human-level system > operating at human-level speed? > > I mean like the way humans reason, there is only concentration on a > small number of objects -- which are only one link away from an almost > inconceivable number of related things -- and then the brain can jump at > least three of these links with lightning rapidity. Once again, the brain > is *massively* parallel and operates with a *huge* sparse matrix. > Activation is *far* more important than truth probabilities and much of > the focus is the other way (and activation is a really tough nut to solve > as you rightly point out with your comments about activation control). > Ask Ben how much actual work has been done on activation control in very > large, very sparse atom spaces in Novamente. He'll tell you that it's a > project for when he's further along. I'll insist (as will Richard) that > if it isn't baked in from the very beginning, you're probably going to > have to go back to the beginning to repair the lack. > > Mark > > P.S. Oh yeah -- if you were public enemy number one, I wouldn't bother > answering you (and I probably should lay off of the fan-boy crap :-). > > _____ > > This list is sponsored by AGIRI: http://www.agiri.org/email > To unsubscribe or change your options, please go to: > http://v2.listbox.com/member/? > <http://v2.listbox.com/member/?& > > & > > ----- > This list is sponsored by AGIRI: http://www.agiri.org/email > To unsubscribe or change your options, please go to: > http://v2.listbox.com/member/?& ----- This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244&id_secret=64923307-824e30
