OK, that's cool but the logical framework that Michalski is talking about is a representation system but not a true logical system. It can be used to represent some interesting relationships of thought-stuff.
Jim Bromer On Mon, May 19, 2014 at 3:00 PM, Piaget Modeler via AGI <[email protected]>wrote: > The forward and backward confidence parameters are adjusted by the type of > knowledge transmutation > performed over the knowledge base. > > My opinion is that Global Consistency is not important for an AGI system. > > Cyc handles consistency by using microtheories, or collections of > propositions and inference rules. > Each microtheory is consistent, but if taken altogether, there will be > global inconsistencies across > microtheories. > > In PAM-P2 we take a similar approach. We have viewpoints which are > similar to Lenat's microtheories, > but we also don't really care if premises are inconsistent. We embrace > inconsistency and rely more on > activation to sort things out. (PAM-P2 is still in process so we'll let > you know how things turn out, and > whether or not we modify our position on this point.) > > But I think Michalski's introduction of merit parameters and probability > into his logical framework has merit, > no pun intended. > > ~PM > > ------------------------------ > Date: Mon, 19 May 2014 14:42:31 -0400 > > Subject: Re: [agi] The Parts Knowledge Can be Used to Make Many > Generalizations > From: [email protected] > To: [email protected] > > > But what was the basis for the forward and backward confidence? The > problem is that this is still a logically inconsistent system posing as a > logically consistent system. I can't create logically consistent AGI > systems, but maybe I am just more honest about it. > > The consequence of this is that his logical system is merely a > representational system. I've known guys who tried to talk about ideas and > then thought they could emphasize them with pseudo-formalization (or maybe > partial-formalization). Nothing wrong with that - unless they thought that > they were actually formalizing their various conjectures. But they were > only simplifying the representation of very narrow ideas by using formal > symbols and stuff. > > So the formalization for these kinds of things are not truly consistent > abstract systems that can be used clearly as the programmatic basis's for > computer programs. It is a notation for an informal system that has limited > applications. Nothing wrong with that, but let's be honest about it. > > Jim Bromer > > > On Mon, May 19, 2014 at 1:09 PM, Piaget Modeler <[email protected] > > wrote: > > > Michalski injected probability into his system with the notion of merit > parameters, > for forward and backward confidence in statements, implying that a purely > logical > system might be insufficient to handle real world phenomena. > > ~PM. > > ------------------------------ > Date: Mon, 19 May 2014 12:25:22 -0400 > Subject: Re: [agi] The Parts Knowledge Can be Used to Make Many > Generalizations > From: [email protected] > To: [email protected]; [email protected] > > > Since false assertions can be mixed in with good assertions, the potential > complexity of an idea (a reference) cannot be neatly or easily categorized. > Michalski did mention that some logical relations are truth-preserving and > some are not but the whole idea of an underlying logical system is that > some important relations may be derived based on the abstractions. (Just as > new mathematical theories are discovered.) The important abstract relations > would typically be discovered by a close study of the applications of these > ideas to real world situations (or to the situations that the mind can > consider). But since references will contain hidden combinations of other > references and since false assertions will tend to be embedded along with > good assertions and since the reasons that would support the insights would > also be based on similar combinations of information, my conclusion is that > the potential benefit that the elaborated logical system might provide may > well be compromised and even fatally flawed by inappropriate assertions and > assumptions. > > So while I would use logic in arbitrarily constrained systems, I feel > strongly that the underlying 'logic' of an AGI system has to be comprised > of the description of the construction of the relationships of the > references. In other words it is a dynamic descriptive system that must > tend to limit the assumption that the systems are based on broad underlying > generalizations. The generalizations that I have in mind will tend to be > specialized (even though I do suppose that similar methods can be used with > them when the methods are fit to the application through trial and error.) > > I really don't have a solid idea what verification will consist of, but I > am supposing that systems of insight that can lead to reliable interactions > will have some value. > > Jim Bromer > > > On Mon, May 19, 2014 at 8:59 AM, Jim Bromer <[email protected]> wrote: > > Thanks for the reference to Inferential Theories of Learning. I found > something on the Internet. http://www.mli.gmu.edu/papers/ > 91-95/MSL4-ITL.pdf > I am glad to see that someone has been interested in looking at learning > as the ability to see how different kinds of inferences may lead to useful > knowledge. I have written (in these groups) about how I believe that > conceptual projection and the integration of different kinds of knowledge > is very important to AGI. So these can reasonably be considered as > different kinds of inferences similar to Michalski's definition. > > My feeling is that an emphasis of the formal - or general - processes that > the author likes to rely on may be a misrepresentation error. Some of his > ideas are good, and the examples are interesting. However, in detailing > some fundamental abstractions (programming abstractions) he is in effect > declaring these as special fundamental abstraction-to-generalization > methods. Maybe I should say it is a fundamental attribution error. > > The problem is that the combination will certainly, and the individual > application will probably lead to contradictions of the theory. In order to > avoid this one would have to create fundamental application definitions > which assert the kind of rule that is being applied to an actual problem. > > In other words, the attempt to rely on a fundamental abstraction or > general rule won't work. I realize that Michalski is aware of this, at > least at some level, but in his assertion that there is some kind of > competency test, (I forget what the test was based on) he is implying that > false assertions can be eliminated. They can't be. > > Sure, I will be using some kind of logic in my model. But, the underlying > principles in my model does not consist of an abstraction of logic but > simply an abstraction of construction that will describe, to some extent, > how the relations of a concept were formed. > > Jim Bromer > > > On Sun, May 18, 2014 at 1:15 PM, Piaget Modeler via AGI > <[email protected]>wrote: > > You may want to read *The Inferential Theory of Learning *by Ryszard > Michalski. > > He and Gheorghe Tecuci of GMU did some very good work in Reasoning. > > It may be helpful in your thinking about this topic. > > ~PM > > ------------------------------ > Date: Sun, 18 May 2014 12:51:40 -0400 > Subject: [agi] The Parts Knowledge Can be Used to Make Many Generalizations > From: [email protected] > To: [email protected] > > > In order to make detailed insights feasible, they need to be generalized. > I bet that almost everyone who will read this in 2014 will misunderstand > what I meant at first. I don't mean that many pieces of knowledge should be > generalized into one idea, but that the parts of many individual pieces of > knowledge can be generalized into many individualized generalizations. I am > sure that this is being implemented in some nlp, but only at a very > rudimentary level. > > The possible abstractions and combinations are uncountable. This > process then would have the capacity for immense individualization. But it > is not as simple as it might seem because computer programs that can keep > track of, refer to and wisely use an immense number of possible > combinations are not simple. > Jim Bromer > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/19999924-4a978ccc> | > Modify <https://www.listbox.com/member/?&> Your Subscription > <http://www.listbox.com> > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/24379807-f5817f28> | > Modify <https://www.listbox.com/member/?&> Your Subscription > <http://www.listbox.com> > > > > > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/24379807-f5817f28> | > Modify<https://www.listbox.com/member/?&>Your Subscription > <http://www.listbox.com> > ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657 Powered by Listbox: http://www.listbox.com
