Coherence is one way for old information to be rejected as well. Belief parameters for coherence in PAM.P2 includes a set of accepted and rejectedpropositions as well as a positive or negative coherence score. As new information is received, it may cause old information to become less coherentand rejected. A cohence propagator continually revises and propagates the coherence score for activated propositions. The coherence propagation agentin PAM.P2 is based on Paul Thagard's work at the University of Waterloo. ~PM From: [email protected] To: [email protected] Subject: RE: [agi] Multiple Conceptual Level Networks Date: Sun, 1 Feb 2015 12:42:06 -0800
So I think what you are referring to is new information overriding old information. This could happen for many reasons but the essence is that some belief parameters(e.g., certainty, arousal, valence, activation, etc.) on the new information cause the new information to be considered before or superior to the older information. In PAM.P2 the basic connection between statements and action is that an intention is formed, representing the desirability of bringing about situation (or proposition), and an action selector (or solver) works to find (or create) a solution that can achieve the situation (or proposition). ~PM Date: Sun, 1 Feb 2015 10:28:18 -0500 Subject: Re: [agi] Multiple Conceptual Level Networks From: [email protected] To: [email protected]; [email protected] So, as I can best understand what you are saying is that:"neural propositions as its knowledge representation""Propositions are disconnected from the underlying agents that refer to and create them. ""There are intention prototypes and solution prototypes. Solutions are tried by the system, the actions of which are attempted." So while your program presumably has many conceptual layers and can create more, there are constraints on the kinds of interactions that can occur between some of the most essential 'conceptual layers' (as I called them.) (Perhaps I should not call them conceptual layers in the context of agi programs. Maybe I should just call them layers or something like that.) Suppose someone convinces a young, somewhat naïve adult that he should forget everything he has ever learned. Of course he can't do that. However, he would be able to start to ignore certain principals which he feels that were taught to him but which he never fully accepted. As he goes on he can teach himself to ignore more and more of those principles. For example, as I recognize that an idea that I thought was my own was actually instilled by advertising I can choose to selectively ignore it. It is my feeling that this kind of example shows that interactions between ideas (or propositions) *and their application to thoughts* are essential to intelligence. I suspect that General Intelligence is impossible if an idea about shaping one's own thinking cannot be applied. The direction for this shaping process may require some kind of justification but that justification will sometimes require a great deal of thought. It can't only come from some external source of verification. There has to be some kind of constraints on the interactions between these levels. The program cannot forget everything it knows just because someone makes an imperative statement to that effect. So there has to be some kind of buffer between the propositional level and the action level. And the argument can be made, especially for a logical system, that the agents that act on the propositional levels are effectively capable of doing the kind of thing that I am talking about. (Or if they are not they can be tweaked so that they are.) However, my point here is that the management of something like that will introduce new kinds of problems (and situations) that require new kinds of sub programs to work on it. I am going to read the paper and watch the video that you referenced. Jim Bromer On Sat, Jan 31, 2015 at 1:27 PM, Piaget Modeler via AGI <[email protected]> wrote: > Date: Sat, 31 Jan 2015 10:45:27 -0500 > Subject: Re: [agi] Multiple Conceptual Level Networks > From: [email protected] > To: [email protected] > > I am reading your (Piaget Modeler's) paper, "The Neural Proposition: > Structures for Cognitive Systems," but I am trying to reread it more > carefully to better understand it. > > So let me ask you a few questions about your project. > Is it an AGI application or an AGI Platform? PAM.P2 is a cognitive architecture ( https://www.academia.edu/9997454/PAM.P2 )that uses neural propositions as its knowledge representation. The ovals in the diagram represent prototypes, instances of which are referenced by the depictedagents. > You know about reification and gerunds. How does your program turn a > statement into an action? There are intention prototypes and solution prototypes. Solutions are tried by the system, the actions of which are attempted. attempts are sent to a device running a psyche application and results are returned as to whetherthe attempt succeeded or failed. > How does your program prevent a statement like, "Forget everything > that you know" from becoming an action that causes it to forget > everything that it knows? Propositions are disconnected from the underlying agents that refer to and create them. That being said, an agent could run amok if incorrectly programmedand delete all the propositions of the system. So the agents have to be carefully programmed. ~PM > > > Jim Bromer > > > On Thu, Jan 29, 2015 at 12:19 PM, Piaget Modeler via AGI > <[email protected]> wrote: > > Do you mean like "Neural Propositions: Structures for Cognitive Systems" ? > > > > ~PM > > > >> Date: Thu, 29 Jan 2015 06:04:02 -0500 > >> Subject: [agi] Multiple Conceptual Level Networks > >> From: [email protected] > >> To: [email protected] > > > >> > >> I came up with a great concept-theory using cross generalizations on > >> logic so I decided to write about it. As I thought about it I > >> remembered seeing some introductory text about network theory > >> somewhere and the first examples that they mentioned used binary > >> nodes. Some of the examples were effectively about kinds of logical > >> cross-generalizations. So what happened to my great new theory? > >> Somehow it fizzled into something that was from some introductory > >> text about networks. The thing is, I don't think current network > >> theory is very interesting. > >> > >> In order to create more interesting networks you have to have multiple > >> layers. Not just multiple processing layers but multiple conceptual > >> layers. But these concept layers should not be associated only by a > >> simplistic associations (on concept nodes for instance) but by the > >> potential for nodes on one layer to interact dramatically with other > >> layers. Of course this can be implemented using contemporary > >> conventions about nodal networks. So why is the idea of multiple > >> concept layers important? Because of the potential of the layered > >> networks to represent cross-categorical relations which might be > >> needed to solve difficult problems and which might be more susceptible > >> to effective methods of analysis. > >> > >> When Internet traffic is being analyzed, for example, the analysis > >> occurs on a different conceptual level than the traffic itself. In > >> this case, there is very limited interaction with the traffic and the > >> analysis. If the analysis is sent to a web manager then the analytical > >> function is itself producing some traffic on the same system. The > >> number of conceptual levels in this example is extremely constricted > >> (there are 2 levels) and the interaction between the levels is tightly > >> constrained as well. > >> > >> But it is easy to imagine systems where there are many different kinds > >> of conceptual levels and a lot of different ways interaction can > >> occur. Can you do this with conventional notions about sub-networks? > >> Ok, but there are times when you need to free your mind from > >> conventional thinking. > >> Jim Bromer > >> > >> > >> ------------------------------------------- > >> AGI > >> Archives: https://www.listbox.com/member/archive/303/=now > >> RSS Feed: https://www.listbox.com/member/archive/rss/303/19999924-4a978ccc > >> Modify Your Subscription: https://www.listbox.com/member/?& > >> Powered by Listbox: http://www.listbox.com > > AGI | Archives | Modify Your Subscription > > > ------------------------------------------- > AGI > Archives: https://www.listbox.com/member/archive/303/=now > RSS Feed: https://www.listbox.com/member/archive/rss/303/19999924-4a978ccc > Modify Your Subscription: https://www.listbox.com/member/?& > Powered by Listbox: http://www.listbox.com AGI | Archives | Modify Your Subscription AGI | Archives | Modify Your Subscription ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657 Powered by Listbox: http://www.listbox.com
