Jim, your prior e-mail reads like you are either a chatbot or are attempting NLP (Neuro-Linguistic Programming) or DHE. Just ask, my answer may be yes or no. My own reason for assisting is that I'd like you to understand my approach. Differentiation IS conditional branching. Observation is receiving sensory stimuli. Coordination means making inferences.Integration is combining different concepts via their attributes akin to crossover or memetic recombination. Please define verification. It may be what I call correlation. Cheers! Imagine, NLP via e-mail. Whooda thunk it?
Date: Wed, 5 Dec 2012 07:54:16 -0500 Subject: Re: [agi] Internal Representation From: [email protected] To: [email protected] I agree with Piaget's remark. I am going to conduct an experiment. I want to see if I can get you to solve a problem for me. So I am going to keep track of our conversation by keeping notes on particular issues related to this experiment. It is unlikely that you would be able to solve a particular problem that is of interest to me, so I am going to be looking for an unexpected solution to some related problem that I will pick up somewhat serendipitously from our conversation. The best way to get you to cooperate with me on this is to get you talk about the thing you are interested in. However, the solutions to the problems of your projects probably will not be the solutions to the problems of my projects, so I have to find a way to get you talk about something that is common to both of our projects. So I have gotten you to describe some ways that your program can apply imagination to problem solving. Your seem to acknowledge that integration is a part of the process, but you haven't acknowledged that complexity is a problem. So now, in order to get you to continue discussing this I have to back off from talking about complexity and emphasize the problems of 'verifying' and integrating internal projections. I will review your message in response to my question of how your program will use imagination, and I will copy that response into my notes. Now that I have reviewed some of your previous messages I see that you mentioned Piaget's comments on coordination before. Coordination seems to be very similar to conceptual integration. I also found that you had told me that Michalski had a fast inferencing method so that must be important to you for some reason. So, to repeat myself for clarity. I am going to run a subjective experiment for a couple of weeks. The goal is to get you to solve a problem for me and I want to be able to note how I personally integrate subject related serendipity into my knowledge structures concerning the subject. It is unlikely that you would be able to solve a problem that I specified in advance so I am going to look for an unexpected serendipitous solution to some problem that I haven't yet completely identified. In order to get you to participate in this experiment I need to encourage you to talk about your project using terms that are relevant to both of us. Since I will be keeping notes I have started by reviewing and collecting some of the comments you made in this thread. I can then use this knowledge to get you to continue talking about things that interest you. I noted that you have not acknowledged that complexity is a problem so I will back off that particular problem and try to shift to integration (coordination) issues that seem challenging for an automated AGI program to use effectively. Now that I have explained this 'experiment' to you I will stop talking about it and get back to the subject. On the list of mental coordination methods, internal simulation methods and inferencing you did not specifically mention conditional branching so there is a chance that you (or Piaget) left that off the list. I would say that is a pretty important concept! On the other hand, running different methods to use in a comparison with perceived events seems to imply a conditional branching. Anyway, the next question I have for you concerns 'verification' and integration (coordination). Without strong verification, coordination is essentially going to tie weak inferences together. If you accept that this could be a problem then how would your program use the products of coordination reliably? Jim Bromer On Tue, Dec 4, 2012 at 11:45 PM, Piaget Modeler <[email protected]> wrote: "The central idea is that knowledge proceeds neither solely from the experience of objects nor from an innate programming performed in the subject, but from successive constructions, the result of constant development of new structures.” ~ Jean Piaget So I think we knit together these insights, piecemeal, until they recur and strengthen, and becomemore predictable and forceful in our minds. Then they integrate and form a larger structure, and eventually they become a subsystem, integrating with other subsystems, until they finally integratewith the totality. Or at least that's how I interpreted it in "The Development of Thought" by J.Piaget. Cheers. ~PM. Date: Tue, 4 Dec 2012 23:12:06 -0500 Subject: Re: [agi] Internal Representation From: [email protected] To: [email protected] Well, I would look at Ryszard Michalski's work on dynamically interlaced hierarchies if it was convenient for me to do so. Nothing about this is mentioned on his home page and the first reference I looked at did not seem like a breakthrough paper. I want to finish something that I was thinking about. We (or a machine) would be able to build strong knowledge if the knowledge that was gained could be used to reliably predict, explain or produce a specific outcome. But often, the outcomes are weak or unreliable indicators of much of value. So instead we are left with a lot of weakly related situation-action-reaction insights that are inexplicably conditional and variant. This is a lot like serendipitous learning. If I try to learn something, I probably won't be able to figure out what I wanted to figure out (unless it is something that other people had already figured out and it was within my field of knowledge). But I would probably learn something new serendipitously. Now can we patch a lot of weak unexpected insights together? Yes, but in order to build something reliable out of a lot of weak structural pieces they have to be integrated pretty thoroughly. The integration does not have to perfect but the matrix of these things have to be strong enough to serve as a foundation for greater insights. Jim Bromer On Tue, Dec 4, 2012 at 9:31 PM, Piaget Modeler <[email protected]> wrote: I would agree that you also need mult-strategy reasoning in addition to correlations. Look at Rysard Michalski's work on dynamically interlaced hierarchies. He has a fast and efficient mechanism for inference. He inspired me. Cheers, ~PM. Date: Tue, 4 Dec 2012 18:36:20 -0500 Subject: Re: [agi] Internal Representation From: [email protected] To: [email protected] I discovered something about logic that I never knew before. It is something that I have thought about for 40 years, but I never stopped to explore the application. Now, shouldn't this new insight give me greater understanding? Well, yeah, but it doesn't work that way. I have a new insight but I haven't got any use for it. So now I have to try to find some practical use for it. Well even though I don't have any use for it, I might pick up some street creds by telling other people about it right? Well no, not really. It is really a turn-the-crank kind of thing and the fact that I thought about it for so long without ever once examining its application is kind of embarrassing. So now, before I can talk about it I have to search for some way to use the idea effectively. If I found some utility for it then I could pick up some credit for it, but until then it is just going to make my work with logic more complicated. The insight was a turn-the-crank kind of insight so it represented the application of a familiar idea onto another familiar idea in a way that was very familiar to me. The only thing I did different was to actually see how it worked in a few examples. When I did that I realized that the effects were not exactly what I expected. However, logic is an artificial field which is well formed so that other logic-based ideas, like something from mathematics, can sometimes be easily integrated into it. In real world examples of ideative projection, the analysis of turn-the-crank imagination cannot easily be achieved just by using other (integrated or related) methods of internal ideative projection. And as I just explained, simple correlation methods are not an easy substitute for insightful methods. Jim Bromer AGI | Archives | Modify Your Subscription AGI | Archives | Modify Your Subscription AGI | Archives | Modify Your Subscription AGI | Archives | Modify Your Subscription AGI | Archives | Modify Your Subscription ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657 Powered by Listbox: http://www.listbox.com
