What is with your diagrams?
On Mon, Apr 15, 2013 at 12:19 PM, Piaget Modeler <[email protected]>wrote: > Can you draw some diagrams to illustrate your theory? > > ~PM > > ------------------------------ > Date: Mon, 15 Apr 2013 11:40:35 -0400 > Subject: [agi] Re: Summary of My Current Theory For an AGI Program. > From: [email protected] > To: [email protected] > > > Part 5 > > Gradual methods seem to be called for. However, by utilizing structural > verification and integration, the gradual method can be augmented by > structural advancements where key pieces of knowledge seem to be able to > better explain a variety of related fragments of knowledge. Of course > even these methods are not absolute so there will always be the problem of > inaccurate knowledge being mixed in with the good. One of the key > problems with contemporary AGI is that ineffective knowledge (in some form) > will interfere with the effort to build even the foundations for an AGI > program. Since I do not believe that there is any method that will work > often enough to allow for a solid foundation to be easily formed, a way to > work with and around inaccurate and inadequate knowledge has to be found. > Structural integration can sometimes enhance a cohesive bunch of > inaccurate fragments of knowledge. But I believe that there are a few > things that can be done to deal with this problem. First of all, the > method of (partial) verification through structural knowledge should > usually work better with effective fragments of knowledge then it would > with inaccurate fragments. Secondly, a few kinds of flaws can often be > found in inaccurate theories. One is that they are often 'circular' or > what I call 'loopy'. Although good paradigms (mini-paradigms) are often > strongly interdependent, nonsensical paradigms do not fit well into systems > external to the central features of the paradigm. This fitting can be > done through the cross-categorization networks across transcendent > boundaries and it is an important part of understanding how good theories > work. The idea of the transcendent boundary is a solvent for the fact > that we don't really form our understanding of the world based on perfect > logic. So by being able to examine cross-categorical relations we should > be able to deal with small logical systems that can be related to other > small systems even though they may not be perfectly integrated. I think > most interested people should be able to get some idea of what I am saying > about this problem and they should be able to find examples of methods to > find flaws in simple systems of theories from real life. But there is > another problem that my theory of the transcendent boundary system would > tend to create. It would be pretty easy to build small systems that > overlay an 'insightful' bounded system and these could even be integrated > with other transcendent systems that were built to overlay other insightful > bounded systems. So a well developed fantasy system could be created on > top of related insightful systems using the methods that I have in mind. This > problem does have a solution. These systems which overlay the insightful > systems can be carefully examined to see if a viable method to tie these > into some IO observations that are directly related to the insightful > systems could be created. If a transcendent system is truly insightful, > it should typically be useful in explaining and predicting some basic > observations. Of course systems like this are not perfect and during the > initial stages of learning the program might create some elaborate systems > of nonsense. And an exhaustive search for inaccurate theories can > interfere with learning since inaccuracies that do not play key roles for > paradigms can act to support the weight of the paradigm while the 'student' > is first learning. For instance, the good student will be aware that the > fact that he does not fully understand the supporting structures (and > transcendent relations) of a paradigm does not mean that he can use his > ignorance to knock the theory down. Similarly, the fantasy that a system > (like an axiomatic system) is sufficient to support an application of the > system would not ruin that student's work with system unless he tried to > apply it to a field where the naive application was not effective (like > trying to use traditional logic to produce AGI). > > > > > On Sun, Apr 14, 2013 at 11:14 PM, Jim Bromer <[email protected]> wrote: > > Part 4 > > Artificial imagination is also necessary for AGI. Imagination can take > place simply by creating associations between concepts but obviously the > best forms of imagination are going to be based on rational meaningfulness. > An association between concepts or (concept objects) which cannot be > interpreted as meaningful is not usually very useful. So it seems that if > the relationship is both imaginative and potentially meaningful it would be > advantageous. An association formed by a categorical substitution is > more likely to be meaningful so I consider this a rational form of > imagination. However, you can find many examples where a categorical > substitution does not produce a meaningful association, so perhaps my claim > that it is a rational process is dependent on the likelihood that the > process will turn up a greater proportion of meaningful relations than > purely random associations. Some imaginative relations may exist just as > entertainment, but I believe that the application of the imagination is one > of the more important steps toward understanding. In fact, I believe > that all understanding is essentially a form of imaginative projection, > where you project previously formed ideas onto an ongoing situation which > is recognized or thought to share some characteristics with the projected > ideas. So from this point of view, the reliance of previously learned > knowledge is really an application of the imagination. Perhaps it is a > special form of imagination but the imagination none the less. Anyway, > once an imaginative association or relation is created it has to be tested. > I feel that relations of understanding cannot be appreciated out of context. > The basic rule of thumb is that it takes knowledge of many things to > understand one thing. This creates a problem when trying to test or > validate an insight which was partially produced by the imagination or > which had to be fitted using imaginative projection. The only way an AGI > program is going to be able to validate a new idea is by seeing how well it > fits and how well it works in a variety of related contexts. This is > what I call a structural integration. It not only represents a single > concept but it also carries a lot of other information with it that can > seemingly explain a lot of other small facts as well. A new idea seems > to make sense if it fits in with a number of insights that were previously > acquired. > > > > > On Sun, Apr 14, 2013 at 3:30 PM, Jim Bromer <[email protected]> wrote: > > Part 3 > > The program will make extensive use of generalizations and > cross-generalization. The program will need to be able to discover > abstractions. These abstractions typically may be used to develop > generalizations. A generalization may be formed from a group in which all > the members share some common characteristics. However, generalizations may > also be formed by various arbitrary processes. And, if the program works, > generalizations may be formed in response to some educational instruction. > The most typical example of cross-generalization may be the consideration > of similarities across individual systems of taxonomies or classes or > subclasses. In this broad definition of generalization, the collections > do not have to be grouped by any common characteristic and the same can go > for cross-categorizations. Although this might be a misuse of the term > generalization, the generalizations that my program will create may not be > trees because they can potentially branch off in different directions. > Indexes > into data for internal searches may be formed in a similar way but I will > have to think about whether the variety of branching makes sense as I am > developing the program. I believe that because of the variety of forms > of generalization or categorization that the program will use it is > necessary for the program to keep track of the different kinds of > categorization and generalization that it develops. And it will put > transcendent boundaries around portions of the generalizations that it > develops as it uses them in particular ways. These boundaries are > transcendent in that overlapping relations may be considered across them > (as in cross-generalization or cross-categorization). Perhaps the terms > relations and categorization are more abstract than the terms of > generalization. So the program will be able to develop abstractions of > relations and then build categorizations from these relations. The > categories that I have in mind may be somewhat free-wheeling. > Cross-categorization > will be important because they will help the program find and consider > similarities across the categorical structures. These categorical > structures may need to be bounded, but since bounded categories may still > be related across a relatively dominant categorical relation that means > that they can be transcended by other associative relations. > > > > > > > On Sat, Apr 13, 2013 at 7:34 AM, Jim Bromer <[email protected]> wrote: > > Part 2 > > I believe that it takes a great deal of knowledge to 'understand' one > thing. A statement has to be integrated into a greater collection of > knowledge in order for the relations of understanding to be formed. And > the knowledge of a single statement has to be integrated into a greater > field of knowledge concerning the central features of the subject for the > intelligent entity to truly understand the statement. While conceptual > integration, by some name, has always been a primary subject in AI/AGI, I > think it was relegated to a subservient position by those who originally > stressed the formal methods of logic, linguistics, psychology, numerics, > probability, and neural networks. Thinking that the details of how ideas > work in actual thinking was either part of some > predawn-of-science-philosophy or the turn-of-the-crank production of the > successful application of formal methods, a focus on the details of how > ideas work in actual problems was seen as naïve. This problem, where the > smartest thinkers would spend lives pursuing the abstract problems without > wasting their time carefully examining many real world cases occurs often > in science. It is amplified by ignorance. If no one knows how to create > a practical application then the experts in the field may become overly > pre-occupied with the proposed formal methods that had been presented to > them. Formal methods are important - but they are each only one kind of > thing. It takes a great deal of knowledge about many different things to > 'understand' one kind of thing. A reasonable rule of thumb is that > formal methods have to be tried and shaped based on exhaustive applications > of the methods to real world problems. > > In order to integrate new knowledge the new idea that is being introduced > usually has to be verified using many steps to show that it holds. Since > there is no absolute insight into truth for this kind of thing, knowledge > has to be integrated in a more thorough trial and error manner. The > program has to create new theories about statements or reactions it is > considering. This would extend to interpretations of observations for > systems where other kinds of sensory systems were used. A single > experiment does not 'prove' a new theory in science. A large number of > experiments are required and most of those experiments have to demonstrate > that the application of the theory can lead to better understanding of > other related effects. It takes a knowledge of a great many things to > verify a statement about one thing. In order for the knowledge > represented by a statement to be verified and comprehended it has to be > related to, and integrated with, a great many other statements concerning > the primary subject matter. It is necessary to see how the primary > subject matter may be used in many different kinds of thoughts to be able > to understand it. > > > On Sat, Apr 13, 2013 at 6:39 AM, Jim Bromer <[email protected]> wrote: > > Part 1 > > I feel that complexity is a major problem facing contemporary AGI. It is > true, that for most human reasoning we do not need to figure out > complicated problems precisely in order to take the first steps toward > competency but so far AGI has not been able to get very far beyond the > narrow-AI barrier. > > I am going to start with a text-based AGI program. I agree that more > kinds of IO modalities would make an effective AGI program better. However, > I am not aware of any evidence that sensory-based AGI or multi-modal > sensory based AGI or robotic based AGI has been able to achieve something > greater than other efforts. The core of AGI is not going to be found in the > peripherals. And it is clear that starting with complicated IO > accessories would make AGI programming more difficult. It seems obvious > that IO is necessary for AI/AGI and this abstraction is a probably more > appropriate basis for the requirements of AGI. > > My AGI program is going to be based on discreet references. I feel that > the argument that only neural networks are able to learn or are able to > incorporate different kinds of data objects into an associative field is > not accurate. I do, however, feel that more attention needs to be paid to > concept integration. And I think that many of us recognize that a good > AGI model is going to create an internal reference model that is a kind of > network. The discreet reference model more easily allows the program to > retain the components of an agglomeration in a way in which the traditional > neural network does not. This means that it is more likely that the > parts of an associative agglomeration can be detected. On the other > hand, since the program will develop its own internal data objects, these > might be formed in such a way so that the original parts might be difficult > to detect. With a more conscious effort to better understand concept > integration I think that the discreet conceptual network model will prove > itself fairly easily. > > I am going to use weighted reasoning and probability but only to a limited > extent. > > > > > > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/19999924-5cfde295> | > Modify <https://www.listbox.com/member/?&> Your Subscription > <http://www.listbox.com> > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/10561250-470149cf> | > Modify<https://www.listbox.com/member/?&>Your Subscription > <http://www.listbox.com> > ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657 Powered by Listbox: http://www.listbox.com
