Well, Wikipedia does give a definition of a scientific theory:

On Mon, Apr 15, 2013 at 5:03 AM, Mike Tintner <[email protected]>wrote:

>   What you have is a v. vague *hypothesis*. A *theory* involves evidence
> as to why it may work..
>
> And you have no Operational Definition of what effect you’re trying to
> achieve. Not even the teeniest weeniest hint of an O.D.
>
> Tch, tch.
>
>  *From:* Jim Bromer <[email protected]>
> *Sent:* Monday, April 15, 2013 4:14 AM
> *To:* AGI <[email protected]>
> *Subject:* [agi] Re: Summary of My Current Theory For an AGI Program.
>
>  Part 4
>
> Artificial imagination is also necessary for AGI.  Imagination can take
> place simply by creating associations between concepts but obviously the
> best forms of imagination are going to be based on rational meaningfulness.
> An association between concepts or (concept objects) which cannot be
> interpreted as meaningful is not usually very useful. So it seems that if
> the relationship is both imaginative and potentially meaningful it would be
> advantageous.  An association formed by a categorical substitution is
> more likely to be meaningful so I consider this a rational form of
> imagination.  However, you can find many examples where a categorical
> substitution does not produce a meaningful association, so perhaps my claim
> that it is a rational process is dependent on the likelihood that the
> process will turn up a greater proportion of meaningful relations than
> purely random associations.  Some imaginative relations may exist just as
> entertainment, but I believe that the application of the imagination is one
> of the more important steps toward understanding.  In fact, I believe
> that all understanding is essentially a form of imaginative projection,
> where you project previously formed ideas onto an ongoing situation which
> is recognized or thought to share some characteristics with the projected
> ideas.  So from this point of view, the reliance of previously learned
> knowledge is really an application of the imagination.  Perhaps it is a
> special form of imagination but the imagination none the less.  Anyway,
> once an imaginative association or relation is created it has to be tested.
> I feel that relations of understanding cannot be appreciated out of context.
> The basic rule of thumb is that it takes knowledge of many things to
> understand one thing.  This creates a problem when trying to test or
> validate an insight which was partially produced by the imagination or
> which had to be fitted using imaginative projection.  The only way an AGI
> program is going to be able to validate a new idea is by seeing how well it
> fits and how well it works in a variety of related contexts.  This is
> what I call a structural integration.  It not only represents a single
> concept but it also carries a lot of other information with it that can
> seemingly explain a lot of other small facts as well.  A new idea seems
> to make sense if it fits in with a number of insights that were previously
> acquired.
>
>
>
>
> On Sun, Apr 14, 2013 at 3:30 PM, Jim Bromer <[email protected]> wrote:
>
>>  Part 3
>>
>> The program will make extensive use of generalizations and
>> cross-generalization. The program will need to be able to discover
>> abstractions.  These abstractions typically may be used to develop
>> generalizations. A generalization may be formed from a group in which all
>> the members share some common characteristics. However, generalizations may
>> also be formed by various arbitrary processes. And, if the program works,
>> generalizations may be formed in response to some educational instruction.
>> The most typical example of cross-generalization may be the consideration
>> of similarities across individual systems of taxonomies or classes or
>> subclasses.  In this broad definition of generalization, the collections
>> do not have to be grouped by any common characteristic and the same can go
>> for cross-categorizations.  Although this might be a misuse of the term
>> generalization, the generalizations that my program will create may not be
>> trees because they can potentially branch off in different directions.  
>> Indexes
>> into data for internal searches may be formed in a similar way but I will
>> have to think about whether the variety of branching makes sense as I am
>> developing the program.  I believe that because of the variety of forms
>> of generalization or categorization that the program will use it is
>> necessary for the program to keep track of the different kinds of
>> categorization and generalization that it develops.  And it will put
>> transcendent boundaries around portions of the generalizations that it
>> develops as it uses them in particular ways. These boundaries are
>> transcendent in that overlapping relations may be considered across them
>> (as in cross-generalization or cross-categorization). Perhaps the terms
>> relations and categorization are more abstract than the terms of
>> generalization.  So the program will be able to develop abstractions of
>> relations and then build categorizations from these relations.  The
>> categories that I have in mind may be somewhat free-wheeling.  
>> Cross-categorization
>> will be important because they will help the program find and consider
>> similarities across the categorical structures. These categorical
>> structures may need to be bounded, but since bounded categories may still
>> be related across a relatively dominant categorical relation that means
>> that they can be transcended by other associative relations.
>>
>>
>>
>>
>>
>>
>> On Sat, Apr 13, 2013 at 7:34 AM, Jim Bromer <[email protected]> wrote:
>>
>>>  Part 2
>>>
>>> I believe that it takes a great deal of knowledge to 'understand' one
>>> thing.  A statement has to be integrated into a greater collection of
>>> knowledge in order for the relations of understanding to be formed.  And
>>> the knowledge of a single statement has to be integrated into a greater
>>> field of knowledge concerning the central features of the subject for the
>>> intelligent entity to truly understand the statement.  While conceptual
>>> integration, by some name, has always been a primary subject in AI/AGI, I
>>> think it was relegated to a subservient position by those who originally
>>> stressed the formal methods of logic, linguistics, psychology, numerics,
>>> probability, and neural networks.  Thinking that the details of how
>>> ideas work in actual thinking was either part of some
>>> predawn-of-science-philosophy or the turn-of-the-crank production of the
>>> successful application of formal methods, a focus on the details of how
>>> ideas work in actual problems was seen as naïve.  This problem, where
>>> the smartest thinkers would spend lives pursuing the abstract problems
>>> without wasting their time carefully examining many real world cases occurs
>>> often in science.  It is amplified by ignorance.  If no one knows how
>>> to create a practical application then the experts in the field may become
>>> overly pre-occupied with the proposed formal methods that had been
>>> presented to them.  Formal methods are important - but they are each
>>> only one kind of thing.  It takes a great deal of knowledge about many
>>> different things to 'understand' one kind of thing.  A reasonable rule
>>> of thumb is that formal methods have to be tried and shaped based on
>>> exhaustive applications of the methods to real world problems.
>>>
>>> In order to integrate new knowledge the new idea that is being
>>> introduced usually has to be verified using many steps to show that it
>>> holds.  Since there is no absolute insight into truth for this kind of
>>> thing, knowledge has to be integrated in a more thorough trial and error
>>> manner.  The program has to create new theories about statements or
>>> reactions it is considering.  This would extend to interpretations of
>>> observations for systems where other kinds of sensory systems were used.
>>> A single experiment does not 'prove' a new theory in science.  A large
>>> number of experiments are required and most of those experiments have to
>>> demonstrate that the application of the theory can lead to better
>>> understanding of other related effects.  It takes a knowledge of a
>>> great many things to verify a statement about one thing.  In order for
>>> the knowledge represented by a statement to be verified and comprehended it
>>> has to be related to, and integrated with, a great many other statements
>>> concerning the primary subject matter.  It is necessary to see how the
>>> primary subject matter may be used in many different kinds of thoughts to
>>> be able to understand it.
>>>
>>>
>>> On Sat, Apr 13, 2013 at 6:39 AM, Jim Bromer <[email protected]> wrote:
>>>
>>>>  Part 1
>>>>
>>>> I feel that complexity is a major problem facing contemporary AGI.  It
>>>> is true, that for most human reasoning we do not need to figure out
>>>> complicated problems precisely in order to take the first steps toward
>>>> competency but so far AGI has not been able to get very far beyond the
>>>> narrow-AI barrier.
>>>>
>>>> I am going to start with a text-based AGI program.  I agree that more
>>>> kinds of IO modalities would make an effective AGI program better.  
>>>> However,
>>>> I am not aware of any evidence that sensory-based AGI or multi-modal
>>>> sensory based AGI or robotic based AGI has been able to achieve something
>>>> greater than other efforts. The core of AGI is not going to be found in the
>>>> peripherals.  And it is clear that starting with complicated IO
>>>> accessories would make AGI programming more difficult.  It seems
>>>> obvious that IO is necessary for AI/AGI and this abstraction is a probably
>>>> more appropriate basis for the requirements of AGI.
>>>>
>>>> My AGI program is going to be based on discreet references. I feel that
>>>> the argument that only neural networks are able to learn or are able to
>>>> incorporate different kinds of data objects into an associative field is
>>>> not accurate. I do, however, feel that more attention needs to be paid to
>>>> concept integration.  And I think that many of us recognize that a
>>>> good AGI model is going to create an internal reference model that is a
>>>> kind of network.  The discreet reference model more easily allows the
>>>> program to retain the components of an agglomeration in a way in which the
>>>> traditional neural network does not.  This means that it is more
>>>> likely that the parts of an associative agglomeration can be detected.
>>>> On the other hand, since the program will develop its own internal data
>>>> objects, these might be formed in such a way so that the original parts
>>>> might be difficult to detect. With a more conscious effort to better
>>>> understand concept integration I think that the discreet conceptual network
>>>> model will prove itself fairly easily.
>>>>
>>>> I am going to use weighted reasoning and probability but only to a
>>>> limited extent.
>>>>
>>>
>>>
>>
>>
>
>   *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
> <https://www.listbox.com/member/archive/rss/303/6952829-59a2eca5> | 
> Modify<https://www.listbox.com/member/?&;>Your Subscription
> <http://www.listbox.com>
>   *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
> <https://www.listbox.com/member/archive/rss/303/10561250-470149cf> |
> Modify<https://www.listbox.com/member/?&;>Your Subscription
> <http://www.listbox.com>
>



-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to