Hi Jim,

This reminds me of Self-organised criticality and Per Bak's simple idea on
how it could be applied to learning in neural networks.

https://en.wikipedia.org/wiki/Self-organized_criticality

https://www.simonsfoundation.org/quanta/20140403-a-fundamental-theory-to-model-the-mind/

I find the simplicity of the idea very attractive. I doubt that it is
enough, but I wouldn't be surprised if it ends up playing a central role in
AGI.

Best,
Telmo.


On Thu, Apr 17, 2014 at 3:31 PM, Jim Bromer <[email protected]> wrote:

> There is a lot of evidence that humans, like other animals, learn
> incrementally. However, my belief is that because we use ideas in different
> ways a new idea can interact with other ideas. There are moments when
> something that is learned incrementally can be leveraged to produce leaps
> of insight. I call this knowledge structural because it means that an idea
> can suddenly provide some greater structure to knowledge related to a
> particular subject. The new increment of knowledge that triggers the
> structural insight may or may not be the key that provides the leverage of
> the structure. It may be that some new piece of knowledge just helps to
> crystalize some structure in a way that helps the learner to better utilize
> other knowledge.
>
> In programming and computational mathematics we find distinctions between
> things like operators and operands and you have to be able to find
> distinctions between other different parts of a computation if you want to
> use mathematics creatively. However, I think it is obvious that the
> situation is more dynamic and more fluid in thought. Some information may
> play some role based on some other information so that it can react with
> some other information and we just cannot categorize how some piece of
> information might be used before hand.  An AGI program has to be able to
> find how information can work together to create greater structures of
> knowledge. But for this to happen, the program has to be designed to
> provide the structure that will ensure that the potential to build learned
> structures is there.
> Jim Bromer
>    *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
> <https://www.listbox.com/member/archive/rss/303/25129130-ee4f7d55> |
> Modify<https://www.listbox.com/member/?&;>Your Subscription
> <http://www.listbox.com>
>



-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to