So, as I can best understand what you are saying is that:
"neural propositions as its knowledge representation"
"Propositions are disconnected from the underlying agents that refer to and
create them. "
"There are intention prototypes and solution prototypes.  Solutions are
tried
by the system, the actions of which are attempted."

So while your program presumably has many conceptual layers and can
create more, there are constraints on the kinds of interactions that can
occur between some of the most essential 'conceptual layers' (as I called
them.) (Perhaps I should not call them conceptual layers in the context of
agi programs. Maybe I should just call them layers or something like that.)

Suppose someone convinces a young, somewhat naïve adult that he should
forget everything he has ever learned. Of course he can't do that. However,
he would be able to start to ignore certain principals which he feels that
were taught to him but which he never fully accepted. As he goes on he can
teach himself to ignore more and more of those principles. For example, as
I recognize that an idea that I thought was my own was actually instilled
by advertising I can choose to selectively ignore it. It is my feeling that
this kind of example shows that interactions between ideas (or
propositions) *and their application to thoughts* are essential to
intelligence. I suspect that General Intelligence is impossible if an idea
about shaping one's own thinking cannot be applied. The direction for this
shaping process may require some kind of justification but that
justification will sometimes require a great deal of thought.  It
can't only come from some external source of verification.

There has to be some kind of constraints on the interactions between these
levels. The program cannot forget everything it knows just because someone
makes an imperative statement to that effect. So there has to be some kind
of buffer between the propositional level and the action level. And the
argument can be made, especially for a logical system, that the agents that
act on the propositional levels are effectively capable of doing the kind
of thing that I am talking about. (Or if they are not they can be tweaked
so that they are.) However, my point here is that the management of
something like that will introduce new kinds of problems (and situations)
that require new kinds of sub programs to work on it.

I am going to read the paper and watch the video that you referenced.


Jim Bromer

On Sat, Jan 31, 2015 at 1:27 PM, Piaget Modeler via AGI <[email protected]>
wrote:

>
>
> > Date: Sat, 31 Jan 2015 10:45:27 -0500
> > Subject: Re: [agi] Multiple Conceptual Level Networks
> > From: [email protected]
> > To: [email protected]
> >
> > I am reading your (Piaget Modeler's) paper, "The Neural Proposition:
> > Structures for Cognitive Systems," but I am trying to reread it more
> > carefully to better understand it.
> >
> > So let me ask you a few questions about your project.
> > Is it an AGI application or an AGI Platform?
>
> PAM.P2  is a cognitive architecture (
> https://www.academia.edu/9997454/PAM.P2 )
> that uses neural propositions as its knowledge representation.  The ovals
> in the
> diagram represent prototypes, instances of which are referenced by the
> depicted
> agents.
>
> > You know about reification and gerunds. How does your program turn a
> > statement into an action?
>
> There are intention prototypes and solution prototypes.  Solutions are
> tried
> by the system, the actions of which are attempted.  attempts are sent to
> a device running a psyche application and results are returned as to
> whether
> the attempt succeeded or failed.
>
> > How does your program prevent a statement like, "Forget everything
> > that you know" from becoming an action that causes it to forget
> > everything that it knows?
>
> Propositions are disconnected from the underlying agents that refer to and
> create them.  That being said, an agent could run amok if incorrectly
> programmed
> and delete all the propositions of the system.  So the agents  have to be
> carefully
> programmed.
>
> ~PM
>
> >
> >
> > Jim Bromer
> >
> >
> > On Thu, Jan 29, 2015 at 12:19 PM, Piaget Modeler via AGI
> > <[email protected]> wrote:
> > > Do you mean like "Neural Propositions: Structures for Cognitive
> Systems" ?
> > >
> > > ~PM
> > >
> > >> Date: Thu, 29 Jan 2015 06:04:02 -0500
> > >> Subject: [agi] Multiple Conceptual Level Networks
> > >> From: [email protected]
> > >> To: [email protected]
> > >
> > >>
> > >> I came up with a great concept-theory using cross generalizations on
> > >> logic so I decided to write about it. As I thought about it I
> > >> remembered seeing some introductory text about network theory
> > >> somewhere and the first examples that they mentioned used binary
> > >> nodes. Some of the examples were effectively about kinds of logical
> > >> cross-generalizations. So what happened to my great new theory?
> > >> Somehow it fizzled into something that was from some introductory
> > >> text about networks. The thing is, I don't think current network
> > >> theory is very interesting.
> > >>
> > >> In order to create more interesting networks you have to have multiple
> > >> layers. Not just multiple processing layers but multiple conceptual
> > >> layers. But these concept layers should not be associated only by a
> > >> simplistic associations (on concept nodes for instance) but by the
> > >> potential for nodes on one layer to interact dramatically with other
> > >> layers. Of course this can be implemented using contemporary
> > >> conventions about nodal networks. So why is the idea of multiple
> > >> concept layers important? Because of the potential of the layered
> > >> networks to represent cross-categorical relations which might be
> > >> needed to solve difficult problems and which might be more susceptible
> > >> to effective methods of analysis.
> > >>
> > >> When Internet traffic is being analyzed, for example, the analysis
> > >> occurs on a different conceptual level than the traffic itself. In
> > >> this case, there is very limited interaction with the traffic and the
> > >> analysis. If the analysis is sent to a web manager then the analytical
> > >> function is itself producing some traffic on the same system. The
> > >> number of conceptual levels in this example is extremely constricted
> > >> (there are 2 levels) and the interaction between the levels is tightly
> > >> constrained as well.
> > >>
> > >> But it is easy to imagine systems where there are many different kinds
> > >> of conceptual levels and a lot of different ways interaction can
> > >> occur. Can you do this with conventional notions about sub-networks?
> > >> Ok, but there are times when you need to free your mind from
> > >> conventional thinking.
> > >> Jim Bromer
> > >>
> > >>
> > >> -------------------------------------------
> > >> AGI
> > >> Archives: https://www.listbox.com/member/archive/303/=now
> > >> RSS Feed:
> https://www.listbox.com/member/archive/rss/303/19999924-4a978ccc
> > >> Modify Your Subscription: https://www.listbox.com/member/?&;
> > >> Powered by Listbox: http://www.listbox.com
> > > AGI | Archives | Modify Your Subscription
> >
> >
> > -------------------------------------------
> > AGI
> > Archives: https://www.listbox.com/member/archive/303/=now
> > RSS Feed:
> https://www.listbox.com/member/archive/rss/303/19999924-4a978ccc
> > Modify Your Subscription: https://www.listbox.com/member/?&;
> > Powered by Listbox: http://www.listbox.com
>    *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
> <https://www.listbox.com/member/archive/rss/303/24379807-653794b5> |
> Modify
> <https://www.listbox.com/member/?&;>
> Your Subscription <http://www.listbox.com>
>



-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to