How can you think of prioritizing goals without considering the
contexts or situations that the various goals are used in? I don't
mean that you have to use explicit examples, but that you really have
to at least define an 'abstract' context or situation - as you did in
your last exposition. I don't understand why this is not essential to
the initial question.

When I reread your reply I realized that I did not quite understand
what you were getting at and when I reread my previous response I
realized that I did not quite understand some of what I had written.

I believe that as we learn using generalizations. 'The next time the
something similar happens I will...' But, how do we determine that
some event in situation A is similar to some event in situation B? My
guess is that it is easier to develop generalizations about different
situations if you are paranoid. 'The people in this group are just
interested in bashing my ideas even though they don't understand
them...'  The advantage of that kind of simplification is that you can
develop generalizations in a more structured way. The disadvantage is
that you are probably missing a lot of basic insights that that might
be useful to you. Having a special skill might also cause you to see
many events in paradigmatic ways. For example, habitually examining
the effects of events in the terms of their potential economic impact
would tend to cause you to develop many generalizations that could be
interrelated using the same measurable objective of economic value.

So then some goals supersede other goals as they tend to create new
markers of those sub-goals. But even this is relative. A sub goal can
supersede a super goal either by directly shaping it or because the
relevant relative relations between the goals are only a subset of all
of the relations between the goals.

So the prioritization of goals should tend to develop as
generalizations and specializations used to characterize and evaluate
the goals develop and as the situations that the goals are employed in
develop. Are the goals met? Are some goals (or sub-goals) interfering?
Are the characterization of the goals accurate? Are the effects of
achieving the goal ineffective in achieving a higher goal? And so on.

Jim Bromer


On Tue, Sep 30, 2014 at 9:23 PM, Piaget Modeler via AGI <[email protected]> 
wrote:
>
> Your exploration is going away from my initial question, why do goals get 
> differing priorities.
>
> But I'll bite:  We can consider a situation to be a list of elements as 
> follows:
>
> (prototype Situation
>   :Items { }
> )
>
> Your question: How with the AGI program differentiate the overgeneralizations 
> that it will tend
> to develop as it responds to various situations.
>
> Let's develop a framework with which to answer it.
>
> Let's assume an action has preconditions and post conditions.
>
> (prototype Action
>   :Preconditions { }
>   :Steps { }
>   :Postconditions { }
> )
>
>
> For now, we'll ignore the steps and postconditions and focus on the 
> preconditions.
>
> So now we can say that a precondition represents a situation just in case the 
> precondition items
> are the same as those in a situation.
>
>   (given [Situation ^ ?S :Items ?items]
>               [Action ^ ?A  (same @:Preconditions ?items) ]
>        do
>              (print 'the preconditions of action ' ?A ' match the situation ' 
> ?S) )
>
> The same goes for the postconditions of action S
>
> Suppose we have a set of situations.  And a set of actions.
>
> We can say that the items in the precondition of an action are more concrete 
> than the items in a
> situation if the the precondition contains more items than the situation.
>
> [Situation ^ situation_1 :Items { quick brown fox } ]
> [Action ^ Action_1 :Preconditions { quick brown fox jumped } ... ]
>
> We can say that the items in the precondition of an action are more abstract 
> than the items in a
> situation if the the precondition contains fewer items than the situation.
>
> [Situation ^ situation_1 :Items { quick brown fox } ]
> [Action ^ Action_1 :Preconditions { quick brown } ... ]
>
> We can say that the items in the precondition of an action are more specific 
> than the items in a
> situation if the the precondition contains  items that are lower in a 
> taxonomy than those in the situation.
>
> [Situation ^ situation_1 :Items { quick brown fox } ]
> [Action ^ Action_1 :Preconditions { quick dark-brown fox } ... ]
> [Type :Subtype dark-brown :Supertype brown ]
>
>
> We can say that the items in the precondition of an action are more general 
> than the items in a
> situation if the the precondition contains  items that are higeher in a 
> taxonomy than those in the situation.
>
> [Situation ^ situation_1 :Items { quick brown fox } ]
> [Action ^ Action_1 :Preconditions { quick brown animal } ... ]
> [Type :Subtype fox :Supertype animal ]
>
>
> So can you please refine your question at this point, so that it may be 
> answered.
>
> How will the AGI program differentiate the overgeneralizations that it will 
> tend
> to develop as it responds to various situations.
>
> Cheers,
>
> ~PM
>
> ________________________________
> Date: Tue, 30 Sep 2014 20:53:01 -0400
> Subject: Re: [agi] Are all goals created equal?
> From: [email protected]
> To: [email protected]
>
>
> On Sun, Sep 28, 2014 at 12:17 AM, Piaget Modeler via AGI <[email protected]> 
> wrote:
>
> You create an AGI and endow it with an initial set of actions and needs.
>
> You prioritize the needs so that, for example, power is more important
> than signal strength. Beyond that, you plug it in and let it run.
>
> The AGI begins to create goals and subgoals in order to satisfy its needs.
> But how does it prioritize these goals and subgoals? We can distinguish
> urgent from important goals by creating separate attributes for urgency
> and priority.
> But why would one goal get a higher priority than another, aside from
> inheriting its priority from a basic need?
>
>
> I think the terms that you are using to characterize the problem are 
> over-generalized.  I guess that when I complain about the necessity of 
> examining questions like this with more detailed characterizations that 
> people who understand what I am trying to say simply dismiss it as obvious. 
> The AGI program would be designed to use (and to learn to use) actions under 
> appropriate conditions and every programmer 'gets' what conditional actions 
> are.. But perhaps the problem that I see is that we need to write these 
> programs so that they will use broad generalizations (just as, according to 
> my point of view, you wrote your message using broad generalizations.)  So 
> then the question of how the AGI program will differentiate the 
> over-generalizations that it will tend to develop as it responds to various 
> situations becomes the more essential question. So, for example, trying to 
> get a better understanding of a particular situation and using that to 
> further shape the characterization of that 'kind' of situation may become a 
> prioritized goal just because it may be (or seem like) a necessary step that 
> the program needs to take to better understand the situation.
>
> To add a little to this view of how an AGI might work notice that a situation 
> may be characterized as being like different -kinds- of situations. This 
> basic insight is something that I never read in other people's comments - Is 
> that because it is so obvious?  A situation is comprised of the 'components' 
> of the situation and each of these may lead to the development of different 
> kinds of situations which the particular situation may be likened to. And the 
> situation may belong to other categories that are based on the projection of 
> some other kind of idea onto it.
> Jim Bromer
>
> On Sun, Sep 28, 2014 at 12:17 AM, Piaget Modeler via AGI <[email protected]> 
> wrote:
>
> You create an AGI and endow it with an initial set of actions and needs.
>
> You prioritize the needs so that, for example, power is more important
> than signal strength. Beyond that, you plug it in and let it run.
>
> The AGI begins to create goals and subgoals in order to satisfy its needs.
>
> But how does it prioritize these goals and subgoals? We can distinguish
> urgent from important goals by creating separate attributes for urgency
> and priority.
>
> But why would one goal get a higher priority than another, aside from
> inheriting its priority from a basic need?
>
> Perhaps I lack imagination.
>
> Your thoughts?
>
> ~PM
> AGI | Archives | Modify Your Subscription
>
>
> AGI | Archives | Modify Your Subscription
> AGI | Archives | Modify Your Subscription


-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to