Would it only be costs and benefits or would we include a full fledged
opportunity "object" with whatever traits that may entail? And does this
notion differ fom Serendipity? Erik T Mueller had a serendipity
component/routine in his Daydreamer system, which recognized when goals
were serendipitously attained.
~PM
> Date: Tue, 12 Jun 2012 07:56:50 -0600
> From: [email protected]
> To: [email protected]
> CC: [email protected]
> Subject: Re: [agi] Attention
>
> PM
>
> A few months ago I mentioned to Jim Bromer that I would build the
> intelligence around the idea of opportunity. This fits directly with
> the idea of goals, in a somewhat obscure way.
>
> If you look at your candidates list below, you see part of what makes
up
> an opportunity. It leaves out important factors that go into choosing
> which opportunity is "greatest" for the moment.
>
> These additional factors get complicated. For example -
> - how reliable are the claims of this opportunity? If we pursue this
> opportunity, can we be somewhat certain we will complete the steps
> needed to consummate and yield the benefit?
>
> - what is the opportunity and how does it compare to other
opportunity.
> In other words, there needs to be a vocabulary of benefit and cost
> that goes with consideration of any opportunity. Here is where the
real
> world is complex. We venture and discover many benefits that we didn't
> expect. And we encounter costs that exceed our expectations. Will
> there be shortages we didn't anticipate...
>
> - where did this opportunity come from? If we want to evaluate
> opportunity, we also need to consider the source. Is this simply an
> idea that we cooked up according to simple logic? Is this an idea that
> comes highly recommended from a reliable source - a master chef.
>
> I understand that the ideas listed above seem off-base from looking at
> goals, but the real world "intelligent" doesn't work via figuring
> everything out. It works from suggestion and observation of what works
> and how it worked out. The intelligence that we want to embody will
one
> day need to be able to choose between opportunity rather than simply
> follow "sub-goals" that make up a recipe.
>
> To get this intelligence going, build the database of opportunity.
>
> Stan
>
> (sorry if I missed a relevant prior post, I'm way behind in list
reading
> - busy time of year... yade yada .)
>
>
>
>
> On 06/11/2012 06:31 PM, Piaget Modeler wrote:
> > Abram, you've characterized it properly. In my vernacular subgoals
= goals.
> >
> >
> > I would say that the job of this particular attention module is to
> > reprioritize the open goal set,
> > given all available information.
> >
> > So the question for me is what should all available information
consist of?
> >
> > Some candidates are: (1) The current context, for sure, (2) alerts,
(3)
> > expectation failures and mismatches,
> > (4) past prioritizations, (5) past episodes.
> >
> > Anything else?
> >
> > Your thoughts?
> >
> >
> > Date: Mon, 11 Jun 2012 11:11:58 -0700
> > Subject: Re: [agi] Attention
> > From: [email protected]
> > To: [email protected]
> >
> > PM,
> >
> > OK. So, in this case, the goal selector is clearly selecting
subgoals to
> > prioritize.
> >
> > It's a difficult question which needs a quickly computable answer,
so
> > the system needs to somehow gather information over time which tells
it
> > what subgoals have been most useful in the past, in what situations.
> > This process can use a wide variety of information; essentially
> > anything. However, to make an efficient choice, the information
> > considered at any particular time needs to be narrowed down
somehow. The
> > space of possible sub-goals is also potentially difficult, and needs
to
> > be narrowed down heuristically...
> >
> > Perhaps the best that I can say at the moment is, this seems like
the
> > sort of problem which requires empirical testing to see what works
and
> > what doesn't!
> >
> > --Abram
> >
> > On Fri, Jun 8, 2012 at 5:49 PM, Piaget Modeler
> > <[email protected] <mailto:[email protected]>>
wrote:
> >
> >
> > Ben
> >
> > Yours is a sufficient response. Thank you.
> >
> > Abram
> >
> > Suppose we decompose a cognitive system down into a few components:
> >
> > 1. A planner (which is fed a goal, a current state and a set of
> > possible actions (i.e., operators, methods, cases, etc.)),
> > 2. An action selector (which is fed the current state, a prioritized
> > set of goals, and a set of methods to choose from),
> > 3. A goal selector / Attention module whose job is to prioritize or
> > select goals for the cognitive system.
> >
> > My question was what would you feed the goal selector to ensure it
> > did its job (prioritizing goals) properly?
> >
> > In a paper I read recently "A Case Study of Goal-Driven Autonomy in
> > Domination Games" by Hector Munoz-Avila and David W. Aha
> > the authors, in their CB-gda system, decompose the cognitive system
> > into two case-based components (a) a planning component,
> > and (b) a mismatch goal [selection] component. The purpose of the
> > latter component was to correct for errors encountered by the
> > planner. The input for the mismatch goal selection component is a
> > mismatch (the difference between the expected state and the
> > goal state).
> >
> > Q: What else would be relevant input for a goal selector / Attention
> > component?
> >
> >
> >
> > Date: Fri, 8 Jun 2012 17:49:15 -0400
> > Subject: Re: [agi] Attention
> > From: [email protected] <mailto:[email protected]>
> > To: [email protected] <mailto:[email protected]>
> >
> >
> >
> > In the OpenCog framework, we supply some hard-coded "top level
> > goals", and then the system learns how to achieve these, which may
> > include learning subgoals...
> >
> > The top level goals are generally of the form "keep so-and-such
> > parameter within range [L,R]"
> >
> > Experience of novelty and discovery of new things are good general
> > top-level goals. For an character in a virtual 3D environment, we
> > add in stuff like getting energy (e.g. from batteries or food),
> > staying safe, and partaking in social interaction....
> >
> > In reference to this sort of framework, I'm unsure if you're talking
> > about top-level goals or learned subgoals...
> >
> > -- Ben G
> >
> >
> >
> > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
> > <https://www.listbox.com/member/archive/rss/303/7190161-766c6f07> |
> > Modify <https://www.listbox.com/member/?&> Your Subscription
> > [Powered by Listbox] <http://www.listbox.com>
> >
> >
> >
> >
> > --
> > Abram Demski
> > http://lo-tho.blogspot.com/
> >
> > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
> > <https://www.listbox.com/member/archive/rss/303/19999924-5cfde295> |
> > Modify <https://www.listbox.com/member/?&> Your Subscription
[Powered
> > by Listbox] <http://www.listbox.com>
> >
> > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
> > <https://www.listbox.com/member/archive/rss/303/9320387-ea529a81> |
> > Modify
> > <https://www.listbox.com/member/?&>
> > Your Subscription [Powered by Listbox] <http://www.listbox.com>
> >
>
>
>
> -------------------------------------------
> AGI
> Archives: https://www.listbox.com/member/archive/303/=now
> RSS Feed:
https://www.listbox.com/member/archive/rss/303/19999924-5cfde295
> Modify Your Subscription:
https://www.listbox.com/member/?&
> Powered by Listbox: http://www.listbox.com
*AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
<https://www.listbox.com/member/archive/rss/303/9320387-ea529a81> |
Modify
<https://www.listbox.com/member/?&>
Your Subscription [Powered by Listbox] <http://www.listbox.com>