PM

A few months ago I mentioned to Jim Bromer that I would build the intelligence around the idea of opportunity. This fits directly with the idea of goals, in a somewhat obscure way.

If you look at your candidates list below, you see part of what makes up an opportunity. It leaves out important factors that go into choosing which opportunity is "greatest" for the moment.

These additional factors get complicated.  For example -
- how reliable are the claims of this opportunity? If we pursue this opportunity, can we be somewhat certain we will complete the steps needed to consummate and yield the benefit?

- what is the opportunity and how does it compare to other opportunity. In other words, there needs to be a vocabulary of benefit and cost that goes with consideration of any opportunity. Here is where the real world is complex. We venture and discover many benefits that we didn't expect. And we encounter costs that exceed our expectations. Will there be shortages we didn't anticipate...

- where did this opportunity come from? If we want to evaluate opportunity, we also need to consider the source. Is this simply an idea that we cooked up according to simple logic? Is this an idea that comes highly recommended from a reliable source - a master chef.

I understand that the ideas listed above seem off-base from looking at goals, but the real world "intelligent" doesn't work via figuring everything out. It works from suggestion and observation of what works and how it worked out. The intelligence that we want to embody will one day need to be able to choose between opportunity rather than simply follow "sub-goals" that make up a recipe.

To get this intelligence going, build the database of opportunity.

Stan

(sorry if I missed a relevant prior post, I'm way behind in list reading - busy time of year... yade yada .)




On 06/11/2012 06:31 PM, Piaget Modeler wrote:
Abram, you've characterized it properly. In my vernacular subgoals = goals.


I would say that the job of this particular attention module is to
reprioritize the open goal set,
given all available information.

So the question for me is what should all available information consist of?

Some candidates are: (1) The current context, for sure, (2) alerts, (3)
expectation failures and mismatches,
(4) past prioritizations, (5) past episodes.

Anything else?

Your thoughts?


Date: Mon, 11 Jun 2012 11:11:58 -0700
Subject: Re: [agi] Attention
From: [email protected]
To: [email protected]

PM,

OK. So, in this case, the goal selector is clearly selecting subgoals to
prioritize.

It's a difficult question which needs a quickly computable answer, so
the system needs to somehow gather information over time which tells it
what subgoals have been most useful in the past, in what situations.
This process can use a wide variety of information; essentially
anything. However, to make an efficient choice, the information
considered at any particular time needs to be narrowed down somehow. The
space of possible sub-goals is also potentially difficult, and needs to
be narrowed down heuristically...

Perhaps the best that I can say at the moment is, this seems like the
sort of problem which requires empirical testing to see what works and
what doesn't!

--Abram

On Fri, Jun 8, 2012 at 5:49 PM, Piaget Modeler
<[email protected] <mailto:[email protected]>> wrote:


    Ben

    Yours is a sufficient response. Thank you.

    Abram

    Suppose we decompose a cognitive system down into a few components:

    1. A planner (which is fed a goal, a current state and a set of
    possible actions (i.e., operators, methods, cases, etc.)),
    2. An action selector (which is fed the current state, a prioritized
    set of goals, and a set of methods to choose from),
    3. A goal selector / Attention module whose job is to prioritize or
    select goals for the cognitive system.

    My question was what would you feed the goal selector to ensure it
    did its job (prioritizing goals) properly?

    In a paper I read recently "A Case Study of Goal-Driven Autonomy in
    Domination Games" by Hector Munoz-Avila and David W. Aha
    the authors, in their CB-gda system, decompose the cognitive system
    into two case-based components (a) a planning component,
    and (b) a mismatch goal [selection] component. The purpose of the
    latter component was to correct for errors encountered by the
    planner. The input for the mismatch goal selection component is a
    mismatch (the difference between the expected state and the
    goal state).

    Q: What else would be relevant input for a goal selector / Attention
    component?



    Date: Fri, 8 Jun 2012 17:49:15 -0400
    Subject: Re: [agi] Attention
    From: [email protected] <mailto:[email protected]>
    To: [email protected] <mailto:[email protected]>



    In the OpenCog framework, we supply some hard-coded "top level
    goals", and then the system learns how to achieve these, which may
    include learning subgoals...

    The top level goals are generally of the form "keep so-and-such
    parameter within range [L,R]"

    Experience of novelty and discovery of new things are good general
    top-level goals. For an character in a virtual 3D environment, we
    add in stuff like getting energy (e.g. from batteries or food),
    staying safe, and partaking in social interaction....

    In reference to this sort of framework, I'm unsure if you're talking
    about top-level goals or learned subgoals...

    -- Ben G



    *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
    <https://www.listbox.com/member/archive/rss/303/7190161-766c6f07> |
    Modify <https://www.listbox.com/member/?&;> Your Subscription
    [Powered by Listbox] <http://www.listbox.com>




--
Abram Demski
http://lo-tho.blogspot.com/

*AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
<https://www.listbox.com/member/archive/rss/303/19999924-5cfde295> |
Modify <https://www.listbox.com/member/?&;> Your Subscription  [Powered
by Listbox] <http://www.listbox.com>

*AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
<https://www.listbox.com/member/archive/rss/303/9320387-ea529a81> |
Modify
<https://www.listbox.com/member/?&;>
Your Subscription       [Powered by Listbox] <http://www.listbox.com>




-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-c97d2393
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-2484a968
Powered by Listbox: http://www.listbox.com

Reply via email to