I don't know that I'd consider that an example of an uncomplicated goal. That seems to me much more complicated than simple responses to sensory inputs. Valuable, yes, and even vital for any significant intelligence, but definitely not at the minimal level of complexity. An example of a minimal goal might be "to cause an extended period of inter-entity communication", or "to find a recharging socket". Note that the second one would probably need to have a hard-coded solution available before the entity was able to start any independent explorations. This doesn't mean that as new answers were constructed the original might not decrease in significance and eventually be garbage collected. It means that it would need to be there as a pre-written answer on the tabula rasa. (I.e., the tablet can't really be blank. You need to start somewhere, even if you leave and never return.) For the first example, I was thinking of "peek-a-boo".

Bob Mottram wrote:

Goals don't necessarily need to be complex or even explicitly defined. One "goal" might just be to minimise the difference between experiences (whether real or simulated) and expectations. In this way the system learns what a normal state of being is, and detect deviations.



On 21/11/06, *Charles D Hixson* <[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>> wrote:

    Bob Mottram wrote:
    >
    >
    > On 17/11/06, *Charles D Hixson* <[EMAIL PROTECTED]
    <mailto:[EMAIL PROTECTED]>
    > <mailto:[EMAIL PROTECTED]
    <mailto:[EMAIL PROTECTED]>>> wrote:
    >
    >     A system understands a situation that it encounters if it
    predictably
    >     acts in such a way as to maximize the probability of
    achieving it's
    >     goals in that situation.
    >
    >
    >
    >
    > I'd say a system "understands" a situation when its internal
    modeling
    > of that situation closely approximates its main salient
    features, such
    > that the difference between expectation and reality is minimised.
    > What counts as salient depends upon goals.  So for example I
    could say
    > that I "understand" how to drive, even if I don't have any detailed
    > knowledge of the workings of a car.
    >
    > When young animals play they're generating and tuning their models,
    > trying to bring them in line with observations and goals.
    That sounds reasonable, but how are you determining the match of the
    internal modeling to the "main salient features".  I propose that
    you do
    this based on it's actions, and thus my definition.  I'll admit,
    however, that this still leaves the problem of how to observe what
    it's
    goals are, but I hypothesize that it will be much simpler to
    examine the
    goals in the code than to examine the internal model.

    -----
    This list is sponsored by AGIRI: http://www.agiri.org/email
    To unsubscribe or change your options, please go to:
    http://v2.listbox.com/member/?list_id=303


------------------------------------------------------------------------
This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?list_id=303

-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?list_id=303

Reply via email to