Reading your response, Jim, brings genetic programming to mind. If the
population, taken as a whole, learns, but the individuals within that
population do not, is it still stigmergy?



Aaron Hosford

> On Jul 25, 2015, at 7:49 PM, Jim Bromer <[email protected]> wrote:
>
> Stigmergy refers to individuals who leave and react to marks (or
> mark-like objects) in the environment that produce group behavior
> rather than forming memories and using reasoning. Their reactions are
> relatively simplistic but the group is able to produce a range of
> variations which may seem surprising. If you are going to use
> stigmergy as a term to describe an AI program or part of an AI program
> then you have to hold the agents (or parts) so that they are not
> learning but just leaving and reacting to marks in the
> pseudo-environment (the blackboard). This suggests that your AI
> program is going to be based on agents capable of simple reactions to
> marks in the blackboard. I would want my AI program to be able to
> learn and I see no reason why shielding the agents from being able to
> learn is going to make the project more likely to succeed. However, it
> is interesting to think about how much could be done this way and it
> is a worthwhile experiment. But, suppose you want to work from this
> stigmergy into some more powerful model without giving up the
> philosophical model of the stigmergy entirely.  You are going to try
> to give the agents some ability to learn but you still want to limit
> the memory store of the agents. What I am saying is that you might
> take that step by saying the agents are endowed with some
> 'abstractons' (or programming) which can then specialize as they are
> needed. Some of the original programming (or the potential range of
> the programming) is going to be filtered out as the agent specializes.
> But, once you take this step, I am saying, it is difficult to justify
> limiting the range of the agents to learn new abstractions (or new
> programming). The specialization is itself equivalent to a kind of
> reprogramming so why stop there? Why not explore other ways that the
> agents can be reprogrammed to deal with the data environment. So the
> 'agents' are not only specializing by filtering out programming steps
> but they could also (for example) be able to try combining programming
> steps in creative ways or even modifying the programming steps in some
> more dramatic (but still well managed) way. So then the simple
> programming of the agents is not a first level abstraction but a
> meta-level abstraction. (The word "abstraction" was originally meant
> to refer to an insight that was derived from learning but then
> Aristotle's redefinition of 'the form' meant the concept of
> abstraction could also be used as a form or formula or a program.)
>
> If you wanted to simulate the behavior of an ant colony you could call
> it stigmergy because the current thinking about the behaviors of ants
> are presumably seen as stigmergic. Now you could take his simulation
> to something more abstract so it is no longer a simulation. Derived
> from a simulation of the stigmergy you can still call it a stigmergic
> model. Then even though you might further change your model you can
> still say that it was derived from stigmergy. But you need to keep
> some kind of reality check on your use of the terminology so you don't
> completely lose track of what the program is doing.
> Jim Bromer
>
>
>> On Fri, Jul 24, 2015 at 3:55 PM, Mike Archbold <[email protected]> wrote:
>>> On 7/24/15, Jim Bromer <[email protected]> wrote:
>>> You can try to use stigmergy as if it were an abstraction that can be
>>> seen as part of a human-like intelligence but then you would, for
>>> example, be forced to declare that the more abstract parts of the
>>> programming were the primitives that were not changing due to the
>>> memories of events and the integration of those event-memories. But,
>>> since you would want a secondary abstraction-generation system be
>>> something that could be learned you would have to reach further into
>>> the abstractions of the abstractions of the programming to find the
>>> truly stigmergic part. It is an interesting philosophical exercise but
>>> can it be used to lead to something new?
>>> Jim Bromer
>>
>> Jim, I really like this paragraph above although I don't know what it
>> means, exactly, but have kind of feel for it...
>> PM, I don't recall you had ideas in your design (apologize if I
>> forgot).  How do you define "idea" in a non formal type way?
>>
>> Mike A
>>
>>
>>>
>>>> On Wed, Jul 22, 2015 at 10:31 AM, Jim Bromer <[email protected]> wrote:
>>>> The definition of stigmergy in Wikipedia is that, "It produces
>>>> complex, seemingly intelligent structures, without need for any
>>>> planning, control, or even direct communication between the agents. As
>>>> such it supports efficient collaboration between extremely simple
>>>> agents, who lack any memory, intelligence or even individual awareness
>>>> of each other."
>>>> So while Facebook, for example, is designed to work based on human
>>>> responses it does also retain 'marks' which are used to determine a
>>>> range of actions that can be subsequently taken in response. However,
>>>> communication between the human agents, who have stores of memories,
>>>> is the whole reason Facebook has succeeded. Can we look at part of a
>>>> distributed active system, even one that relies on human IO, and say
>>>> that part of it is stigmergic? OK, but the next question is why? What
>>>> can you do with that point of view? I think (it is obvious that) human
>>>> beings are sometimes reacting without fully realizing what is going on
>>>> and instead base their responses on prevailing commonalities of
>>>> insight (like prevailing memes). This kind of reaction might be
>>>> likened to a stigmergic reaction. Subsequent interactions can then be
>>>> used to refine the first attempts to understand what is going on (or
>>>> what someone else is trying to say.) So perhaps by looking at
>>>> foundational or simple methods that can combine stigmergy with more
>>>> traditional AI methods so that stigmergic reactions can be integrated
>>>> with previous reactions (for example successive statements) someone
>>>> might be able to gain a little more insight in AGI. However, this
>>>> implies that simple reactions must be context-sensitive to different
>>>> combinations of events and they have to be sensitive to hidden parts
>>>> that need to be inferred and discovered in order to appreciate special
>>>> meanings (or to invoke special reactions) related to individuation of
>>>> the agents. So I can see one way how this extension of the definition
>>>> of stigmergy might be used to yield some novel experimental results.
>>>> If I only had the time...
>>>
>>>
>>> -------------------------------------------
>>> AGI
>>> Archives: https://www.listbox.com/member/archive/303/=now
>>> RSS Feed: https://www.listbox.com/member/archive/rss/303/11943661-d9279dae
>>> Modify Your Subscription:
>>> https://www.listbox.com/member/?&;
>>> Powered by Listbox: http://www.listbox.com
>>
>>
>> -------------------------------------------
>> AGI
>> Archives: https://www.listbox.com/member/archive/303/=now
>> RSS Feed: https://www.listbox.com/member/archive/rss/303/24379807-653794b5
>> Modify Your Subscription: https://www.listbox.com/member/?&;
>> Powered by Listbox: http://www.listbox.com
>
>
> -------------------------------------------
> AGI
> Archives: https://www.listbox.com/member/archive/303/=now
> RSS Feed: https://www.listbox.com/member/archive/rss/303/23050605-2da819ff
> Modify Your Subscription: https://www.listbox.com/member/?&;
> Powered by Listbox: http://www.listbox.com


-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to