Greetings Telmo,
 I've responded to your comments below.
Are you working on an ontology based AGI approach?

Stan

On 04/28/2014 02:30 PM, Telmo Menezes via AGI wrote:
Hi Stanley,


On Mon, Apr 28, 2014 at 9:23 PM, Stanley Nilsen via AGI <[email protected] <mailto:[email protected]>> wrote:

    Hi PM,

    A few thoughts -

    One might try to come up with methods to generalize situations -
put in categories and sub categories and sub sub categories... This sounds logical, but also terribly tedious.

    My alternative is to look at the world as sets of triggers.   A
    trigger initiates an action - maybe to assert a new fact.  The new
    fact might then trigger something else...


Ok, but I don't see how this removes the need for an ontology.
As I understand it, there are several efforts to create massive ontology. And, we can all see the "value" of it. The struggle is in finding the mechanisms that can cash in on that value - the magic sauce?

I focus on how to become more intelligent when you start at next to nothing. What's the bootstrap look like? At what point does a computer begin to build it's intelligence? And, what do the construction elements resemble?

It could be implicit or explicit, but you still have to be able to make more and more distinctions between triggers or actions. I tell the AI to book me a trip to Cambridge. What Cambridge, UK or USA? And then, to book the ticket I have to know that Cambridge is a town, and that I already know something about how to book travels into towns, and so on.

Software "assistants" are pretty popular now. I understand Microsoft is planning one to compete with Siri. Maybe this is the way to the future. Start out assisting and one day take over :)


You need some way to generalise, and this leads to some hierarchy of types. I bet our brain encodes a huge one. But how does it encode it?


    What is triggered depends on what our "understanding" makes of
    triggers.  Pretty much a Rube Goldberg contraption, but gets
    interesting quickly. Understanding isn't that vague, it's whatever
    can be coded into rules.


So you would say that a thermostat understands temperature?
No, I would say that whatever is reading and setting the thermostat needs to understand the effect they want to achieve. The "user" chooses the thermostat based on understanding of outcomes that are expected.

The thermostat is simply a "see" mechanism - it triggers something else. If you wrote a rule to act like a thermostat, I would say that the rule understands an aspect of a thermostat - e.g. numbers change over time and there is a trigger point. I don't think the rule needs to know about atomic vibrations, or the cost of a barrel of oil.

I'm not downplaying ontology, it will be useful. I just don't put it as first priority in building an AGI.

Stan





-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to