There is no difference in *what* can be treated as an object or a situation, but they are different *treatments *of those things. Treating something as an object is a holistic perspective, while treating it as a situation is a reductionist perspective. Each has its own advantages and disadvantages, depending on what you are attempting to accomplish. It would probably be ideal if a system could look at things automatically as objects, and then "zoom in" as needed, converting them to situations. This would lend a sort of fractal structure to problem solving, allowing relationships and interactions to be inspected at the most appropriate level of detail.
On Tue, Apr 29, 2014 at 5:30 AM, Jim Bromer via AGI <[email protected]> wrote: > There is no difference between an object and a situation, because a > situation can be treated as an 'object' (of thought or otherwise be treated > as object-like. And of course situations occur inside of situations. That > is true even in traditional uses of the terms. > > Jim Bromer > > > On Mon, Apr 28, 2014 at 8:25 PM, Piaget Modeler via AGI > <[email protected]>wrote: > >> Can one have situations inside situations? >> >> What's the difference between an object and a situation? >> >> Kindly advise. >> >> ~PM >> >> ------------------------------ >> Date: Mon, 28 Apr 2014 16:50:24 -0600 >> From: [email protected] >> To: [email protected] >> Subject: Re: [agi] Situations >> >> >> Greetings Telmo, >> I've responded to your comments below. >> Are you working on an ontology based AGI approach? >> >> Stan >> >> On 04/28/2014 02:30 PM, Telmo Menezes via AGI wrote: >> >> Hi Stanley, >> >> >> On Mon, Apr 28, 2014 at 9:23 PM, Stanley Nilsen via AGI >> <[email protected]>wrote: >> >> Hi PM, >> >> A few thoughts - >> >> One might try to come up with methods to generalize situations - put in >> categories and sub categories and sub sub categories... This sounds >> logical, but also terribly tedious. >> >> My alternative is to look at the world as sets of triggers. A trigger >> initiates an action - maybe to assert a new fact. The new fact might then >> trigger something else... >> >> >> Ok, but I don't see how this removes the need for an ontology. >> >> As I understand it, there are several efforts to create massive >> ontology. And, we can all see the "value" of it. The struggle is in >> finding the mechanisms that can cash in on that value - the magic sauce? >> >> I focus on how to become more intelligent when you start at next to >> nothing. What's the bootstrap look like? At what point does a computer >> begin to build it's intelligence? And, what do the construction elements >> resemble? >> >> It could be implicit or explicit, but you still have to be able to >> make more and more distinctions between triggers or actions. I tell the AI >> to book me a trip to Cambridge. What Cambridge, UK or USA? And then, to >> book the ticket I have to know that Cambridge is a town, and that I already >> know something about how to book travels into towns, and so on. >> >> >> Software "assistants" are pretty popular now. I understand Microsoft is >> planning one to compete with Siri. Maybe this is the way to the future. >> Start out assisting and one day take over :) >> >> >> You need some way to generalise, and this leads to some hierarchy of >> types. I bet our brain encodes a huge one. But how does it encode it? >> >> >> >> What is triggered depends on what our "understanding" makes of triggers. >> Pretty much a Rube Goldberg contraption, but gets interesting quickly. >> Understanding isn't that vague, it's whatever can be coded into rules. >> >> >> So you would say that a thermostat understands temperature? >> >> No, I would say that whatever is reading and setting the thermostat needs >> to understand the effect they want to achieve. The "user" chooses the >> thermostat based on understanding of outcomes that are expected. >> >> The thermostat is simply a "see" mechanism - it triggers something else. >> If you wrote a rule to act like a thermostat, I would say that the rule >> understands an aspect of a thermostat - e.g. numbers change over time and >> there is a trigger point. I don't think the rule needs to know about >> atomic vibrations, or the cost of a barrel of oil. >> >> I'm not downplaying ontology, it will be useful. I just don't put it as >> first priority in building an AGI. >> >> Stan >> >> >> *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> >> <https://www.listbox.com/member/archive/rss/303/19999924-4a978ccc> | >> Modify <https://www.listbox.com/member/?&> Your Subscription >> <http://www.listbox.com> >> *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> >> <https://www.listbox.com/member/archive/rss/303/24379807-f5817f28> | >> Modify <https://www.listbox.com/member/?&> Your Subscription >> <http://www.listbox.com> >> > > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/23050605-2da819ff> | > Modify<https://www.listbox.com/member/?&>Your Subscription > <http://www.listbox.com> > ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657 Powered by Listbox: http://www.listbox.com
