PM: You must say why it cannot be done to be helpful, which you haven't. I would have thought a man with your imagination wouldn’t need to be told – the principal reason you cannot have a complex set here, is that you cannot define or calculate a) the trajectories or lines of action a body must take to “go across a room”” – esp. given b) the infinite diversity of starting points and positions of the body, and c) the infinite diversity of obstacles that may lie in the way, and d) the infinite natures and configurations of room terrains, and e) the endless combinations of limbs and body positions that may be necessary to move from the infinite diversity of starting points.
The solution lies in the direction of having truly “fluid concepts” of lines of action, with which the body can be fluidly aligned, and proceeding not by prior consideration of a set of trajectories (which is absolutely impossible) but adventurously “putting your best foot forward” towards the goal, one step at a time, and seeing what happens. Logan’s formulations lie intuitively in that direction – but they look like purely symboiic formulations – whereas internal representations must be literally linear in nature. Concepts provide literally “lines of action” - and involve literally “thinking along these lines”. From: Piaget Modeler Sent: Sunday, December 16, 2012 4:54 PM To: AGI Subject: RE: [agi] Internal Representation According to Confucius, Mike Tinter, you are correct. "The person who says it can't be done, and the person who says it can be done are both usually correct." ~ Confucius. If you want to use some imagination, and figure out how it can be done, then by all means. If you just want to nay say and say it can't be done, then that's not helpful. You must say why it cannot be done to be helpful, which you haven't. I think it is already being done, and you simply may not know about it, because certain people aren't revealing their results just yet. ~PM. -------------------------------------------------------------------------------- From: [email protected] To: [email protected] Subject: Re: [agi] Internal Representation Date: Sun, 16 Dec 2012 16:06:49 +0000 “Complexity” is a substitute for thought here. You couldn’t begin to specify what are the complex elements here. The reality is that living systems can translate these desires/goals, however internally represented, into initial actions in a second, and extended courses of action in just a few seconds. The idea that there is some systematic consideration of sets of sets of alternative courses of action and environments – “sets of sets” because the infant/animal could be in an infinite diversity of situations – is quite, quite mad, i.e. divorced from any reality whatsoever other than narrow AI programs which are incapable of this kind of intelligence and action. From: Jim Bromer Sent: Sunday, December 16, 2012 3:09 PM To: AGI Subject: Re: [agi] Internal Representation On Sun, Dec 16, 2012 at 4:43 AM, Mike Tintner <[email protected]> wrote: How do the words translate into a physical course of action? How do you get from “go” to its first and subsequent movements of limbs? Bear in mind, that the same wish – the same formulation – could apply to the infant in vastly diverse physical situations - initial physical positions llike lying, sitting, lying on side, standing against some object etc - and vastly different room configurations. ---------------------------- There is a problem here, but the only true problem is one of complexity. The same problem occurs when the program tries to make sense out of the IO data that is input to the program so that it can recognize what kind of situation that it is in or responding to. The problem also occurs when it has to select the best kind of reaction to the situation when there are many otions that it can select from which are related to the complexities of the situation. There are no serious problems with implementing general AI other then the complexity of these kinds of problems. A situation can be recognized based on many components and there are many 'ideas' (about different kinds of situations) that can be recognized based on some of the components that may occur in the situation. It is a many-to-many kind of interpretation problem. Some problems are much simpler, but the complications is what makes general AI so complex. Jim Bromer On Sun, Dec 16, 2012 at 4:43 AM, Mike Tintner <[email protected]> wrote: Logan, You’re thinking of converting words into more words, here? How do the words translate into a physical course of action? How (i.e. in what forms of representation) would an infant AGI think that hadn’t yet learned language, when it say wants to “go across the room”... crawl, roll, whatever across the carpet to a toy? How do you get from “go” to its first and subsequent movements of limbs? Bear in mind, that the same wish – the same formulation – could apply to the infant in vastly diverse physical situations - initial physical positions llike lying, sitting, lying on side, standing against some object etc - and vastly different room configurations. AGI | Archives | Modify Your Subscription AGI | Archives | Modify Your Subscription AGI | Archives | Modify Your Subscription ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657 Powered by Listbox: http://www.listbox.com
