“Complexity” is a substitute for thought here. You couldn’t begin to specify 
what are the complex elements here.

The reality is that living systems can translate these desires/goals, however 
internally represented, into initial actions in a second, and extended courses 
of action in just a few seconds.

The idea that there is some systematic consideration of sets of sets of 
alternative courses of action and environments  – “sets of sets” because the 
infant/animal could be in an infinite diversity of situations – is quite, quite 
mad, i.e. divorced from any reality whatsoever other than narrow AI programs 
which are incapable of this kind of intelligence and action.

From: Jim Bromer 
Sent: Sunday, December 16, 2012 3:09 PM
To: AGI 
Subject: Re: [agi] Internal Representation

On Sun, Dec 16, 2012 at 4:43 AM, Mike Tintner <[email protected]> wrote:
How do the words translate into a physical course of action?

How do you get from “go” to its first and subsequent movements of limbs? Bear 
in mind, that the same wish – the same formulation – could apply to the infant 
in vastly diverse physical situations - initial physical positions llike lying, 
sitting, lying on side, standing against some object etc - and vastly different 
room configurations.
----------------------------

There is a problem here, but the only true problem is one of complexity.  The 
same problem occurs when the program tries to make sense out of the IO data 
that is input to the program so that it can recognize what kind of situation 
that it is in or responding to.  The problem also occurs when it has to select 
the best kind of reaction to the situation when there are many otions that it 
can select from which are related to the complexities of the situation.  There 
are no  serious problems with implementing general AI other then the complexity 
of these kinds of problems.  A situation can be recognized based on many 
components and there are many 'ideas' (about different kinds of situations) 
that can be recognized based on some of the components that may occur in the 
situation.  It is a many-to-many kind of interpretation problem. Some problems 
are much simpler, but the complications is what makes general AI so complex.
Jim Bromer

On Sun, Dec 16, 2012 at 4:43 AM, Mike Tintner <[email protected]> wrote:

  Logan,
  You’re thinking of converting words into more words, here?
  How do the words translate into a physical course of action?
  How (i.e. in what forms of representation) would an infant AGI think that 
hadn’t yet learned language, when it say wants to
  “go across the room”...
  crawl, roll, whatever across the carpet to a toy?
  How do you get from “go” to its first and subsequent movements of limbs? Bear 
in mind, that the same wish – the same formulation – could apply to the infant 
in vastly diverse physical situations - initial physical positions llike lying, 
sitting, lying on side, standing against some object etc - and vastly different 
room configurations.
      AGI | Archives  | Modify Your Subscription   



-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to