Folks who are obsessed with the historical failure of AI to achieve AGI,
really amuse me....

I mean: the field has only been around 60 years.  This is minimal in the
scope of human history; and the available hardware is still very weak in
important respects compared to a mouse brain, let alone a human brain...

The failure of AI tech to achieve AGI so far, is certainly not compelling
evidence that all the underlying concepts are basically wrong ;-p

-- ben G

On Sun, Dec 30, 2012 at 1:34 PM, Mike Tintner <[email protected]>wrote:

>      1) Ben:  I, personally, think it will be best to connect AGI systems
> to robots in order to achieve human-like general intelligence....  In such
> a system, the logic and math underlying the AGI's algorithms will be used
> to interpret the data from the robot's sensors, and to control the data fed
> to the robot's actuators....
>
> All very well in broad principle. But the *evidence* is that
> robots/computers cannot yet produce any AGI-worthy percepts or concepts of
> the world whatsoever – cannot really make any AGI-worthy sensory contact
> with the world – and are as vast a distance away from understanding even
> the simplest of real world scenes, let alone individual objects, as they
> are from understanding variations on the simplest of two/three word
> sentences (as I have explained elsewhere). This has proved an
> extraordinarily intractable set of problems.
>
> Not knowing where the solutions lie – what  form successful percepts and
> concepts of the world will take, -  you are nevertheless proceeding on the
> basis that logic and maths will provide the solution, and are central to an
> AGI architecture,  despite their extreme and consistent lack of success so
> far. Far from being a troll in this area, I repeatedly try to engage with
> the problem here – with the forms that successful perceptualisation and
> conceptualisation must take. You simply aren’t interested in any solution
> that isn’t logicomathematical – isn’t within your current competency – and
> don’t have any ideas to offer about how logic & maths or anything else will
> directly solve these problems.
>
> I would say that you are the true AGI troll – by not engaging directly
> with the real problems,you are holding the field back.
>
> 2) Ditto re your rocket analogy. You are actually a rocket builder who has
> defied every principle of serious creativity by building an extremely
> complex machine, without the slightest evidence or proof of concept, that
> your machine is equipped for flight – for its principal function  – in this
> case for AGI. You have designed a lot of engine/architecture, so to speak,
> but you don’t have a take-off mechanism period – and cannot explain in any
> way how your machine will achieve takeoff or airborne thrust – how it will
> be either general or creative. Wow.
>
> Again I would say you are the real troll – this is not the way to go about
> a serious project, and absolutely the wrong example to set  -  you are
> again characteristically not engaging with the main problem first – which
> is what must be done. And many – many AGI-ers, -  not just me, have told
> you this. (Partly in your defence, this is a field-wide phenomenon, but it
> is still terrible).
>
> 3) Because I am engaging with the problems, I will in the not too distant
> future, explain the exact nature of real world reasoning – and how it is
> totally different from logicomathematical induction/deduction etc. (To talk
> of it being “imaginative reasoning” is to identify the medium of its
> reasoning, but not the main ways in which it differs as a form of
> inference).
>
>
>
>
>
>
>
>
>  *From:* Ben Goertzel <[email protected]>
> *Sent:* Sunday, December 30, 2012 4:56 PM
> *To:* AGI <[email protected]>
> *Subject:* Re: [agi] Why Logic & Maths Have Sweet FA to do with Real
> world reasoning
>
>
>
> On Sun, Dec 30, 2012 at 5:31 AM, Mike Tintner <[email protected]>wrote:
>
>> You & others think that an AGI computer is going to be wise about the
>> world by sifting through textual documents, and working out logically
>> whether “JOHN LOVES MARY”.
>
>
>
>
> You're battling a straw man here, hmm?   I've already been explicit that
> I, personally, think it will be best to connect AGI systems to robots in
> order to achieve human-like general intelligence....  In such a system, the
> logic and math underlying the AGI's algorithms will be used to interpret
> the data from the robot's sensors, and to control the data fed to the
> robot's actuators....
>
> Of course, the proof of the pudding will be in whether the completed
> system manages to learn how to control the robot intelligently...
>
> I imagine you as a troll watching a team of engineers and scientists build
> a rocket.   You, the troll, deny the theory of rocketry has any relevance
> to building real rockets, because that theory uses math, and the world is
> after all not obviously made of math.  Anyway they didn't teach you much
> about math back in troll school.  Meanwhile, the engineers and scientists
> are about 25% done with building their rocket.  And you're wasting your
> time, and causing annoyance, standing there pointing your trolly little
> finger at them, screeching "Hahaha!!  You idiots, can't you see that your
> supposedly 25% complete rocket can't fly 25% of the way to the moon??
> Don't you understand that if you want anyone to believe your completed
> rocket will fly to the moon, you need to show success according to
> quantifiable metrics of incremental progress; you need to show real
> evidence!!!"
>
> Eventually, when the rocket blasts off, the troll brushes the ashes off
> his beard and somehow convinces himself it was just an illusion...
>
> Duuuhhh...
>
> -- Ben G
>
>   *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
> <https://www.listbox.com/member/archive/rss/303/6952829-59a2eca5> | 
> Modify<https://www.listbox.com/member/?&;>Your Subscription
> <http://www.listbox.com>
>   *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
> <https://www.listbox.com/member/archive/rss/303/212726-deec6279> | 
> Modify<https://www.listbox.com/member/?&;>Your Subscription
> <http://www.listbox.com>
>



-- 
Ben Goertzel, PhD
http://goertzel.org

"My humanity is a constant self-overcoming" -- Friedrich Nietzsche



-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to