I'm told it's due at 11:11am UTC tomorrow. This should be fun... :)

On Thu, Dec 20, 2012 at 9:30 AM, A. T. Murray <[email protected]> wrote:

> Brett Ward wrote:
> >
> >Hi All,
> >
> >It's been more than a couple of decades now that I've believed AGI is
> >just around the corner. I've begun to explore the 'negative'. Here's
> >some of my thought experiment, a bit of fun ; )
> >
> >1) We don't know what General Intelligence really 'is'. The definition
> >of intelligence is a blurry concept. Are the definitions that we have
> >now as accurate as what Dalton had for the atom? How much detail is
> >required before it can be built?
> >
> >2) Even Human Intelligence could be defined as 'Narrow'. Can any random
> >human baby be trained to be both a concert pianist and a theoretical
> >physicist? Perhaps narrow A.I. will gradually expand to cover our needs,
> >without ever needing to become 'general'. There was a profession of
> >'human computer' which I guess is one of the first to be replaced by
> >very narrow 'intelligence'.
> >
> >3) A 'soul' is required before intelligence can be obtained. Those
> >without religious beliefs, feel free to replace 'soul' with
> >'unknown/uncompromisable part' (perhaps parallel quantum reactions with
> >other dimensions). Consider a chimpanzee with the belief that it can
> >build a boat, except all its efforts are put towards shaping clay to
> >look like a water bird.
> >
> >4) Human genetics/evolution play an 'unobtainable' part in general
> >intelligence. A significant proportion of human behavior is set by
> >genetics (reference identical twin studies). The beginnings of language
> >have a genetic beginning ('ma' and 'da' appear as parental names across
> >different languages). Perhaps the 'fine tuning' performed by evolution
> >is not obtainable in an artificial manner?
> >
> >5) Mental stability.  People aren't stable. Individuals commit suicide,
> >groups cluster and wage wars. If this stability is an integral part of
> >general intelligence, perhaps it's not achievable with our level of
> >understanding/technology.
> >
> >6) Effectiveness. Alchemists believed that they could make themselves
> >rich converting lead into gold. The materials were very similar, but the
> >techniques attempted were mostly futile (but did spin off some other
> >nice discoveries). It is possible to perform this transmutation, but the
> >process is not economically viable. By the time we have the
> >understanding to create general intelligence, we'll have other
> >technology (maybe brain enhancement implants) that are far more effective?
> >
> >What are your 'beliefs'?
>
> I believe a Singularity is due tomorrow on Fri.21.DEC.2012,
> but I'm not sure just what time of day :-)
>
> Today before breakfast I have been coding my AGI at
>
> http://www.scn.org/~mentifex/mindforth.txt
>
> and I have just uploaded it to the Web
> even though it is still kind of rough and imperfect,
> because for the first time it introduces the
>
> http://code.google.com/p/mindforth/wiki/InFerence
>
> module for machine reasoning.
>
> Thanks for posting your musings, Brett.
>
> Mentifex (Arthur)
>
>
> -------------------------------------------
> AGI
> Archives: https://www.listbox.com/member/archive/303/=now
> RSS Feed: https://www.listbox.com/member/archive/rss/303/23050605-2da819ff
> Modify Your Subscription:
> https://www.listbox.com/member/?&;
> Powered by Listbox: http://www.listbox.com
>



-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to