‘The intuitive mind is a sacred gift and the rational mind is a faithful servant. We have created a society that honours the servant and has forgotten the gift.’
‘The intellect has little to do on the road to discovery. There comes a leap in consciousness, call it intuition or what you will, and the solution comes to you and you don’t know how or why.’ — Albert Einstein We are here talking like programmers who needs to build a new system; Just divide the problem, solve it one by one, arrange the pieces and voila. We are missing something fundamentally here. That I believe has to come as a stroke of genius to someone. thanks, Deepak On Mon, Jul 19, 2010 at 4:10 PM, Mike Tintner <[email protected]>wrote: > No, Dave & I vaguely agree here that you have to start simple. To think > of movies is massively confused - rather like saying: "when we have created > an entire new electric supply system for cars, we will have solved the > problem of replacing gasoline" - first you have to focus just on inventing a > radically cheaper battery, before you consider the possibly hundreds to > thousands of associated inventions and innovations.involved in creating a > major new supply system." > > Here it would be much simpler to focus on understanding a single > photographic scene - or real, directly-viewed scene - of objects, rather > than the many thousands involved in a movie. > > In terms of language, it would be simpler to focus on understanding just > two consecutive sentences of a text or section of dialogue - or even as > I've already suggested, just the flexible combinations of two words - rather > than the hundreds of lines and many thousands of words involved in a movie > or play script. > > And even this is probably all too evolved, for humans only came to use > formal representations of the world v. recently in evolution. > > The general point - a massively important one - is that AGI-ers cannot > continue to think of AGI in terms of massively complex and evolved > intelligent systems, as you are doing. You have to start with the simplest > possible systems and gradually evolve them. Anything else is a defiance of > all the laws of technology - and will see AGI continuing to go absolutely > nowhere. > > *From:* deepakjnath <[email protected]> > *Sent:* Monday, July 19, 2010 5:19 AM > *To:* agi <[email protected]> > *Subject:* Re: [agi] Of definitions and tests of AGI > > Exactly my point. So if I show a demo of an AGI system that can see two > movies and understand that the plot of the movies are same even though they > are 2 entirely different movies, you would agree that we have created a true > AGI. > > Yes there are always lot of things we need to do before we reach that > level. Its just good to know the destination so that we will know it when it > arrives. > > > > > On Mon, Jul 19, 2010 at 2:18 AM, Mike Tintner <[email protected]>wrote: > >> Jeez, no AI program can understand *two* consecutive *sentences* in a >> text - can understand any text period - can understand language, period. And >> you want an AGI that can understand a *story*. You don't seem to understand >> that requires cognitively a fabulous, massively evolved, highly educated, >> hugely complex set of powers . >> >> No AI can understand a photograph of a scene, period - a crowd scene, a >> house by the river. Programs are hard put to recognize any objects other >> than those in v. standard positions. And you want an AGI that can understand >> a *movie*. >> >> You don't seem to realise that we can't take the smallest AGI *step* yet >> - and you're fantasying about a superevolved AGI globetrotter. >> >> That's why Benjamin & I tried to focus on v. v. simple tests - & they're >> still way too complex & they (or comparable tests) will have to be refined >> down considerably for anyone who is interested in practical vs sci-fi >> fantasy AGI. >> >> I recommend looking at Packbots and other military robots and hospital >> robots and the like, and asking how we can free them from their human >> masters and give them the very simplest of capacities to rove and handle the >> world independently - like handling and travelling on rocks. >> >> Anyone dreaming of computers or robots that can follow "Gone with The >> Wind" or become a child (real) scientist in the foreseeable future pace Ben, >> has no realistic understanding of what is involved. >> *From:* deepakjnath <[email protected]> >> *Sent:* Sunday, July 18, 2010 9:04 PM >> *To:* agi <[email protected]> >> *Subject:* Re: [agi] Of definitions and tests of AGI >> >> Let me clarify. As you all know there are somethings computers are good at >> doing and somethings that Humans can do but a computer cannot. >> >> One of the test that I was thinking about recently is to have to movies >> show to the AGI. Both movies will have the same story but it would be a >> totally different remake of the film probably in different languages and >> settings. If the AGI is able to understand the sub plot and say that the >> story line is similar in the two movies then it could be a good test for AGI >> structure. >> >> The ability of a system to understand its environment and underlying sub >> plots is an important requirement of AGI. >> >> Deepak >> >> On Mon, Jul 19, 2010 at 1:14 AM, Mike Tintner >> <[email protected]>wrote: >> >>> Please explain/expound freely why you're not "convinced" - and indicate >>> what you expect, - and I'll reply - but it may not be till tomorrow. >>> >>> Re your last point, there def. is no consensus on a general problem/test >>> OR a def. of AGI. >>> >>> One flaw in your expectations seems to be a desire for a single test - >>> almost by definition, there is no such thing as >>> >>> a) a single test - i.e. there should be at least a dual or serial test - >>> having passed any given test, like the rock/toy test, the AGI must be >>> presented with a new "adjacent" test for wh. it has had no preparation, >>> like say building with cushions or sand bags or packing with fruit. (and >>> neither rock/toy test state that clearly) >>> >>> b) one kind of test - this is an AGI, so it should be clear that if it >>> can pass one kind of test, it has the basic potential to go on to many >>> different kinds, and it doesn't really matter which kind of test you start >>> with - that is partly the function of having a good.definition of AGI . >>> >>> >>> *From:* deepakjnath <[email protected]> >>> *Sent:* Sunday, July 18, 2010 8:03 PM >>> *To:* agi <[email protected]> >>> *Subject:* Re: [agi] Of definitions and tests of AGI >>> >>> So if I have a system that is close to AGI, I have no way of really >>> knowing it right? >>> >>> Even if I believe that my system is a true AGI there is no way of >>> convincing the others irrefutably that this system is indeed a AGI not just >>> an advanced AI system. >>> >>> I have read the toy box problem and rock wall problem, but not many >>> people will still be convinced I am sure. >>> >>> I wanted to know that if there is any consensus on a general problem >>> which can be solved and only solved by a true AGI. Without such a test bench >>> how will we know if we are moving closer or away from our quest. There is no >>> map. >>> >>> Deepak >>> >>> >>> >>> On Sun, Jul 18, 2010 at 11:50 PM, Mike Tintner <[email protected] >>> > wrote: >>> >>>> I realised that what is needed is a *joint* definition *and* range of >>>> tests of AGI. >>>> >>>> Benamin Johnston has submitted one valid test - the toy box problem. >>>> (See archives). >>>> >>>> I have submitted another still simpler valid test - build a rock wall >>>> from rocks given, (or fill an earth hole with rocks). >>>> >>>> However, I see that there are no valid definitions of AGI that explain >>>> what AGI is generally , and why these tests are indeed AGI. Google - there >>>> are v. few defs. of AGI or Strong AI, period. >>>> >>>> The most common: AGI is human-level intelligence - is an >>>> embarrassing non-starter - what distinguishes human intelligence? No >>>> explanation offered. >>>> >>>> The other two are also inadequate if not as bad: Ben's "solves a variety >>>> of complex problems in a variety of complex environments". Nope, so does a >>>> multitasking narrow AI. Complexity does not distinguish AGI. Ditto Pei's - >>>> something to do with "insufficient knowledge and resources..." >>>> "Insufficient" is open to narrow AI interpretations and reducible to >>>> mathematically calculable probabilities.or uncertainties. That doesn't >>>> distinguish AGI from narrow AI. >>>> >>>> The one thing we should all be able to agree on (but who can be sure?) >>>> is that: >>>> >>>> ** an AGI is a general intelligence system, capable of independent >>>> learning** >>>> >>>> i.e. capable of independently learning new activities/skills with >>>> minimal guidance or even, ideally, with zero guidance (as humans and >>>> animals >>>> are) - and thus acquiring a "general", "all-round" range of intelligence.. >>>> >>>> This is an essential AGI goal - the capacity to keep entering and >>>> mastering new domains of both mental and physical skills WITHOUT being >>>> specially programmed each time - that crucially distinguishes it from >>>> narrow >>>> AI's, which have to be individually programmed anew for each new task. >>>> Ben's >>>> AGI dog exemplified this in a v simple way - the dog is supposed to be >>>> able >>>> to learn to fetch a ball, with only minimal instructions, as real dogs do - >>>> they can learn a whole variety of new skills with minimal instruction. But >>>> I am confident Ben's dog can't actually do this. >>>> >>>> However, the independent learning def. while focussing on the >>>> distinctive AGI goal, still is not detailed enough by itself. >>>> >>>> It requires further identification of the **cognitive operations** which >>>> distinguish AGI, and wh. are exemplified by the above tests. >>>> >>>> [I'll stop there for interruptions/comments & continue another time]. >>>> >>>> P.S. Deepakjnath, >>>> >>>> It is vital to realise that the overwhelming majority of AGI-ers do not >>>> * want* an AGI test - Ben has never gone near one, and is merely typical >>>> in >>>> this respect. I'd put almost all AGI-ers here in the same league as the US >>>> banks, who only want mark-to-fantasy rather than mark-to-market tests of >>>> their assets. >>>> *agi* | Archives <https://www.listbox.com/member/archive/303/=now> >>>> <https://www.listbox.com/member/archive/rss/303/> | >>>> Modify<https://www.listbox.com/member/?&>Your Subscription >>>> <http://www.listbox.com> >>>> >>> >>> >>> >>> -- >>> cheers, >>> Deepak >>> *agi* | Archives <https://www.listbox.com/member/archive/303/=now> >>> <https://www.listbox.com/member/archive/rss/303/> | >>> Modify<https://www.listbox.com/member/?&>Your Subscription >>> <http://www.listbox.com> >>> *agi* | Archives <https://www.listbox.com/member/archive/303/=now> >>> <https://www.listbox.com/member/archive/rss/303/> | >>> Modify<https://www.listbox.com/member/?&>Your Subscription >>> <http://www.listbox.com> >>> >> >> >> >> -- >> cheers, >> Deepak >> *agi* | Archives <https://www.listbox.com/member/archive/303/=now> >> <https://www.listbox.com/member/archive/rss/303/> | >> Modify<https://www.listbox.com/member/?&>Your Subscription >> <http://www.listbox.com> >> *agi* | Archives <https://www.listbox.com/member/archive/303/=now> >> <https://www.listbox.com/member/archive/rss/303/> | >> Modify<https://www.listbox.com/member/?&>Your Subscription >> <http://www.listbox.com> >> > > > > -- > cheers, > Deepak > *agi* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/> | > Modify<https://www.listbox.com/member/?&>Your Subscription > <http://www.listbox.com> > *agi* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/> | > Modify<https://www.listbox.com/member/?&>Your Subscription > <http://www.listbox.com> > -- cheers, Deepak ------------------------------------------- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c Powered by Listbox: http://www.listbox.com
