While I seem to be in a mood to waste more time than I usually would on AGI list discussions, I'm not going to sink so low as to have yet another debate about the meaning of "pattern" with Mike Tintner ;p
Mikey my friend, we will just have to agree to disagree. You can have it your way, and I will have it the right way ;) -- Ben G On Thu, Dec 27, 2012 at 7:18 PM, Mike Tintner <[email protected]>wrote: > You seem to have forgotten Ben that “magic sauce” is a term used by > many AGI-ers for the “missing but crucial ingredient of AGI* wh. an AGI > projectbuilder claims he will produce at a later stage from his present > outline but never does. > > You’ve also forgotten how many times you have indeed promised a missing > but crucial ingredient of AGI would be produced at a later stage with some > publication of yours – and never have produced it. > > You have done a QED by now indicating another magic ingredient of AGI, > this time a pattern recognition system . you have never produced any > actual examples of how patterns are relevant to AGI problems, and never > will – just waffled generally. This is complete (if fairly widely-held) > nonsense – > > By definition, an essential requirement of AGI is to solve problems about > actions and environments that do NOT fit existing patterns. By definition > an AGI must acquire new skills and undertake new kinds of actions – > something that presently defies all narrow AI programs. This is a > problematic business – precisely because the new skill/action is NOT like > already known actions – does NOT fit any existing, known pattern. Mastering > a new skill like say the actions of table tennis after tennis, or vice > versa, presents difficulties precisely because the arm and body actions do > not fit the same patterns. Mastering/understanding physics after chemistry > is difficult precisely because the laws of one do not easily fit the > laws/patterns of the other. And so it is with every subject area in > organized knowledge. > > Being able to have a conversation with one person is difficult precisely > he does NOT fit the conversational patterns of others. Every one is > different. Talking to Jim about AGI is different from talking to Matt, or > Pei, or Aaron, etc etc – because each one has a different approach to AGI, > each one is individual, and each one’s idiosyncrasies have to be gradually > identified in order to talk to them. They may share some common elements, > but overall they are very different. > > Nor do the conversations or posts of any individual taken altogether fit a > distinct pattern. We may have distinctive “styles” of conversation, but > those styles are fluid schemas at best, and nothing like the precise > patterns you are talking about – and result in very diverse, multiform > posts. Check your own posts in this or any other thread. I defy you to > identify overall patterns. > > I know that you have never actually applied your pattern theories to > actual AGI problems, just as it was clear from your book on creativity that > you had never applied your creativity theories to any actual creative > problems that you had independently researched. > > You work by adapting other people’s theories – and unfortunately for you, > none of them apply to building a real AGI., esp. patterns and pattern > recognition. > > P.S. Some indication of how AGI’s actually adapt to the new is given in > our talking of “a period of adjustment” being required, of “getting the > hang” of things, and “finding our feet” after stumbling and groping around. > All of this adapting to the new has nothing to do with pattern recognition. > > A GOS must be designed to enable a robot to endlessly *develop* its > actions – endlessly move along *new* lines - not fit its actions to the > same old patterns and lines. That’s narrow AI. > > > > *From:* Ben Goertzel <[email protected]> > *Sent:* Thursday, December 27, 2012 10:56 PM > *To:* AGI <[email protected]> > *Subject:* Re: [agi] The Only Test of AGI > > > >> >> Neither Ben nor anyone else in AGI is directly addressing the problem >> of a take-off system – or indeed has a clue – wh. is why you can >> immediately write off Opencog and other such efforts. They have absolutely >> nothing to do with AGI/ take-off – wh. is also why Ben et al have always >> resisted any form of test – they always have and always will fail any test >> of take-off/generality. (It’s not just me BTW – many have remarked that >> Ben et al’s “magic sauce” is not there – not even the idea of one) >> >> > > The idea that a "magic sauce" is needed for AGI is a mystical delusion, > redolent of vitalism in biology... > > The statement that I resist any form of test is a bald-faced lie. I > don't think that testing a completed human-level AGI is a particularly hard > problem, and I think it's a useful thing to do. The Turing Test is an OK > one (if it does on for an hour or more), or the test of having a robot pass > the third grade, etc. etc. I am skeptical of quantitative metrics for > early-stage partial progress toward human-level AGI, because I haven't yet > seen any that aren't either > > -- requiring a system already 80% of the way to human-level AGI > > -- too easily game-able by narrow-AI systems written especially to pass > the test > > ... > > The magic of general intelligence is simply this: A pattern recognition > system that can recognize patterns in its environment and *itself*, > including patterns regarding which actions tend to achieve which objectives > in which contexts. > > The challenge of general intelligence is: Recognizing a sufficient scope > of patterns, within a relevant and broad set of contexts, within the > limited compute resources available.... > > Meeting this challenge seems, so far as I can tell, to require a fairly > complex and multifaceted software system with interdependent parts; which > makes building AGI a major engineering and algorithmic challenge. > > That's not as romantic as daydreaming about some "magic sauce" that you > can just pour into your robot's head to make its wiring or its software get > smart -- but it's the reality... > > -- Ben G > > > > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/6952829-59a2eca5> | > Modify<https://www.listbox.com/member/?&>Your Subscription > <http://www.listbox.com> > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/212726-deec6279> | > Modify<https://www.listbox.com/member/?&>Your Subscription > <http://www.listbox.com> > -- Ben Goertzel, PhD http://goertzel.org "My humanity is a constant self-overcoming" -- Friedrich Nietzsche ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657 Powered by Listbox: http://www.listbox.com
