What about DESTIN? Jim has talked about video. Could DESTIN be generalized to 3 dimensions, or even n dimensions?
- Ian Parker On 9 August 2010 07:16, John G. Rose <johnr...@polyplexic.com> wrote: > Actually this is quite critical. > > > > Defining a chair - which would agree with each instance of a chair in the > supplied image - is the way a chair should be defined and is the way the > mind processes it. > > > > It can be defined mathematically in many ways. There is a particular one I > would go for though... > > > > John > > > > *From:* Mike Tintner [mailto:tint...@blueyonder.co.uk] > *Sent:* Sunday, August 08, 2010 7:28 AM > *To:* agi > *Subject:* Re: [agi] How To Create General AI Draft2 > > > > You're waffling. > > > > You say there's a pattern for chair - DRAW IT. Attached should help you. > > > > Analyse the chairs given in terms of basic visual units. Or show how any > basic units can be applied to them. Draw one or two. > > > > You haven't identified any basic visual units - you don't have any. Do > you? Yes/no. > > > > No. That's not "funny", that's a waste.. And woolly and imprecise through > and through. > > > > > > > > *From:* David Jones <davidher...@gmail.com> > > *Sent:* Sunday, August 08, 2010 1:59 PM > > *To:* agi <agi@v2.listbox.com> > > *Subject:* Re: [agi] How To Create General AI Draft2 > > > > Mike, > > We've argued about this over and over and over. I don't want to repeat > previous arguments to you. > > You have no proof that the world cannot be broken down into simpler > concepts and components. The only proof you attempt to propose are your > example problems that *you* don't understand how to solve. Just because > *you* cannot solve them, doesn't mean they cannot be solved at all using a > certain methodology. So, who is really making wild assumptions? > > The mere fact that you can refer to a "chair" means that it is a > recognizable pattern. LOL. That fact that you don't realize this is quite > funny. > > Dave > > On Sun, Aug 8, 2010 at 8:23 AM, Mike Tintner <tint...@blueyonder.co.uk> > wrote: > > Dave:No... it is equivalent to saying that the whole world can be modeled > as if everything was made up of matter > > > > And "matter" is... ? Huh? > > > > You clearly don't realise that your thinking is seriously woolly - and you > will pay a heavy price in lost time. > > > > What are your "basic world/visual-world analytic units" wh. you are > claiming to exist? > > > > You thought - perhaps think still - that *concepts* wh. are pretty > fundamental intellectual units of analysis at a certain level, could be > expressed as, or indeed, were patterns. IOW there's a fundamental pattern > for "chair" or "table." Absolute nonsense. And a radical failure to > understand the basic nature of concepts which is that they are *freeform* > schemas, incapable of being expressed either as patterns or programs. > > > > You had merely assumed that concepts could be expressed as patterns,but had > never seriously, visually analysed it. Similarly you are merely assuming > that the world can be analysed into some kind of visual units - but you > haven't actually done the analysis, have you? You don't have any of these > basic units to hand, do you? If you do, I suggest, reply instantly, naming a > few. You won't be able to do it. They don't exist. > > > > Your whole approach to AGI is based on variations of what we can call > "fundamental analysis" - and it's wrong. God/Evolution hasn't built the > world with any kind of geometric, or other consistent, bricks. He/It is a > freeform designer. You have to start thinking outside the > box/brick/"fundamental unit". > > > > *From:* David Jones <davidher...@gmail.com> > > *Sent:* Sunday, August 08, 2010 5:12 AM > > *To:* agi <agi@v2.listbox.com> > > *Subject:* Re: [agi] How To Create General AI Draft2 > > > > Mike, > > I took your comments into consideration and have been updating my paper to > make sure these problems are addressed. > > See more comments below. > > On Fri, Aug 6, 2010 at 8:15 PM, Mike Tintner <tint...@blueyonder.co.uk> > wrote: > > 1) You don't define the difference between narrow AI and AGI - or make > clear why your approach is one and not the other > > > I removed this because my audience is for AI researchers... this is AGI > 101. I think it's clear that my design defines general as being able to > handle the vast majority of things we want the AI to handle without > requiring a change in design. > > > > > 2) "Learning about the world" won't cut it - vast nos. of progs. claim > they can learn about the world - what's the difference between narrow AI and > AGI learning? > > > The difference is in what you can or can't learn about and what tasks you > can or can't perform. If the AI is able to receive input about anything it > needs to know about in the same formats that it knows how to understand and > analyze, it can reason about anything it needs to. > > > > > 3) "Breaking things down into generic components allows us to learn about > and handle the vast majority of things we want to learn about. This is what > makes it general!" > > > > Wild assumption, unproven or at all demonstrated and untrue. > > > You are only right that I haven't demonstrated it. I will address this in > the next paper and continue adding details over the next few drafts. > > As a simple argument against your counter argument... > > If that were true that we could not understand the world using a limited > set of rules or concepts, how is it that a human baby, with a design that is > predetermined to interact with the world a certain way by its DNA, is able > to deal with unforeseen things that were not preprogrammed? That’s right, > the baby was born with a set of rules that robustly allows it to deal with > the unforeseen. It has a limited set of rules used to learn. That is > equivalent to a limited set of “concepts” (i.e. rules) that would allow a > computer to deal with the unforeseen. > > > Interesting philosophically because it implicitly underlies AGI-ers' > fantasies of "take-off". You can compare it to the idea that all science can > be reduced to physics. If it could, then an AGI could indeed take-off. But > it's demonstrably not so. > > > No... it is equivalent to saying that the whole world can be modeled as if > everything was made up of matter. Oh, I forgot, that is the case :) It is a > limited set of "concepts", yet it can create everything we know. > > > > > You don't seem to understand that the problem of AGI is to deal with the > NEW - the unfamiliar, that wh. cannot be broken down into familiar > categories, - and then find ways of dealing with it ad hoc. > > > You don't seem to understand that even the things you think cannot be > broken down, can be. > > > Dave > > *agi* | Archives <https://www.listbox.com/member/archive/303/=now> | > Modify Your Subscription <https://www.listbox.com/member/archive/rss/303/> > > <https://www.listbox.com/member/archive/rss/303/> > > *agi | Archives | Modify Your > Subscription<https://www.listbox.com/member/archive/rss/303/> > * > > <https://www.listbox.com/member/archive/rss/303/> > > <https://www.listbox.com/member/archive/rss/303/> > > *agi | Archives | Modify Your Subscription > <https://www.listbox.com/member/archive/rss/303/>* > > <https://www.listbox.com/member/archive/rss/303/> > > *agi | Archives | Modify Your > Subscription<https://www.listbox.com/member/archive/rss/303/> > * > > <https://www.listbox.com/member/archive/rss/303/> > > <https://www.listbox.com/member/archive/rss/303/> > *agi* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/> | > Modify<https://www.listbox.com/member/?&>Your Subscription > <http://www.listbox.com> > ------------------------------------------- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c Powered by Listbox: http://www.listbox.com