PM, You do understand that that’s a total non-answer? “We will give this problem to the computer, and hopefully the computer will solve it..”
That’s waffle. The computer has the same problem as you or any agent - HOW will it solve the problem? - what are the COMMON ELEMENTS - and COMMON RELATIONSHIPS OF THOSE ELEMENTS – that will enable you or the computer to identify these different figures as belonging to the same class of “chair” and not “collages of wood” or “piles of assorted forms” or “computer desk” or “collections of tools”? ARE there any common elements? You haven’t identified any. [I am using both you guys here as representative figures – it isn’t personal – what you are doing, everyone is doing – totally evading the problem and any analysis of the problem. And yet waffling on confidently about how of course semantic nets or some other form of narrow AI will definitely solve the problem. B & B are saying in effect that those chairs share a “pattern”. But nothing being mooted works – or is in any way relevant to the problem. This IS the problem of AGI – whether you define it as creativity/concepts/visual object recognition/metaphor – the chair problem I have set can be reframed to fit them all. Do you want to face it or evade it? From: Piaget Modeler Sent: Tuesday, October 23, 2012 6:14 PM To: AGI Subject: RE: [agi] Re: Superficiality Produces Misunderstanding - Not Good Enough We would teach the system, PAM-P2 for example, the same way we would teach an infant or toddler. We would show the picture, and then say the word "Chair" or have the word "chair" written under the picture. We would also teach the cognitive system to say the word associated with the picture. We could do this for some number of training examples t < 25. we would then later prompt the system with a test image, and ask what it is, and hopefully the system will respond "Chair". Pretty much that's how it should happen. The cognitive system should learn to associate visual, auditory, proprioceptive, and other modalities within its current, forward, and episodic models in the same manner as children. ~PM. -------------------------------------------------------------------------------- From: [email protected] To: [email protected] Subject: Re: [agi] Re: Superficiality Produces Misunderstanding - Not Good Enough Date: Tue, 23 Oct 2012 17:29:23 +0100 PM & Aaron, You do realise that whatever semantic net system you use must apply to not just one chair, but chair after chair – image after image? Bearing that in mind, explain the elements of your semantic net which you will use to analyse these fairly simple figures as **chairs**:: http://image.shutterstock.com/display_pic_with_logo/95781/95781,1218564477,2/stock-vector-modern-chair-vector-16059484.jpg Let’s label these chairs 1-25 (going L to R from the top down, row after row) Start with just 1. and 2. top left and explain how your net will recognize 2 as another example of 1. How IOW do you define a “chair” in terms of simple abstract forms? Then we can apply your system, successively, to 3. 4. etc. This is the problem that has defeated all AGI-ers and all psychologists and philosophers so far. But Aaron (and PM?) has a semantic net solution to it - if you can solve jungle scenes, this should be a piece of cake. I am saying, Aaron, you do not understand this problem – the problem of visual object recognition/conceptualisation//applicability of semantic nets. You are saying you do – and it’s me who is confused. Show me. From: Piaget Modeler Sent: Tuesday, October 23, 2012 4:41 PM To: AGI Subject: RE: [agi] Re: Superficiality Produces Misunderstanding - Not Good Enough Mike, When you type "Chair" what should happen is the AGI's model should activate the chair concept first at a perceptual level to form the pixels into the words, then at a linguistic level to form letters into a word, then at a conceptual level, then at a simulation level where images of chair instances are evoked. This is just simple activation. Semantic networks tied into perception and simulation would achieve the necessary effect you seek. Transformations on these perception-simulation-semantic networks is what much of Piaget's work was about. ~PM. -------------------------------------------------------------------------------- From: [email protected] To: [email protected] Subject: Re: [agi] Re: Superficiality Produces Misunderstanding - Not Good Enough Date: Tue, 23 Oct 2012 15:09:30 +0100 CHAIR ... It should be able to handle any transformation of the concept, as in DRAW ME (or POINT TO/RECOGNIZE) A CHAIR IN TWO PIECES –.. ..SQUASHED ..IN PIECES -HALF VISIBLE ..WITH AN ARM MISSING ...WITH NO SEAT ..IN POLKA DOTS ...WITH RED STRIPES Concepts are designed for a world of everchanging, everevolving multiform objects (and actions). Semantic networks have zero creativity or adaptability – are applicable only to a uniform set of objects, (basically a database) - and also, crucially, have zero ability to physically recognize or interact with the relevant objects. I’ve been into it at length recently. You’re the one not paying attention. The suggestion that networks or similar can handle concepts is completely absurd. This is yet another form of the central problem of AGI, which you clearly do not understand – and I’m not trying to be abusive – I’ve been realising this again recently – people here are culturally punchdrunk with concepts like *concept* and *creativity*, and just don’t understand them in terms of AGI. From: Jim Bromer Sent: Tuesday, October 23, 2012 2:04 PM To: AGI Subject: Re: [agi] Re: Superficiality Produces Misunderstanding - Not Good Enough Mike Tintner <[email protected]> wrote: AI doesn’t handle concepts. Give me one example to prove that AI doesn't handle concepts. Jim Bromer On Tue, Oct 23, 2012 at 4:24 AM, Mike Tintner <[email protected]> wrote: Jim: Mike refuses to try to understand what I am saying because he would have to give up his sense of a superior point of view in order to understand it Concepts have nothing to do with semantic networks. AI doesn’t handle concepts. That is the challenge for AGI. The form of concepts is graphics. The referents of concepts are infinite realms.. What are you saying that is relevant to this, or that can challenge this – from any evidence? AGI | Archives | Modify Your Subscription AGI | Archives | Modify Your Subscription AGI | Archives | Modify Your Subscription AGI | Archives | Modify Your Subscription AGI | Archives | Modify Your Subscription AGI | Archives | Modify Your Subscription ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-c97d2393 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-2484a968 Powered by Listbox: http://www.listbox.com
