Hi Peter, Yep I kind of figured that'd be your take. I know your view of things. This IGI is specifically designed to embark on incorporating brain-mimetic natural 'computation' on substrates. In an IGI version of AGI the role of computers in the operation of the intelligence will be determined by the needs of a core substrate that forms the physical seat of the AGI. Computation is central. The computer as an origin of that computation is an experimental variable.
A summary ....we are looking at the idea that there are 2 fundamental kinds of putative AGI (1) & (3), and their hybrid (2) that forms a third approach as follows: (1) C-AGI computer substrate only. Neuromorphic equivalents of it. (2) H-AGI hybrid of (1) and (3). The inorganic version is a new kind of neuromorphic chip. The organic version has ... erm... organics in it. (3) S-AGI synthetic AGI. organic or inorganic. Natural brain physics only. No computer. (aside: S-AGI just came out of my fingers. I hope this is OK, Dorian!) Think this way: What we have now is 100% computer. S-AGI is 100% natural physics (organic or inorganic). H-AGI is set somewhere in between. It's the level of computer computation/natural computation that is at issue. All are computation. The human brain is a natural version of (3) with a neuronal/astrocyte substrate. (3) has no computer whatever in it. it retains all the natural physics (whatever that is). H-AGI targets the inclusion of the essential natural brain physics in the substrate of (2) and to incorporate (1) computer-substrates and software to an extent to be determined. In my case an H-AGI would be inorganic. Others see differently. Where you might have a stake in this? The history of AGI can be summed up as an experiment that seeks to see if the role of (1) C-AGI as a brain is fundamentally indistinguishable from (3) S-AGI under all conditions. That is the hypothesis. The 65 year old bet that has attracted 100% of the investment to date. H-AGI does not make that presupposition and seeks to contrast (1) and (3) in revealing ways that then allow us to speak authoritatively about the (1)/(3) relationship in AGI potential. Only then will we really understand the difference between (1) and (3). So far that difference is entirely and intuition. A good one. But only intuition. Its time for that intuition to be turned into science. Experiments in (1) have ruled to date. Now we seek to do some (2)... E.E. we have 65 years of 'control' subject. H-AGI builds the first 'test' subject. How about this? What would be super cool is if this mighty AGI beast you intend making could be turned into the brain of a robot. Then we could contrast what it does with what an IGI candidate brain does in an identical robot in the same test. That kind of testing vision (as far off as it may seem) is a potential way your work and the IGI might interface. Which candidate robot best encounters radical novelty, without any human intervention/involvement whatever? .... is a really good question. To do this test you'd not need to reveal anything about its workings. Observed robot behaviour is decisive. It seems to me that whatever venture you plan, it might be wise to keep an eye on any (2)/(3) approaches. IGI or not. Because it is directly informing expectations of outcomes in (1). We are currently asking the question "*If H-AGI were to be championed into existence, what would the first vehicle for that look like?*" If the enthusiasm maintains it will be sketched into a web page and we'll see what it tells us and what to do next. It may halt. It may go. I don't know. Worth a shot? You bet. With your (1) C-AGI glasses firmly strapped to your head, your wisdom at all stages in this would be well received, whatever the messages. So if you have time to keep an eye on happenings, I for one would appreciate it. regards Colin Hales On Wed, May 20, 2015 at 6:58 AM, Peter Voss <[email protected]> wrote: > Thanks for asking. Haven’t followed the IGI discussions. > > > > Is this about non-computer based approaches to AGI? If so, I don’t think > I have anything positive to contribute. > > > > More generally, non-profit orgs need strong focus and champions. And > specific goals. > > > > *From:* Benjamin Kapp [mailto:[email protected]] > *Sent:* Tuesday, May 19, 2015 12:23 PM > *To:* AGI > *Subject:* Re: [agi] Institute of General Intelligence (IGI) > > > > Mr. Voss, > > Given your understanding of the AGI community do you believe an IGI would > be redundant? Would your organization be open to collaborating with the > IGI? Do you have any advice for how we could be successful in starting up > this organization? Perhaps you would be open to being a member of the > board? > > > > On Tue, May 19, 2015 at 2:03 PM, Peter Voss <[email protected]> wrote: > > Not something that can be adequately covered in a few words, but…. “We’re > building a fully integrated, top-down & bottom-up, real-time, adaptive > knowledge (& skill) representation, learning and reasoning engine. We’re > using a combination of graph representation and NN techniques overlaid with > fuzzy, adaptive rule systems” – ha! > > > > Here again are links for some clues: > > > > > http://www.kurzweilai.net/essentials-of-general-intelligence-the-direct-path-to-agi > > http://www.realagi.com/index.html > > https://www.facebook.com/groups/RealAGI/ > > > > > > *From:* Benjamin Kapp [mailto:[email protected]] > > > > Mr. Voss, > > Since you are the founder I'm certain you know what agi-3's methodology > is. In a few words (maybe more?) could you share with us what that is? > > > > On Tue, May 19, 2015 at 1:24 PM, Peter Voss <[email protected]> wrote: > > *>*http://www.agi-3.com They just glue together anything and everything > that works. > > Actually, no. We have a very specific theory of AGI and architecture > > *Peter Voss* > > *Founder, AGI Innovations Inc.* > > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/26973278-698fd9ee>| Modify > <https://www.listbox.com/member/?&> Your Subscription > > <http://www.listbox.com> > > > > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/231420-b637a2b0>| Modify > <https://www.listbox.com/member/?&> Your Subscription > > <http://www.listbox.com> > > > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/11721311-20a65d4a> | > Modify > <https://www.listbox.com/member/?&> > Your Subscription <http://www.listbox.com> > ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657 Powered by Listbox: http://www.listbox.com
