Hi Peter,
Yep I kind of figured that'd be your take. I know your view of things. This
IGI is specifically designed to embark on incorporating brain-mimetic
natural 'computation' on substrates. In an IGI version of AGI the role
of computers in the operation of the intelligence will be determined by the
needs of a core substrate that forms the physical seat of the AGI.
Computation is central. The computer as an origin of that computation is an
experimental variable.

A summary ....we are looking at the idea that there are 2 fundamental kinds
of putative AGI (1) & (3), and their hybrid (2) that forms a third approach
as follows:

(1) C-AGI      computer substrate only. Neuromorphic equivalents of it.
(2) H-AGI      hybrid of (1) and (3). The inorganic version is a new kind
of neuromorphic chip. The organic version has ... erm... organics in it.
(3) S-AGI      synthetic AGI. organic or inorganic. Natural brain physics
only. No computer.

(aside: S-AGI just came out of my fingers. I hope this is OK, Dorian!)

Think this way: What we have now is 100% computer. S-AGI is 100% natural
physics (organic or inorganic). H-AGI is set somewhere in between.  It's
the level of computer computation/natural computation that is at issue. All
are computation.

The human brain is a natural version of (3) with a neuronal/astrocyte
 substrate. (3) has no computer whatever in it. it retains all the natural
physics (whatever that is). H-AGI targets the inclusion of the essential
natural brain physics in the substrate of (2) and to incorporate (1)
computer-substrates and software to an extent to be determined. In my case
an H-AGI would be inorganic. Others see differently.

Where you might have a stake in this?

The history of AGI can be summed up as an experiment that seeks to see if
the role of (1) C-AGI as a brain is fundamentally indistinguishable from
(3) S-AGI under all conditions. That is the hypothesis. The 65 year old bet
that has attracted 100% of the investment to date. H-AGI does not make that
presupposition and seeks to contrast (1) and (3) in revealing ways that
then allow us to speak authoritatively about the (1)/(3) relationship in
AGI potential. Only then will we really understand the difference between
(1) and (3). So far that difference is entirely and intuition. A good one.
But only intuition. Its time for that intuition to be turned into science.
Experiments in (1) have ruled to date. Now we seek to do some (2)... E.E.
we have 65 years of 'control' subject. H-AGI builds the first 'test'
subject.

How about this?

What would be super cool is if this mighty AGI beast you intend making
could be turned into the brain of a robot. Then we could contrast what it
does with what an IGI candidate brain does in an identical robot in the
same test. That kind of testing vision (as far off as it may seem) is a
potential way your work and the IGI might interface. Which candidate robot
best encounters radical novelty, without any human intervention/involvement
whatever? .... is a really good question. To do this test you'd not need to
reveal anything about its workings. Observed robot behaviour is decisive.

It seems to me that whatever venture you plan, it might be wise to keep an
eye on any (2)/(3) approaches. IGI or not. Because it is directly informing
expectations of outcomes in (1). We are currently asking the question "*If
H-AGI were to be championed into existence, what would the first vehicle
for that look like?*" If the enthusiasm maintains it will be sketched into
a web page and we'll see what it tells us and what to do next. It may halt.
It may go. I don't know. Worth a shot? You bet.

With your (1) C-AGI glasses firmly strapped to your head, your wisdom at
all stages in this would be well received, whatever the messages. So if you
have time to keep an  eye on happenings, I for one would appreciate it.

regards

Colin Hales



On Wed, May 20, 2015 at 6:58 AM, Peter Voss <[email protected]> wrote:

> Thanks for asking. Haven’t followed the IGI discussions.
>
>
>
> Is this about non-computer based approaches to AGI?  If so, I don’t think
> I have anything positive to contribute.
>
>
>
> More generally, non-profit orgs need strong focus and champions.  And
> specific goals.
>
>
>
> *From:* Benjamin Kapp [mailto:[email protected]]
> *Sent:* Tuesday, May 19, 2015 12:23 PM
> *To:* AGI
> *Subject:* Re: [agi] Institute of General Intelligence (IGI)
>
>
>
> Mr. Voss,
>
> Given your understanding of the AGI community do you believe an IGI would
> be redundant?  Would your organization be open to collaborating with the
> IGI?  Do you have any advice for how we could be successful in starting up
> this organization?  Perhaps you would be open to being a member of the
> board?
>
>
>
> On Tue, May 19, 2015 at 2:03 PM, Peter Voss <[email protected]> wrote:
>
> Not something that can be adequately covered in a few words, but…. “We’re
> building a fully integrated, top-down & bottom-up, real-time, adaptive
> knowledge (& skill) representation, learning and reasoning engine. We’re
> using a combination of graph representation and NN techniques overlaid with
> fuzzy, adaptive rule systems” – ha!
>
>
>
> Here again are links for some clues:
>
>
>
>
> http://www.kurzweilai.net/essentials-of-general-intelligence-the-direct-path-to-agi
>
> http://www.realagi.com/index.html
>
> https://www.facebook.com/groups/RealAGI/
>
>
>
>
>
> *From:* Benjamin Kapp [mailto:[email protected]]
>
>
>
> Mr. Voss,
>
> Since you are the founder I'm certain you know what agi-3's methodology
> is.  In a few words (maybe more?) could you share with us what that is?
>
>
>
> On Tue, May 19, 2015 at 1:24 PM, Peter Voss <[email protected]> wrote:
>
> *>*http://www.agi-3.com  They just glue together anything and everything
> that works.
>
> Actually, no.  We have a very specific theory of AGI and architecture
>
> *Peter Voss*
>
> *Founder, AGI Innovations Inc.*
>
> *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
> <https://www.listbox.com/member/archive/rss/303/26973278-698fd9ee>| Modify
> <https://www.listbox.com/member/?&;> Your Subscription
>
> <http://www.listbox.com>
>
>
>
> *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
> <https://www.listbox.com/member/archive/rss/303/231420-b637a2b0>| Modify
> <https://www.listbox.com/member/?&;> Your Subscription
>
> <http://www.listbox.com>
>
>
>   *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
> <https://www.listbox.com/member/archive/rss/303/11721311-20a65d4a> |
> Modify
> <https://www.listbox.com/member/?&;>
> Your Subscription <http://www.listbox.com>
>



-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to