Ben,

Thanks for that. Your explanation makes the whole thing a lot clearer. 
I'll come back to this thread again after Eliezer's discussion on AGI 
friendliness has progressed a bit further.

Cheers, Philip

From:                   "Ben Goertzel" <[EMAIL PROTECTED]>
To:                     <[EMAIL PROTECTED]>
Subject:                RE: [agi] Novamente: how crtical is self-improvement to 
getting human parity?
Date sent:              Sun, 16 Feb 2003 12:13:16 -0500
Send reply to:          [EMAIL PROTECTED]



Hi,

As we're thinking about it now, Novamente Version 1 will not have 
feature 4. It will involve Novamente learning a lot of small programs to 
use within its overall architecture, but not modifying its overall 
architecture.

Technically speaking: Novamente Version 1 will be C++ code, and 
within this C++ code, there will small programs running in a language 
called Sasha. Novamente will write its own Sasha code to run in its 
C++ "Mind OS", but will not modify its C++ source.

The plan for Novamente Version 2 is still sketchy, because we're 
focusing on Version 1, which still has a long way to go. One possible 
path is to write a fast, scalable Sasha compiler and write the whole 
thing in Sasha. Then the Sasha-programming skills of Novamente 
Version 1 will fairly easily translate into skills at deeper-level self-
modification. (Of course, the Sasha compiler will be in C++ ... so 
eventually you can't escape teaching Novamente C++ ;-).

How intelligent Novamente Version 1 will be -- well ... hmmm ... who 
knows!! 

Among the less sexy benefits of the Novamente Version 2 architecture, 
I really like the idea of having Novamente correct bugs in its own 
source code. It is really hard to get a complex system like this truly 
bug-free..... An AGI should be a lot better at debugging very complex 
code than humans are! 

So the real answer to your question is, I'm not sure. My hope, and my 
guess, is that Novamente Version 1 will --- with ample program learning 
and self-modification on the Sasha level -- be able to achieve levels of 
intelligence that seem huge by human standards. 

Of course, a lot of sci-fi scenarios suggest themselves: What happens 
when we have a super-smart Version 1system and it codes Version 2 
and finds a security hole in Linux and installs Version 2 in place of half 
of itself, then all of itself... etc. 


-- Ben G



    -----Original Message-----
    From: [EMAIL PROTECTED] [mailto:owner-
    [EMAIL PROTECTED]]On Behalf Of Philip Sutton
    Sent: Sunday, February 16, 2003 10:55 AM
    To: [EMAIL PROTECTED]
    Subject: [agi] Novamente: how crtical is self-improvement to 
    getting human parity?
    
    Hi Ben,

As far as I can work out, there are four things that could conceivably 
contribute to a Novamente reaching human intelligence parity:

1 the cleverness/power of the original architecture 

2 the intensity, length and effectiveness of the Novamente learning
 after being booted up

3 the upgrading of the achitecture/code base by humans as a result 
of
 learning by anyone (including Novamentes). 

4 the self-improvement of the achitecture/code base by the 
Novamente
 as a result of learning by anyone (humans and Novamentes). 

To what extend is the learning system of the current Novamente 
system (current or planned for the first switched on version) dependent 
on or intertwined with the capacity for a Novamente to alter its own 
fundamental architecture?

It seems to me that the risk of getting to the sigularity (or even a 
dangerous earlier stage) without the human plus AGI community being 
adequately prepared and sufficiently ethically mature lies in the 
possiblity of AGIs self-improving on an unhalted exponential trajectory.

If you could get Novamentes to human parity using strategies 1-3 only 
then you might be able to control the process of moving beyond human 
parity sufficiently to make it safe.

If getting to human parity relies on strategy 4 then the safety strategy 
could well be very problematic - Eliezer's full Friendly AI program might 
need to be applied in full (ie. developing the theory of friendlieness first 
and then applying "Supersaturated Friendliness" (as Eliezer calls it).

What do you reckon?

Cheers, Philip

-------
To unsubscribe, change your address, or temporarily deactivate your subscription, 
please go to http://v2.listbox.com/member/?[EMAIL PROTECTED]

Reply via email to