Ben, If you want to argue that recursive self improvement is a special case of learning, then I have no disagreement with the rest of your argument.
But is this really a useful approach to solving AGI? A group of humans can generally make better decisions (more accurate predictions) by voting than any member of the group can. Did these humans improve themselves? My point is that a single person can't create much of anything, much less an AI smarter than himself. If it happens, it will be created by an organization of billions of humans. Without this organization, you would probably not think to create spears out of sticks and rocks. That is my problem with the seed AI approach. The seed AI depends on the knowledge and resources of the economy to do anything. An AI twice as smart as a human could not do any more than 2 people could. You need to create an AI that is billions of times smarter to get anywhere. We are already doing that. Human culture is improving itself by accumulating knowledge, by becoming better organized through communication and specialization, and by adding more babies and computers. -- Matt Mahoney, [EMAIL PROTECTED] --- On Mon, 10/13/08, Ben Goertzel <[EMAIL PROTECTED]> wrote: From: Ben Goertzel <[EMAIL PROTECTED]> Subject: Re: [agi] Updated AGI proposal (CMR v2.1) To: agi@v2.listbox.com Date: Monday, October 13, 2008, 11:46 PM On Mon, Oct 13, 2008 at 11:30 PM, Matt Mahoney <[EMAIL PROTECTED]> wrote: Ben, Thanks for the comments on my RSI paper. To address your comments, You seem to be addressing minor lacunae in my wording, while ignoring my main conceptual and mathematical point!!! 1. I defined "improvement" as achieving the same goal (utility) in less time or achieving greater utility in the same time. I don't understand your objection that I am ignoring run time complexity. OK, you are not "ignoring run time completely" ... BUT ... in your measurement of the benefit achieved by RSI, you're not measuring the amount of run-time improvement achieved, you're only measuring algorithmic information. What matters in practice is, largely, the amount of run-time improvement achieved. This is the point I made in the details of my reply -- which you have not counter-replied to. I contend that, in my specific example, program P2 is a *huge* improvement over P1, in a way that is extremely important to practical AGI yet is not captured by your algorithmic-information-theoretic measurement method. What is your specific response to my example?? 2. I agree that an AIXI type interactive environment is a more appropriate model than a Turing machine receiving all of its input at the beginning. The problem is how to formally define improvement in a way that distinguishes it from learning. I am open to suggestions. To see why this is a problem, consider an agent that after a long time, guesses the environment's program and is able to achieve maximum reward from that point forward. The agent could "improve" itself by hard-coding the environment's program into its successor and thereby achieve maximum reward right from the beginning. Recursive self-improvement **is** a special case of learning; you can't completely distinguish them. 3. A computer's processor speed and memory have no effect on the algorithmic complexity of a program running on it. Yes, I can see I didn't phrase that point properly, sorry. I typed that prior email too hastily as I'm trying to get some work done ;-) The point I *wanted* to make in my third point, was that if you take a program with algorithmic information K, and give it the ability to modify its own hardware, then it can achieve algorithmic information M > K. However, it is certainly true that this can happen even without the program modifying its own hardware -- especially if you make fanciful assumptions like Turing machines with huge tapes ... but even without such fanciful assumptions. The key point, which I did not articulate properly in my prior message, is that: ** by engaging with the world, the program can intake new information, which can increase its algorithmic information ** The new information a program P1 takes in from the **external world** may be random with regard to P1, yet may not be random with regard to {P1 + the new information taken in}. As self-modification may cause the intake of new information causing algorithmic information to increase arbitrarily much, your argument does not hold in the case of a program interacting with a world that has much higher algorithmic information than it does. And this of course is exactly the situation people are in. For instance, a program may learn that "In the past, on 10 occasions, I have taken in information from Bob that was vastly beyond my algorithmic information content at that time. In each case this process helped me to achieve my goals, though in ways I would not have been able to understand before taking in the information. So, once again, I am going to trust Bob to alter me with info far beyond my current comprehension and algorithmic information content." Sounds a bit like a child trusting their parent, eh? This is a separate point from my point about P1 and P2 in point 1. But the two phenomena intersect, of course. -- Ben G This intake agi | Archives | Modify Your Subscription ------------------------------------------- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=117534816-b15a34 Powered by Listbox: http://www.listbox.com