To put it bluntly: Isn't the concept of reinforcement learning (as typically described) exactly the kind of simplistic AI models that AGI tries to distance itself from? Included below is a relevant section of a mail I previously sent to this mailing list. In short, I do not believe AGI is possible based on one single reinfocement loop, so to study them in isolation is of little use (for more detail, read my section about parallell adaptation). Also I do not know wether that document you showed gives any clues about how to scale the method described, but it is scalability that would be most important for any AGI learning algorithms.
I also do not agree that learning about unscalable AI methods is a good way to start learning about AGI. When learning a more scalable methodology, it is sometimes necessary to force oneself not to think in terms of unscalable methods. For example, when artists learn do draw portrait, they need to aquire a drawing skill that scales to the whole face and the whole body they are about to draw. It is not enough to get an ear and nose right if they are not in the correct relative position! So to do this, arts teachers needs to force their students to stop think about the details, and there are various methods for doing this. For example in model drawing there is often a time limit that forces the student to draw very quickly on the entire canvas, giving no time to care about details. Arts students also use more crude tools, like charcoal that gives little room for expressing detail. When fully trained, the student can abstractly disposition a complex scene, and only then it is time to care about details! So if anything, I feel reinforcement learning is the completley wrong thing to start with for someone wanting to learn about AGI. Reading about systems like Novamente (which I found very interesting by the way, maybe I will write some comments on it later) would be a much better start I feel. /Robert Wensman ... *The principle of parallel adaptation* As any programmer knows, premature optimisation is often a bad idea, but on the other hand some optimisation principles are of such importance, that it is necessary to take them into consideration from the very outset. One such principle is that of *parallel adaptation*. If we are to build an adaptive system, working on any domain complex enough, it might be infeasible if we cannot divide the feedback streams into smaller ones that can be considered independently. Assume for example that a certain system has a state of 32 bits that are to be adapted based on some feedback stream. Assume also that there are very little we can assume about the feedback when creating the system, so basically we just have to test different states at random and analyse the feedback stream. The time to adapt such a system could hence be in the order of 2^32. But if instead we can divide the feedback into two steams, each giving feedback to 16 bits correspondingly, then the time to adaptation will suddenly be in the order of 2^17. Hence this is a performance gain of such magnitude that we cannot ignore it, and if evolution creates anywhere near optimal solutions, we can be almost sure of that our own brain utilizes this principle if possible somehow. As you might have noticed my definition of AGI and the assumptions I make about the world are influenced by this principle. For example, I find it necessary for an AGI system to make a clear distinction between knowledge about the world, and ideas about how to act, partially because mixing them could lead to an infeasible adaptation problem. Also, I think assumption (3) I make about the world, namely that the world consists out of a number of objects that could be understood independently, is a key point in any attempt to build an effective AGI. When I was a student I often met AI researchers and other studens interested in AI, who believed it would be impossible to separate behaviour from knowledge in a true AGI, and many held the belief that if we were to build a true AGI, it would be impossible for us to understand its internal structure anyway. Essentially their approach would point towards one single feedback stream for the entire AGI, much like the simple neural networks that were so popular in media a couple of years ago. But there is a risk that if we do not take paralellization of adaptation seriously enough the adaptation process for a larger class of AGI becomes infeasible, and nothing is going to happen in a million years! When it comes to assumption 3 about the world, I have only roughly envisioned some ways to utilize it. It seems that if memes are used to describe the objects of the world, then we need some kind of mechanism that tries to guess what objects are present in a given situation, and which are not, so any adaptation of the world model that occurs, is targeted to the right objects. This is what I would call *spatial meme adaptation*. To elaborate further on this idea, assume that the AGI keeps track of *n*situations called *spatials*, and for each such spatial, there is a *frequency function *that maps memes to a frequency, a real value between 0 and 1 that represents how likely different objects are to occur in this specific situation, in effect defining the profile of a population. Another way could be that each spatial explicitly contains its own population of memeplexes. For any given input the AGI would then perform the following steps. 1. Choose spatial. 2. Perform evolutionary adaptation of the current spatial. There might be a lot of research threads to follow here, for example, is spatial meme adaptation feasible at all, find out different methods to organize the spatials, and would it make sense to have the concept of recursive spatials? (Based on the human language nesting capability, some recursive mechanism in the brain is limited to 2-3 steps, could it be the recursive spatials?). It can be noted that the idea of recursive spatial meme adaptation has structural similarities with Marvin Minsky's "Society of Mind" idea, but focuses more on adaptation and world modelling rather than behavior. ... 2007/6/22, Lukasz Stafiniak <[EMAIL PROTECTED]>:
Obligatory reading: http://www.cs.ualberta.ca/~sutton/book/ebook/the-book.html Cheers. ----- This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?&
----- This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=231415&user_secret=e9e40a7e
