Re: [agi] AGI's Philosophy of Learning

2008-08-20 Thread BillK
On Tue, Aug 19, 2008 at 2:56 PM, Abram Demski [EMAIL PROTECTED] wrote: Wow, sorry about that. I am using firefox and had no problems. The site was just the first reference I was able to find using google. Wikipedia references the same fact:

Re: [agi] AGI's Philosophy of Learning

2008-08-19 Thread Brad Paulsen
Abram, Just FYI... When I attempted to access the Web page in your message, http://www.learnartificialneuralnetworks.com/ (that's without the backpropagation.html part), my virus checker, AVG, blocked the attempt with a message similar to the following: Threat detected! Virus found:

Re: [agi] AGI's Philosophy of Learning

2008-08-19 Thread BillK
On Tue, Aug 19, 2008 at 8:42 AM, Brad Paulsen wrote: Abram, Just FYI... When I attempted to access the Web page in your message, http://www.learnartificialneuralnetworks.com/ (that's without the backpropagation.html part), my virus checker, AVG, blocked the attempt with a message similar to

Re: [agi] AGI's Philosophy of Learning

2008-08-18 Thread Mike Tintner
Abram:I am worried-- worried that an AGI system based on anything less than the one most powerful logic will be able to fool AGI researchers for a long time into thinking that it is capable of general intelligence. Can you explain this to me? (I really am interested in understanding your

Re: [agi] AGI's Philosophy of Learning

2008-08-18 Thread Abram Demski
Mike, There are at least 2 ways this can happen, I think. The first way is that a mechanism is theoretically proven to be complete, for some less-than-sufficient formalism. The best example of this is one I already mentioned: the neural nets of the nineties (specifically, feedforward neural nets

Re: [agi] AGI's Philosophy of Learning

2008-08-18 Thread Mike Tintner
Abram, The key distinction here is probably that some approach to AGI may be widely accepted as having great *promise*. That has certainly been the case, although I doubt actually that it could happen again. There were also no robots of note in the past. Personally, I can't see any approach

Re: [agi] AGI's Philosophy of Learning

2008-08-18 Thread Abram Demski
Mike, But this is horrible! If what you are saying is true, then research will barely progress. On Mon, Aug 18, 2008 at 11:46 AM, Mike Tintner [EMAIL PROTECTED] wrote: Abram, The key distinction here is probably that some approach to AGI may be widely accepted as having great *promise*. That

Re: [agi] AGI's Philosophy of Learning

2008-08-18 Thread Abram Demski
Charles, I find this perspective interesting. Given what logicians know so far, it is more plausible that there is not one right logic, but merely a hierarchy of better/worse/different logics. My search for the top is somewhat unjustified (but I cannot help myself from thinking that there must be

Re: [agi] AGI's Philosophy of Learning

2008-08-15 Thread Jim Bromer
On Thu, Aug 14, 2008 at 5:05 PM, Abram Demski [EMAIL PROTECTED] wrote: But, I am looking for a system that is me. You, like everyone else's me, has it's limitations. So there is a difference between the potential of the system and the actual system. This point of stressing potentiality rather

Re: [agi] AGI's Philosophy of Learning

2008-08-15 Thread Abram Demski
That made more sense to me. Responses follow. On Fri, Aug 15, 2008 at 10:57 AM, Jim Bromer [EMAIL PROTECTED] wrote: On Thu, Aug 14, 2008 at 5:05 PM, Abram Demski [EMAIL PROTECTED] wrote: But, I am looking for a system that is me. You, like everyone else's me, has it's limitations. So there

Re: [agi] AGI's Philosophy of Learning

2008-08-15 Thread Jim Bromer
Our ability to think about abstractions and extrapolations off of abstractions comes because we are able to create game boundaries around the systems that we think about. So yes you can talk about infinite resources and compare it to the domain of the lambda calculus, but this kind of thinking is

Re: [agi] AGI's Philosophy of Learning

2008-08-15 Thread Abram Demski
I don't think the problems of a self-referential paradox is significantly more difficult than the problems of general reference. Not only are there implicit boundaries, some of which have to be changed in an instant as the conversation develops, there are also multiple levels of

Re: [agi] AGI's Philosophy of Learning

2008-08-15 Thread Jim Bromer
On Fri, Aug 15, 2008 at 3:40 PM, Abram Demski [EMAIL PROTECTED] wrote: The paradox seems trivial, of course. I generally agree with your analysis (describing how we consider the sentence, take into account its context, and so on. But the big surprise to logicians was that the paradox is not

Re: [agi] AGI's Philosophy of Learning

2008-08-14 Thread Jim Bromer
One of the worst problems of early AI was that it over-generalized when it tried to use a general rule on a specific case. Actually they over-generalized, under-generalized, and under-specified problem solutions, but over-generalization was the most notable because they relied primarily on word

Re: [agi] AGI's Philosophy of Learning

2008-08-14 Thread Jim Bromer
I realized that I made a very important error in my brief description of prejudice. Prejudice is the application of over-generalizations, typically critical, that are inappropriately applied to a group. The cause of the prejudice is based on a superficial characteristic that most of the members

Re: [agi] AGI's Philosophy of Learning

2008-08-14 Thread Abram Demski
This looks like it could be an interesting thread. However, I disagree with your distinction between ad hoc and post hoc. The programmer may see things from the high-level maze view, but the program itself typically deals with the mess. So, I don't think there is a real distinction to be made

Re: [agi] AGI's Philosophy of Learning

2008-08-14 Thread Jim Bromer
On Thu, Aug 14, 2008 at 12:59 PM, Abram Demski [EMAIL PROTECTED] wrote: A more worrisome problem is that B may be contradictory in and of itself. If (1) I can as a human meaningfully explain logical system X, and (2) logical system X can meaningfully explain anything that humans can, then (3)

Re: [agi] AGI's Philosophy of Learning

2008-08-14 Thread Abram Demski
Jim, You are right to call me on that. I need to provide an argument that, if no logic satisfying B exists, human-level AGI is impossible. B1: A foundational logic for a human-level intelligence should be capable of expressing any concept that a human can meaningfully express. If a broad enough

Re: [agi] AGI's Philosophy of Learning

2008-08-14 Thread Jim Bromer
On Thu, Aug 14, 2008 at 3:06 PM, Abram Demski [EMAIL PROTECTED] wrote: Jim, You are right to call me on that. I need to provide an argument that, if no logic satisfying B exists, human-level AGI is impossible. I don't know why I am being so aggressive these days. I don't start out intending

Re: [agi] AGI's Philosophy of Learning

2008-08-14 Thread Jim Bromer
The paradox (I assume that is what you were pointing to) is based on your idealized presentation. Not only was your presentation idealized, but it was also exaggerated. I sometimes wonder why idealizations can be so effective in some cases. An idealization is actually an imperfect way of

Re: [agi] AGI's Philosophy of Learning

2008-08-14 Thread Abram Demski
On Thu, Aug 14, 2008 at 4:26 PM, Jim Bromer [EMAIL PROTECTED] wrote: On Thu, Aug 14, 2008 at 3:06 PM, Abram Demski [EMAIL PROTECTED] wrote: Jim, You are right to call me on that. I need to provide an argument that, if no logic satisfying B exists, human-level AGI is impossible. I don't know