Re: [agi] Approximations of Knowledge

2008-06-23 Thread Steve Richfield
On 6/23/08, J. Andrew Rogers <[EMAIL PROTECTED]> wrote: > > Or it could simply mean that the vast majority of programmers and software > monkeys are mediocre at best such that the handful of people you will meet > with deep talent won't constitute a useful sample size. Hell, even Brooks > suggest

Re: [agi] Approximations of Knowledge

2008-06-23 Thread J. Andrew Rogers
On Jun 23, 2008, at 7:53 PM, Steve Richfield wrote: Andy, The use of diminutives is considered rude in many parts of anglo- culture if the individual does not use it to identify themselves, though I realize it is common practice in some regions of the US. When in doubt, use the given for

Re: [agi] Approximations of Knowledge

2008-06-23 Thread Steve Richfield
Andy, This is a PERFECT post, because it so perfectly illustrates a particular point of detachment from reality that is common among AGIers. In the real world we do certain things to achieve a good result, but when we design politically correct AGIs, we banish the very logic that allows us to func

Re: [agi] Approximations of Knowledge

2008-06-23 Thread Richard Loosemore
Jim Bromer wrote: Loosemore said, "It is very important to understand that the paper I wrote was about the methodology of AGI research, not about specific theories/models/systems within AGI. It is about the way that we come up with ideas for systems and the way that we explore those systems,

Re: [agi] Approximations of Knowledge

2008-06-23 Thread Richard Loosemore
Abram Demski wrote: Thanks for the comments. My replies: It does happen to be the case that I believe that logic-based methods are mistaken, but I could be wrong about that, and it could turn out that the best way to build an AGI is with a completely logic-based AGI, along with just one small

Re: [agi] Equivalent ..P.S.I just realised - how can you really understand what I'm talking about - without supplementary images/evidence?

2008-06-23 Thread Mike Tintner
I just realised - how can you really understand what I'm talking about - without supplementary images/evidence? So here's simple evidence - look at the following foto - and note that you can distinguish each individual in it immediately. And you can only do it imagistically. No maths, no langu

Re: [agi] Equivalent of the bulletin for atomic scientists or CRN for AI?

2008-06-23 Thread Russell Wallace
On Mon, Jun 23, 2008 at 11:57 PM, Mike Tintner <[EMAIL PROTECTED]> wrote: > Oh yes, it can be proven. It requires an extended argument to do so > properly, which I won't attempt here. Fair enough, I'd be interested to see your attempted proof if you ever get it written up. --

Re: [agi] Equivalent of the bulletin for atomic scientists or CRN for AI?

2008-06-23 Thread Mike Tintner
Russell:quite a few very smart people (myself among them) have tried hard to design something that could enhance its intelligence divorced from the real world, and all such attempts have failed. Obviously I can't _prove_ the impossibility of this - in the same way that I can't prove the impossi

Re: [agi] Equivalent of the bulletin for atomic scientists or CRN for AI?

2008-06-23 Thread Vladimir Nesov
On Tue, Jun 24, 2008 at 1:29 AM, Russell Wallace <[EMAIL PROTECTED]> wrote: > On Mon, Jun 23, 2008 at 8:48 PM, Vladimir Nesov <[EMAIL PROTECTED]> wrote: >> There are only evolution-built animals, which is a very limited >> repertoir of intelligences. You are saying that if no apple tastes >> like a

Re: [agi] Approximations of Knowledge

2008-06-23 Thread Jim Bromer
Loosemore said, "It is very important to understand that the paper I wrote was about the methodology of AGI research, not about specific theories/models/systems within AGI. It is about the way that we come up with ideas for systems and the way that we explore those systems, not about the conten

Re: [agi] Coin-flipping duplicates (was: Breaking Solomonoff induction (really))

2008-06-23 Thread Matt Mahoney
--- On Mon, 6/23/08, Kaj Sotala <[EMAIL PROTECTED]> wrote: > a) Perform the experiment several times. If, on any of the trials, > copies are created, then have all of them partake in the next trial as > well, flipping a new coin and possibly being duplicated again (and > quickly leading to an expon

Re: [agi] Equivalent of the bulletin for atomic scientists or CRN for AI?

2008-06-23 Thread Russell Wallace
On Mon, Jun 23, 2008 at 8:48 PM, Vladimir Nesov <[EMAIL PROTECTED]> wrote: > There are only evolution-built animals, which is a very limited > repertoir of intelligences. You are saying that if no apple tastes > like a banana, therefore no fruit tastes like a banana, even banana. I'm saying if no

Re: [agi] Equivalent of the bulletin for atomic scientists or CRN for AI?

2008-06-23 Thread Mike Tintner
Vlad, You seem to be arguing in a logical vacuum in denying the essential nature of evidence to most real-world problem-solving. Let's keep it real, bro. Science - bear in mind science deals with every part of the world - from the cosmos to the earth to living organisms, animals, humans, so

Re: [agi] Equivalent of the bulletin for atomic scientists or CRN for AI?

2008-06-23 Thread Vladimir Nesov
On Mon, Jun 23, 2008 at 9:35 PM, Russell Wallace <[EMAIL PROTECTED]> wrote: > On Mon, Jun 23, 2008 at 5:58 PM, Vladimir Nesov <[EMAIL PROTECTED]> wrote: > >> Evidence is an indication that depends on the >> referred event: evidence is there when referred event is there, but >> evidence is not there

Re: [agi] Approximations of Knowledge

2008-06-23 Thread Abram Demski
Thanks for the comments. My replies: > It does happen to be the case that I > believe that logic-based methods are mistaken, but I could be wrong about > that, and it could turn out that the best way to build an AGI is with a > completely logic-based AGI, along with just one small mechanism that

Re: [agi] Equivalent of the bulletin for atomic scientists or CRN for AI?

2008-06-23 Thread Russell Wallace
On Mon, Jun 23, 2008 at 5:58 PM, Vladimir Nesov <[EMAIL PROTECTED]> wrote: > On Mon, Jun 23, 2008 at 8:32 PM, Russell Wallace >> Why do you think that? All the evidence is to the contrary - the >> examples we have of figuring out efficient learning, from evolution to >> childhood play to formal edu

Re: [agi] Equivalent of the bulletin for atomic scientists or CRN for AI?

2008-06-23 Thread Vladimir Nesov
On Mon, Jun 23, 2008 at 8:32 PM, Russell Wallace <[EMAIL PROTECTED]> wrote: > On Mon, Jun 23, 2008 at 5:22 PM, Vladimir Nesov <[EMAIL PROTECTED]> wrote: >> But it can just work with a static corpus. When you need to figure out >> efficient learning, you only need to know a little about the overall

Re: [agi] Equivalent of the bulletin for atomic scientists or CRN for AI?

2008-06-23 Thread Russell Wallace
On Mon, Jun 23, 2008 at 5:22 PM, Vladimir Nesov <[EMAIL PROTECTED]> wrote: > But it can just work with a static corpus. When you need to figure out > efficient learning, you only need to know a little about the overall > structure of your data (which can be described by a reasonably small > number

Re: [agi] Equivalent of the bulletin for atomic scientists or CRN for AI?

2008-06-23 Thread Vladimir Nesov
On Mon, Jun 23, 2008 at 7:52 PM, Russell Wallace <[EMAIL PROTECTED]> wrote: > On Mon, Jun 23, 2008 at 4:34 PM, Vladimir Nesov <[EMAIL PROTECTED]> wrote: >> On Mon, Jun 23, 2008 at 6:52 PM, Russell Wallace >>> Indeed, but becoming more efficient at processing evidence is >>> something that requires

Re: [agi] Equivalent of the bulletin for atomic scientists or CRN for AI?

2008-06-23 Thread Richard Loosemore
William Pearson wrote: While SIAI fills that niche somewhat, it concentrates on the Intelligence explosion scenario. Is there a sufficient group of researchers/thinkers with a shared vision of the future of AI coherent enough to form an organisation? This organisation would discus, explore and di

Re: [agi] Equivalent of the bulletin for atomic scientists or CRN for AI?

2008-06-23 Thread Russell Wallace
On Mon, Jun 23, 2008 at 4:34 PM, Vladimir Nesov <[EMAIL PROTECTED]> wrote: > On Mon, Jun 23, 2008 at 6:52 PM, Russell Wallace >> Indeed, but becoming more efficient at processing evidence is >> something that requires being embedded in the environment to which the >> evidence pertains. > > Why is t

Re: [agi] Approximations of Knowledge

2008-06-23 Thread Abram Demski
Since combinatorial search problems are so common to artificial intelligence, it has obvious applications. If such an algorithm can be made, it seems like it could be used *everywhere* inside an AGI: deduction (solve for cases consistent with constraints), induction (search for the best model), pla

Re: [agi] Approximations of Knowledge

2008-06-23 Thread Richard Loosemore
Abram Demski wrote: To be honest, I am not completely satisfied with my conclusion on the post you refer to. I'm not so sure now that the fundamental split between logical/messy methods should occur at the line between perfect & approximate methods. This is one type of messiness, but one only. I

Re: [agi] Equivalent of the bulletin for atomic scientists or CRN for AI?

2008-06-23 Thread Vladimir Nesov
On Mon, Jun 23, 2008 at 6:52 PM, Russell Wallace <[EMAIL PROTECTED]> wrote: > On Mon, Jun 23, 2008 at 3:43 PM, Vladimir Nesov <[EMAIL PROTECTED]> wrote: >> We are very inefficient in processing evidence, there is "plenty of >> room at the bottom" in this sense alone. Knowledge doesn't come from >>

Re: [agi] Equivalent of the bulletin for atomic scientists or CRN for AI?

2008-06-23 Thread William Pearson
2008/6/23 Vladimir Nesov <[EMAIL PROTECTED]>: > On Mon, Jun 23, 2008 at 12:50 AM, William Pearson <[EMAIL PROTECTED]> wrote: >> 2008/6/22 Vladimir Nesov <[EMAIL PROTECTED]>: >> >>> >>> Two questions: >>> 1) Do you know enough to estimate which scenario is more likely? >> >> Well since intelligence

Re: [agi] Equivalent of the bulletin for atomic scientists or CRN for AI?

2008-06-23 Thread Russell Wallace
On Mon, Jun 23, 2008 at 3:43 PM, Vladimir Nesov <[EMAIL PROTECTED]> wrote: > We are very inefficient in processing evidence, there is "plenty of > room at the bottom" in this sense alone. Knowledge doesn't come from > just feeding the system with data - try to read machine learning > textbooks to a

Re: [agi] Equivalent of the bulletin for atomic scientists or CRN for AI?

2008-06-23 Thread Vladimir Nesov
On Mon, Jun 23, 2008 at 5:22 PM, Russell Wallace <[EMAIL PROTECTED]> wrote: > If we step back and think about it, we really knew this already. In > every case where humans, machines or biological systems exhibit > anything that could be called an intelligence improvement - biological > evolution, a

Re: [agi] Equivalent of the bulletin for atomic scientists or CRN for AI?

2008-06-23 Thread Mike Tintner
Russell:The mistake of trying to reach truth by pure armchair thought was understandable in ancient Greece. We now know better.So attractive as the image of a Transcendent Power popping out of a basement may be to us geeks, it doesn't have anything to do with reality. Making smarter machines in

Re: [agi] Equivalent of the bulletin for atomic scientists or CRN for AI?

2008-06-23 Thread Russell Wallace
Philosophically, "intelligence explosion" in the sense being discussed here is akin to ritual magic - the primary fallacy is the attribution to symbols alone of powers they simply do not possess. The argument is that an initially somewhat intelligent program A can generate a more intelligent progr

[agi] Coin-flipping duplicates (was: Breaking Solomonoff induction (really))

2008-06-23 Thread Kaj Sotala
On 6/23/08, Matt Mahoney <[EMAIL PROTECTED]> wrote: > --- On Sun, 6/22/08, Kaj Sotala <[EMAIL PROTECTED]> wrote: > > On 6/21/08, Matt Mahoney <[EMAIL PROTECTED]> wrote: > > > > > > Eliezer asked a similar question on SL4. If an agent > > flips a fair quantum coin and is copied 10 times if it >

Re: [agi] Equivalent of the bulletin for atomic scientists or CRN for AI?

2008-06-23 Thread Vladimir Nesov
On Mon, Jun 23, 2008 at 12:50 AM, William Pearson <[EMAIL PROTECTED]> wrote: > 2008/6/22 Vladimir Nesov <[EMAIL PROTECTED]>: > >> >> Two questions: >> 1) Do you know enough to estimate which scenario is more likely? > > Well since intelligence explosions haven't happened previously in our > light c

Re: [agi] Equivalent of the bulletin for atomic scientists or CRN for AI?

2008-06-23 Thread William Pearson
2008/6/23 Bob Mottram <[EMAIL PROTECTED]>: > 2008/6/22 William Pearson <[EMAIL PROTECTED]>: >> 2008/6/22 Vladimir Nesov <[EMAIL PROTECTED]>: >> Well since intelligence explosions haven't happened previously in our >> light cone, it can't be a simple physical pattern > > Probably the last "intellige

Re: [agi] Equivalent of the bulletin for atomic scientists or CRN for AI?

2008-06-23 Thread Bob Mottram
2008/6/22 William Pearson <[EMAIL PROTECTED]>: > 2008/6/22 Vladimir Nesov <[EMAIL PROTECTED]>: > Well since intelligence explosions haven't happened previously in our > light cone, it can't be a simple physical pattern Probably the last "intelligence explosion" - a relatively rapid increase in the