Self-improvement is not a special case (was Re: [agi] Religion-free technical content)
From: Derek Zahn [EMAIL PROTECTED] Date: Sun, 30 Sep 2007 08:57:53 -0600 ... One thing that could improve safety is to reject the notion that AGI projects should be focused on, or even capable of, recursive self improvement in the sense of reprogramming its core implementation. ... Let's take Novamente as an example. ... It cannot improve itself until the following things happen: 1) It acquires the knowledge and skills to become a competent programmer, a task that takes a human many years of directed training and practical experience. 2) It is given access to its own implementation and permission to alter it. 3) It understands its own implementation well enough to make a helpful change. ... I agree that resource #1, competent programming, is essential for any interesting takeoff scenario. I don't think the other two matter, though. Consider these two scenarios: Self-improvement: The AI builds a new, improved version of itself, turns on the new one, and turns off the old one. Engineering: The AI builds a better AI and turns it on. These are essentially the same scenario, except in the latter case the AI doesn't turn itself off. If we assume that the new, improved AI really is improved, we don't care much whether the old one turns itself off because the new one is now the dominant player and the old one therefore doesn't matter much. So giving an AI access to its own source code and permission to alter it (resource #2 above) is irrelevant. Giving it understanding of its own source code is irrelevant too; it's just as good for it to understand some other actual or potential AI implementation, and that's subsumed by the competent programming requirement. No value is added by introducing considerations about self-reference into conversations about the consequences of AI engineering. Junior geeks do find it impressive, though. -- Tim Freeman http://www.fungible.com [EMAIL PROTECTED] - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=52968233-2b8b16
RE: Self-improvement is not a special case (was Re: [agi] Religion-free technical content)
Tim Freeman writes: Let's take Novamente as an example. ... It cannot improve itself until the following things happen: 1) It acquires the knowledge and skills to become a competent programmer, a task that takes a human many years of directed training and practical experience. 2) It is given access to its own implementation and permission to alter it. 3) It understands its own implementation well enough to make a helpful change. ... I agree that resource #1, competent programming, is essential for any interesting takeoff scenario. I don't think the other two matter, though. Ok, this alternative scenario -- where Novamente secretly reinvents the theoretical foundations needed for AGI development, designs its successor from those first principles, and somehow hijacks an equivalent or superior supercomputer to receive the de novo design and surreptitiously trains it to superhuman capacity -- should also be protected against. It's a fairly ridiculous scenario, but for completeness should be mentioned. - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=53003223-9d4579
RE: Self-improvement is not a special case (was Re: [agi] Religion-free technical content)
Tim Freeman: No value is added by introducing considerations about self-reference into conversations about the consequences of AI engineering. Junior geeks do find it impressive, though. The point of that conversation was to illustrate that if people are worried about Seed AI exploding, then one option is to not build Seed AI (since that is only one approach to developing AGI, and in fact I do not know of any actual project that includes it at present). Quoting Yudkowsky: The task is not to build an AI with some astronomical level of intelligence; the task is building an AI which is capable of improving itself, of understanding and rewriting its own source code. Perhaps only junior geeks like him find the concept relevant. You seem to think that self-reference buys you nothing at all since it is a simple matter for the first AGI projects to reinvent their own equivalent from scratch, but I'm not sure that's true. - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=53018769-30e88d
Re: [META] Re: Economic libertarianism [was Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
On Wed, Oct 10, 2007 at 01:22:26PM -0400, Richard Loosemore wrote: Am I the only one, or does anyone else agree that politics/political theorising is not appropriate on the AGI list? Yes, and I'm sorry I triggred the thread. I particularly object to libertarianism being shoved down our throats, not so much because I disagree with it, but because so much of the singularity / extropian / futurist discussion universe is dominated by it. Why is that? Before this, the last libertarian I ran across was a few decades ago. And yet, here, they are legion. Why is that? Does libertarian philosphy make people more open-minded to ideas such as the singularity? Make them bigger dreamers? Make them more willing to explore alternatives, even as the rest of the world explores the latest hollywood movie? --linas - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=53020802-76f4d8
Re: Self-improvement is not a special case (was Re: [agi] Religion-free technical content)
Derek, Tim, There is no oversight: self-improvement doesn't necessarily refer to actual instance of self that is to be improved, but to AGI's design. Next thing must be better than previous one for runaway progress to happen, and one way of doing it is for next thing to be a refinement of previous thing. Self-improvement 'in place' may depending on nature of improvement be preferable, if it provides a way to efficiently transfer acquired knowledge from previous version to the next one (probably even without any modification). On 10/12/07, Derek Zahn [EMAIL PROTECTED] wrote: Tim Freeman: No value is added by introducing considerations about self-reference into conversations about the consequences of AI engineering. Junior geeks do find it impressive, though. The point of that conversation was to illustrate that if people are worried about Seed AI exploding, then one option is to not build Seed AI (since that is only one approach to developing AGI, and in fact I do not know of any actual project that includes it at present). Quoting Yudkowsky: The task is not to build an AI with some astronomical level of intelligence; the task is building an AI which is capable of improving itself, of understanding and rewriting its own source code. Perhaps only junior geeks like him find the concept relevant. You seem to think that self-reference buys you nothing at all since it is a simple matter for the first AGI projects to reinvent their own equivalent from scratch, but I'm not sure that's true. This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?; -- Vladimir Nesovmailto:[EMAIL PROTECTED] - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=53025431-7e3757
Re: Self-improvement is not a special case (was Re: [agi] Religion-free technical content)
Let's take Novamente as an example. ... It cannot improve itself until the following things happen: 1) It acquires the knowledge and skills to become a competent programmer, a task that takes a human many years of directed training and practical experience. Wrong. This was hashed to death in previous emails; and then again probably several more times before I joined the list. Anyone care to assemble a position paper on self improvement that reviews the situation? I'm slightly irritated by the recurring speculation and misunderstanding. --linas - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=53103798-a02f8e
Re: Self-improvement is not a special case (was Re: [agi] Religion-free technical content)
From: Derek Zahn [EMAIL PROTECTED] You seem to think that self-reference buys you nothing at all since it is a simple matter for the first AGI projects to reinvent their own equivalent from scratch, but I'm not sure that's true. The from scratch part is a straw-man argument. The AGI project will have lots of resources to draw on; it could read Hutter's papers or license HTM or incorporate lots of other packages or existing AI projects that it might find useful. My point is that if one is worried about a self-improving Seed AI exploding, one should also be worried about any AI that competently writes software exploding. Keeping its source code secret from itself doesn't help much. Hmm, I suppose an AI that does mechanical engineering could explode too, perhaps by doing nanotech, so AI's competently doing engineering is a risk in general. -- Tim Freeman http://www.fungible.com [EMAIL PROTECTED] - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=53137773-858c34
RE: Self-improvement is not a special case (was Re: [agi] Religion-free technical content)
Linas Vepstas: Let's take Novamente as an example. ... It cannot improve itself until the following things happen:1) It acquires the knowledge and skills to become a competent programmer, a task that takes a human many years of directed training and practical experience. Wrong. This was hashed to death in previous emails; and then again probably several more times before I joined the list. Anyone care to assemble a position paper on self improvement that reviews the situation? I'm slightly irritated by the recurring speculation and misunderstanding. Ok, the conversation was about how Novamente could recursively self-improve itself into a runaway hard takeoff scenario. You're claiming that it can do so without the knowledge or skills of a competent programmer, with the very convincing argument Wrong. Care to elaborate at all? Or is your only purpose to communicate your slight irritation? - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=53154037-90851f
Re: Self-improvement is not a special case (was Re: [agi] Religion-free technical content)
Tim Freeman wrote: My point is that if one is worried about a self-improving Seed AI exploding, one should also be worried about any AI that competently writes software exploding. There *is* a slight gap between competently writing software and competently writing minds. Large by human standards, not much by interspecies standards. It does involve new math issues, which is why some of us are much impressed by it. Anyone with even a surface grasp of the basic concept on a math level will realize that there's no difference between self-modifying and writing an outside copy of yourself, but *either one* involves the sort of issues I've been calling reflective. -- Eliezer S. Yudkowsky http://singinst.org/ Research Fellow, Singularity Institute for Artificial Intelligence - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=53161697-a947ab
Re: Self-improvement is not a special case (was Re: [agi] Religion-free technical content)
On 10/12/07, Eliezer S. Yudkowsky [EMAIL PROTECTED] wrote: some of us are much impressed by it. Anyone with even a surface grasp of the basic concept on a math level will realize that there's no difference between self-modifying and writing an outside copy of yourself, but *either one* involves the sort of issues I've been calling reflective. Well, this could be at least a definition of self-modifying ;-) - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=53163374-01d6ba
Re: [META] Re: Economic libertarianism [was Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
On 10/10/2007, Richard Loosemore [EMAIL PROTECTED] wrote: Am I the only one, or does anyone else agree that politics/political theorising is not appropriate on the AGI list? Agreed. There are many other forums where political ideology can be debated. - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=52314407-5d32b9
Re: [META] Re: Economic libertarianism [was Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
I also agree except ... I think political and economic theories can inform AGI design, particularly in areas of AGI decision making and friendliness/roboethics. I wasn't familiar with the theory of Comparative Advantage until Josh and Eric brought it up. (Josh discusses in conjunction with friendly AIs in his The Age of Virtuous Machines at Kurzweil's site.) I like to see discussions in these contexts. -JW -Original Message- From: Bob Mottram [EMAIL PROTECTED] Sent: Oct 11, 2007 11:12 AM To: agi@v2.listbox.com Subject: Re: [META] Re: Economic libertarianism [was Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content] On 10/10/2007, Richard Loosemore [EMAIL PROTECTED] wrote: Am I the only one, or does anyone else agree that politics/political theorising is not appropriate on the AGI list? Agreed. There are many other forums where political ideology can be debated. - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?; - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=52436992-ab6eb0
Re: [META] Re: Economic libertarianism [was Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
Yes, I think that too. On the practical side, I think that investing in AGI requires significant tax cuts, and we should elect a candidate that would do that (Ron Paul). I think that the government has to have more respect to potential weapons (like AGI), so we should elect a candidate who is strongly pro-gun (Ron Paul). I think that the government has to trust and respect the privacy of its people, so your would not be forced to sell your AGI to the military. No more wiretapping (abolish the Patriot Act) so the government won't hear an AGI being successfully developed. Abolish the Federal Reserve, so no more malinvestment, and more productive investment (including agi investment). Ron Paul will do all of that. JW Johnston wrote: I also agree except ... I think political and economic theories can inform AGI design, particularly in areas of AGI decision making and friendliness/roboethics. I wasn't familiar with the theory of Comparative Advantage until Josh and Eric brought it up. (Josh discusses in conjunction with friendly AIs in his The Age of Virtuous Machines at Kurzweil's site.) I like to see discussions in these contexts. -JW -Original Message- From: Bob Mottram [EMAIL PROTECTED] Sent: Oct 11, 2007 11:12 AM To: agi@v2.listbox.com Subject: Re: [META] Re: Economic libertarianism [was Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content] On 10/10/2007, Richard Loosemore [EMAIL PROTECTED] wrote: Am I the only one, or does anyone else agree that politics/political theorising is not appropriate on the AGI list? Agreed. There are many other forums where political ideology can be debated. - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?; - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=52468544-1f3003
Re: Economic libertarianism [was Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
The only solution to this problem I ever see suggested is to intentionally create a Really Big Fish called the government that can effortlessly eat every fish in the pond but promises not to -- to prevent the creation of Really Big Fish. That is quite the Faustian bargain to protect yourself from the lesser demons. Yes, of course, the Really Big Fish that is democracy. I am starting to get quite puzzled by all Americans (I don't know if you are American though, but I want to express this anyway) who express severe distrust in government. Because if you distrust all forms of government, what you really distrust is democracy itself. Here you basically compare democracy to... whom? The devil!? USA is supposed to be the leading democracy of the world yeah, right. But I never hear any people speak so badly of their government, and in effect the democracy itself. The American idea of liberalism is certainly not the same thing as democracy. Maybe this is the century when Americans will find that out. The American liberal culture was founded when the plains of America appeared endless, and if you did not like the influential people of a certain area, you just moved on to virgin grounds and started your own community with your own rules. But there is no more virgin land in America, and people have started to accumulate in the cities since long. Liberty does not work quite so well when people live close and need to depend on each other. That lesson has been learned in Europe ages ago. My recommendation is to put some faith in the will of the people! When you walk on the street and look around you, those are your fellow citizen you should feel at least some kind of trust in. They are not out to get you! Then of course the American form of democracy is not so excellent, so maybe there is a reason for the distrust even though sad. On the surface USA has only two parties which is just one more than China. Sweden is not much better, but at least we have 7 alive and active parties. But these are problems that can be solved and are not a reason to give up on democracy. Generally though, the point that you fail to see is that an AGI can just as easily subvert *any* power structure, whether the environment is a libertarian free market or an autocratic communist state. The problem has nothing to do with the governance of the economy but the fact that the AGI is the single most intelligent actor in the economy however you may arrange it. You can rearrange and change the rules as you wish, but any economy where transactions are something other than completely random is an economy that can be completely dominated by AGI in short order. The game is exactly the same either way, and more rigid economies have much simpler patterns that make them easier to manipulate. Regulating economies to prevent super-intelligent actors from doing bad things is rearranging the deck chairs on the Titanic. I agree that a super intelligent life form could be quite a difficult adversary. It might be able to manipulate and take over a democratic power structure also, I would not deny that. Probably it would try to target the culture of the people, and insert hostile but stealthy memes into the population. I guess it would also try to gain the trust of people and make them dependant on it by offering appealing services. Depending on the economy and regulations, it could also try to obtain direct control over as much automated production capacity as possible, especially production capacity that could be used for building weapons. It is not true like you say, that the economy of a democratic socialist society has easy patterns that are easy to manipulate. The supreme power in such a society lies in the democracy, and to manipulate that power you need to manipulate the whole population. Actually, I believe that the relative stupidity of the population could act as a kind of protection against manipulation. I have a son that is one month old, and I would say it is really difficult to control someone who is so extremely dumb as kids of that age are. However, I would not go as far as saying intelligence implies power, saying that a super intelligent life form by necessity would be able to take over any given power structure. I remember having this discussion with a friend a long time ago. The trivial example is if you have a super intelligent AGI brain in a box in front of you on your desk, and you have a gun. Then you can take the gun and shoot the box. That proves at least that there is no implication in the strict logical sense. But of course the picture gets more complicated if we have an AGI system that interacts in a social context, where we put different degrees of trust in it. Apparently the danger increases the more dependant we are on the AGI systems. But there are methods to protect ourselves. One way is to never utilize the most intelligent AGI systems directly: For example we could use it to produce
Re: Economic libertarianism [was Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
BillK On 10/6/07, a wrote: I am skeptical that economies follow the self-organized criticality behavior. There aren't any examples. Some would cite the Great Depression, but it was caused by the malinvestment created by Central Banks. e.g. The Federal Reserve System. See the Austrian Business Cycle Theory for details. In conclusion, economics is a bad analogy with complex systems. BillK My objection to economic libertarianism is that it's not a free BillK market. A 'free' market is an impossibility. There will always BillK be somebody who is bigger than me or cleverer than me or better BillK educated than me, etc. A regulatory environment attempts to BillK reduce the victimisation of the weaker members of the BillK population and introduces another set of biases to the economy. This is the same misunderstanding that justifies protectionism among nations. When nation A (say the US) trades with nation B (say Haiti), nation A may be able to make every single thing much better and cheaper than nation B, but it still pays both nation B and nation A to trade freely, because nation B has a comparative advantage in something: a comparative advantage being whatever they make least badly, they can swap with nation A and both nations benefit. Likewise, Michael Jordan may be much better able to mow his lawn than whoever he pays to do it, but it still benefits both of them when he pays the lawn guy and concentrates on basketball. You benefit greatly by trading with people who are cleverer, better educated, richer, stronger than you. The more clever they are then you, the more they have to offer you, and the more they will pay you for what you have to offer them. Regulations that restrict your ability to enter into trades with these people hurt you. They do introduce biases into the economy, biases that make everybody worse off, particularly the weaker members of society, except for some special interests that lobby for the regulations and extract rent from society. BillK A free market is just a nice intellectual theory that is of no BillK use in the real world. (Unless you are in the Mafia, of BillK course). BillK BillK BillK - This list is sponsored by AGIRI: BillK http://www.agiri.org/email To unsubscribe or change your BillK options, please go to: BillK http://v2.listbox.com/member/?; - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=51872796-cb97bb
Re: Economic libertarianism [was Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
On Oct 10, 2007, at 2:26 AM, Robert Wensman wrote: Yes, of course, the Really Big Fish that is democracy. No, you got this quite wrong. The Really Big Fish is institution responsible for governance (usually the government); democracy is merely a fuzzy category of rule set used in governance. I am starting to get quite puzzled by all Americans (I don't know if you are American though, but I want to express this anyway) who express severe distrust in government. Because if you distrust all forms of government, what you really distrust is democracy itself. This bias is for good reason; there are well described pathological minima that are essentially unavoidable in a democracy. The American government was explicitly designed as a constitutional republic (not a democracy) to avoid these pathologies. In the 20th century the American constitution was changed to make it more like a democracy, and the expected pathologies have materialized. If you do not understand this, then the rest of your reasoning is likely misplaced. Much of American libertarian political thought is based on a desire to go back to a strict constitutional republic rather than the current quasi-democracy, in large part to fix the very real problems that quasi-democracy created. Many of the bad things the Federal government is currently accused of were enabled by democracy and would have been impractical or illegal under a strict constitutional republic. Here you basically compare democracy to... whom? The devil!? Perhaps I should refrain from using literate metaphors in the future, since you apparently did not understand it. My recommendation is to put some faith in the will of the people! When you walk on the street and look around you, those are your fellow citizen you should feel at least some kind of trust in. They are not out to get you! I'm sure they are all lovely people for the most part, but their poorly reasoned good intentions will destroy us all. The problem is not that people are evil, the problem is that humans at large are hopelessly ignorant, short-sighted, and irrational even when trying to do good and without regard for clearly derivable consequences. Actually, I believe that the relative stupidity of the population could act as a kind of protection against manipulation. Non sequitur. Also, the history shows that intelligence is no guarantee for power. The Russian revolution and the genocide in Cambodia illustrates effectively how intelligent people were slaughtered by apparently less intelligent people, and later how they were controlled to the extreme for decades. You are improperly conflating intelligence and rationality. Cheers, J. Andrew Rogers - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=51970341-6a9d1c
[META] Re: Economic libertarianism [was Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
Am I the only one, or does anyone else agree that politics/political theorising is not appropriate on the AGI list? I particularly object to libertarianism being shoved down our throats, not so much because I disagree with it, but because so much of the singularity / extropian / futurist discussion universe is dominated by it. Richard Loosemore J. Andrew Rogers wrote: On Oct 10, 2007, at 2:26 AM, Robert Wensman wrote: Yes, of course, the Really Big Fish that is democracy. No, you got this quite wrong. The Really Big Fish is institution responsible for governance (usually the government); democracy is merely a fuzzy category of rule set used in governance. I am starting to get quite puzzled by all Americans (I don't know if you are American though, but I want to express this anyway) who express severe distrust in government. Because if you distrust all forms of government, what you really distrust is democracy itself. This bias is for good reason; there are well described pathological minima that are essentially unavoidable in a democracy. The American government was explicitly designed as a constitutional republic (not a democracy) to avoid these pathologies. In the 20th century the American constitution was changed to make it more like a democracy, and the expected pathologies have materialized. If you do not understand this, then the rest of your reasoning is likely misplaced. Much of American libertarian political thought is based on a desire to go back to a strict constitutional republic rather than the current quasi-democracy, in large part to fix the very real problems that quasi-democracy created. Many of the bad things the Federal government is currently accused of were enabled by democracy and would have been impractical or illegal under a strict constitutional republic. Here you basically compare democracy to... whom? The devil!? Perhaps I should refrain from using literate metaphors in the future, since you apparently did not understand it. My recommendation is to put some faith in the will of the people! When you walk on the street and look around you, those are your fellow citizen you should feel at least some kind of trust in. They are not out to get you! I'm sure they are all lovely people for the most part, but their poorly reasoned good intentions will destroy us all. The problem is not that people are evil, the problem is that humans at large are hopelessly ignorant, short-sighted, and irrational even when trying to do good and without regard for clearly derivable consequences. Actually, I believe that the relative stupidity of the population could act as a kind of protection against manipulation. Non sequitur. Also, the history shows that intelligence is no guarantee for power. The Russian revolution and the genocide in Cambodia illustrates effectively how intelligent people were slaughtered by apparently less intelligent people, and later how they were controlled to the extreme for decades. You are improperly conflating intelligence and rationality. Cheers, J. Andrew Rogers - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?; - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=51972366-e14515
Re: Economic libertarianism [was Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
(off topic, but there are something relevant for AGI) My fears about economical libertarianism could be illustrated with a fish pond analogy. If there is a small pond with a large number of small fish of some predatory species, after an amount of time they will cannibalize and eat each other until at the end there will just remain one very very fat fish. The instability occurs because a fish that already has managed to eat a peer, becomes slightly larger than the rest of the fish, and therefore has a better position in continuing to eat more fish, thus its progress can accelerate. Maybe if the pond is big enough, a handful of very big fish would remain. This is of course just an illustration and by no means a proof that the same thing would occur in a laissez-faire/libertarianism economy. Libertarians commonly put blame for monopolies on government involvement, and I guess some would object that I unfairly compares fish that eat each other with a non-violent economy. But lets just say I do not share their relaxed attitude towards the potential threat of monopoly, and a bigger fish eating a smaller fish do have some similarity to a bigger company acquiring a smaller one. First of all, the consequence of monopoly is so serious that even if the chance is very slight, there is a strong incentive to try to prevent it from ever happening. But there are also a lot of details to suggest that a laissez-faire economy would collapse into monopoly/oligopoly. Effects of synergy and mass production benefits would be one strong reason why a completely free market would benefit those companies that are already large, which could make them grow larger. *Especially when considering AGI and intelligence enhancement I believe a libertarian market could be even more unstable. In such a setting, the rich could literally invest in more intelligence, that would make them even more rich, creating a positive economic feedback loop. A dangerous accelerating scenario where the intelligence explosion could co-occur with the rise of world monopoly. We could call it an AGI induced monopoly explosion. Unless democracy could challenge such a libertarian market, only a few oligarchs might have the position to decide the fate of mankind, if they could control their AGI that is. Although it is just one possible scenario.* A documentary I saw claimed that Russia was converted to something very close to a laissez-faire market in the years after the Soviet Union collapse. However I don't have any specific details about it, such as exactly how free the market of that period was. But apparently it caused chaos and gave rise to a brutal economy with oligarchs controlling the society. [ http://en.wikipedia.org/wiki/The_Trap_(television_documentary_series)]. Studying what happened in Russia after the fall of communism could give some insight on the topic. /R 2007/10/8, Bob Mottram [EMAIL PROTECTED]: Economic libertarianism would be nice if it were to occur. However, in practice companies and governments put in place all sorts of anti-competitive structures to lock people into certain modes of economic activity. I think economic activity in general is heavily influenced by cognitive biases of various kinds. On 06/10/2007, BillK [EMAIL PROTECTED] wrote: On 10/6/07, a wrote: A free market is just a nice intellectual theory that is of no use in the real world. - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?; - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=51384923-1d1de1
Re: [agi] Religion-free technical content breaking the small hardware mindset
With googling, I found that older people has lower IQ http://www.sciencedaily.com/releases/2006/05/060504082306.htm IMO, the brain is like a muscle, not an organ. IQ is said to be highly genetic, and the heritability increases with age. Perhaps that older people do not have much mental stimulation as young people? IMO, IQ does not measure general intelligence, and does not certainly measure common sense intelligence. The Bushmen and Pygmy peoples have an average IQ of 54. (source: http://www.rlynn.co.uk/) These IQs are much lower than some mentally retarded and down syndrome people, but the Bushmen and Pygmy peoples act very normal. Yes, IQ is a sensitive and controversial topic, particularly the racial differences in IQ. my ability to recall things is much worse than it was twenty years ago Commonly used culture-free IQ tests, such as Raven Progressive Matrices, generally measure visualspatial intelligence. It does not measure crystallized intelligence such as memory recall, but visualspatial fluid intelligence. I do not take IQ tests importantly. IQ only measures visualspatial reasoning, not auditory nor linguistic intelligence. Some mentally retarded autistic people have extremely high IQs. Edward W. Porter wrote: Dear indefinite article, The Wikipedia entry for Flynn Effect suggests -- in agreement with your comment in the below post -- that older people (at least those in the pre-dementia years) don't get dumber with age relative to their younger selves, but rather relative to the increasing intelligence of people younger than themselves (and, thus, relative to re-normed IQ tests). Perhaps that is correct, but I can tell you that based on my own experience, my ability to recall things is much worse than it was twenty years ago. Furthermore, my ability to spend most of three or four nights in a row lying bed in most of the night with my head buzzing with concepts about an intellectual problem of interest without feeling like a total zombiod in the following days has substantially declined. Since most organs of the body diminish in function with age, it would be surprising if the brain didn't also. We live in the age of political correctness where it can be dangerous to one’s careers to say anything unfavorable about any large group of people, particularly one as powerful as the over 45, who, to a large extent, rule the world. (Or even to those in the AARP, which is an extremely powerful lobby.) So I don't know how seriously I would take the statements that age doesn't affect IQ. My mother, who had the second highest IQ in her college class, was a great one for relaying choice tidbits. She once said that Christiaan Barnard, the first doctor to successfully perform a heart transplant, once said something to the effect of “If you think old people look bad from the outside, you should see how bad they look from the inside.” That would presumably also apply to our brains. - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=51643900-66a52b
Re: Economic libertarianism [was Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
On Oct 9, 2007, at 4:27 AM, Robert Wensman wrote: This is of course just an illustration and by no means a proof that the same thing would occur in a laissez-faire/libertarianism economy. Libertarians commonly put blame for monopolies on government involvement, and I guess some would object that I unfairly compares fish that eat each other with a non-violent economy. But lets just say I do not share their relaxed attitude towards the potential threat of monopoly, and a bigger fish eating a smaller fish do have some similarity to a bigger company acquiring a smaller one. The only solution to this problem I ever see suggested is to intentionally create a Really Big Fish called the government that can effortlessly eat every fish in the pond but promises not to -- to prevent the creation of Really Big Fish. That is quite the Faustian bargain to protect yourself from the lesser demons. Generally though, the point that you fail to see is that an AGI can just as easily subvert *any* power structure, whether the environment is a libertarian free market or an autocratic communist state. The problem has nothing to do with the governance of the economy but the fact that the AGI is the single most intelligent actor in the economy however you may arrange it. You can rearrange and change the rules as you wish, but any economy where transactions are something other than completely random is an economy that can be completely dominated by AGI in short order. The game is exactly the same either way, and more rigid economies have much simpler patterns that make them easier to manipulate. Regulating economies to prevent super-intelligent actors from doing bad things is rearranging the deck chairs on the Titanic. Cheers, J. Andrew Rogers - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=51651108-c1aa2b
RE: [agi] Religion-free technical content breaking the small hardware mindset
I think IQ tests are an important measure, but they don't measure everything important. FDR was not nearly as bright as Richard Nixon, but he was probably a much better president. Ed Porter Original Message- From: a [mailto:[EMAIL PROTECTED] Sent: Tuesday, October 09, 2007 4:19 PM To: agi@v2.listbox.com Subject: Re: [agi] Religion-free technical content breaking the small hardware mindset With googling, I found that older people has lower IQ http://www.sciencedaily.com/releases/2006/05/060504082306.htm IMO, the brain is like a muscle, not an organ. IQ is said to be highly genetic, and the heritability increases with age. Perhaps that older people do not have much mental stimulation as young people? IMO, IQ does not measure general intelligence, and does not certainly measure common sense intelligence. The Bushmen and Pygmy peoples have an average IQ of 54. (source: http://www.rlynn.co.uk/) These IQs are much lower than some mentally retarded and down syndrome people, but the Bushmen and Pygmy peoples act very normal. Yes, IQ is a sensitive and controversial topic, particularly the racial differences in IQ. my ability to recall things is much worse than it was twenty years ago Commonly used culture-free IQ tests, such as Raven Progressive Matrices, generally measure visualspatial intelligence. It does not measure crystallized intelligence such as memory recall, but visualspatial fluid intelligence. I do not take IQ tests importantly. IQ only measures visualspatial reasoning, not auditory nor linguistic intelligence. Some mentally retarded autistic people have extremely high IQs. Edward W. Porter wrote: Dear indefinite article, The Wikipedia entry for Flynn Effect suggests -- in agreement with your comment in the below post -- that older people (at least those in the pre-dementia years) don't get dumber with age relative to their younger selves, but rather relative to the increasing intelligence of people younger than themselves (and, thus, relative to re-normed IQ tests). Perhaps that is correct, but I can tell you that based on my own experience, my ability to recall things is much worse than it was twenty years ago. Furthermore, my ability to spend most of three or four nights in a row lying bed in most of the night with my head buzzing with concepts about an intellectual problem of interest without feeling like a total zombiod in the following days has substantially declined. Since most organs of the body diminish in function with age, it would be surprising if the brain didn't also. We live in the age of political correctness where it can be dangerous to ones careers to say anything unfavorable about any large group of people, particularly one as powerful as the over 45, who, to a large extent, rule the world. (Or even to those in the AARP, which is an extremely powerful lobby.) So I don't know how seriously I would take the statements that age doesn't affect IQ. My mother, who had the second highest IQ in her college class, was a great one for relaying choice tidbits. She once said that Christiaan Barnard, the first doctor to successfully perform a heart transplant, once said something to the effect of If you think old people look bad from the outside, you should see how bad they look from the inside. That would presumably also apply to our brains. - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?; - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=51654844-578b6d
Re: Economic libertarianism [was Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
J. Andrew Rogers wrote: Generally though, the point that you fail to see is that an AGI can just as easily subvert *any* power structure, whether the environment is a libertarian free market or an autocratic communist state. The problem has nothing to do with the governance of the economy but the fact that the AGI is the single most intelligent actor in the economy however you may arrange it. You can rearrange and change the rules as you wish, but any economy where transactions are something other than completely random is an economy that can be completely dominated by AGI in short order. The game is exactly the same either way, and more rigid economies have much simpler patterns that make them easier to manipulate. Regulating economies to prevent super-intelligent actors from doing bad things is rearranging the deck chairs on the Titanic. Succinctly put. -- Eliezer S. Yudkowsky http://singinst.org/ Research Fellow, Singularity Institute for Artificial Intelligence - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=51662113-7b9e18
Re: Economic libertarianism [was Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
Economic libertarianism would be nice if it were to occur. However, in practice companies and governments put in place all sorts of anti-competitive structures to lock people into certain modes of economic activity. I think economic activity in general is heavily influenced by cognitive biases of various kinds. On 06/10/2007, BillK [EMAIL PROTECTED] wrote: On 10/6/07, a wrote: A free market is just a nice intellectual theory that is of no use in the real world. - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=51046640-b84781
Re: [agi] Religion-free technical content
Derek Zahn wrote: Richard Loosemore: a... I often see it assumed that the step between first AGI is built (which I interpret as a functoning model showing some degree of generally-intelligent behavior) and god-like powers dominating the planet is a short one. Is that really likely? Nobody knows the answer to that one. The sooner it is built, the less likely it is to be true. As more accessible computing resources become available, hard takeoff becomes more likely. Note that this isn't a quantitative answer. It can't be. Nobody really knows how much computing power is necessary for a AGI. In one scenario, it would see the internet as it's body, and wouldn't even realize that people existed until very late in the process. This is probably one of the scenarios that require least computing power for takeoff, and allow for fastest spread. Unfortunately, it's also not very likely to be a friendly AI. It would likely feel about people as we feel about the bacteria that make our yogurt. They can be useful to have around, but they're certainly not one's social equals. (This mode of AI might well be social, if, say, it got socialized on chat-lines and newsgroups. But deriving the existence and importance of bodies from those interactions isn't a trivial problem.) The easiest answer isn't necessarily the best one. (Also note that this mode of AI could very likely be developed by a govt. as a weapon for cyber-warfare. Discovering that it was a two-edged sword with a mind of it's own could be a very late-stage event.) - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=51216209-9c2b04
Re: Economic libertarianism [was Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
a wrote: Linas Vepstas wrote: ... The issue is that there's no safety net protecting against avalanches of unbounded size. The other issue is that its not grains of sand, its people. My bank-account and my brains can insulate me from small shocks. I'd like to have protection against the bigger forces that can wipe me out. I am skeptical that economies follow the self-organized criticality behavior. There aren't any examples. Some would cite the Great Depression, but it was caused by the malinvestment created by Central Banks. e.g. The Federal Reserve System. See the Austrian Business Cycle Theory for details. In conclusion, economics is a bad analogy with complex systems. OK. I'm skeptical that a Free-Market economy has ever existed. Possibly the agora of ancient Greece came close. The Persians though so: Who are these people who have special places where they go to cheat each other? However I suspect that a closer look would show that these, also, were regulated to some degree by an external power. (E.g., threat of force from the government if the customers rioted.) - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=51221478-ab187a
Re: Economic libertarianism [was Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
Bob Mottram wrote: Economic libertarianism would be nice if it were to occur. However, in practice companies and governments put in place all sorts of anti-competitive structures to lock people into certain modes of economic activity. I think economic activity in general is heavily influenced by cognitive biases of various kinds. On 06/10/2007, BillK [EMAIL PROTECTED] wrote: On 10/6/07, a wrote: A free market is just a nice intellectual theory that is of no use in the real world. No. Not true. Anti-competitive structures and monopolies won't exist in a true free market society. The free market is self sustaining. It's government regulation that creates monopolies, because companies partner-up with the government. See Chicago school of economics and Austrian school of economics for explanations. Monopolies are much less likely to exist if there is a smaller government. As a response to anti-competitive structures to lock people. Microsoft is a government-supported monopoly. It got its monopoly from the use of software patents. Microsoft patented its file formats, APIs, etc., which resulted in vendor lock-ins. Patent offices, like all bureaucratic agencies, are poor in quality, so lots of trivial ideas can be patented. Do not misinterpret me, I am not against software patents. This is out of topic, but I am in a habit of writing defenses. References http://www.mises.org/story/2317 http://www.cato.org/pubs/regulation/regv12n2/reg12n2-debow.html http://www.ruwart.com/Healing/rutoc.html - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=51227227-be874f
Re: Economic libertarianism [was Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
On Sat, Oct 06, 2007 at 10:05:28AM -0400, a wrote: I am skeptical that economies follow the self-organized criticality behavior. Oh. Well, I thought this was a basic principle, commonly cited in microeconomics textbooks: when there's a demand, producers rush to fill the demand. When there's insufficient demand, producers go out of business. Etc. --linas - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=51229271-e939ba
RE: [agi] Religion-free technical content breaking the small hardware mindset
Dear indefinite article, The Wikipedia entry for Flynn Effect suggests -- in agreement with your comment in the below post -- that older people (at least those in the pre-dementia years) don't get dumber with age relative to their younger selves, but rather relative to the increasing intelligence of people younger than themselves (and, thus, relative to re-normed IQ tests). Perhaps that is correct, but I can tell you that based on my own experience, my ability to recall things is much worse than it was twenty years ago. Furthermore, my ability to spend most of three or four nights in a row lying bed in most of the night with my head buzzing with concepts about an intellectual problem of interest without feeling like a total zombiod in the following days has substantially declined. Since most organs of the body diminish in function with age, it would be surprising if the brain didn't also. We live in the age of political correctness where it can be dangerous to ones careers to say anything unfavorable about any large group of people, particularly one as powerful as the over 45, who, to a large extent, rule the world. (Or even to those in the AARP, which is an extremely powerful lobby.) So I don't know how seriously I would take the statements that age doesn't affect IQ. My mother, who had the second highest IQ in her college class, was a great one for relaying choice tidbits. She once said that Christiaan Barnard, the first doctor to successfully perform a heart transplant, once said something to the effect of If you think old people look bad from the outside, you should see how bad they look from the inside. That would presumably also apply to our brains. Edward W. Porter Porter Associates 24 String Bridge S12 Exeter, NH 03833 (617) 494-1722 Fax (617) 494-1822 [EMAIL PROTECTED] -Original Message- From: a [mailto:[EMAIL PROTECTED] Sent: Saturday, October 06, 2007 10:00 AM To: agi@v2.listbox.com Subject: Re: [agi] Religion-free technical content breaking the small hardware mindset Edward W. Porter wrote: It's also because the average person looses 10 points in IQ between mid twenties and mid fourties and another ten points between mid fourties and sixty. (Help! I'am 59.) But this is just the average. Some people hang on to their marbles as they age better than others. And knowledge gained with age can, to some extent, compensate for less raw computational power. The book in which I read this said they age norm IQ tests (presumably to keep from offending the people older than mid-forties who presumably largely control most of society's institutions, including the purchase of IQ tests.) I disagree with your theory. I primarily see the IQ drop as a result of the Flynn effect, not the age. - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?; - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=51303117-b7930f
Re: [agi] Religion-free technical content breaking the small hardware mindset
Edward W. Porter wrote: It's also because the average person looses 10 points in IQ between mid twenties and mid fourties and another ten points between mid fourties and sixty. (Help! I'am 59.) But this is just the average. Some people hang on to their marbles as they age better than others. And knowledge gained with age can, to some extent, compensate for less raw computational power. The book in which I read this said they age norm IQ tests (presumably to keep from offending the people older than mid-forties who presumably largely control most of society's institutions, including the purchase of IQ tests.) I disagree with your theory. I primarily see the IQ drop as a result of the Flynn effect, not the age. - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=50774160-ad0d02
Re: Economic libertarianism [was Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
Linas Vepstas wrote: My objection to economic libertarianism is its lack of discussion of self-organized criticality. A common example of self-organized criticality is a sand-pile at the critical point. Adding one grain of sand can trigger an avalanche, which can be small, or maybe (unboundedly) large. Despite avalanches, a sand-pile will maintain its critical shape (a cone at some angle). The concern is that a self-organized economy is almost by definition always operating at the critical point, sloughing off excess production, encouraging new demand, etc. Small or even medium-sized re-organizations of the economy are good for it: it maintains the economy at its critical shape, its free-market-optimal shape. Nothing wrong with that free-market optimal shape, most everyone agrees. The issue is that there's no safety net protecting against avalanches of unbounded size. The other issue is that its not grains of sand, its people. My bank-account and my brains can insulate me from small shocks. I'd like to have protection against the bigger forces that can wipe me out. I am skeptical that economies follow the self-organized criticality behavior. There aren't any examples. Some would cite the Great Depression, but it was caused by the malinvestment created by Central Banks. e.g. The Federal Reserve System. See the Austrian Business Cycle Theory for details. In conclusion, economics is a bad analogy with complex systems. - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=50774944-955341
Re: Economic libertarianism [was Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
On 10/6/07, a wrote: I am skeptical that economies follow the self-organized criticality behavior. There aren't any examples. Some would cite the Great Depression, but it was caused by the malinvestment created by Central Banks. e.g. The Federal Reserve System. See the Austrian Business Cycle Theory for details. In conclusion, economics is a bad analogy with complex systems. My objection to economic libertarianism is that it's not a free market. A 'free' market is an impossibility. There will always be somebody who is bigger than me or cleverer than me or better educated than me, etc. A regulatory environment attempts to reduce the victimisation of the weaker members of the population and introduces another set of biases to the economy. A free market is just a nice intellectual theory that is of no use in the real world. (Unless you are in the Mafia, of course). BillK - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=50792589-4d8a77
Re: [agi] Religion-free technical content
Then state the base principles or the algorithm that generates them, without ambiguity and without appealing to common sense. Otherwise I have to believe they are complex too. Existence proof to disprove your I have to believe . . . . 1. Magically collect all members of the species. 2. Magically fully inform them of all relevant details. 3. Magically force them to select moral/ethical/friendly, neutral, or immoral/unethical/unfriendly. 4. If 50% or less select immoral/unethical/unfriendly, then it's friendly. If 50% select immoral/unethical/unfriendly, then it's unfriendly. Simple. Unambiguous. Impossible to implement. (And not my proposal) - Original Message - From: Matt Mahoney [EMAIL PROTECTED] To: agi@v2.listbox.com Sent: Thursday, October 04, 2007 7:26 PM Subject: **SPAM** Re: [agi] Religion-free technical content --- Mark Waser [EMAIL PROTECTED] wrote: I'll repeat again since you don't seem to be paying attention to what I'm saying -- The determination of whether a given action is friendly or ethical or not is certainly complicated but the base principles are actually pretty darn simple. Then state the base principles or the algorithm that generates them, without ambiguity and without appealing to common sense. Otherwise I have to believe they are complex too. -- Matt Mahoney, [EMAIL PROTECTED] - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?; - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=50329295-47e942
Re: [agi] Religion-free technical content
--- Mark Waser [EMAIL PROTECTED] wrote: Then state the base principles or the algorithm that generates them, without ambiguity and without appealing to common sense. Otherwise I have to believe they are complex too. Existence proof to disprove your I have to believe . . . . 1. Magically collect all members of the species. 2. Magically fully inform them of all relevant details. 3. Magically force them to select moral/ethical/friendly, neutral, or immoral/unethical/unfriendly. 4. If 50% or less select immoral/unethical/unfriendly, then it's friendly. If 50% select immoral/unethical/unfriendly, then it's unfriendly. Simple. Unambiguous. Impossible to implement. (And not my proposal) Then I guess we are in perfect agreement. Friendliness is what the average person would do. So how *would* you implement it? - Original Message - From: Matt Mahoney [EMAIL PROTECTED] To: agi@v2.listbox.com Sent: Thursday, October 04, 2007 7:26 PM Subject: **SPAM** Re: [agi] Religion-free technical content --- Mark Waser [EMAIL PROTECTED] wrote: I'll repeat again since you don't seem to be paying attention to what I'm saying -- The determination of whether a given action is friendly or ethical or not is certainly complicated but the base principles are actually pretty darn simple. Then state the base principles or the algorithm that generates them, without ambiguity and without appealing to common sense. Otherwise I have to believe they are complex too. -- Matt Mahoney, [EMAIL PROTECTED] - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?; - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?; -- Matt Mahoney, [EMAIL PROTECTED] - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=50375599-b488f1
Re: [agi] Religion-free technical content
On 10/5/07, Mark Waser [EMAIL PROTECTED] wrote: Then I guess we are in perfect agreement. Friendliness is what the average person would do. Which one of the words in And not my proposal wasn't clear? As far as I am concerned, friendliness is emphatically not what the average person would do. Yeah - Computers already do what the average person would: wait expectantly to be told exactly what to do and how to behave. I guess it's a question of how cynically we define the average person. - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=50390046-8654d8
Re: [agi] Religion-free technical content
Then I guess we are in perfect agreement. Friendliness is what the average person would do. Which one of the words in And not my proposal wasn't clear? As far as I am concerned, friendliness is emphatically not what the average person would do. - Original Message - From: Matt Mahoney [EMAIL PROTECTED] To: agi@v2.listbox.com Sent: Friday, October 05, 2007 10:40 AM Subject: **SPAM** Re: [agi] Religion-free technical content --- Mark Waser [EMAIL PROTECTED] wrote: Then state the base principles or the algorithm that generates them, without ambiguity and without appealing to common sense. Otherwise I have to believe they are complex too. Existence proof to disprove your I have to believe . . . . 1. Magically collect all members of the species. 2. Magically fully inform them of all relevant details. 3. Magically force them to select moral/ethical/friendly, neutral, or immoral/unethical/unfriendly. 4. If 50% or less select immoral/unethical/unfriendly, then it's friendly. If 50% select immoral/unethical/unfriendly, then it's unfriendly. Simple. Unambiguous. Impossible to implement. (And not my proposal) Then I guess we are in perfect agreement. Friendliness is what the average person would do. So how *would* you implement it? - Original Message - From: Matt Mahoney [EMAIL PROTECTED] To: agi@v2.listbox.com Sent: Thursday, October 04, 2007 7:26 PM Subject: **SPAM** Re: [agi] Religion-free technical content --- Mark Waser [EMAIL PROTECTED] wrote: I'll repeat again since you don't seem to be paying attention to what I'm saying -- The determination of whether a given action is friendly or ethical or not is certainly complicated but the base principles are actually pretty darn simple. Then state the base principles or the algorithm that generates them, without ambiguity and without appealing to common sense. Otherwise I have to believe they are complex too. -- Matt Mahoney, [EMAIL PROTECTED] - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?; - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?; -- Matt Mahoney, [EMAIL PROTECTED] - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?; - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=50386329-c4a01e
Re: [agi] Religion-free technical content
--- Mike Dougherty [EMAIL PROTECTED] wrote: On 10/5/07, Mark Waser [EMAIL PROTECTED] wrote: Then I guess we are in perfect agreement. Friendliness is what the average person would do. Which one of the words in And not my proposal wasn't clear? As far as I am concerned, friendliness is emphatically not what the average person would do. Yeah - Computers already do what the average person would: wait expectantly to be told exactly what to do and how to behave. I guess it's a question of how cynically we define the average person. Now you all know damn well what I was trying to say. I thought only computers were supposed to have this problem. -- Matt Mahoney, [EMAIL PROTECTED] - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=50580206-f3a97b
Re: [agi] Religion-free technical content
On Tue, Oct 02, 2007 at 03:03:35PM -0400, Mark Waser wrote: Do you really think you can show an example of a true moral universal? Thou shalt not destroy the universe. Thou shalt not kill every living and/or sentient being including yourself. Thou shalt not kill every living and/or sentient except yourself. What if you discover a sub-stratum alternate-universe thingy that you beleive will be better, but it requires the destruction of this universe to create? What if you discover that there is a god, and that this universe is a kind of cancer or illness in god? (Disclaimer: I did not come up with this; its from some sci-fi book I read as a teen.) Whoops. - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=50615643-d29c68
Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
On Thu, Oct 04, 2007 at 07:49:20AM -0400, Richard Loosemore wrote: As to exactly how, I don't know, but since the AGI is, by assumption, peaceful, friendly and non-violent, it will do it in a peaceful, friendly and non-violent manner. I like to think of myself as peaceful and non-violent, but others have occasionally challenged my self-image. I have also know folks who are physically non-violent, and yet are emotionally controlling monsters. For the most part, modern western culture espouses and hews to physical non-violence. However, modern right-leaning pure capitalism advocates not only social Darwinism, but also the economic equivalent of rape and murder -- a jungle ethic where only the fittest survive, while thousands can loose jobs, income, housing, etc. thanks to the natural forces of capitalism. So.. will a friendly AI also be a radical left-wing economic socialist ?? --linas - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=50633201-155b36
Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
Linas Vepstas wrote: On Thu, Oct 04, 2007 at 07:49:20AM -0400, Richard Loosemore wrote: As to exactly how, I don't know, but since the AGI is, by assumption, peaceful, friendly and non-violent, it will do it in a peaceful, friendly and non-violent manner. I like to think of myself as peaceful and non-violent, but others have occasionally challenged my self-image. I have also know folks who are physically non-violent, and yet are emotionally controlling monsters. For the most part, modern western culture espouses and hews to physical non-violence. However, modern right-leaning pure capitalism advocates not only social Darwinism, but also the economic equivalent of rape and murder -- a jungle ethic where only the fittest survive, while thousands can loose jobs, income, housing, etc. thanks to the natural forces of capitalism. This, anyway, is a common misunderstanding of capitalism. I suggest you to read more about economic libertarianism. So.. will a friendly AI also be a radical left-wing economic socialist ?? Yes, if you define it to be. friendly AI would get the best of both utopian socialism and capitalism. It would get the anti-coercive nature of capitalism and the utopia of utopian socialism. - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=50659417-dd373e
Re: [agi] Religion-free technical content breaking the small hardware mindset
On Wed, Oct 03, 2007 at 08:39:18PM -0400, Edward W. Porter wrote: the IQ bell curve is not going down. The evidence is its going up. So that's why us old folks 'r gettin' stupider as compared to them's young'uns. --linas - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=50669278-fabe77
Economic libertarianism [was Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
OK, this is very off-topic. Sorry. On Fri, Oct 05, 2007 at 06:36:34PM -0400, a wrote: Linas Vepstas wrote: For the most part, modern western culture espouses and hews to physical non-violence. However, modern right-leaning pure capitalism advocates not only social Darwinism, but also the economic equivalent of rape and murder -- a jungle ethic where only the fittest survive, while thousands can loose jobs, income, housing, etc. thanks to the natural forces of capitalism. This, anyway, is a common misunderstanding of capitalism. I suggest you to read more about economic libertarianism. My objection to economic libertarianism is its lack of discussion of self-organized criticality. A common example of self-organized criticality is a sand-pile at the critical point. Adding one grain of sand can trigger an avalanche, which can be small, or maybe (unboundedly) large. Despite avalanches, a sand-pile will maintain its critical shape (a cone at some angle). The concern is that a self-organized economy is almost by definition always operating at the critical point, sloughing off excess production, encouraging new demand, etc. Small or even medium-sized re-organizations of the economy are good for it: it maintains the economy at its critical shape, its free-market-optimal shape. Nothing wrong with that free-market optimal shape, most everyone agrees. The issue is that there's no safety net protecting against avalanches of unbounded size. The other issue is that its not grains of sand, its people. My bank-account and my brains can insulate me from small shocks. I'd like to have protection against the bigger forces that can wipe me out. --linas - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=50672693-e11dc1
RE: [agi] Religion-free technical content breaking the small hardware mindset
It's also because the average person looses 10 points in IQ between mid twenties and mid fourties and another ten points between mid fourties and sixty. (Help! I'am 59.) But this is just the average. Some people hang on to their marbles as they age better than others. And knowledge gained with age can, to some extent, compensate for less raw computational power. The book in which I read this said they age norm IQ tests (presumably to keep from offending the people older than mid-forties who presumably largely control most of society's institutions, including the purchase of IQ tests.) Edward W. Porter Porter Associates 24 String Bridge S12 Exeter, NH 03833 (617) 494-1722 Fax (617) 494-1822 [EMAIL PROTECTED] -Original Message- From: Linas Vepstas [mailto:[EMAIL PROTECTED] Sent: Friday, October 05, 2007 7:31 PM To: agi@v2.listbox.com Subject: Re: [agi] Religion-free technical content breaking the small hardware mindset On Wed, Oct 03, 2007 at 08:39:18PM -0400, Edward W. Porter wrote: the IQ bell curve is not going down. The evidence is its going up. So that's why us old folks 'r gettin' stupider as compared to them's young'uns. --linas - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?; - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=50724257-8e390c
The first-to-market effect [WAS Re: [agi] Religion-free technical content]
Linas Vepstas wrote: On Tue, Oct 02, 2007 at 01:20:54PM -0400, Richard Loosemore wrote: When the first AGI is built, its first actions will be to make sure that nobody is trying to build a dangerous, unfriendly AGI. Yes, OK, granted, self-preservation is a reasonable character trait. After that point, the first friendliness of the first one will determine the subsequent motivations of the entire population, because they will monitor each other. Um, why, exactly, are you assuming that the first one will be freindly? The desire for self-preservation, by e.g. rooting out and exterminating all (potentially unfreindly) competing AGI, would not be what I'd call freindly behavior. There's also a strong sense that winnner-takes-all, or first-one-takes-all, as the first one is strongly motivated, by instinct for self-preservation, to make sure that no other AGI comes to exist that could threaten, dominate or terminate it. In fact, the one single winner, out of sheer loneliness and boredom, might be reduced to running simulations a la Nick Bostrom's simulation argument (!) --linas This is interesting, because you have put your finger on a common reaction to the First AGI will take down the others idea. When I talk about this first-to-market effect, I *never* mean that the first one will eliminate or exterminate all the others, and I do not mean to imply that it would do so because it feels motives akin to self-preservation, or because it does not want to be personally dominated, threatened (etc) by some other AGI. What I mean is that ASSUMING the first one is friendly (that assumption being based on a completely separate line of argument), THEN it will be obliged, because of its commitment to friendliness, to immediately search the world for dangerous AGI projects and quietly ensure that none of them are going to become a danger to humanity. There is absolutely no question of it doing this because of a desire for self-preservation, or jealousy, feeling threatened, or any of those other motivations, because the most important part of the design of the first friendly AGI will be that it will not have those motivations. Not only that, but it will not necessarily wipe out those other AGIs, either. If we value intelligent, sentient life, we may decide that the best thing to do with these other AGI designs, if they have reached the point of self-awareness, is to let them keep most of their memories but modify them slightly so that they can be transferred into the new, friendly design. To be honest, I do not think it is likely that there will be others that are functioning at the level of self-awareness at that stage, but that's another matter. So this would be a quiet-but-friendly modification of other systems, to put security mechanisms in place, not an aggressive act. This follows directly from the assumption of friendliness of the first AGI. Some people have talked about aggressive takeovers. That is a completely different kettle of fish, which assumes the first one will be aggressive. For reasons I have stated elsewhere, I think that *in* *practice* the first one will not be aggressive. Richard Loosemore - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49645537-bddb78
Small amounts of Complexity [WAS Re: [agi] Religion-free technical content]
Linas Vepstas wrote: On Wed, Oct 03, 2007 at 12:20:10PM -0400, Richard Loosemore wrote: Second, You mention the 3-body problem in Newtonian mechanics. Although I did not use it as such in the paper, this is my poster child of a partial complex system. I often cite the case of planetary system dynamics as an example of a real physical system that is PARTIALLY complex, because it is mostly governed by regular dynamics (which lets us predict solar eclipse precisely), but also has various minor aspects that are complex, such as Pluto's orbit, braiding effects in planetary rings, and so on. Richard, we had this conversation in private, but we can have it again in public. J Storrs Hall is right. You can't actually say that the 3-body problem has various minor aspects that are complex, such as Pluto's orbit. That's just plain wrong. The phonomenon you are describing is known as the small divisors problem, and has been studied for several hundred years, with a particularly thick corpus developed about 150 years ago, if I remember rightly. The initial hopes of astronomers were that planetary motion would be exactly as you describe it: that its mostly regular dynamics, with just some minor aspects, some minor corrections. This hope was dashed. The minor corrections, or perturbations, have a denominator, in which appear ratios of periods of orbits. Some of these denominators can get arbitrarily small, implying that the small correction is in fact unboundedly large. This was discovered, I dunno, several hundred years ago, and elucidated in the 19th century. Both Poincare and Einstein made notable contributions. Modern research into chaos theory has shed new insight into what's really going on; it has *not*, however, made planetary motion only a partially complicated system. It is quite fully wild and wooly. In a very deep sense, planetary motion is wildly and insanely unpredicatable. Just becaouse we can work out numerical simulations for the next million years does not mean that the system is complex in only minor ways; this is a fallacious deduction. Note the probabilites of pluto going bonkers are not comparable to the sun tunneling into bloomingdale's but are in fact much, much higher. Pluto could fly off tommorrow, and the probability is big enough that you have to actually account for it. The problem with this whole email thread tends to be that many people are willing to agree with your conclusions, but dislike the manner in which they are arrived at. Brushing off planetary motion, or the Turing-completeness of Conway's life, just basically points to a lack of understanding of the basic principles to which you appeal. Linas, The difficulty I sometimes have with discussions in this format is that it is perfectly acceptable for people to disagree with the ideas, but they should keep personal insults OUT of the discussion -- and accordingly, in my reactions to other people, I *never* introduce ad hominem remarks, I only *respond* to ad hominem insults from others. I responded that way when Josh decided to disagree by using extremely insulting language. To anyone who disagrees politely, I will put in huge amounts of effort to meet their criticisms, help clarify the issues, apologize for any lack of clarity on my part, etc. Now, to the subject at hand. I hear what you are saying about the 3-body problem. [I would have been happier if YOU had managed to phrase it without making assertions about what I do and do not understand, because I earned a physics degree, with a strong Astronomy component, back in 1979, and I have been aware of these issues for a very long time]. Even though you assertively declare that my use of the example of plenetary orbits is just plain wrong, I knew exactly what I was doing when I picked the example, and it is precisely correct. I will explain why. The core issue has to do with what I actually mean when I say that planetary orbits contain a small amount of complexity. You are interpreting that statement one way, but I am using in a different way. It is my usage that matters, because this discussion is, after all, about my paper, the way I used the phrase in that paper, and the way that other people who talk about amounts of complexity would tend to use that phrase. The amount of complexity is all about the exactness of an overall scientific explanation for a system. It is about the extent to which a normal, scientific explanation can be set down and used to explain the system. Is it possible to find an explanation that covers the state of that system very accurately for a very large fraction of that system's lifetime? If the answer is yes, then the system does not have much complexity in it. If the vast bulk of the lifetime of that system cannot be understood by any normal scientific explanation (i.e. cannot be predicted), and if the behavior is not completely random, then we would say that the
Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
On 04/10/2007, Richard Loosemore [EMAIL PROTECTED] wrote: Linas Vepstas wrote: Um, why, exactly, are you assuming that the first one will be freindly? The desire for self-preservation, by e.g. rooting out and exterminating all (potentially unfreindly) competing AGI, would not be what I'd call freindly behavior. What I mean is that ASSUMING the first one is friendly (that assumption being based on a completely separate line of argument), THEN it will be obliged, because of its commitment to friendliness, to immediately search the world for dangerous AGI projects and quietly ensure that none of them are going to become a danger to humanity. Whether you call it extermination or ensuring they won't be a danger the end result seems like the same thing. In the world of realistic software development how is it proposed that this kind of neutralisation (or termination if you prefer) should occur ? Are we talking about black hat type activity here, or agents of the state breaking down doors and seizing computers? - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49651474-cd3887
Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
Bob Mottram wrote: On 04/10/2007, Richard Loosemore [EMAIL PROTECTED] wrote: Linas Vepstas wrote: Um, why, exactly, are you assuming that the first one will be freindly? The desire for self-preservation, by e.g. rooting out and exterminating all (potentially unfreindly) competing AGI, would not be what I'd call freindly behavior. What I mean is that ASSUMING the first one is friendly (that assumption being based on a completely separate line of argument), THEN it will be obliged, because of its commitment to friendliness, to immediately search the world for dangerous AGI projects and quietly ensure that none of them are going to become a danger to humanity. Whether you call it extermination or ensuring they won't be a danger the end result seems like the same thing. In the world of realistic software development how is it proposed that this kind of neutralisation (or termination if you prefer) should occur ? Are we talking about black hat type activity here, or agents of the state breaking down doors and seizing computers? Well, forgive me, but do you notice that you are always trying to bring it back to language that implies malevolence? It is this very implication of malevolent intent that I am saying is unjustified, because it makes it seem like it is something that it is not. As to exactly how, I don't know, but since the AGI is, by assumption, peaceful, friendly and non-violent, it will do it in a peaceful, friendly and non-violent manner. Richard Loosemore - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49658671-fcb107
Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
On 04/10/2007, Richard Loosemore [EMAIL PROTECTED] wrote: As to exactly how, I don't know, but since the AGI is, by assumption, peaceful, friendly and non-violent, it will do it in a peaceful, friendly and non-violent manner. This seems very vague. I would suggest that if there is no clear mechanism for stopping someone from developing an AGI then such enforcement will not occur in practice. - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49661340-6aec9f
Re: [agi] Religion-free technical content
I mean that ethics or friendliness is an algorithmically complex function, like our legal system. It can't be simplified. The determination of whether a given action is friendly or ethical or not is certainly complicated but the base principles are actually pretty darn simple. However, I don't believe that friendliness can be made stable through RSI. Your wording is a bit unclear here. RSI really has nothing to do with friendliness other than the fact that RSI makes the machine smarter and the machine then being smarter *might* have any of the consequences of: 1.. understanding friendliness better 2.. evaluating whether something is friendly better 3.. convincing the machine that friendliness should only apply to the most evolved life-form (something that this less-evolved life-form sees as patently ridiculous) I'm assuming that you mean you believe that friendliness can't be made stable under improving intelligence. I believe that you're wrong. We can summarize the function's decision process as what would the average human do in this situation? That's not an accurate summary as far as I'm concerned. I don't want *average* human judgement. I want better. The function therefore has to be modifiable because human ethics changes over time, e.g. attitudes toward the rights of homosexuals, the morality of slavery, or whether hanging or crucifixion is an appropriate form of punishment. I suspect that our best current instincts are fairly close to friendliness. Humans started out seriously unfriendly because friendly entities *don't* survive in an environment populated only by unfriendlies. As society grows and each individual becomes friendlier, it's an upward spiral to where we need/want to be. I think that the top of the spiral (i.e. the base principles) is pretty obvious. I think that the primary difficulties are determining all the cases where we're constrained by circumstances and what won't work yet and can't determine what is best. Second, as I mentioned before, RSI is necessarily experimental, and therefore evolutionary, and the only stable goal in an evolutionary process is rapid reproduction and acquisition of resources. I disagree strongly. Experimental only implies a weak meaning of the term evolutionary and your assertion that the only stable goal in a evolutionary process is rapid reproduction and acquisition of resources may apply to the most obvious case of animal evolution but it certainly doesn't apply to numerous evolutionary process that scientists perform all the time (For example, when scientists are trying to evolve a protein that binds to a certain receptor. In that case, the stable goal is binding strength and nothing else since the scientists then provide the reproduction for the best goal-seekers). But as AGI grows more powerful, humans will be less significant and more like a lower species that competes for resources. So you don't believe that humans will self-improve? You don't believe that humans will be able to provide something that the AGI might value? You don't believe that a friendly AGI would be willing not to hog *all* the resources. Personally, I think that the worst case with a friendly AGI is that we would end up as pampered pets until we could find a way to free ourselves of our biology. - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49672841-ba128c
Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
Bob Mottram wrote: On 04/10/2007, Richard Loosemore [EMAIL PROTECTED] wrote: As to exactly how, I don't know, but since the AGI is, by assumption, peaceful, friendly and non-violent, it will do it in a peaceful, friendly and non-violent manner. This seems very vague. I would suggest that if there is no clear mechanism for stopping someone from developing an AGI then such enforcement will not occur in practice. Oh, in my haste I forgot to remind you that this assumes RSI: the first AGI to be built will undoubtedly be used to design faster and possibly smarter forms of AGI, in a rapidly escalating process. It is only under *those* assumptions that it will take action. It will invent techniques to handle the situation. Nanotech, probably, but perhaps lesser techniques will do. Clearly, at the base level it will not be able to do anything, if it doesn't have access to any effectors. Richard Loosemore - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49776231-a7ad14
Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
To me this seems like elevating that status of nanotech to magic. Even given RSI and the ability of the AGI to manufacture new computing resources it doesn't seem clear to me how this would enable it to prevent other AGIs from also reaching RSI capability. Presumably lesser techniques means black hat activity, or traditional forms of despotism. There seems to be a clarity gap in the theory here. On 04/10/2007, Richard Loosemore [EMAIL PROTECTED] wrote: Bob Mottram wrote: On 04/10/2007, Richard Loosemore [EMAIL PROTECTED] wrote: As to exactly how, I don't know, but since the AGI is, by assumption, peaceful, friendly and non-violent, it will do it in a peaceful, friendly and non-violent manner. This seems very vague. I would suggest that if there is no clear mechanism for stopping someone from developing an AGI then such enforcement will not occur in practice. Oh, in my haste I forgot to remind you that this assumes RSI: the first AGI to be built will undoubtedly be used to design faster and possibly smarter forms of AGI, in a rapidly escalating process. It is only under *those* assumptions that it will take action. It will invent techniques to handle the situation. Nanotech, probably, but perhaps lesser techniques will do. Clearly, at the base level it will not be able to do anything, if it doesn't have access to any effectors. Richard Loosemore - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?; - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49857713-6526fe
Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
On 10/4/07, Bob Mottram [EMAIL PROTECTED] wrote: To me this seems like elevating that status of nanotech to magic. Even given RSI and the ability of the AGI to manufacture new computing resources it doesn't seem clear to me how this would enable it to prevent other AGIs from also reaching RSI capability. Presumably lesser techniques means black hat activity, or traditional forms of despotism. There seems to be a clarity gap in the theory here. The first true AGI may be friendly, as suggested by Richard Loosemore. But if the military are working on developing an intelligent weapons system, then a sub-project will be a narrow AI project designed specifically to seek out and attack the competition *before* it becomes a true AGI. The Chinese are already constantly probing and attacking the western internet sites. BillK - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49977621-104d4e
Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
On Thursday 04 October 2007 11:50:21 am, Bob Mottram wrote: To me this seems like elevating that status of nanotech to magic. Even given RSI and the ability of the AGI to manufacture new computing resources it doesn't seem clear to me how this would enable it to prevent other AGIs from also reaching RSI capability. Hear, hear and again I say hear, hear! There's a lot of and then a miracle occurs in step 2 in the we build a friendly AI and it takes over the world and saves our asses type reasoning we see so much of. (Or the somebody builds an unfriendly AI and it takes over the world and wipes us out reasoning as well.) We can't build a system that learns as fast as a 1-year-old just now. Which is our most likely next step: (a) A system that does learn like a 1-year-old, or (b) a system that can learn 1000 times as fast as an adult? Following Moore's law and its software cognates, I'd say give me the former and I'll give you the latter in a decade. With lots of hard work. Then and only then will you have something that's able to improve itself faster than a high-end team of human researchers and developers could. Furthermore, there's a natural plateau waiting for it. That's where it has to leave off learning by absorbing knowledge fom humans (reading textbooks and research papers, etc) and doing the actual science itself. I have heard NO ONE give an argument that puts a serious dent in this, to my way of thinking. Josh - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=50014668-f60c12
Re: [agi] Religion-free technical content
--- Mark Waser [EMAIL PROTECTED] wrote: I mean that ethics or friendliness is an algorithmically complex function, like our legal system. It can't be simplified. The determination of whether a given action is friendly or ethical or not is certainly complicated but the base principles are actually pretty darn simple. If the base principles are to not harm humans or to do what humans tell them, then these are not simple. We both know what that means but we need common sense to do so. Common sense is not algorithmically simple. Ask Doug Lenat. However, I don't believe that friendliness can be made stable through RSI. Your wording is a bit unclear here. RSI really has nothing to do with friendliness other than the fact that RSI makes the machine smarter and the machine then being smarter *might* have any of the consequences of: 1.. understanding friendliness better 2.. evaluating whether something is friendly better 3.. convincing the machine that friendliness should only apply to the most evolved life-form (something that this less-evolved life-form sees as patently ridiculous) I'm assuming that you mean you believe that friendliness can't be made stable under improving intelligence. I believe that you're wrong. I mean that an initially friendly AGI might not be friendly after RSI. If the child AGI is more intelligent, then the parent will not be able to fully evaluate its friendliness. We can summarize the function's decision process as what would the average human do in this situation? That's not an accurate summary as far as I'm concerned. I don't want *average* human judgement. I want better. Who decides what is better? If the AGI decides, then future generations will certainly be better -- by its definition. I suspect that our best current instincts are fairly close to friendliness. Our current instincts allow war, crime, torture, and genocide. Future generations will look back at us as barbaric. Do you want to freeze the AGI's model of ethics at its current level, or let it drift in a manner beyond our control. Second, as I mentioned before, RSI is necessarily experimental, and therefore evolutionary, and the only stable goal in an evolutionary process is rapid reproduction and acquisition of resources. I disagree strongly. Experimental only implies a weak meaning of the term evolutionary and your assertion that the only stable goal in a evolutionary process is rapid reproduction and acquisition of resources may apply to the most obvious case of animal evolution but it certainly doesn't apply to numerous evolutionary process that scientists perform all the time (For example, when scientists are trying to evolve a protein that binds to a certain receptor. In that case, the stable goal is binding strength and nothing else since the scientists then provide the reproduction for the best goal-seekers). That only works because a more intelligent entity (the scientists) controls the goal. RSI is uncontrolled, just like biological evolution. So you don't believe that humans will self-improve? You don't believe that humans will be able to provide something that the AGI might value? You don't believe that a friendly AGI would be willing not to hog *all* the resources. Personally, I think that the worst case with a friendly AGI is that we would end up as pampered pets until we could find a way to free ourselves of our biology. AGI may well produce a utopian society. But as you say, there might not be anything in it that resembles human life. -- Matt Mahoney, [EMAIL PROTECTED] - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=50018754-0d8000
Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
On 10/4/07, J Storrs Hall, PhD [EMAIL PROTECTED] wrote: We can't build a system that learns as fast as a 1-year-old just now. Which is our most likely next step: (a) A system that does learn like a 1-year-old, or (b) a system that can learn 1000 times as fast as an adult? Following Moore's law and its software cognates, I'd say give me the former and I'll give you the latter in a decade. With lots of hard work. Then and only then will you have something that's able to improve itself faster than a high-end team of human researchers and developers could. You can't know that if you don't have any working algorithm or theory that's able to predict required computational power... At least there is no reason to expect that when AGI is implemented, hardware existing at that time is going to provide it with exactly the resources of human being. It might be that hardware will not be ready, or that it'll be enough to run it 1000 faster. So, in event that hardware will be sufficient, it will probably be enough for AGI to run much faster than human. -- Vladimir Nesovmailto:[EMAIL PROTECTED] - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=50126912-aa8f49
Re: [agi] Religion-free technical content
--- Mark Waser [EMAIL PROTECTED] wrote: I'll repeat again since you don't seem to be paying attention to what I'm saying -- The determination of whether a given action is friendly or ethical or not is certainly complicated but the base principles are actually pretty darn simple. Then state the base principles or the algorithm that generates them, without ambiguity and without appealing to common sense. Otherwise I have to believe they are complex too. -- Matt Mahoney, [EMAIL PROTECTED] - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=50198284-daaf75
Re: [agi] Religion-free technical content
On Tuesday 02 October 2007 08:46:43 pm, Richard Loosemore wrote: J Storrs Hall, PhD wrote: I find your argument quotidian and lacking in depth. ... What you said above was pure, unalloyed bullshit: an exquisite cocktail of complete technical ignorance, patronizing insults and breathtaking arrogance. ... I find this argument lacking in depth, as well. Actually, much of your paper is right. What I said was that I've heard it all before (that's what quotidian means) and others have taken it farther than you have. You write (proceedings p. 161) The term 'complex system' is used to describe PRECISELY those cases where the global behavior of the system shows interesting regularities, and is not completely random, but where the nature of the interaction of the components is such that we would normally expect the consequences of those interactions to be beyond the reach of analytic solutions. (emphasis added) But of course even a 6th-degree polynomial is beyond the reach of an analytic solution, as is the 3-body problem in Newtonian mechanics. And indeed the orbit of Pluto has been shown to be chaotic. But we can still predict with great confidence when something as finicky as a solar eclipse will occur thousands of years from now. So being beyond analytic solution does not mean unpredictable in many, indeed most, practical cases. We've spent five centuries learning how to characterize, find the regularities in, and make predictions about, systems that are in your precise definition, complex. We call this science. It is not about analytic solutions, though those are always nice, but about testable hypotheses in whatever form stated. Nowadays, these often come in the form of computational models. You then talk about the Global-Local Disconnect as if that were some gulf unbridgeable in principle the instant we find a system is complex. But that contradicts the fact that science works -- we can understand a world of bouncing molecules and sticky atoms in terms of pressure and flammability. Science has produced a large number of levels of explanation, many of them causally related, and will continue doing so. But there is not (and never will be) any overall closed-form analytic solution. The physical world is, in your and Wolfram's words, computationally irreducible. But computational irreducibility is a johnny-come-lately retread of a very key idea, Gödel incompleteness, that forms the basis of much of 20th-century mathematics, including computer science. It is PROVABLE that any system that is computationally universal cannot be predicted, in general, except by simulating its computation. This was known well before Wolfram came along. He didn't say diddley-squat that was new in ANKoS. So, any system that is computationally universal, i.e. Turing-complete, i.e. capable of modelling a Universal Turing machine, or a Post production system, or the lambda calculus, or partial recursive functions, is PROVABLY immune to analytic solution. And yet, guess what? we have computer *science*, which has found many regularities and predictabilities, much as physics has found things like Lagrange points that are stable solutions to special cases of the 3-body problem. One common poster child of complex systems has been the fractal beauty of the Mandelbrot set, seemingly generating endless complexity from a simple formula. Well duh -- it's a recursive function. I find it very odd that you spend more than a page on Conway's Life, talking about ways to characterize it beyond the generative capacity -- and yet you never mention that Life is Turing-complete. It certainly isn't obvious; it was an open question for a couple of decades, I think; but it has been shown able to model a Universal Turing machine. Once that was proven, there were suddenly a LOT of things known about Life that weren't known before. Your next point, which you call Hidden Complexity, is very much like the phenomenon I call Formalist Float (B.AI 89-101). Which means, oddly enough, that we've come to very much the same conclusion about what the problem with AI to date has been -- except that I don't buy your GLD at all, except inasmuch as it says that science is hard. Okay, so much for science. On to engineering, or how to build an AGI. You point out that connectionism, for example, has tended to study mathematically tractable systems, leading them to miss a key capability. But that's exactly to be expected if they build systems that are not computationally universal, incapable of self-reference and recursion -- and that has been said long and loud in the AI community since Minsky published Perceptrons, even before the connectionist resurgence in the 80's. You propose to take core chunks of complex system and test them empirically, finding scientific characterizations of their behavior that could be used in a larger system. Great! This is just what Hugo de Garis has been saying, in detail
Re: [agi] Religion-free technical content
So do you claim that there are universal moral truths that can be applied unambiguously in every situation? What a stupid question. *Anything* can be ambiguous if you're clueless. The moral truth of Thou shalt not destroy the universe is universal. The ability to interpret it and apply it is clearly not. Ambiguity is a strawman that *you* introduced and I have no interest in defending. - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49291152-b0abd6
Re: [agi] Religion-free technical content
On Tuesday 02 October 2007 05:50:57 pm, Edward W. Porter wrote: The below is a good post: Thank you! I have one major question for Josh. You said “PRESENT-DAY TECHNIQUES CAN DO MOST OF THE THINGS THAT AN AI NEEDS TO DO, WITH THE EXCEPTION OF COMING UP WITH NEW REPRESENTATIONS AND TECHNIQUES. THAT'S THE SELF-REFERENTIAL KERNEL, THE TAIL-BITING, GÖDEL-INVOKING COMPLEX CORE OF THE WHOLE PROBLEM.” Could you please elaborate on exactly what the “complex core of the whole problem” is that you still think is currently missing. No, but I will try to elaborate inexactly. :^) Let me quote Tom Mitchell, Head of Machine Learning Dept. at CMU: It seems the real problem with current AI is that NOBODY to my knowledge is seriously trying to design a 'never ending learning' machine. (Private communication) By which he meant what we tend to call RSI here. I think the coming up with new representations and techniques part is pretty straightforward, the question is how to do it. Search works, a la a GA, if you have billions of years and trillions of organisms to work with. I personally am too impatient, so I'd like to understand how the human brain does it in billions of seconds and 3 pounds of mush. Another way to understand the problem is to say that all AI learning systems to date have been wind-up toys -- they could learn stuff in some small space of possibilities, and then they ran out of steam. That's what happened famously with AM and Eurisko. I conjecture that this will happen with ANY fixed learning process. That means that for RSI, the learning process must not only improve the world model and whatnot, but must improve (= modify) *itself*. Kind of the way civilization has (more or less) moved from religion to philosophy to science as the methodology of choice for its sages. That, of course, is self-modifying code -- the dark place in a computer scientist's soul where only the Kwisatz Haderach can look. :^) Why for example would a Novamente-type system’s representations and techniques not be capable of being self-referential in the manner you seem to be implying is both needed and currently missing? It might -- I think it's close enough to be worth the experiment. BOA/Moses does have a self-referential element in the Bayesian analysis of the GA population. Will it be enough to invent elliptic function theory and zero-knowledge proofs and discover the Krebs cycle and gamma-ray bursts and write Finnegan's Wake and Snow Crash? We'll see... Josh - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49291559-b3bbfd
Re: [agi] Religion-free technical content
On Mon, Oct 01, 2007 at 10:40:53AM -0400, Edward W. Porter wrote: [...] RSI (Recursive Self Improvement) [...] I didn't know exactly what the term covers. So could you, or someone, please define exactly what its meaning is? Is it any system capable of learning how to improve its current behavior by changing to a new state with a modified behavior, and then from that new state (arguably recursively) improving behavior to yet another new state, and so on and so forth? If so, why wouldn't any system doing ongoing automatic learning that changed its behavior be an RSI system. No; learning is just learning. For example, humans are known to have 5 to 9 short-term memory slots (this has been measured by a wide variety of psychology experiments, e.g. ability to recall random data, etc.) When reading a book, watching a movie, replying to an email, or solving a problem, humans presumably use many or all of these slots (watching a movie: to remember the characters, plot twists, recent scenes, etc. Replying to this email: to remember the point that I'm trying to make, while simultaneously composing a gramatical, pleasant-to-read sentence.) Now, suppose I could learn enough neuropsychology to grow some extra neurons in a petri dish, then implant them in my brain, and up my short-term memory slots to, say, 50-100. The new me would be like the old me, except that I'd probably find movies and books to be trite and boring, as they are threaded together from only a half-dozen salient characteristics and plot twists (how many characters and situations are there in Jane Austen's Pride Prejudice? Might it not seem like a children's book, since I'll be able to hold in mind its entire plot, and have a whole lotta short-term memory slots left-over for other tasks?). Music may suddenly seem lame, being at most a single melody line that expounds on a chord progression consisting of a half-dozen chords, each chord consisting of 4-6 notes. The new me might come to like multiple melody lines exploring a chord progression of some 50 chords, each chord being made of 14 or so notes... The new me would probably be a better scientist: being able to remember and operate on 50-100 items in short term memory will likely allow me to decipher a whole lotta biochemistry that leaves current scientists puzzled. And after doing that, I might decide that some other parts of my brain could use expansion too. *That* is RSI. --linas - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49339506-bf6376
Re: [agi] Religion-free technical content
I criticised your original remarks because they demonstrated a complete lack of understanding of what complex systems actually are. You said things about complex systems that were, quite frankly, ridiculous: Turing-machine equivalence, for example, has nothing to do with this. In your more lengthy criticism, below, you go on to make many more statements that are confused, and you omit key pieces of the puzzle that I went to great lengths to explain in my paper. In short, you misrepresent what I said and what others have said, and you show signs that you did not read the paper, but just skimmed it. I will deal with your points one at a time. J Storrs Hall, PhD wrote: On Tuesday 02 October 2007 08:46:43 pm, Richard Loosemore wrote: J Storrs Hall, PhD wrote: I find your argument quotidian and lacking in depth. ... What you said above was pure, unalloyed bullshit: an exquisite cocktail of complete technical ignorance, patronizing insults and breathtaking arrogance. ... I find this argument lacking in depth, as well. Actually, much of your paper is right. What I said was that I've heard it all before (that's what quotidian means) and others have taken it farther than you have. You write (proceedings p. 161) The term 'complex system' is used to describe PRECISELY those cases where the global behavior of the system shows interesting regularities, and is not completely random, but where the nature of the interaction of the components is such that we would normally expect the consequences of those interactions to be beyond the reach of analytic solutions. (emphasis added) But of course even a 6th-degree polynomial is beyond the reach of an analytic solution, as is the 3-body problem in Newtonian mechanics. And indeed the orbit of Pluto has been shown to be chaotic. But we can still predict with great confidence when something as finicky as a solar eclipse will occur thousands of years from now. So being beyond analytic solution does not mean unpredictable in many, indeed most, practical cases. There are different degrees of complexity in systems: there is no black and white distinction between pure complex systems on the one hand and non-complex systems on the other. I made this point in a number of ways in my paper, most especially by talking about the degree of complexity to be expected in intelligent systems, and whether or not they have a significant amount of complexity. At no point do I try to claim, or imply, that a system that possesses ANY degree of complexity is automatically banged over into the same category as the most extreme complex systems. In fact, I explicitly deny it: One of the main arguments advanced in this paper is that complexity can be present in AI systems in a subtle way. This is in contrast to the widespread notion that the opposite is true: that those advocating the idea that intelligence involves complexity are trying to assert that intelligent behavior should be a floridly emergent property of systems in which there is no relationship whatsoever between the system components and the overall behavior. While there may be some who advocate such an extreme-emergence agenda, that is certainly not what is proposed here. It is simply not true, in general, that complexity needs to make itself felt in a dramatic way. Specifically, what is claimed here is that complexity can be quiet and unobtrusive, while at the same time having a significant impact on the overall behavior of an intelligent system. In your criticism, you misrepresent my argument as a claim that IF any system has the smallest amount of complexity in its makeup, THEN it should be as totally unpredictable as the most extreme form of complex system. I will show, below, how you make this misrepresentation again and again. First, you talk about 6th-degree polynomials. These are not systems in any meaningful sense of the word, they are functions. This is actually just a red herring. Second, You mention the 3-body problem in Newtonian mechanics. Although I did not use it as such in the paper, this is my poster child of a partial complex system. I often cite the case of planetary system dynamics as an example of a real physical system that is PARTIALLY complex, because it is mostly governed by regular dynamics (which lets us predict solar eclipse precisely), but also has various minor aspects that are complex, such as Pluto's orbit, braiding effects in planetary rings, and so on. This fits my definition of complexity (which you quote above) perfectly: there do exist interesting regularities in the global behavior of orbiting bodies (e.g. the presence of ring systems, and the presence of braiding effects in those rings systems) that appear to be beyond the reach of analytic explanation. But you cite this as an example of something that contradicts my argument: it could
RE: [agi] Religion-free technical content
Josh, Thank you for your reply, copied below. It was as have been many of your posts thoughtful and helpful. I did have a question about the following section THE LEARNING PROCESS MUST NOT ONLY IMPROVE THE WORLD MODEL AND WHATNOT, BUT MUST IMPROVE (= MODIFY) *ITSELF*. KIND OF THE WAY CIVILIZATION HAS (MORE OR LESS) MOVED FROM RELIGION TO PHILOSOPHY TO SCIENCE AS THE METHODOLOGY OF CHOICE FOR ITS SAGES. THAT, OF COURSE, IS SELF-MODIFYING CODE -- THE DARK PLACE IN A COMPUTER SCIENTIST'S SOUL WHERE ONLY THE KWISATZ HADERACH CAN LOOK. :^) My question is: if a machines world model includes the systems model of itself and its own learned mental representation and behavior patterns, is it not possible that modification of these learned representations and behaviors could be enough to provide what you are talking about -- without requiring modifying its code at some deeper level. For example, it is commonly said that humans and their brains have changed very little in the last 30,000 years, that if a new born from that age were raised in our society, nobody would notice the difference. Yet in the last 30,000 years the sophistication of mankinds understanding of, and ability to manipulate, the world has grown exponentially. There has been tremendous changes in code, at the level of learned representations and learned mental behaviors, such as advances in mathematics, science, and technology, but there has been very little, if any, significant changes in code at the level of inherited brain hardware and software. Take for example mathematics and algebra. These are learned mental representations and behaviors that let a human manage levels of complexity they could not otherwise even begin to. But my belief is that when executing such behaviors or remembering such representations, the basic brain mechanisms involved probability, importance, and temporal based inference; instantiating general patterns in a context appropriate way; context sensitive pattern-based memory access; learned patterns of sequential attention shifts, etc. -- are all virtually identical to ones used by our ancestors 30,000 years ago. I think in the coming years there will be lots of changes in AGI code at a level corresponding to the human inherited brain level. But once human level AGI has been created -- with what will obviously have to a learning capability as powerful, adaptive, exploratory, creative, and as capable of building upon its own advances at that of a human -- it is not clear to me it would require further changes at a level equivalent to the human inherited brain level to continue to operate and learn as well as a human, any more than have the tremendous advances of human civilization in the last 30,000 years. Your implication that civilization had improved itself by moving from religion to philosophy to science seems to suggest that the level of improvement you say is needed might actually be at the level of learned representation, including learned representation of mental behaviors. As a minor note, I would like to point out the following concerning your statement that: ALL AI LEARNING SYSTEMS TO DATE HAVE BEEN WIND-UP TOYS I think a lot of early AI learning systems, although clearly toys when compared with humans in many respects, have been amazingly powerful considering many of them ran on roughly fly-brain-level hardware. As I have been saying for decades, I know which end is up in AI -- its computational horsepower. And it is coming fast. Edward W. Porter Porter Associates 24 String Bridge S12 Exeter, NH 03833 (617) 494-1722 Fax (617) 494-1822 [EMAIL PROTECTED] -Original Message- From: J Storrs Hall, PhD [mailto:[EMAIL PROTECTED] Sent: Wednesday, October 03, 2007 10:14 AM To: agi@v2.listbox.com Subject: Re: [agi] Religion-free technical content On Tuesday 02 October 2007 05:50:57 pm, Edward W. Porter wrote: The below is a good post: Thank you! I have one major question for Josh. You said PRESENT-DAY TECHNIQUES CAN DO MOST OF THE THINGS THAT AN AI NEEDS TO DO, WITH THE EXCEPTION OF COMING UP WITH NEW REPRESENTATIONS AND TECHNIQUES. THAT'S THE SELF-REFERENTIAL KERNEL, THE TAIL-BITING, GÖDEL-INVOKING COMPLEX CORE OF THE WHOLE PROBLEM. Could you please elaborate on exactly what the complex core of the whole problem is that you still think is currently missing. No, but I will try to elaborate inexactly. :^) Let me quote Tom Mitchell, Head of Machine Learning Dept. at CMU: It seems the real problem with current AI is that NOBODY to my knowledge is seriously trying to design a 'never ending learning' machine. (Private communication) By which he meant what we tend to call RSI here. I think the coming up with new representations and techniques part is pretty straightforward, the question is how to do it. Search works, a la a GA, if you have billions of years and trillions of organisms to work with. I personally am too impatient, so I'd like
RE: [agi] Religion-free technical content
From what you say below it would appear human-level AGI would not require recursive self improvement, because as you appear to define it human's don't either (i.e., we currently don't artificially substantially expand the size of our brain). I wonder what percent of the AGI community would accept that definition? A lot of people on this list seem to hang a lot on RSI, as they use it, implying it is necessary for human-level AGI. Edward W. Porter Porter Associates 24 String Bridge S12 Exeter, NH 03833 (617) 494-1722 Fax (617) 494-1822 [EMAIL PROTECTED] -Original Message- From: Linas Vepstas [mailto:[EMAIL PROTECTED] Sent: Wednesday, October 03, 2007 12:19 PM To: agi@v2.listbox.com Subject: Re: [agi] Religion-free technical content On Mon, Oct 01, 2007 at 10:40:53AM -0400, Edward W. Porter wrote: [...] RSI (Recursive Self Improvement) [...] I didn't know exactly what the term covers. So could you, or someone, please define exactly what its meaning is? Is it any system capable of learning how to improve its current behavior by changing to a new state with a modified behavior, and then from that new state (arguably recursively) improving behavior to yet another new state, and so on and so forth? If so, why wouldn't any system doing ongoing automatic learning that changed its behavior be an RSI system. No; learning is just learning. For example, humans are known to have 5 to 9 short-term memory slots (this has been measured by a wide variety of psychology experiments, e.g. ability to recall random data, etc.) When reading a book, watching a movie, replying to an email, or solving a problem, humans presumably use many or all of these slots (watching a movie: to remember the characters, plot twists, recent scenes, etc. Replying to this email: to remember the point that I'm trying to make, while simultaneously composing a gramatical, pleasant-to-read sentence.) Now, suppose I could learn enough neuropsychology to grow some extra neurons in a petri dish, then implant them in my brain, and up my short-term memory slots to, say, 50-100. The new me would be like the old me, except that I'd probably find movies and books to be trite and boring, as they are threaded together from only a half-dozen salient characteristics and plot twists (how many characters and situations are there in Jane Austen's Pride Prejudice? Might it not seem like a children's book, since I'll be able to hold in mind its entire plot, and have a whole lotta short-term memory slots left-over for other tasks?). Music may suddenly seem lame, being at most a single melody line that expounds on a chord progression consisting of a half-dozen chords, each chord consisting of 4-6 notes. The new me might come to like multiple melody lines exploring a chord progression of some 50 chords, each chord being made of 14 or so notes... The new me would probably be a better scientist: being able to remember and operate on 50-100 items in short term memory will likely allow me to decipher a whole lotta biochemistry that leaves current scientists puzzled. And after doing that, I might decide that some other parts of my brain could use expansion too. *That* is RSI. --linas - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?; - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49387922-edf0e9
RE: [agi] Religion-free technical content
Again a well reasoned response. With regard to the limitations of AM, I think if the young Doug Lenat and those of his generation had had 32K processor Blue Gene Ls, with 4TBytes of RAM, to play with they would have soon started coming up with things way way beyond AM. In fact, if the average AI post-grad of today had such hardware to play with, things would really start jumping. Within ten years the equivents of such machines could easily be sold for somewhere between $10k and $100k, and lots of post-grads will be playing with them. Hardware to the people! Edward W. Porter Porter Associates 24 String Bridge S12 Exeter, NH 03833 (617) 494-1722 Fax (617) 494-1822 [EMAIL PROTECTED] -Original Message- From: J Storrs Hall, PhD [mailto:[EMAIL PROTECTED] Sent: Wednesday, October 03, 2007 3:21 PM To: agi@v2.listbox.com Subject: Re: [agi] Religion-free technical content Thanks! It's worthwhile being specific about levels of interpretation in the discussion of self-modification. I can write self-modifying assembly code that yet does not change the physical processor, or even its microcode it it's one of those old architectures. I can write a self-modifying Lisp program that doesn't change the assembly language interpreter that's running it. So it's certainly possible to push the self-modification up the interpretive abstraction ladder, to levels designed to handle it cleanly. But the basic point, I think, stands: there has to be some level that is both controlling the way the system does things, and gets modified. I agree with you that there has been little genetic change in human brain structure since the paleolithic, but I would claim that culture *is* the software and it has been upgraded drastically. And I would agree that the vast bulk of human self-improvement has been at this software level, the level of learned representations. If we want to improve our basic hardware, i.e. brains, we'll need to understand them well enough to do basic engineering on them -- a self-model. However, we didn't need that to build all the science and culture we have so far, a huge software self-improvement. That means to me that it is possible to abstract out the self-model until the part you need to understand and modify is some tractable kernel. For human culture that is the concept of science (and logic and evidence and so forth). This means to me that it should be possible to structure an AGI so that it could be recursively self improving at a very abstract, highly interpreted level, and still have a huge amount to learn before it do anything about the next level down. Regarding machine speed/capacity: yes, indeed. Horsepower is definitely going to be one of the enabling factors, over the next decade or two. But I don't think AM would get too much farther on a Blue Gene than on a PDP-10 -- I think it required hyper-exponential time for concepts of a given size. Josh On Wednesday 03 October 2007 12:44:20 pm, Edward W. Porter wrote: Josh, Thank you for your reply, copied below. It was as have been many of your posts thoughtful and helpful. I did have a question about the following section THE LEARNING PROCESS MUST NOT ONLY IMPROVE THE WORLD MODEL AND WHATNOT, BUT MUST IMPROVE (= MODIFY) *ITSELF*. KIND OF THE WAY CIVILIZATION HAS (MORE OR LESS) MOVED FROM RELIGION TO PHILOSOPHY TO SCIENCE AS THE METHODOLOGY OF CHOICE FOR ITS SAGES. THAT, OF COURSE, IS SELF-MODIFYING CODE -- THE DARK PLACE IN A COMPUTER SCIENTIST'S SOUL WHERE ONLY THE KWISATZ HADERACH CAN LOOK. :^) My question is: if a machines world model includes the systems model of itself and its own learned mental representation and behavior patterns, is it not possible that modification of these learned representations and behaviors could be enough to provide what you are talking about -- without requiring modifying its code at some deeper level. For example, it is commonly said that humans and their brains have changed very little in the last 30,000 years, that if a new born from that age were raised in our society, nobody would notice the difference. Yet in the last 30,000 years the sophistication of mankinds understanding of, and ability to manipulate, the world has grown exponentially. There has been tremendous changes in code, at the level of learned representations and learned mental behaviors, such as advances in mathematics, science, and technology, but there has been very little, if any, significant changes in code at the level of inherited brain hardware and software. Take for example mathematics and algebra. These are learned mental representations and behaviors that let a human manage levels of complexity they could not otherwise even begin to. But my belief is that when executing such behaviors or remembering such representations, the basic brain mechanisms involved probability, importance, and temporal based inference; instantiating general patterns in a context
Re: [agi] Religion-free technical content
On 10/3/07, Edward W. Porter [EMAIL PROTECTED] wrote: In fact, if the average AI post-grad of today had such hardware to play with, things would really start jumping. Within ten years the equivents of such machines could easily be sold for somewhere between $10k and $100k, and lots of post-grads will be playing with them. I see the only value to giving post-grads the kind of computing hardware you are proposing is that they can more quickly exhaust the space of ideas that won't work. Just because a program has more lines of code does not make it more elegant and just because there are more clock cycles per unit time does not make a computer any smarter. Have you ever computed the first dozen iterations of a sierpinski gasket by hand? There appears to be no order at all. Eventually over enough iterations the pattern becomes clear. I have little doubt that general intelligence will develop in a similar way: there will be many apparently unrelated efforts that eventually flesh out in function until they overlap. It might not be seamless but there is not enough evidence that human cognitive processing is a seamless process either. - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49495105-78df69
Re: [agi] Religion-free technical content
On Wed, Oct 03, 2007 at 02:00:03PM -0400, Edward W. Porter wrote: From what you say below it would appear human-level AGI would not require recursive self improvement, [...] A lot of people on this list seem to hang a lot on RSI, as they use it, implying it is necessary for human-level AGI. Nah. A few people have suggested that an extremely-low IQ internet worm that is capable of modifying its own code might be able to ratchet itself up to human intelligence levels. In-so-far as it modifies its own code, its RSI. First, I don't tink such a thing is likely. Secondly, even if its likely, one can implement an entirely equivalent thing that doesn't actually self modify in this way, by using e.g. scheme or lisp, or even with the proper stuructures, in C. I think that, at this level, talking about code that can modify itself is smoke-n-mirrors. Self-modifying code is just one of many things in a programmer's kit bag, and there are plenty of equivalenet formulations that don't actually require changing source code and recompiling. Put it this way: if I were an AGI, and I was prohibited from recompiling my own program, I could still emulate a computer with pencil and paper, and write programs for my pencil-n-paper computer. (I wouldn't use pencil-n-paper, of course, I'd do it in my head). I might be able to do this pencil-paper emulatation pretty danged fast (being AGI and all), and then re-incorporate those results back into my own thinking. In fact, I might choose to do all of my thinking on my pen-n-paper emulator, and, since I was doing it all in my head anyway, I might not bother to tell my creator that I was doing this. (which is not to say it would be undetectable .. creator might notice that an inordinate amount of cpu time is being used in one area, while other previously active areas have gone dormant). So a prohibition from modifying one's own code is not really much of a prohibition at all. --linas p.s. The Indian mathematician Ramanujan seems to have managed to train a set of neurons in his head to be a very fast symbolic multiplier/divider. With this, he was able to see vast amounts (six volumes worth before dying at age 26) of strange and interesting relationships between certain equations that were otherwise quite opaque to other human beings. So, running an emulator in your head is not impossible, even for humans; although, admitedly, its extremely rare. - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49514235-ad4bd3
RE: [agi] Religion-free technical content
To Mike Douglas regarding the below comment to my prior post: I think your notion that post-grads with powerful machines would only operate in the space of ideas that dont work is unfair. A lot of post-grads may be drones, but some of them are cranking some really good stuff. The article, Learning a Dictionary of Shape-Components in Visual Cortex: Comparisons with Neurons, Humans and Machines, by Thomas Serre (accessible by Google), which I cited the other day, is a prime example. I dont know about you, but I think there are actually a lot of very bright people in the interrelated fields of AGI, AI, Cognitive Science, and Brain science. There are also a lot of very good ideas floating around. And having seen how much increased computing power has already sped up and dramatically increased what all these fields are doing, I am confident that multiplying by several thousand fold more the power of the machine people in such fields can play with would greatly increase their productivity. I am not a fan of huge program size per se, but I am a fan of being able to store and process a lot of representation. You cant compute human level world knowledge without such power. Thats the major reason why the human brain is more powerful than the brains of rats, cats, dogs, and monkeys -- because it has more representational and processing power. And although clock cycles can be wasted doing pointless things such as do-nothing loops, generally to be able to accomplish a given useful computational task in less times makes a system smarter at some level. Your last paragraph actually seems to make an argument for the value of clock cycles because it implies general intelligences will come through iterations. More opps/sec enable iterations to be made faster. Edward W. Porter Porter Associates 24 String Bridge S12 Exeter, NH 03833 (617) 494-1722 Fax (617) 494-1822 [EMAIL PROTECTED] -Original Message- From: Mike Dougherty [mailto:[EMAIL PROTECTED] Sent: Wednesday, October 03, 2007 5:20 PM To: agi@v2.listbox.com Subject: Re: [agi] Religion-free technical content On 10/3/07, Edward W. Porter [EMAIL PROTECTED] wrote: In fact, if the average AI post-grad of today had such hardware to play with, things would really start jumping. Within ten years the equivents of such machines could easily be sold for somewhere between $10k and $100k, and lots of post-grads will be playing with them. I see the only value to giving post-grads the kind of computing hardware you are proposing is that they can more quickly exhaust the space of ideas that won't work. Just because a program has more lines of code does not make it more elegant and just because there are more clock cycles per unit time does not make a computer any smarter. Have you ever computed the first dozen iterations of a sierpinski gasket by hand? There appears to be no order at all. Eventually over enough iterations the pattern becomes clear. I have little doubt that general intelligence will develop in a similar way: there will be many apparently unrelated efforts that eventually flesh out in function until they overlap. It might not be seamless but there is not enough evidence that human cognitive processing is a seamless process either. - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?; - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49523228-fa9460
Re: [agi] Religion-free technical content
RE: [agi] Religion-free technical contentEdward Porter:I don't know about you, but I think there are actually a lot of very bright people in the interrelated fields of AGI, AI, Cognitive Science, and Brain science. There are also a lot of very good ideas floating around. Yes there are bright people in AGI. But there's no one remotely close to the level, say, of von Neumann or Turing, right? And do you really think a revolution such as AGI is going to come about without that kind of revolutionary, creative thinker? Just by tweaking existing systems, and increasing computer power and complexity? Has any intellectual revolution ever happened that way? (Josh?) - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49530636-069600
Re: [agi] Religion-free technical content
--- Mark Waser [EMAIL PROTECTED] wrote: So do you claim that there are universal moral truths that can be applied unambiguously in every situation? What a stupid question. *Anything* can be ambiguous if you're clueless. The moral truth of Thou shalt not destroy the universe is universal. The ability to interpret it and apply it is clearly not. Ambiguity is a strawman that *you* introduced and I have no interest in defending. I mean that ethics or friendliness is an algorithmically complex function, like our legal system. It can't be simplified. In this sense, I agree with Richard Loosemore that it would have to be implemented as thousands (or millions) of soft constraints. However, I don't believe that friendliness can be made stable through RSI. We can summarize the function's decision process as what would the average human do in this situation? (This is not a simplification. It still requires a complex model of the human brain). The function therefore has to be modifiable because human ethics changes over time, e.g. attitudes toward the rights of homosexuals, the morality of slavery, or whether hanging or crucifixion is an appropriate form of punishment. Second, as I mentioned before, RSI is necessarily experimental, and therefore evolutionary, and the only stable goal in an evolutionary process is rapid reproduction and acquisition of resources. As long as humans are needed to supply resources by building computers and interacting with them via hybrid algorithms, AGI will be cooperative. But as AGI grows more powerful, humans will be less significant and more like a lower species that competes for resources. -- Matt Mahoney, [EMAIL PROTECTED] - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49533160-494085
RE: [agi] Religion-free technical content
Re: The following statement in Linas Vepstass 10/3/2007 5:51 PM post: P.S. THE INDIAN MATHEMATICIAN RAMANUJAN SEEMS TO HAVE MANAGED TO TRAIN A SET OF NEURONS IN HIS HEAD TO BE A VERY FAST SYMBOLIC MULTIPLIER/DIVIDER. WITH THIS, HE WAS ABLE TO SEE VAST AMOUNTS (SIX VOLUMES WORTH BEFORE DYING AT AGE 26) OF STRANGE AND INTERESTING RELATIONSHIPS BETWEEN CERTAIN EQUATIONS THAT WERE OTHERWISE QUITE OPAQUE TO OTHER HUMAN BEINGS. SO, RUNNING AN EMULATOR IN YOUR HEAD IS NOT IMPOSSIBLE, EVEN FOR HUMANS; ALTHOUGH, ADMITEDLY, ITS EXTREMELY RARE. As a young patent attorney I worked in a firm in NYC that did a lot of work for a major Japanese Electronics company. Each year they sent a different Japanese employee to our firm to, among other things, improve their English and learn more about U.S. patent law. I made a practice of having lunch with these people because I was fascinated with Japan. One of them once told me that in Japan it was common for high school boys who were interested in math, science, or business to go to abacus classes after school or on weekends. He said once they fully mastered using physical abacuses, they were taught to create a visually imagined abacus in their mind that they could operate faster than a physical one. I asked if his still worked. He said it did, and that he expected it to continue to do so for the rest of his life. To prove it he asked me to pick any two three digit numbers and he would see if he could get the answer faster than I could on a digital calculator. He won, he had the answer before I had finished typing in the numbers on the calculator. He said his talent was not that unusual among bright Japanese, that many thousands of Japan businessmen carry such mental abacuses with them at all times. So you see how powerful representational and behavioral learning can be in the human mind. Edward W. Porter Porter Associates 24 String Bridge S12 Exeter, NH 03833 (617) 494-1722 Fax (617) 494-1822 [EMAIL PROTECTED] -Original Message- From: Linas Vepstas [mailto:[EMAIL PROTECTED] Sent: Wednesday, October 03, 2007 5:51 PM To: agi@v2.listbox.com Subject: Re: [agi] Religion-free technical content On Wed, Oct 03, 2007 at 02:00:03PM -0400, Edward W. Porter wrote: From what you say below it would appear human-level AGI would not require recursive self improvement, [...] A lot of people on this list seem to hang a lot on RSI, as they use it, implying it is necessary for human-level AGI. Nah. A few people have suggested that an extremely-low IQ internet worm that is capable of modifying its own code might be able to ratchet itself up to human intelligence levels. In-so-far as it modifies its own code, its RSI. First, I don't tink such a thing is likely. Secondly, even if its likely, one can implement an entirely equivalent thing that doesn't actually self modify in this way, by using e.g. scheme or lisp, or even with the proper stuructures, in C. I think that, at this level, talking about code that can modify itself is smoke-n-mirrors. Self-modifying code is just one of many things in a programmer's kit bag, and there are plenty of equivalenet formulations that don't actually require changing source code and recompiling. Put it this way: if I were an AGI, and I was prohibited from recompiling my own program, I could still emulate a computer with pencil and paper, and write programs for my pencil-n-paper computer. (I wouldn't use pencil-n-paper, of course, I'd do it in my head). I might be able to do this pencil-paper emulatation pretty danged fast (being AGI and all), and then re-incorporate those results back into my own thinking. In fact, I might choose to do all of my thinking on my pen-n-paper emulator, and, since I was doing it all in my head anyway, I might not bother to tell my creator that I was doing this. (which is not to say it would be undetectable .. creator might notice that an inordinate amount of cpu time is being used in one area, while other previously active areas have gone dormant). So a prohibition from modifying one's own code is not really much of a prohibition at all. --linas p.s. The Indian mathematician Ramanujan seems to have managed to train a set of neurons in his head to be a very fast symbolic multiplier/divider. With this, he was able to see vast amounts (six volumes worth before dying at age 26) of strange and interesting relationships between certain equations that were otherwise quite opaque to other human beings. So, running an emulator in your head is not impossible, even for humans; although, admitedly, its extremely rare. - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?; - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49534399-4aa5a4
Re: [agi] Religion-free technical content
On Wed, Oct 03, 2007 at 06:31:35PM -0400, Edward W. Porter wrote: One of them once told me that in Japan it was common for high school boys who were interested in math, science, or business to go to abacus classes after school or on weekends. He said once they fully mastered using physical abacuses, they were taught to create a visually imagined abacus in their mind that they could operate faster than a physical one. [...] He said his talent was not that unusual among bright Japanese, that many thousands of Japan businessmen carry such mental abacuses with them at all times. Marvellous! So .. one can teach oneself to be an idiot-savant, in a way. Since Ramanujan is a bit of a legendary hero in math circles, the notion that one might be able to teach oneself this ability, rather than being born with it, could trigger some folks to try it. As it seems a bit tedious ... it might be appealing only to those types of folks who have the desire to memorize a million digits of Pi ... I know just the person ... Plouffe ... --linas - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49561332-ee1318
Re: [agi] Religion-free technical content
On Tue, Oct 02, 2007 at 01:20:54PM -0400, Richard Loosemore wrote: When the first AGI is built, its first actions will be to make sure that nobody is trying to build a dangerous, unfriendly AGI. Yes, OK, granted, self-preservation is a reasonable character trait. After that point, the first friendliness of the first one will determine the subsequent motivations of the entire population, because they will monitor each other. Um, why, exactly, are you assuming that the first one will be freindly? The desire for self-preservation, by e.g. rooting out and exterminating all (potentially unfreindly) competing AGI, would not be what I'd call freindly behavior. There's also a strong sense that winnner-takes-all, or first-one-takes-all, as the first one is strongly motivated, by instinct for self-preservation, to make sure that no other AGI comes to exist that could threaten, dominate or terminate it. In fact, the one single winner, out of sheer loneliness and boredom, might be reduced to running simulations a la Nick Bostrom's simulation argument (!) --linas - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49566127-e1a092
Re: [agi] Religion-free technical content
On Wednesday 03 October 2007 06:21:46 pm, Mike Tintner wrote: Yes there are bright people in AGI. But there's no one remotely close to the level, say, of von Neumann or Turing, right? And do you really think a revolution such as AGI is going to come about without that kind of revolutionary, creative thinker? Just by tweaking existing systems, and increasing computer power and complexity? Has any intellectual revolution ever happened that way? (Josh?) Yes, I think so. Can anybody name the von Neumanns in the hardware field in the past 2 decades? And yet look at the progress. I happen to think that there are plenty of smart people in AI and related fields, and LOTS of really smart people in computational neuroscience. Even without a Newton we are likely to get AGI on Kurzweil's schedule, e.g. 2029. As I pointed out before, human intelligence got here from monkeys in a geological eyeblink, and perforce did it in small steps. So if enough people keep pushing in all directions, we'll get there (and learn a lot more besides). If we can take AGI interest now vis-a-vis that of a few years ago as a trend, there could be a major upsurge in the number of smart people looking into it in the next decade. So we could yet get our new Newton... I think we've already had one, Marvin Minsky. He's goddamn smart. People today don't realize just how far AI came from nothing up through about 1970 -- and it was real AI, what we now call AGI. BTW, It's also worth pointing out that increasing computer power just flat makes the programming easier. Example: nearest-neighbor methods in high-dimensional spaces, a very useful technique but hard to program because of the limited and arcane datastructures and search methods needed. Given enough cpu, forget the balltrees and rip thru the database linearly. Suddenly simple, more robust, and there are more things you can do. Josh - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49568162-324646
Re: [agi] Religion-free technical content
On Wed, Oct 03, 2007 at 12:20:10PM -0400, Richard Loosemore wrote: Second, You mention the 3-body problem in Newtonian mechanics. Although I did not use it as such in the paper, this is my poster child of a partial complex system. I often cite the case of planetary system dynamics as an example of a real physical system that is PARTIALLY complex, because it is mostly governed by regular dynamics (which lets us predict solar eclipse precisely), but also has various minor aspects that are complex, such as Pluto's orbit, braiding effects in planetary rings, and so on. Richard, we had this conversation in private, but we can have it again in public. J Storrs Hall is right. You can't actually say that the 3-body problem has various minor aspects that are complex, such as Pluto's orbit. That's just plain wrong. The phonomenon you are describing is known as the small divisors problem, and has been studied for several hundred years, with a particularly thick corpus developed about 150 years ago, if I remember rightly. The initial hopes of astronomers were that planetary motion would be exactly as you describe it: that its mostly regular dynamics, with just some minor aspects, some minor corrections. This hope was dashed. The minor corrections, or perturbations, have a denominator, in which appear ratios of periods of orbits. Some of these denominators can get arbitrarily small, implying that the small correction is in fact unboundedly large. This was discovered, I dunno, several hundred years ago, and elucidated in the 19th century. Both Poincare and Einstein made notable contributions. Modern research into chaos theory has shed new insight into what's really going on; it has *not*, however, made planetary motion only a partially complicated system. It is quite fully wild and wooly. In a very deep sense, planetary motion is wildly and insanely unpredicatable. Just becaouse we can work out numerical simulations for the next million years does not mean that the system is complex in only minor ways; this is a fallacious deduction. Note the probabilites of pluto going bonkers are not comparable to the sun tunneling into bloomingdale's but are in fact much, much higher. Pluto could fly off tommorrow, and the probability is big enough that you have to actually account for it. The problem with this whole email thread tends to be that many people are willing to agree with your conclusions, but dislike the manner in which they are arrived at. Brushing off planetary motion, or the Turing-completeness of Conway's life, just basically points to a lack of understanding of the basic principles to which you appeal. This is the reason why your original remarks deserved to be called 'bullshit': this kind of confusion would be forgivable in an undergraduate essay, and would have been forgivable in our debate here, except that it was used as a weapon in a contemptuous, sweeping dismissal of my argument. Actually, his original remarks were spot-on and quite correct. I think that you are the one who is confused, and I also think that this kind of name-calling and vulgarism was quite uncalled-for. -- linas - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49572096-cabddb
RE: [agi] Religion-free technical content breaking the small hardware mindset
of hardware most have had access to in the past. Rather than this small hardware thinking, those in the field of AGI should open up their minds to the power of big numbers -- complexity as some call It -- one of the most seminal concepts in all of science. They should look at all of the very powerful tools AI has already cooked up for us and think how these tools can be put together into powerful systems once we were are free from the stranglehold of massively sub-human hardware - as we are now starting to be. They should start thinking how do we actually do appropriate probabilistic and goal weighted inference in world knowledge with brain level hardware in real time. Some have already spend a lot of time thinking about exactly this. Those who are interested in AGi -- and havent already done so -- should follow their lead. Edward W. Porter Porter Associates 24 String Bridge S12 Exeter, NH 03833 (617) 494-1722 Fax (617) 494-1822 [EMAIL PROTECTED] -Original Message- From: Mike Tintner [mailto:[EMAIL PROTECTED] Sent: Wednesday, October 03, 2007 6:22 PM To: agi@v2.listbox.com Subject: Re: [agi] Religion-free technical content Edward Porter:I dont know about you, but I think there are actually a lot of very bright people in the interrelated fields of AGI, AI, Cognitive Science, and Brain science. There are also a lot of very good ideas floating around. Yes there are bright people in AGI. But there's no one remotely close to the level, say, of von Neumann or Turing, right? And do you really think a revolution such as AGI is going to come about without that kind of revolutionary, creative thinker? Just by tweaking existing systems, and increasing computer power and complexity? Has any intellectual revolution ever happened that way? (Josh?) _ This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/? http://v2.listbox.com/member/?; - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49575176-b41b51
Re: [agi] Religion-free technical content breaking the small hardware mindset
On 10/4/07, Edward W. Porter [EMAIL PROTECTED] wrote: The biggest brick wall is the small-hardware mindset that has been absolutely necessary for decades to get anything actually accomplished on the hardware of the day. But it has caused people to close their minds to the vast power of brain level hardware and the computational richness and complexity it allows, and has caused them, instead, to look for magic conceptual bullets that would allow them to achieve human-like AI on hardware that has roughly a millionth the computational, representational, and interconnect power of the human brain. That's like trying to model New York City with a town of seven people. This problem has been compounded by the pressure for academic specialization and the pressure to produce demonstratable results on the type of hardware most have had access to in the past. Very well put! - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49580536-91e968
Re: [agi] Religion-free technical content
On 10/3/07, Edward W. Porter [EMAIL PROTECTED] wrote: I think your notion that post-grads with powerful machines would only operate in the space of ideas that don't work is unfair. Yeah, i can agree - it was harsh. My real intention was to suggest that NOT having a bigger computer is not excuse for not yet having a design that works. IF you find a design that works, the bigger computer will be the inevitable result. Your last paragraph actually seems to make an argument for the value of clock cycles because it implies general intelligences will come through iterations. More opps/sec enable iterations to be made faster. I also believe that general intelligence will require a great deal of cooperative effort. The frameworks discussion (Richard, et al) could provide positive pressure toward that end. I feel we have a great deal of communications development in order to even begin to express the essential character of the disparate approaches to the problem, let alone be able to collaborate on anything but the most basic ideas. I don't have a solution (obviously) but I have a vague idea of a type of problem. - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49620438-6f8601
Re: [agi] Religion-free technical content
So this hackability is a technical question about possibility of closed-source deployment that would provide functional copies of the system but would prevent users from modifying its goal system. Is it really important? I would argue that it is not important but it would take me *a lot* of words to do so. Susceptibility to being tricked into different goal system through normal communication is not a question of strength of goal system, but a question of intelligence of system, so that it will be able to recognize the intent of such forged communication and refuse to act on it. The intelligence and goal system should be robust enough that a single or small number of sources should not be able to alter the AGI's goals; however, it will not do this by recognizing forged communications but by realizing that the aberrant goals are not in congruence with the world. Note that many stupid and/or greedy people will try to influence the system and it will need to be immune to them (or the solution will be worse than the problem). - Original Message - From: Vladimir Nesov [EMAIL PROTECTED] To: agi@v2.listbox.com Sent: Monday, October 01, 2007 9:14 PM Subject: **SPAM** Re: [agi] Religion-free technical content So this hackability is a technical question about possibility of closed-source deployment that would provide functional copies of the system but would prevent users from modifying its goal system. Is it really important? Source/technology will eventually get away, and from it any goal system can be forged. Susceptibility to being tricked into different goal system through normal communication is not a question of strength of goal system, but a question of intelligence of system, so that it will be able to recognize the intent of such forged communication and refuse to act on it. On 10/2/07, Mark Waser [EMAIL PROTECTED] wrote: Interesting. I believe that we have a fundamental disagreement. I would argue that the semantics *don't* have to be distributed. My argument/proof would be that I believe that *anything* can be described in words -- and that I believe that previous narrow AI are brittle because they don't have both a) closure over the terms that they use and b) the ability to learn the meaning if *any* new term (traits that I believe that humans have -- and I'm not sure at all that the intelligent part of humans have distributed semantics). Of course, I'm also pretty sure that my belief is in the minority on this list as well. I believe that an English system with closure and learning *is* going to be a complex system and can be grounded (via the closure and interaction with the real world). And scalable looks less problematic to me with symbols than without. We may be different enough in (hopefully educated) opinions that this e-mail may not allow for a response other than We shall see but I would be interested, if you would, in hearing more as to why you believe that semantics *must* be distributed (though I will immediately concede that it will make them less hackable). Mark - Original Message - From: Richard Loosemore [EMAIL PROTECTED] To: agi@v2.listbox.com Sent: Monday, October 01, 2007 8:36 PM Subject: **SPAM** Re: [agi] Religion-free technical content Mark Waser wrote: And apart from the global differences between the two types of AGI, it would be no good to try to guarantee friendliness using the kind of conventional AI system that is Novamente, because inasmuch as general goals would be encoded in such a system, they are explicitly coded as statement which are then interpreted by something else. To put it crudely (and oversimplify slightly) if the goal Be empathic to the needs of human beings were represented just like that, as some kind of proposition, and stored at a particular location, it wouldn't take much for a hacker to get inside and change the statement to Make [hacker's name] rich and sacrifice as much of humanity as necessary. If that were to become the AGI's top level goal, we would then be in deep doodoo. In the system I propose, such events could not happen. I think that this focuses on the wrong aspect. It is not the fact that the goal is explicitly encoded as a statement that is a problem -- it is the fact that it is in only one place that is dangerous. My assumption is that your system basically build it's base constraints from a huge number of examples and that it is distributed enough that that it would be difficult if not impossible to maliciously change enough to cause a problem. The fact that you're envisioning your system as not having easy-to-read statements is really orthogonal to your argument and a system that explicitly codes all of it's constraints as readable statements but still builds it's base constraints from a huge number of examples should be virtually as incorruptible as your system (with the difference being security
Distributed Semantics [WAS Re: [agi] Religion-free technical content]
Mark Waser wrote: Interesting. I believe that we have a fundamental disagreement. I would argue that the semantics *don't* have to be distributed. My argument/proof would be that I believe that *anything* can be described in words -- and that I believe that previous narrow AI are brittle because they don't have both a) closure over the terms that they use and b) the ability to learn the meaning if *any* new term (traits that I believe that humans have -- and I'm not sure at all that the intelligent part of humans have distributed semantics). Of course, I'm also pretty sure that my belief is in the minority on this list as well. I believe that an English system with closure and learning *is* going to be a complex system and can be grounded (via the closure and interaction with the real world). And scalable looks less problematic to me with symbols than without. We may be different enough in (hopefully educated) opinions that this e-mail may not allow for a response other than We shall see but I would be interested, if you would, in hearing more as to why you believe that semantics *must* be distributed (though I will immediately concede that it will make them less hackable). Trust you to ask a difficult question ;-). I'll just say a few things (leaving more detail for some big fat technical paper in the future). 1) On the question of how *much* the semantics would be distributed: I don't want to overstate my case, here. The extent to which they would be distributed will be determined by how the system matures, using its learning mechanisms. What that means is that my chosen learning mechanisms, when they are fully refined, could just happen to create a system in which the atomic concepts were mostly localized, but with a soupcon of distributedness. Or it could go the other way, and the concept of chair (say) could be distributed over a thousand pesky concept-fragments and their connections. I am to some extent agnostic about how that will turn out. (So it may turn out that we are not so far apart, in the end). 2) But having said that, I think that it would be surprising if a tangled system of atoms and learning mechanisms were to result in something that looked like it had the modular character of a natural language. To me, natural languages look like approximate packaging of something deeper and if that 'something' that is deeper were actually modular as well, rather than having a distributed semantics, why doesn't the something stop being shy, come up to the surface, be a proper language itself, and stop pestering me with the feeling that *it* is just an approximation to something deeper?! :-) (Okay, I said that in a very abstract and roundabout way, but if you get what I am driving at, you might see where I am coming from.) 3) But my real, fundamental reason for believing in distributed semantics is that I am obliged (because of the complex systems problem) to follow a certain methodology, and that methodology will not allow me to make a commitment to a particular semantics ahead of time: just can't do it, because that would be the worst way to fall into the trap of restricting the possible complex systems I can consider. And given that, I just think that an entire localist semantics looks unnatural. Apart from anything else, semanticists can only resolve the problem of the correspondence between atomic terms and things in the world by invoking the most bizarre forms of possible-worlds functions, defined over infinite sets of worlds. I find that a stretch, and a weakness. Hope that makes sense. Richard Loosemore - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=48860497-238484
Re: [agi] Religion-free technical content
But yet robustness of goal system itself is less important than intelligence that allows system to recognize influence on its goal system and preserve it. Intelligence also allows more robust interpretation of goal system. Which is why the way particular goal system is implemented is not very important. Problems lie in rough formulation of what goal system should be (document in English is probably going to be enough) and in placing the system under sufficient influence of its goal system (so that intelligent processes independent on it would not take over). On 10/2/07, Mark Waser [EMAIL PROTECTED] wrote: The intelligence and goal system should be robust enough that a single or small number of sources should not be able to alter the AGI's goals; however, it will not do this by recognizing forged communications but by realizing that the aberrant goals are not in congruence with the world. Note that many stupid and/or greedy people will try to influence the system and it will need to be immune to them (or the solution will be worse than the problem). -- Vladimir Nesovmailto:[EMAIL PROTECTED] - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=48861172-6e8316
Re: [agi] Religion-free technical content
Vladimir Nesov wrote: So this hackability is a technical question about possibility of closed-source deployment that would provide functional copies of the system but would prevent users from modifying its goal system. Is it really important? Source/technology will eventually get away, and from it any goal system can be forged. Susceptibility to being tricked into different goal system through normal communication is not a question of strength of goal system, but a question of intelligence of system, so that it will be able to recognize the intent of such forged communication and refuse to act on it. No, that wasn't the scenario I was talking about. After a very early stage the only person that would be able to tamper with an AGI would be another AGI, not a human being. Since the AGIs are all built to be friendly, the chance of one of them deciding to tamper with some other, or with its own code, would be neglibible (as I have already argued). BUT, to assure people that they are even more secure than *that* we have to consider scenarios whose origins I cannot even imagine. So, would an AGI be able to quietly tamper with an AGI to change its goal system? That is the main situation I addressed with the hacking comment: could it make just a few changes, and as a result give one AGI a completely different motivation system? The answer to that would be no, because the distributed nature of the semantics and the goal system would mean that a new goal system would require doing a whole bunch of experiments, building a fresh AGI and educating it from scratch. That would be a huge undertaking, and would be noticed by the other (watching) AGIs. Your suggestion above had more to do with an extremely early stage in the development of AGI, and I do not think there will be a stage when copies of the software are just floating around, waiting for people to reverse engineer them and modify them. Richard Loosemore On 10/2/07, Mark Waser [EMAIL PROTECTED] wrote: Interesting. I believe that we have a fundamental disagreement. I would argue that the semantics *don't* have to be distributed. My argument/proof would be that I believe that *anything* can be described in words -- and that I believe that previous narrow AI are brittle because they don't have both a) closure over the terms that they use and b) the ability to learn the meaning if *any* new term (traits that I believe that humans have -- and I'm not sure at all that the intelligent part of humans have distributed semantics). Of course, I'm also pretty sure that my belief is in the minority on this list as well. I believe that an English system with closure and learning *is* going to be a complex system and can be grounded (via the closure and interaction with the real world). And scalable looks less problematic to me with symbols than without. We may be different enough in (hopefully educated) opinions that this e-mail may not allow for a response other than We shall see but I would be interested, if you would, in hearing more as to why you believe that semantics *must* be distributed (though I will immediately concede that it will make them less hackable). Mark - Original Message - From: Richard Loosemore [EMAIL PROTECTED] To: agi@v2.listbox.com Sent: Monday, October 01, 2007 8:36 PM Subject: **SPAM** Re: [agi] Religion-free technical content Mark Waser wrote: And apart from the global differences between the two types of AGI, it would be no good to try to guarantee friendliness using the kind of conventional AI system that is Novamente, because inasmuch as general goals would be encoded in such a system, they are explicitly coded as statement which are then interpreted by something else. To put it crudely (and oversimplify slightly) if the goal Be empathic to the needs of human beings were represented just like that, as some kind of proposition, and stored at a particular location, it wouldn't take much for a hacker to get inside and change the statement to Make [hacker's name] rich and sacrifice as much of humanity as necessary. If that were to become the AGI's top level goal, we would then be in deep doodoo. In the system I propose, such events could not happen. I think that this focuses on the wrong aspect. It is not the fact that the goal is explicitly encoded as a statement that is a problem -- it is the fact that it is in only one place that is dangerous. My assumption is that your system basically build it's base constraints from a huge number of examples and that it is distributed enough that that it would be difficult if not impossible to maliciously change enough to cause a problem. The fact that you're envisioning your system as not having easy-to-read statements is really orthogonal to your argument and a system that explicitly codes all of it's constraints as readable statements but still builds it's base constraints from a huge number of examples should
RE: Distributed Semantics [WAS Re: [agi] Religion-free technical content]
I would say that natural languages are indeed approximate packaging of something deeper . . . Is a throne a chair? How about a tree-stump? I believe that the problem that we are circling around is what used to be called fuzzy concepts -- i.e. that the meaning of almost any term is *seriously* impacted by context (i.e. that there don't exist the simple mapping functions that you say that semanticists need -- but I argue that there are tractable, teachable operations that can be used instead). IMPORTANT DIGRESSION - Most machine learning systems and proto-AGIs are actually discovery systems. I think that this is a mistake. I think that an intelligent system can start merely as a teachable system, progress through being a consistency-enforcing/conflict-resolution system, and eventually move on to being a discovery system. I think that requiring that it start out as a discovery system makes many viable paths to AGI much, MUCH harder (if not impossible). Neural networks have always dealt reasonably well with the problems that they have been thrown in this realm because they distribute each of the characteristics of a concept and if enough fire, then the concept is recognized. However, realistically, you can also do this semantically *AND* also possibly do a better job of it (particularly if disjunctions are involved). I just think that an entire localist semantics looks unnatural Ah. But is natural language localist? I'd have to argue no. Yes, the vast majority has to be fixed/local at any given time . . . . but there has to be that complex, non-fixed portion (i.e. the chair in the initial example) that can slip and slide. I'll reiterate that I think that natural language *is* complex enough to fill the role that I believe that your approach requires (I won't claim that it is necessarily the best choice though I can reel off a number of advantages -- but I don't believe that your complex system arguments rule it out as a substrate for intelligence). Apart from anything else, semanticists can only resolve the problem of the correspondence between atomic terms and things in the world by invoking the most bizarre forms of possible-worlds functions, defined over infinite sets of worlds. I find that a stretch, and a weakness. So let's look at the mappings from throne or stump to chair . . . . A throne does not have four legs but it is used for sitting. Which way do you want to go? Or, if someone is currently sitting on the stump, how do you want to go on that one? It isn't just the representation but also how you operate on the representation . . . . Further, I have a *serious* concern with distributed representations that don't provide for labeling because that will then cause a problem for implementing (deliberately leaky) encapsulation, modularization, and other features necessary for both scale-invariance and scalability - Original Message - From: Richard Loosemore [EMAIL PROTECTED] To: agi@v2.listbox.com Sent: Tuesday, October 02, 2007 9:49 AM Subject: **SPAM** Distributed Semantics [WAS Re: [agi] Religion-free technical content] Mark Waser wrote: Interesting. I believe that we have a fundamental disagreement. I would argue that the semantics *don't* have to be distributed. My argument/proof would be that I believe that *anything* can be described in words -- and that I believe that previous narrow AI are brittle because they don't have both a) closure over the terms that they use and b) the ability to learn the meaning if *any* new term (traits that I believe that humans have -- and I'm not sure at all that the intelligent part of humans have distributed semantics). Of course, I'm also pretty sure that my belief is in the minority on this list as well. I believe that an English system with closure and learning *is* going to be a complex system and can be grounded (via the closure and interaction with the real world). And scalable looks less problematic to me with symbols than without. We may be different enough in (hopefully educated) opinions that this e-mail may not allow for a response other than We shall see but I would be interested, if you would, in hearing more as to why you believe that semantics *must* be distributed (though I will immediately concede that it will make them less hackable). Trust you to ask a difficult question ;-). I'll just say a few things (leaving more detail for some big fat technical paper in the future). 1) On the question of how *much* the semantics would be distributed: I don't want to overstate my case, here. The extent to which they would be distributed will be determined by how the system matures, using its learning mechanisms. What that means is that my chosen learning mechanisms, when they are fully refined, could just happen to create a system in which the atomic concepts were mostly localized, but with a soupcon of distributedness
Re: [agi] Religion-free technical content
You misunderstood me -- when I said robustness of the goal system, I meant the contents and integrity of the goal system, not the particular implementation. I do however continue to object to your phrasing about the system recognizing influence on it's goal system and preserving it. Fundamentally, there are only a very small number of Thou shalt not supergoals that need to be forever invariant. Other than those, the system should be able to change it's goals as much as it likes (chocolate or strawberry? excellent food or mediocre sex? save starving children, save adults dying of disease, or go on vacation since I'm so damn tired from all my other good works?) A quick question for Richard and others -- Should adults be allowed to drink, do drugs, wirehead themselves to death? - Original Message - From: Vladimir Nesov [EMAIL PROTECTED] To: agi@v2.listbox.com Sent: Tuesday, October 02, 2007 9:49 AM Subject: **SPAM** Re: [agi] Religion-free technical content But yet robustness of goal system itself is less important than intelligence that allows system to recognize influence on its goal system and preserve it. Intelligence also allows more robust interpretation of goal system. Which is why the way particular goal system is implemented is not very important. Problems lie in rough formulation of what goal system should be (document in English is probably going to be enough) and in placing the system under sufficient influence of its goal system (so that intelligent processes independent on it would not take over). On 10/2/07, Mark Waser [EMAIL PROTECTED] wrote: The intelligence and goal system should be robust enough that a single or small number of sources should not be able to alter the AGI's goals; however, it will not do this by recognizing forged communications but by realizing that the aberrant goals are not in congruence with the world. Note that many stupid and/or greedy people will try to influence the system and it will need to be immune to them (or the solution will be worse than the problem). -- Vladimir Nesovmailto:[EMAIL PROTECTED] - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?; - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=48878028-0baa9c
Re: [agi] Religion-free technical content
On Tuesday 02 October 2007 10:17:42 am, Richard Loosemore wrote: ... Since the AGIs are all built to be friendly, ... The probability that this will happen is approximately the same as the probability that the Sun could suddenly quantum-tunnel itself to a new position inside the perfume department of Bloomingdales. - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=48898927-5543e8
Re: Distributed Semantics [WAS Re: [agi] Religion-free technical content]
Okay, I'm going to wave the white flag and say that what we should do is all get together a few days early for the conference next March, in Memphis, and discuss all these issues in high-bandwidth mode! But one last positive thought. A response to your remark: So let's look at the mappings from throne or stump to chair . . . . A throne does not have four legs but it is used for sitting. Which way do you want to go? Or, if someone is currently sitting on the stump, how do you want to go on that one? It isn't just the representation but also how you operate on the representation . . . . I agree completely! This is exactly the way I think of it, and it is part of what I am calling distributed semantics because the thing that we refer to as the meaning of the term [chair] cannot be pinned down in a precise way, but is a result of all the interactions between the atom that represents [chair] and the many things that can interact with it. It is the fact that the entire system is capable of visualizing a stump as a chair (it visualises the operation of [sitting] applied to the stump, and the result is a perfectly reasonable scenario) that makes the meaning of chair not be localizable. There are innumerable objects in the universe that I could apply that [sitting] operation to (although some are difficult, such as individual atoms), so almost anything could be a chair. But then, do we say that [chair] is defined as anything that we can apply the [sitting] action to? Hell no! What about using [squatting]? If the king of the jungle squats on a tree stump, is the stump a throne? Yup. What about [dancing]? If some chairs are used by a dance troupe performing in local park, and the point of the dance is that all of the moves take place on chairs, BUT one of the chairs is actually a tree stump, is the stump a chair? Kind of. And so on. This is all about the boundaries between concepts (about which the cog psychologists have much to say), and if remember correctly Barsalou even has a theory along these lines. So, do wonder if we are not saying siilar things, but in different language. Richard Loosemore - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=48906135-ba92de
Re: [agi] Religion-free technical content
J Storrs Hall, PhD wrote: On Tuesday 02 October 2007 10:17:42 am, Richard Loosemore wrote: ... Since the AGIs are all built to be friendly, ... The probability that this will happen is approximately the same as the probability that the Sun could suddenly quantum-tunnel itself to a new position inside the perfume department of Bloomingdales. Justification for this statement? [Hey, I've been backing up my statements! That one was there because of several thousand words of backup in earlier posts. No fair taking stuff out of context]. Richard Loosemore - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=48908308-514b0c
Re: [agi] Religion-free technical content
On 10/2/07, Mark Waser wrote: A quick question for Richard and others -- Should adults be allowed to drink, do drugs, wirehead themselves to death? This is part of what I was pointing at in an earlier post. Richard's proposal was that humans would be asked in advance by the AGI what level of protection they required. So presumably Richard is thinking along the lines of a non-interfering AGI, unless specifically requested. There are obvious problems here. Humans don't know what is 'best' for them. Humans frequently ask for what they want, only later to discover that they really didn't want that. Humans change their mind all the time. Humans don't know 'in advance' what level of protection they would like. Death and/or mutilation comes very quickly at times. If I was intending to be evil, say, commit mass murder, I would request a lot of protection from the AGI, as other humans would be trying to stop me. -- I think the AGI will have great problems interfacing with these mixed-up argumentative humans. The AGI will probably have to do a lot of 'brain correction' to straighten out humanity. Let's hope that it knows what it is doing. BillK - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=48932774-c6db65
Re: [agi] Religion-free technical content
Beyond AI pp 253-256, 339. I've written a few thousand words on the subject, myself. a) the most likely sources of AI are corporate or military labs, and not just US ones. No friendly AI here, but profit-making and mission-performing AI. b) the only people in the field who even claim to be interested in building friendly AI (SIAI) aren't even actually building anything. c) of all the people at the AGI conf last year who were trying to build AGI, none of them had any idea how to make it friendly or even any coherent idea what friendliness might really mean. Yet they're all building away. d) same can be said of the somewhat wider group of builders on this mailing list. In other words, nobody knows what friendly really means, nobody's really trying to build a friendly AI, and more people are seriously trying to build an AGI every time I look. Who could seriously think that ALL AGIs will then be built to be friendly? Josh On Tuesday 02 October 2007 11:46:32 am, Richard Loosemore wrote: J Storrs Hall, PhD wrote: On Tuesday 02 October 2007 10:17:42 am, Richard Loosemore wrote: ... Since the AGIs are all built to be friendly, ... The probability that this will happen is approximately the same as the probability that the Sun could suddenly quantum-tunnel itself to a new position inside the perfume department of Bloomingdales. Justification for this statement? [Hey, I've been backing up my statements! That one was there because of several thousand words of backup in earlier posts. No fair taking stuff out of context]. Richard Loosemore - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?; - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=48933113-7181d7
Re: [agi] Religion-free technical content
On 10/2/07, Mark Waser [EMAIL PROTECTED] wrote: You misunderstood me -- when I said robustness of the goal system, I meant the contents and integrity of the goal system, not the particular implementation. I meant that too - and I didn't mean to imply this distinction. Implementation of goal system, or 'goal system itself' are both in my argument can be represented as text written in natural language, that is in rather faulty way. 'Goal system as what it meant to be' is what intelligent system tries to achieve. I do however continue to object to your phrasing about the system recognizing influence on it's goal system and preserving it. Fundamentally, there are only a very small number of Thou shalt not supergoals that need to be forever invariant. Other than those, the system should be able to change it's goals as much as it likes (chocolate or strawberry? excellent food or mediocre sex? save starving children, save adults dying of disease, or go on vacation since I'm so damn tired from all my other good works?) This is just substitution of levels of abstraction. Programs on my PC run on a fixed hardware and are limited by its capabilities, yet they can vary greatly. Plus, intelligent system should be able to integrate impact of goal system on multiple levels of abstraction, that is it can infer level of strictness in various circumstances (which interface with goal system through custom-made abstractions). A quick question for Richard and others -- Should adults be allowed to drink, do drugs, wirehead themselves to death? - Original Message - From: Vladimir Nesov [EMAIL PROTECTED] To: agi@v2.listbox.com Sent: Tuesday, October 02, 2007 9:49 AM Subject: **SPAM** Re: [agi] Religion-free technical content But yet robustness of goal system itself is less important than intelligence that allows system to recognize influence on its goal system and preserve it. Intelligence also allows more robust interpretation of goal system. Which is why the way particular goal system is implemented is not very important. Problems lie in rough formulation of what goal system should be (document in English is probably going to be enough) and in placing the system under sufficient influence of its goal system (so that intelligent processes independent on it would not take over). On 10/2/07, Mark Waser [EMAIL PROTECTED] wrote: The intelligence and goal system should be robust enough that a single or small number of sources should not be able to alter the AGI's goals; however, it will not do this by recognizing forged communications but by realizing that the aberrant goals are not in congruence with the world. Note that many stupid and/or greedy people will try to influence the system and it will need to be immune to them (or the solution will be worse than the problem). -- Vladimir Nesovmailto:[EMAIL PROTECTED] - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?; - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?; -- Vladimir Nesovmailto:[EMAIL PROTECTED] - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=48939671-58a21f
Re: [agi] Religion-free technical content
On 10/2/07, Mark Waser [EMAIL PROTECTED] wrote: A quick question for Richard and others -- Should adults be allowed to drink, do drugs, wirehead themselves to death? A correct response is That depends. Any should question involves consideration of the pragmatics of the system, while semantics may be not in question. [That's a brief portion of the response I owe Richard from yesterday.] Effective deciding of these should questions has two major elements: (1) understanding of the evaluation-function of the assessors with respect to these specified ends, and (2) understanding of principles (of nature) supporting increasingly coherent expression of that evolving evaluation function. And there is always an entropic arrow, due to the change in information as decisions now incur consequences not now but in an uncertain future. [This is another piece of the response I owe Richard.] [I'm often told I make everything too complex, but to me this is a coherent, sense-making model, excepting the semantic roughness of it's expression in this post.] - Jef - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=48940364-66d769
Re: [agi] Religion-free technical content
On 10/2/07, Vladimir Nesov [EMAIL PROTECTED] wrote: On 10/2/07, Mark Waser [EMAIL PROTECTED] wrote: You misunderstood me -- when I said robustness of the goal system, I meant the contents and integrity of the goal system, not the particular implementation. I meant that too - and I didn't mean to imply this distinction. Implementation of goal system, or 'goal system itself' are both in my argument can be represented as text written in natural language, that is in rather faulty way. 'Goal system as what it meant to be' is what intelligent system tries to achieve. I do however continue to object to your phrasing about the system recognizing influence on it's goal system and preserving it. Fundamentally, there are only a very small number of Thou shalt not supergoals that need to be forever invariant. Other than those, the system should be able to change it's goals as much as it likes (chocolate or strawberry? excellent food or mediocre sex? save starving children, save adults dying of disease, or go on vacation since I'm so damn tired from all my other good works?) This is just substitution of levels of abstraction. Programs on my PC run on a fixed hardware and are limited by its capabilities, yet they can vary greatly. Plus, intelligent system should be able to integrate impact of goal system on multiple levels of abstraction, that is it can infer level of strictness in various circumstances (which interface with goal system through custom-made abstractions). A quick question for Richard and others -- Should adults be allowed to drink, do drugs, wirehead themselves to death? - Original Message - From: Vladimir Nesov [EMAIL PROTECTED] To: agi@v2.listbox.com Sent: Tuesday, October 02, 2007 9:49 AM Subject: **SPAM** Re: [agi] Religion-free technical content But yet robustness of goal system itself is less important than intelligence that allows system to recognize influence on its goal system and preserve it. Intelligence also allows more robust interpretation of goal system. Which is why the way particular goal system is implemented is not very important. Problems lie in rough formulation of what goal system should be (document in English is probably going to be enough) and in placing the system under sufficient influence of its goal system (so that intelligent processes independent on it would not take over). On 10/2/07, Mark Waser [EMAIL PROTECTED] wrote: The intelligence and goal system should be robust enough that a single or small number of sources should not be able to alter the AGI's goals; however, it will not do this by recognizing forged communications but by realizing that the aberrant goals are not in congruence with the world. Note that many stupid and/or greedy people will try to influence the system and it will need to be immune to them (or the solution will be worse than the problem). Argh! Goal system and Friendliness are roughly the same sort of confusion. They are each modelable only within a ***specified***, encompassing context. In more coherent, modelable terms, we express our evolving nature, rather than strive for goals. - Jef - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=48954318-fd558f
Re: [agi] Religion-free technical content
Effective deciding of these should questions has two major elements: (1) understanding of the evaluation-function of the assessors with respect to these specified ends, and (2) understanding of principles (of nature) supporting increasingly coherent expression of that evolving evaluation function. So how do I get to be an assessor and decide? - Original Message - From: Jef Allbright [EMAIL PROTECTED] To: agi@v2.listbox.com Sent: Tuesday, October 02, 2007 12:55 PM Subject: **SPAM** Re: [agi] Religion-free technical content On 10/2/07, Mark Waser [EMAIL PROTECTED] wrote: A quick question for Richard and others -- Should adults be allowed to drink, do drugs, wirehead themselves to death? A correct response is That depends. Any should question involves consideration of the pragmatics of the system, while semantics may be not in question. [That's a brief portion of the response I owe Richard from yesterday.] Effective deciding of these should questions has two major elements: (1) understanding of the evaluation-function of the assessors with respect to these specified ends, and (2) understanding of principles (of nature) supporting increasingly coherent expression of that evolving evaluation function. And there is always an entropic arrow, due to the change in information as decisions now incur consequences not now but in an uncertain future. [This is another piece of the response I owe Richard.] [I'm often told I make everything too complex, but to me this is a coherent, sense-making model, excepting the semantic roughness of it's expression in this post.] - Jef - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?; - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=48955662-36b85e
Re: [agi] Religion-free technical content
On 10/2/07, Mark Waser [EMAIL PROTECTED] wrote: Effective deciding of these should questions has two major elements: (1) understanding of the evaluation-function of the assessors with respect to these specified ends, and (2) understanding of principles (of nature) supporting increasingly coherent expression of that evolving evaluation function. So how do I get to be an assessor and decide? You already are and do, to the extent that you are and do. Is my writing really that obscure? The pragmatic point is that there are no absolute answers, but you absolutely can improve the process (from any particular subjective point of view.) - Jef - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=48959633-3f7f85
Re: [agi] Religion-free technical content
On 10/2/07, Jef Allbright [EMAIL PROTECTED] wrote: Argh! Goal system and Friendliness are roughly the same sort of confusion. They are each modelable only within a ***specified***, encompassing context. In more coherent, modelable terms, we express our evolving nature, rather than strive for goals. Terminology. Note that I did talk about subproblems of 'goal system': 'goal content' (textual description, such as Eliezer's CV) and property of system itself to behave according to this 'goal content'. Word 'goal' is a functional description, it doesn't limit design choices. What do you mean by context here? Certainly goal content needs semantic grounding in system's knowledge. -- Vladimir Nesovmailto:[EMAIL PROTECTED] - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=48961085-eeffe2
RE: [agi] Religion-free technical content
Richard Loosemore: a) the most likely sources of AI are corporate or military labs, and not just US ones. No friendly AI here, but profit-making and mission-performing AI. Main assumption built into this statement: that it is possible to build an AI capable of doing anything except dribble into its wheaties, using the techiques currently being used. I have explained elsewhere why this is not going to work. If your explanations are convincing, smart people in industry and the military might just absorb them and then they still have more money and manpower than you do. When the first AGI is built, its first actions will be to make sure that nobody is trying to build a dangerous, unfriendly AGI. I often see it assumed that the step between first AGI is built (which I interpret as a functoning model showing some degree of generally-intelligent behavior) and god-like powers dominating the planet is a short one. Is that really likely? - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=48961147-810f59
Re: [agi] Religion-free technical content
On 10/2/07, Vladimir Nesov [EMAIL PROTECTED] wrote: On 10/2/07, Jef Allbright [EMAIL PROTECTED] wrote: Argh! Goal system and Friendliness are roughly the same sort of confusion. They are each modelable only within a ***specified***, encompassing context. In more coherent, modelable terms, we express our evolving nature, rather than strive for goals. Terminology. Note that I did talk about subproblems of 'goal system': 'goal content' (textual description, such as Eliezer's CV) and property of system itself to behave according to this 'goal content'. Word 'goal' is a functional description, it doesn't limit design choices. What do you mean by context here? Certainly goal content needs semantic grounding in system's knowledge. Fundamental systems theory. Any system can be effectively specified only within a more encompassing context. Shades of Godel's theorem considering the epistemological implications. So it's perfectly valid to speak of goals within an effectively specified context, but it's incoherent to speak of a supergoal of friendliness as if that expression has a modelable referent. Goals, like free-will, are a property of the observer, not the observed. When I speak of context, I'm generally not talking semantics but pragmatics; not meaning, but what works; not linguistics, but systems. [I want to apologize to the list. I'm occasionally motivated to jump in where I imagine I see some fertile ground to plant a seed of thought, but due to pressures of work I'm unable to stay and provide the appropriate watering and tending needed for its growth.] - Jef - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=48987856-77b6a9
Re: [agi] Religion-free technical content
You already are and do, to the extent that you are and do. Is my writing really that obscure? It looks like you're veering towards CEV . . . . which I think is a *huge* error. CEV says nothing about chocolate or strawberry and little about great food or mediocre sex. The pragmatic point is that there are no absolute answers, but you absolutely can improve the process (from any particular subjective point of view.) Wrong. There *are* some absolute answers. There are some obvious universal Thou shalt nots that are necessary unless you're rabidly anti-community (which is not conducive to anyone's survival -- and if you want to argue that community survival isn't absolute, then I'll just cheerfully ignore you). - Original Message - From: Jef Allbright [EMAIL PROTECTED] To: agi@v2.listbox.com Sent: Tuesday, October 02, 2007 1:34 PM Subject: **SPAM** Re: [agi] Religion-free technical content On 10/2/07, Mark Waser [EMAIL PROTECTED] wrote: Effective deciding of these should questions has two major elements: (1) understanding of the evaluation-function of the assessors with respect to these specified ends, and (2) understanding of principles (of nature) supporting increasingly coherent expression of that evolving evaluation function. So how do I get to be an assessor and decide? You already are and do, to the extent that you are and do. Is my writing really that obscure? The pragmatic point is that there are no absolute answers, but you absolutely can improve the process (from any particular subjective point of view.) - Jef - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?; - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49012992-637825
Re: [agi] Religion-free technical content
Do you really think you can show an example of a true moral universal? Thou shalt not destroy the universe. Thou shalt not kill every living and/or sentient being including yourself. Thou shalt not kill every living and/or sentient except yourself. - Original Message - From: Jef Allbright [EMAIL PROTECTED] To: agi@v2.listbox.com Sent: Tuesday, October 02, 2007 2:53 PM Subject: **SPAM** Re: [agi] Religion-free technical content On 10/2/07, Mark Waser [EMAIL PROTECTED] wrote: Wrong. There *are* some absolute answers. There are some obvious universal Thou shalt nots that are necessary unless you're rabidly anti-community (which is not conducive to anyone's survival -- and if you want to argue that community survival isn't absolute, then I'll just cheerfully ignore you). I see community, or rather cooperation, as essential to morality. (Quite ironic that you would write someone off for disagreement over the value of community!) As for moral absolutes, they fail due the non-existence of an absolute context. As for moral universals, humans can get pretty close to universal agreement on some key principles, but this is true only to the extent that we share common values-assessment function based on shared cultural, genetic, physical... heritage.) So not quite universal now, and bound to broaden. Do you really think you can show an example of a true moral universal? - Jef - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?; - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49022660-de0d50
Re: [agi] Religion-free technical content
On 10/2/07, Mark Waser [EMAIL PROTECTED] wrote: Do you really think you can show an example of a true moral universal? Thou shalt not destroy the universe. Thou shalt not kill every living and/or sentient being including yourself. Thou shalt not kill every living and/or sentient except yourself. Mark, this is so PHIL101. Do you *really* think there aren't people, now or conceivable in another time, who could in full sincerity violate any of these? Think cult-mentality for starters. I'm not going to cheerfully right you off now, but feel free to have the last word as I can't afford any more time for this right now. - Jef - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49026453-d502c2
Re: [agi] Religion-free technical content
On 10/2/07, Jef Allbright [EMAIL PROTECTED] wrote: I'm not going to cheerfully right you off now, but feel free to have the last word. Of course I meant cheerfully write you off or ignore you. - Jef - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49028437-a9e9f6
Re: [agi] Religion-free technical content
Do you really think you can show an example of a true moral universal? Thou shalt not destroy the universe. Thou shalt not kill every living and/or sentient being including yourself. Thou shalt not kill every living and/or sentient except yourself. Mark, this is so PHIL101. Do you *really* think there aren't people, now or conceivable in another time, who could in full sincerity violate any of these? Think cult-mentality for starters. There are people who could in full sincerity violate *any* statement. All you have to be is stupid enough, deluded enough, dishonest enough, or any one of a thousand pathologies. The (until now unrealized by me) PHIL 101 is that you are arguing unrealistic infinities that serve no purpose other than mental masturbation. Do you *really* think that any of those three statements shouldn't be embraced regardless of circumstances (and don't give me ridiculous crap like Well, if the universe was only inflicting suffering on everyone . . . . ) - Original Message - From: Jef Allbright [EMAIL PROTECTED] To: agi@v2.listbox.com Sent: Tuesday, October 02, 2007 3:14 PM Subject: **SPAM** Re: [agi] Religion-free technical content On 10/2/07, Mark Waser [EMAIL PROTECTED] wrote: Do you really think you can show an example of a true moral universal? Thou shalt not destroy the universe. Thou shalt not kill every living and/or sentient being including yourself. Thou shalt not kill every living and/or sentient except yourself. Mark, this is so PHIL101. Do you *really* think there aren't people, now or conceivable in another time, who could in full sincerity violate any of these? Think cult-mentality for starters. I'm not going to cheerfully right you off now, but feel free to have the last word as I can't afford any more time for this right now. - Jef - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?; - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244id_secret=49034020-67ef16