[agi] The Metamorphosis of Prime Intellect

2003-01-14 Thread Philip . Sutton
I've just read the first chapter of The Metamorphosis of Prime Intellect.

http://www.kuro5hin.org/prime-intellect

It makes you realise that Ben's notion that ethical structures should be 
based on a hierarchy going from general to specific is very valid - if 
Prime Intellect had been programmed to respect all *life* and not just 
humans then the 490 worlds with sentient life not to mention the 14,623 
worlds with life of some type might have been spared.

It also makes it clear that when we talk about building AGIs for 'human 
friendliness' we are using language that does not follow Ben's 
recommended ethical goal structure.

I'm wondering (seriously) whether the AGI movement needs to change 
it short hand language (human friendly) in this case - in other arenas 
people talk about the need for ethical behaviour.  Would that term 
suffice?

Cheers, Philip

---
To unsubscribe, change your address, or temporarily deactivate your subscription, 
please go to http://v2.listbox.com/member/?[EMAIL PROTECTED]



Re: [agi] The Metamorphosis of Prime Intellect

2003-01-14 Thread Eliezer S. Yudkowsky
Eliezer S. Yudkowsky wrote:

There may be additional rationalization mechanisms I haven't identified 
yet which are needed to explain anosognosia and similar disorders. 
Mechanism (4) is the only one deep enough to explain why, for example, 
the left hemisphere automatically and unconsciously rationalizes the 
actions of the left hemisphere; and mechanism (4) doesn't necessarily 
explain that, it only looks like it might someday do so.

That is, the left hemisphere automatically and unconsciously rationalizes 
the actions of the right hemisphere in split-brain patients.  Sorry.

--
Eliezer S. Yudkowsky  http://singinst.org/
Research Fellow, Singularity Institute for Artificial Intelligence

---
To unsubscribe, change your address, or temporarily deactivate your subscription, 
please go to http://v2.listbox.com/member/?[EMAIL PROTECTED]


Re: [agi] The Metamorphosis of Prime Intellect

2003-01-14 Thread Eliezer S. Yudkowsky
Ben Goertzel wrote:

This is exactly why I keep trying to emphasize that we all should forsake
those endlessly fascinating, instinctively attractive political arguments
over our favorite moralities, and instead focus on the much
harder problem
of defining an AI architecture which can understand that its morality is
wrong in various ways; wrong definitions, wrong reinforcement
procedures, wrong source code, wrong Friendliness architecture, wrong
definition of wrongness, and many others.  These are nontrivial
problems!  Each turns out to require nonobvious structural qualities in
the architecture of the goal system.


Hmmm.  It seems to me the ability to recognize one's own potential wrongness
comes along automatically with general intelligence...


Ben, I've been there, 1996-2000, and that turned out to be the WRONG 
ANSWER.  There's an enormous amount of moral complexity that does *not* 
come along with asymptotically increasing intelligence.  Thankfully, 
despite the tremendous emotional energy I put into believing that 
superintelligences are inevitably moral, and despite the amount of 
published reasoning I had staked on it, I managed to spot this mistake 
before I pulled a Lawrence on the human species.  Please, please, please 
don't continue where I left off.

The problem here is the imprecision of words.  *One* form of wrongness, 
such as factual error, or wrong source code which is wrong because it 
is inefficient or introduces factual errors, is readily conceivable by a 
general intelligence without extra moral complexity.  You do, indeed, get 
recognition of *that particular* kind of wrongness for free.  It does 
not follow that all the things we recognize as wrong, in moral domains 
especially, can be recognized by a general intelligence without extra 
moral complexity.

If it is the case that a general intelligence necessarily has the ability 
to conceive of a wrongness in a top-level goal definition and has a 
mechanism for correcting it, this is not obvious to me - not for any 
definition of wrongness at all.  Prime Intellect, with its total 
inability to ask any moral question except how desirable is X, under the 
Three Laws as presently defined, seems to me quite realistic.

Note also that the ability to identify *a* kind of wrongness, does not 
necessarily mean the ability to see - as a human would - the specific 
wrongness of your own programmer standing by and screaming That's not 
what I meant!  Stop!  Stop!  If this realization is a necessary ability 
of all minds-in-general it is certainly not clear why.

Recognizing wrong source code requires a codic modality, of course, and
recognizing wrong Friendliness architecture requires an intellectual
knowledge of philosophy and software design.

What is there about recognizing one's wrongness in the ways you mention,
that doesn't come for free with general cognition and appropriate
perception?


So... you think a real-life Prime Intellect would have, for free, 
recognized that it should not lock Lawrence out?  But why?

I guess there is an attitude needed to recognize one's own wrongness: a lack
of egoistic self-defensive certainty in one's own correctness  A
skeptical attitude even about one's own most deeply-held beliefs.

In Novamente, this skeptical attitude has two aspects:

1) very high level schemata that must be taught not programmed
2) some basic parameter settings that will statistically tend to incline the
system toward skepticism of its own conclusions [but you can't turn the dial
too far in the skeptical direction either...]


That's for skepticism about facts.  I agree you get that for free with 
general intelligence.  If *all* questions of morality, means and ends and 
ultimate goals, were reducible to facts and deducible by logic or 
observation, then the issue would end right there.  That was my position 
1996-2000.  Is this your current position?

--
Eliezer S. Yudkowsky  http://singinst.org/
Research Fellow, Singularity Institute for Artificial Intelligence

---
To unsubscribe, change your address, or temporarily deactivate your subscription, 
please go to http://v2.listbox.com/member/?[EMAIL PROTECTED]


RE: [agi] The Metamorphosis of Prime Intellect

2003-01-14 Thread Ben Goertzel

  In Novamente, this skeptical attitude has two aspects:
 
  1) very high level schemata that must be taught not programmed
  2) some basic parameter settings that will statistically tend
 to incline the
  system toward skepticism of its own conclusions [but you can't
 turn the dial
  too far in the skeptical direction either...]

 That's for skepticism about facts.  I agree you get that for free with
 general intelligence.  If *all* questions of morality, means and ends and
 ultimate goals, were reducible to facts and deducible by logic or
 observation, then the issue would end right there.  That was my position
 1996-2000.  Is this your current position?

Not exactly, no... that is not my current position.

For example: Of course, there is no logical way to deduce that killing
chimpanzees is morally worse than killing fleas, from no assumptions.

If one assumes that killing humans is morally bad, then from this
assumption, reasoning (probabilistic analogical reasoning, for instance)
leads one to conclude that killing chimpanzees is morally worse than killing
fleas...

My view is fairly subtle, and has progressed a bit since I wrote Thoughts
on AI Morality.  I don't have time to write a long e-mail on it today, but
I will do so tomorrow.  I think that will be better than writing something
hasty and hard-to-understand right now.

Ben


---
To unsubscribe, change your address, or temporarily deactivate your subscription, 
please go to http://v2.listbox.com/member/?[EMAIL PROTECTED]



RE: [agi] The Metamorphosis of Prime Intellect

2003-01-14 Thread Billy Brown
Eliezer S. Yudkowsky wrote:
 I don't think we are the beneficiaries of massive evolutionary debugging.
   I think we are the victims of massive evolutionary warpage to win
 arguments in adaptive political contexts.  I've identified at least four
 separate mechanisms of rationalization in human psychology so far:

Well, yes. Human minds are tuned for fitness in the ancestral environment,
not for correspondence with objective reality. But just getting to the point
where implementing those rationalizations is possible would be a huge leap
forward from current AI systems.

In any case, I think your approach to the problem is a step in the right
direction. We need a theory of AI ethics before we can test it, and we need
lots of experimental testing before we start building things that have any
chance of taking off. Sometimes I think it is a good thing that AI is still
stuck in a mire of wishful thinking, because we aren't ready to build AGI
safely.

Billy Brown

---
To unsubscribe, change your address, or temporarily deactivate your subscription, 
please go to http://v2.listbox.com/member/?[EMAIL PROTECTED]