Edward W. Porter wrote:
Richard in your November 02, 2007 11:15 AM post you stated:

“If AI systems are built with motivation systems that are stable, then we could predict that they will remain synchronized with the goals of the human race until the end of history.”

and

“I can think of many, many types of non-goal-stack motivational systems for which [Matt’s statement about the inherent instability of goal systems of recursively self improving AGIs] is a complete falsehood.”

In your 11/3/2007 1:17 PM post you described what I assume to be such a suppostedly stable “non-goal-stack motivational system.” as follows:

“ consider the motivational system of the
best kind of AGI:  it is motivated by a balanced set of desires that
include the desire to explore and learn, and empathy for the human
species.  By definition, I would think, this simple cluster of desires
and empathic motivations *are* the things that "give it pleasure".

and

“I think that in general, making the AGI as similar to us as possible
(but without the aggressive and dangerous motivations that we are
victims of) would be a good idea simply because we want them to start
out with a strong empathy for us, and we want them to stay that way.”

I think this type of motivational system makes a lot of sense, but for all the reasons stated in my Fri 11/2/2007 2:07 PM post (arguments you have not responded to) as well as many other reasons, it does appear at all certain such a motivational system would reliably remain stable and “synchronized with the goals of the human race until the end of history,” as you claim.

For example, humans might for short sighted personal gain (such as when using them in weapon systems)

Whoaa! You assume that it would be possible to "use" an AGI for personal gain, or in a weapon system. If it starts out with the supposed empathic motivational system, it would not allow this.


 or accidentally alter such a motivational
system.

Again, under the assumption, it would not allow such 'accidental" alteration.

 Or over time the inherent biases that were designed to make
AGI’s have empathy for humans, might cause it to have empathy for some humans more than others

As part of its initial (assumed) empathy, it will set up mechanisms to monitor such things. It could not possibly start having "more empathy for some humans more than others" without *also* being aware of the fact that, by being so biassed, it was in conflict with its general motivation. So it would not do such a thing. (We have to remember not to assume it would be both superintelligent, and also suscetible to such easily-caught problems).

, or might cause them to make decisions that they
think are in our best interest, but would not.

Again, this assumes (implicitly) that it would both be generally and broadly empathic -- which means sensitive to our wishes -- and at the same time, for some inexplicable reason, decide to do something that it thinks is in our best interests, without consulting us. Effectively assumed that it would spontaneously *stop* being empathic, without explaining how it could happen.



  Or perhaps AGI robots
would begin to embody the "human features" that they have been taught to be empathetic to better than people. Etc.

Does this mean things like beginning to get aggressive, or jealous, etc? this is where the technical characteristics of a motivational system become important: this kind of drift would be impossible unless the motivational system already had aggressiveness modules built in (which is not the case, by assumption).



The world is too complicated and is going to change too rapidly in the next one hundred, one thousand, or ten thousand years for any goal system designed circa 2015 to remain appropriate until the end of history – unless history ends pretty soon.

Not true:  the statement was that it would stay empathic to our motivations.

Only if the goal system were particularly rigid would this be a problem, and by assumption I am talking about motivational systems that are stable (diffuse systems, along the lines of my previously mentioned post).



If I am wrong I would appreciate the enlightenment and increased hope that would come with being shown how I am wrong.

I apologize for giving too brief answers to these questions. I have too much stuff that is not written out in long form.


Richard Loosemore



Ed Porter

-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?member_id=8660244&id_secret=60958578-e87007

Reply via email to