J. Storrs Hall, PhD. wrote:
On Friday 01 December 2006 23:42, Richard Loosemore wrote:

It's a lot easier than you suppose.  The system would be built in two
parts:  the motivational system, which would not change substantially
during RSI, and the "thinking part" (for want of a better term), which
is where you do all the improvement.

For concreteness, I have called these the Utility Function and World Model in my writings on the subject...

Well .... I am avoiding "Utility Function" precisely because it has a specific meaning in the context of the type of AI that I have been lambasting as the "goal stack approach" to motivation.

A plan that says "Let RSI consist of growing the WM and not the UF" suffers from the problem that the sophistication of the WM's understanding soon makes the UF look crude and stupid. Human babies want food, proximity to their mothers, and are frightened of strangers. That's good for babies but a person with greater understanding and capabilities is better off (and the rest of us are better off if the person has) a more sophisticated UF as well.

I don't want to take the bait on your baby-motivation analogy because I do not believe the difference between human baby and adult is the same as the difference between adult AI and even-smarter-adult-AI. Some, including myself, are of the opinion that there is a threshold of sentience above which things settle down a lot, so the AI would never look back on its earlier motivational system and call it "crude and stupid".

Also, implicit in your description of the UF and WM are some ideas that I have been explicitly avoiding in my discussion of "diffuse" motivational systems. That would make some of your points not applicable.

No time to spell it out right now. If you look back at the root of this thread you might see why, or you can wait until I get the thing written up properly.


It is not quite a contradiction, but certainly this would be impossible:
  deciding to make a modification that clearly was going to leave it
wanting something that, if it wanted that thing today, would contradict
its current priorities.  Do you see why?  The motivational mechanism IS
what the system wants, it is not what the system is considering wanting.

This is a good first cut at the problem, and is taken by e.g. Nick Bostrom in a widely cited paper at http://www.nickbostrom.com/ethics/ai.html

Well, Nick Bostrum is not the origin of this idea:  it is kind of obvious.


The system is not protecting current beliefs, it is believing its
current beliefs.  Becoming more capable of understanding the "reality"
it is immersed in?  You have implicitly put a motivational priority in
your system when you suggest that that is important to it ... does that
rank higher than its empathy with the human race?

You see where I am going:  there is nothing god-given about the desire
to "understand reality" in a better way.  That is just one more
candidate for a motivational priority.

Ah, but consider: knowing more about how the world works is often a valuable asset to the attempt to increase the utility of the world, *no matter* what else the utility function might specify.

Whoa: "increase the utility of the world"? Again, your terms do not map onto a viewpoint of motivation that dumps the idea of a crude UF. In essence, you have restated the idea that I was attacking: that "increase the utility of the world" is a motivation that trumps others. It is not necessarily the case that this is the system's primary motivation.


Thus, a system's self-modification (or evolution in general) is unlikely to remove curiosity / thirst for knowledge / desire to improve one's WM as a high utility even as it changes other things.

Yes and no.  I am going to have to get back to you on this.

Here is an idea to try to fit into that worldview. After the Singularity, I would love to go into a closed domain in which I get to live in a replica of 17th century England, growing up there from childhood with my memories put on ice for the duration of a (then normal) lifetime, and with the goal of experiencing what it would have been like to be a Natural Philosopher discovering the wonder of science for the first time. I want to discover things that are known in this era, after temporarily removing them from my mind.

So I would be.... what? Contradicting my utility function by deliberately removing knowledge? Seeking to do what? Get the knowledge back a different way? Am I seeking knowledge, or just seeking a new "experience"?

I claim the latter: but that idea of seeking new experience just does not map onto the kind of silly :-) utility functions that AI people play games with today. They cannot even represent the goal of "having interesting subjective experiences", as far as I can see.


Richard Loosemore


There are several such properties of a utility function that are likely to be invariant under self-improvement or evolution. It is by the use of such invariants that we can design self-improving AIs with reasonable assurance of their continued beneficence.

--Josh



-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?list_id=303

Reply via email to