rg wrote:
Hi

I made some responses below.

Richard Loosemore wrote:
rg wrote:
Hi

Is anyone discussing what to do in the future when we
have made AGIs? I thought that was part of why
the singularity institute was made ?

Note, that I am not saying we should not make them!
Because someone will regardless of what we decide.

I am asking for what should do to prepare for it!
and also how we should affect the creation of AGIs?

Here's some questions, I hope I am not the first to come up with.

* Will they be sane?
* Will they just be smart enough to pretend to be sane?
   until...they do not have to anymore.

* Should we let them decide for us ?
 If not should we/can we restrict them ?

* Can they feel any empathy for us ?
  If not, again should we try to manipulate/force them to
  act like they do?

* Our society is very dependent on computer systems
 everywhere and its increasing!!!
  Should we let the AGIs have access to the internet ?
 If not is it even possible to restrict an AGI that can think super fast
 is a super genious and also has a lot of raw computer power?
 That most likely can find many solutions to get internet access...
 (( I can give many crazy examples on how if anyone doubts))

* What should we "stupid" organics do to prepare ?
  Reduce our dependency?

* Should a scientist, that do not have true ethical values be allowed to do AGI research ? Someone that just pretend to be ethical, someone that just wants the glory and the Nobel price....someone that answers the statement: It is insane With: Oh its just needs
 some adjustment, don't worry :)
* What is the military doing ? Should we raise public awareness to gain insight?
   I guess all can imagine why this is important..

The only answers I have found to what can truly control/restrict an AGI smarter than us
are few..

- Another AGI
- Total isolation

So anyone thinking about this?

Hi

You should know that there are many people who indeed are deeply concerned about these questions, but opinions differ greatly over what the dangers are and how to deal with them.

This sounds good :)
I have been thinking about these questions for at least the last 20 years, and I am also an AGI developer and cognitive psychologist. My own opinion is based on a great deal of analysis of the motivations of AI systems in general, and AGI systems in particular.

I have two conclusions to offer you.

1) Almost all of the discussion of this issue is based on assumptions about how an AI would behave, and the depressing truth is that most of those assuptions are outrageously foolish. I say this, not to be antagonistic, but because the degree of nonsense talked on this subject is quite breathtaking, and I feel at a loss to express just how ridiculous the situation has become.

It is not just that people make wrong assumptions, it is that people make wrong assumptions very, very loudly: declaring these wrog assumptions to be "obviously true". Nobody does this out of personal ignorance, it is just that our culture is saturated with crazy ideas on the subject.

This is probably true.
Therefore I try to make very few assumptions, except one: They will eventually be much smarter than us.
(If you want I can justify this, based on scalability.)

Your comments are interesting, because they give me some opportunities to illustrate the extreme difficulty of analysing these questions without making hidden assumptions.

To begin with your above remark: it is fair to assume that they will be much smarter than us, but the consequences of this are not as obvious as they might appear.

For example: what if the inevitable outcome were that "they" would give us the option of elevating our intelligence up to their level, at will (albeit with the proviso that when going up to their level we would leave the dangerous human motivations on ice for that time)? Under these circumstances there would not be any meaningful "them" and "us" but actually one population of beings, some of whom would be superintelligent some of the time, but with a flexibility in the level of intelligence of any given individual that is completely impossible today.

Second, we have to consider not their intelligence level as such, but their motivations. More on this in a moment.

2) I believe it is entirely possible to build a completely safe AGI. I also beelieve that this completely safe AGI would be the simplest one to build, so it is likley to be built first. Lastly, I believe that it will not matter a great deal who builds the first AGI (within limits) because an AGI will "self-stabilize" toward a benevolent state.

Why is it simplest to make a safe AGI?

A long argument, the shortest version of which is: you have to give a motivation system of some sort (NOT a conventional goal stack, which does not work for full AGI systems) and the motivation system will have a set of drives .... if you try to make it violent or aggressive, this will tend to destabilize it, meaning that some other team which is building a simpler AGI with neutral motivations will be able to get their system to work first. But even they have to make the AGI want to do something, and I submit that they will give it simple desire to want to be empathic to the human race, because this will facilitate the learning process. After all, we know a good deal about how to do the teacher-student thing when the teacher and student are in harmony with one another (e.g. patient adult and cheerful, affectionate, cooperative child), but we find it very difficult to imagine teaching something that does not relate to us in any way whatsoever (ever tried to teach a wild cat how to do algebra?).

This is a complex subject, so my point in a short post like this is to show you that:

a) Some people (i.e. me) have thought some of these issues through in great depth,

b) The factors that determine the outcome are, at the very least, not quite as obvious as they seem, and

c) There is some prima facie argument to back up the idea that the simplest type of AGI might (suprisingly, perhaps) also be the friendliest and therefore safest.

Is it not more difficult to make something that is guaranteed to be in some way?

This is true of a "technology" that is not able to think for itself at a high level of intelligence. But the crux of my argument is that this is not really "technology" we are talking about here, so the assumption that it is always *intrinsically* more difficult to make something guaranteed safe does not necessarily apply.

Two factors contribute to the safety feature. One is that because the AGI can think, it will be able to understand its own safety. No mere machine could ever do that (which is why machines are not safe). The second (and perhaps more important) is that the design of the motivation system can be carried out in such a way that it is "stable" in the sense that a system constrained by a very large number of constraints is stable. For example: the sun is stable with respect to its physical location because, to go anywhere, rather a lot of atoms have to simultaneously decide to move in the same direction.... its position is constrained by the fact that an impossibly large number of 'accidents' would need to occur for it to go anywhere spontaneously.


Is it not easier to just make something that can potentially be safe, unsafe and whatever.

Only for ordinary 'technology'. The usual arguments, as I say, do not apply for a 'technology' that thinks.


When you say it will "self-stabilize" toward a benevolent state do you not make a large assumption. It will exist in the same world as humans, do we all stabilize into benevolent states?

This is the most interesting of your questions. We are humans. We have motivations systems designed by evolution to be aggressive, selfish, etc etc. When you generalize from our behavior to the behavior of an AGI, you implicitly assume that the AGI would have those same motivations. This is, at the very least, a questionable assumption. In fact I have argued that it is much more than just questionable, it is very likely to be 100% wrong (see above).

The self-stabilization idea has many components to it (some outlined above). One that I have not mentioned so far is an intriguing effect that occurs in humans with severe, but episodic mental disorders: in a nutshell, if your brain periodically makes you insane, then during the periods when you are not suffering an insane episode you "want" to do something to not be the insane person any more. If you had a switch to get rid of the pathoogical 'you', you would use it. So if someone made a pathologically violent AGI, that AGI would have to have some episodes during which it was calm, otherwise it would never develop into an intelligent (let alone superintelligent) entity. But during the quiet episodes it would know enough to modify itself to eliminate the pathological side.

This is one important aspect of what I mean by self-stabilization.



Unless you introduce this artificially during the evolutionary process of said AGI. Rewarding certain behavior. But what happens when the AGI realizes it has been designed in an evolutionary process with this goals in mind, what will it do?
We can not know can we ?

Suppose you knew that evolution designed you in such a way that you have a predesigned module in your motivation system that makes you feel affection for your parents? Suppose also that you are not afflicated by violent emotions or motivations.

Would your knowledge of the *fact* of this innate desiugn feature make you suddenly feel inclined to rip it out and replace it with one that gave you the opposite motivation?

I submit that, if that is what drives you, you will feel no such inclination. An AGI would be the same way. It could understand everything about our intentions and actions in designing it, but it would *be* the peaceful, empathic creature that we designed, so it would feel no inclination to abandon that motivation system just because we put it there.


Richard Loosemore

-------------------------------------------
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244&id_secret=95818715-a78a9b
Powered by Listbox: http://www.listbox.com

Reply via email to