rg wrote:
Hi

Is anyone discussing what to do in the future when we
have made AGIs? I thought that was part of why
the singularity institute was made ?

Note, that I am not saying we should not make them!
Because someone will regardless of what we decide.

I am asking for what should do to prepare for it!
and also how we should affect the creation of AGIs?

Here's some questions, I hope I am not the first to come up with.

* Will they be sane?
* Will they just be smart enough to pretend to be sane?
   until...they do not have to anymore.

* Should we let them decide for us ?
 If not should we/can we restrict them ?

* Can they feel any empathy for us ?
  If not, again should we try to manipulate/force them to
  act like they do?

* Our society is very dependent on computer systems
 everywhere and its increasing!!!
  Should we let the AGIs have access to the internet ?
 If not is it even possible to restrict an AGI that can think super fast
 is a super genious and also has a lot of raw computer power?
 That most likely can find many solutions to get internet access...
 (( I can give many crazy examples on how if anyone doubts))

* What should we "stupid" organics do to prepare ?
  Reduce our dependency?

* Should a scientist, that do not have true ethical values be allowed to do AGI research ? Someone that just pretend to be ethical, someone that just wants the glory and the Nobel price....someone that answers the statement: It is insane With: Oh its just needs
 some adjustment, don't worry :)
* What is the military doing ? Should we raise public awareness to gain insight?
   I guess all can imagine why this is important..

The only answers I have found to what can truly control/restrict an AGI smarter than us
are few..

- Another AGI
- Total isolation

So anyone thinking about this?

Hi

You should know that there are many people who indeed are deeply concerned about these questions, but opinions differ greatly over what the dangers are and how to deal with them.

I have been thinking about these questions for at least the last 20 years, and I am also an AGI developer and cognitive psychologist. My own opinion is based on a great deal of analysis of the motivations of AI systems in general, and AGI systems in particular.

I have two conclusions to offer you.

1) Almost all of the discussion of this issue is based on assumptions about how an AI would behave, and the depressing truth is that most of those assuptions are outrageously foolish. I say this, not to be antagonistic, but because the degree of nonsense talked on this subject is quite breathtaking, and I feel at a loss to express just how ridiculous the situation has become.

It is not just that people make wrong assumptions, it is that people make wrong assumptions very, very loudly: declaring these wrog assumptions to be "obviously true". Nobody does this out of personal ignorance, it is just that our culture is saturated with crazy ideas on the subject.

2) I believe it is entirely possible to build a completely safe AGI. I also believe that this completely safe AGI would be the simplest one to build, so it is likley to be built first. Lastly, I believe that it will not matter a great deal who builds the first AGI (within limits) because an AGI will "self-stabilize" toward a benevolent state.




Richard Loosemore













-------------------------------------------
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244&id_secret=95818715-a78a9b
Powered by Listbox: http://www.listbox.com

Reply via email to