rg wrote:
Hi
Is anyone discussing what to do in the future when we
have made AGIs? I thought that was part of why
the singularity institute was made ?
Note, that I am not saying we should not make them!
Because someone will regardless of what we decide.
I am asking for what should do to prepare for it!
and also how we should affect the creation of AGIs?
Here's some questions, I hope I am not the first to come up with.
* Will they be sane?
* Will they just be smart enough to pretend to be sane?
until...they do not have to anymore.
* Should we let them decide for us ?
If not should we/can we restrict them ?
* Can they feel any empathy for us ?
If not, again should we try to manipulate/force them to
act like they do?
* Our society is very dependent on computer systems
everywhere and its increasing!!!
Should we let the AGIs have access to the internet ?
If not is it even possible to restrict an AGI that can think super fast
is a super genious and also has a lot of raw computer power?
That most likely can find many solutions to get internet access...
(( I can give many crazy examples on how if anyone doubts))
* What should we "stupid" organics do to prepare ?
Reduce our dependency?
* Should a scientist, that do not have true ethical values be allowed to
do AGI research ?
Someone that just pretend to be ethical, someone that just wants the
glory and the
Nobel price....someone that answers the statement: It is insane With:
Oh its just needs
some adjustment, don't worry :)
* What is the military doing ? Should we raise public awareness to
gain insight?
I guess all can imagine why this is important..
The only answers I have found to what can truly control/restrict an AGI
smarter than us
are few..
- Another AGI
- Total isolation
So anyone thinking about this?
Hi
You should know that there are many people who indeed are deeply
concerned about these questions, but opinions differ greatly over what
the dangers are and how to deal with them.
I have been thinking about these questions for at least the last 20
years, and I am also an AGI developer and cognitive psychologist. My
own opinion is based on a great deal of analysis of the motivations of
AI systems in general, and AGI systems in particular.
I have two conclusions to offer you.
1) Almost all of the discussion of this issue is based on assumptions
about how an AI would behave, and the depressing truth is that most of
those assuptions are outrageously foolish. I say this, not to be
antagonistic, but because the degree of nonsense talked on this subject
is quite breathtaking, and I feel at a loss to express just how
ridiculous the situation has become.
It is not just that people make wrong assumptions, it is that people
make wrong assumptions very, very loudly: declaring these wrog
assumptions to be "obviously true". Nobody does this out of personal
ignorance, it is just that our culture is saturated with crazy ideas on
the subject.
2) I believe it is entirely possible to build a completely safe AGI. I
also believe that this completely safe AGI would be the simplest one to
build, so it is likley to be built first. Lastly, I believe that it
will not matter a great deal who builds the first AGI (within limits)
because an AGI will "self-stabilize" toward a benevolent state.
Richard Loosemore
-------------------------------------------
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription:
http://www.listbox.com/member/?member_id=8660244&id_secret=95818715-a78a9b
Powered by Listbox: http://www.listbox.com