Hi Jason,

On Mon, Nov 27, 2017 at 11:09 PM, Jason Resch <jasonre...@gmail.com> wrote:
> I think there might be two ways of interpreting this, each with different
> answers.
>
> The first question: Does AI create more threats that never existed before?
>
> I think the answer is most definitely yes. Some examples:
> - Large scale unemployment/disempowerment of people who cannot compete with
> increasing machine intelligence
> - Algorithms that identify and wipe out dissent / control opposition
> - New and terrifying weapons (e.g.
> https://www.youtube.com/watch?v=HipTO_7mUOw )
> - More infrastructure and systems that can be hacked or introduce defects
> (air traffic control systems, self-driving cars, etc.)
>
> The second question: Will super intelligence ultimately decide to eliminate
> us (as meaningless, redundant, to make room for more computation, etc.)?
>
> This question is more interesting. I tend to fall in the camp that we
> exercise little control over the ultimate decision made by such a super
> intelligence, but I am optimistic that a super intelligence will, during the
> course of its ascension, discover and formalize a system of ethics, and this
> may lead to it deciding not to wipe out other life forms.  For example, it
> might discover the same ideas expressed here (
> https://www.researchgate.net/profile/Arnold_Zuboff/publication/233329805_One_Self_The_Logic_of_Experience/links/54adcdb60cf2213c5fe419ec/One-Self-The-Logic-of-Experience.pdf
> ) and therefore determine something like the golden rule is rationally
> justified.

On a more serious note, the paper above is very interesting, thanks.

If you have some time/patience, let me know what you think of my arguments here:
https://arxiv.org/abs/1609.02009

Telmo.

> Jason
>
>
>
> On Mon, Nov 27, 2017 at 3:32 PM, <agrayson2...@gmail.com> wrote:
>>
>> IIRC, this is the view of Hawking and Musk.
>>
>> --
>> You received this message because you are subscribed to the Google Groups
>> "Everything List" group.
>> To unsubscribe from this group and stop receiving emails from it, send an
>> email to everything-list+unsubscr...@googlegroups.com.
>> To post to this group, send email to everything-list@googlegroups.com.
>> Visit this group at https://groups.google.com/group/everything-list.
>> For more options, visit https://groups.google.com/d/optout.
>
>
> --
> You received this message because you are subscribed to the Google Groups
> "Everything List" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to everything-list+unsubscr...@googlegroups.com.
> To post to this group, send email to everything-list@googlegroups.com.
> Visit this group at https://groups.google.com/group/everything-list.
> For more options, visit https://groups.google.com/d/optout.

-- 
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to everything-list+unsubscr...@googlegroups.com.
To post to this group, send email to everything-list@googlegroups.com.
Visit this group at https://groups.google.com/group/everything-list.
For more options, visit https://groups.google.com/d/optout.

Reply via email to