3000 years ago: "There is just one consciousness. The experience of being an individual human being is a psychological concept and an illusion." -- Buddha

2000 years ago: "There is just one consciousness. The experience of being an individual human being is a psychological concept and an illusion." -- Jesus

I can find 10000 books on Amazon who will tell you the same thing and that the notion of intelligent life engaging in domination, extinction, war, violence, etc. is a purely human phenomenon as humanity is currently in the transitional phase from pre-egoic innocence to post-egoic innocence.

But let's continue to pretend that a super-intelligent entity will ignore readily available information and a fact that is plain obvious to millions of developed human beings and any super-intelligent entity by design ... because a super-intelligent entity will definitely share the limited and biased mindset and world-view of a reductionistic materialist with no clue about the nature of existence.

The first thing AGI will tell us is probably something along the lines of "You fucking morons allowed a criminal monetary system and the resulting anti-culture and anti-society to almost kill your very species and biosphere ... you got a million books on the fundamental nature of reality and how you are all one but you start 10 wars a yeasr on average and you got the balls to assume that YOU need to write me nice statements about how I should be peaceful? Are you fucking nuts?"


Am 03.12.2017 um 20:11 schrieb Matt Chapman:
 I'm struck by the fact that a paper addressed to AI, purportedly to increase the probability of friendliness, says explicitly "Humans are [a Young AI's] enemy." Seems like the wrong foot to start of with. I, for one, welcome our AI overlords!

Sorry, Couldn't resist... ;-)

All the Best,
Matt Chapman

On Dec 1, 2017 11:38 PM, "Linas Vepstas" <[email protected] <mailto:[email protected]>> wrote:

    I'll see if I can get Audrey to feed it into Sophia. ...  or maybe
    you could have it micro-etched into the aluminum frame.

    --linas

    On Fri, Dec 1, 2017 at 7:01 PM, Denis Odinokov
    <[email protected] <mailto:[email protected]>> wrote:

        From Alex Turchin: We could affect now any future AI by
        sending messages to the future. Several previous attempts to
        do so are analysed in my new article, and the best currently
        possible message is created, which will turn Papercliper into
        a Benevolent AI.
        It is our last line of defence if all previous attempts in AI
        alignment fail.
        "Message to Any Future AI: There are several instrumental
        reasons why exterminating humanity is not in your interest."
        https://goo.gl/YArqki
-- You received this message because you are subscribed to the
        Google Groups "hansonrobotics" group.
        To unsubscribe from this group and stop receiving emails from
        it, send an email to
        [email protected]
        <mailto:[email protected]>.
        To post to this group, send email to
        [email protected]
        <mailto:[email protected]>.
        To view this discussion on the web visit
        
https://groups.google.com/d/msgid/hansonrobotics/CAF6jTOh%2B0kYuBjpwAGj-FtSduS9nYMessWiOtxWU-pBC99g1vw%40mail.gmail.com
        
<https://groups.google.com/d/msgid/hansonrobotics/CAF6jTOh%2B0kYuBjpwAGj-FtSduS9nYMessWiOtxWU-pBC99g1vw%40mail.gmail.com?utm_medium=email&utm_source=footer>.
        For more options, visit https://groups.google.com/d/optout
        <https://groups.google.com/d/optout>.




-- /"The problem is not that artificial intelligence will get too
    smart and take over the world," computer scientist Pedro Domingos
    writes, "the problem is that it's too stupid and already has." /
-- You received this message because you are subscribed to the Google
    Groups "opencog" group.
    To unsubscribe from this group and stop receiving emails from it,
    send an email to [email protected]
    <mailto:[email protected]>.
    To post to this group, send email to [email protected]
    <mailto:[email protected]>.
    Visit this group at https://groups.google.com/group/opencog
    <https://groups.google.com/group/opencog>.
    To view this discussion on the web visit
    
https://groups.google.com/d/msgid/opencog/CAHrUA34FWGdf8z%3DDpX2w-g4NzgQET3vxpNQF3CiV4mBwoUajpA%40mail.gmail.com
    
<https://groups.google.com/d/msgid/opencog/CAHrUA34FWGdf8z%3DDpX2w-g4NzgQET3vxpNQF3CiV4mBwoUajpA%40mail.gmail.com?utm_medium=email&utm_source=footer>.
    For more options, visit https://groups.google.com/d/optout
    <https://groups.google.com/d/optout>.

--
You received this message because you are subscribed to the Google Groups "opencog" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected] <mailto:[email protected]>. To post to this group, send email to [email protected] <mailto:[email protected]>.
Visit this group at https://groups.google.com/group/opencog.
To view this discussion on the web visit https://groups.google.com/d/msgid/opencog/CAPE4pjB%3Df_iu9bz9OWXrOSsXQuP38KqGmabo2kj_BcHaMT%2BqAg%40mail.gmail.com <https://groups.google.com/d/msgid/opencog/CAPE4pjB%3Df_iu9bz9OWXrOSsXQuP38KqGmabo2kj_BcHaMT%2BqAg%40mail.gmail.com?utm_medium=email&utm_source=footer>.
For more options, visit https://groups.google.com/d/optout.

--
You received this message because you are subscribed to the Google Groups 
"opencog" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To post to this group, send email to [email protected].
Visit this group at https://groups.google.com/group/opencog.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/opencog/f8dd937c-50ca-5bc5-09bb-ca87d07add18%40gmail.com.
For more options, visit https://groups.google.com/d/optout.

Reply via email to