This is probably an important discussion, independent of the event that
prompted it, but it turns out the protest at SXSW was a hoax/ viral
marketing campaign staged to promote a new dating site.

On Tue, Mar 17, 2015 at 8:06 AM, Calum Chace via AGI <[email protected]>
wrote:

> Very well said.  (And Pandora's Brain does say it, as it happens...)
>
> But the pro-AGI community also needs to convince the public that the AGI
> we'll get will be a Friendly one.
>
> Calum
>
> On 17 March 2015 at 15:56, Ben Goertzel via AGI <[email protected]> wrote:
>
>>
>> A problem is that careful, balanced discussions of difficult issues are
>> boring and don't attract media attention
>>
>> Joel Pitt and I wrote a fairly thoughtful discussion of AGI safety issues
>> a few years ago,
>>
>> http://jetpress.org/v22/goertzel-pitt.htm
>>
>> but of course our thoughts are more complex and nuanced, whereas a tweet
>> from a billionaire comparing AI research to demon-summoning is a lot
>> sexier...
>>
>> IMO, to get media attention sufficient to counteract the media's love of
>> alarmism and doomsaying, the pro-AGI community would need to come forward
>> very
>> aggressively with the message that AGI is important for SAVING AND
>> IMPROVING  HUMAN LIVES ... for designing the next generation of
>> medicines, for creating elder-care robots to make old age more
>> livable, for extending healthspan for those who want it, for aiding
>> the invention of new energy sources, for aiding in the fight against
>> physical and cyber terrorism, and so forth....   "Don't worry too
>> much, we'll be careful" is not a convincing counterargument -- a better
>> counterargument to the Musks, Hawkings, Bostroms and Yudkowskys of the
>> world is more like  "Hey, I don't want your fear of science fiction bad
>> guys to
>> deny my grandma her life-extending, health-improving medicine and her
>> robot friend, to eliminate my future of virtually unlimited energy and
>> to put me at risk from terrorist attacks...."   I.e. "DON'T LET THE
>> LUDDITES KILL YOUR GRANDMA AND TAKE YOUR TOYS AWAY!!   EMBRACE AI AND
>> ROBOTS LIKE YOU'VE EMBRACED SMARTPHONES, AC POWER, THE INTERNET AND BIRTH
>> CONTROL PILLS -- AND YOUR LIFE WILL BE BETTER -- " ....
>>
>> OK I'm semi-joking ;) ;p ... but unfortunately I think it's a mistake to
>> overestimate the general public's appetite for rational, balanced
>> discussion and thinking ;p ...  When careful nuanced thinking on difficult
>> issues it put out there, it tends to be vigorously ignored...
>>
>> -- Ben
>>
>> On Tue, Mar 17, 2015 at 8:01 PM, Benjamin Kapp via AGI <[email protected]>
>> wrote:
>>
>>> If you think of governments as an artificial man (as was done by
>>> Aristotle and Hobbes amongst others) which is composed of humans who are
>>> the muscles (military, police), the intelligence (spys, scientists), the
>>> judging and planning (judges, politicians), etc..  In a way the state is a
>>> leviathan (a thing which has power to overawe any individual or group of
>>> individuals).  And in this way AGI (or a super intelligence) already
>>> exists.
>>>
>>> On Tue, Mar 17, 2015 at 6:29 AM, Calum Chace via AGI <[email protected]>
>>> wrote:
>>>
>>>> Thanks Basile.  I agree with Pitrat, although I might dial up the
>>>> consideration of the downside possibility a touch.
>>>>
>>>> Hawking usually gets slightly mis-represented.  He said that AGI could
>>>> be either the best or the worst thing ever to happen to humanity.  The
>>>> "best" bit seems to get missed by both sides of the debate.
>>>>
>>>> So, my question is, what is the best way for people who think along
>>>> these lines to try and steer the public debate on AGI?  Alarmism is
>>>> unhelpful, and hard to avoid.  Secrecy won't work.  Ben is tackling the
>>>> issue head-on (as in the video he posted just now), but it's a hard debate
>>>> to get right.
>>>>
>>>> Calum
>>>>
>>>> On 17 March 2015 at 11:17, Basile Starynkevitch <
>>>> [email protected]> wrote:
>>>>
>>>>> On Tue, Mar 17, 2015 at 09:33:22AM +0100, Calum Chace via AGI wrote:
>>>>> > Steve
>>>>> >
>>>>> > I sympathise with your very understandable preference not to be
>>>>> targeted by
>>>>> > anti-AI crazies!
>>>>> >
>>>>> > What do you think is the best way to try and shape the growing public
>>>>> > debate about AGI?  Following Bostrom's book, and the comments by
>>>>> Hawking,
>>>>> > Musk and Gates, a fair proportion of the general public is now aware
>>>>> that
>>>>> > AGI might arrive in the medium term, and that it will have a very big
>>>>> > impact.
>>>>> >
>>>>> > Some AI researchers seem to be responding by saying, "Don't worry,
>>>>> it can't
>>>>> > happen for centuries, if ever".  No doubt some of them genuinely
>>>>> believe
>>>>> > that, but I wonder whether some are saying it in the (forlorn?) hope
>>>>> the
>>>>> > debate will go away. It won't.  In fact I suspect that the new
>>>>> Avengers
>>>>> > movie will kick it up a level.
>>>>> >
>>>>> > Others are saying, "Don't worry, AGI cannot and will not harm
>>>>> humans."  To
>>>>> > my mind (and I realise I may be in a small minority here on this)
>>>>> that is
>>>>> > hard to be certain about - as Bostrom demonstrated.
>>>>> >
>>>>> > Yet others are saying, "AI researcher will solve the problem long
>>>>> before
>>>>> > AGI arrives, and it's best not to worry everyone else in the
>>>>> meantime."
>>>>> >  That seems a dangerous approach to me.  If the public ever feels
>>>>> (rightly
>>>>> > or wrongly) that things have been hidden from them, they will react
>>>>> badly.
>>>>> >
>>>>> > But I do definitely sympathise with the desire not to be targeted by
>>>>> > crazies, or to be vilified by journalists who have half-understood
>>>>> the
>>>>> > situation!
>>>>> >
>>>>>
>>>>> [...]
>>>>>
>>>>> > >> > -------------------------------------------
>>>>> > >> > AGI
>>>>> > >> > Archives: https://www.listbox.com/member/archive/303/=now
>>>>>
>>>>> [....]
>>>>>
>>>>>
>>>>> I would suggest reading J.Pitrat's december 2014 blog entry on that
>>>>> subject.
>>>>> J.Pitrat is probably not subscribing to that list, i
>>>>> so I am blind-carbon-copying him.
>>>>>
>>>>>
>>>>> http://bootstrappingartificialintelligence.fr/WordPress3/2014/12/not-developing-an-advanced-artificial-intelligence-could-spell-the-end-of-the-human-race/
>>>>>
>>>>> He is explaining that
>>>>>
>>>>>  "Not developing an advanced artificial intelligence
>>>>>   could spell the end of the human race"
>>>>>
>>>>> and I believe he has a point. Of course AGI researchers should be
>>>>> careful.
>>>>>
>>>>> Regards
>>>>>
>>>>> --
>>>>> Basile Starynkevitch   http://starynkevitch.net/Basile/
>>>>>
>>>>>
>>>>
>>>>
>>>> --
>>>> Regards
>>>>
>>>> Calum
>>>>
>>>>    *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
>>>> <https://www.listbox.com/member/archive/rss/303/26973278-698fd9ee> |
>>>> Modify <https://www.listbox.com/member/?&;> Your Subscription
>>>> <http://www.listbox.com>
>>>>
>>>
>>>    *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
>>> <https://www.listbox.com/member/archive/rss/303/212726-deec6279> |
>>> Modify <https://www.listbox.com/member/?&;> Your Subscription
>>> <http://www.listbox.com>
>>>
>>
>>
>>
>> --
>> Ben Goertzel, PhD
>> http://goertzel.org
>>
>> "The reasonable man adapts himself to the world: the unreasonable one
>> persists in trying to adapt the world to himself. Therefore all progress
>> depends on the unreasonable man." -- George Bernard Shaw
>>    *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
>> <https://www.listbox.com/member/archive/rss/303/26879140-5b8435c3> |
>> Modify <https://www.listbox.com/member/?&;> Your Subscription
>> <http://www.listbox.com>
>>
>
>
>
> --
> Regards
>
> Calum
>
>    *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
> <https://www.listbox.com/member/archive/rss/303/18769370-bddcdfdc> |
> Modify
> <https://www.listbox.com/member/?&;>
> Your Subscription <http://www.listbox.com>
>



-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to