Very well said. (And Pandora's Brain does say it, as it happens...) But the pro-AGI community also needs to convince the public that the AGI we'll get will be a Friendly one.
Calum On 17 March 2015 at 15:56, Ben Goertzel via AGI <[email protected]> wrote: > > A problem is that careful, balanced discussions of difficult issues are > boring and don't attract media attention > > Joel Pitt and I wrote a fairly thoughtful discussion of AGI safety issues > a few years ago, > > http://jetpress.org/v22/goertzel-pitt.htm > > but of course our thoughts are more complex and nuanced, whereas a tweet > from a billionaire comparing AI research to demon-summoning is a lot > sexier... > > IMO, to get media attention sufficient to counteract the media's love of > alarmism and doomsaying, the pro-AGI community would need to come forward > very > aggressively with the message that AGI is important for SAVING AND > IMPROVING HUMAN LIVES ... for designing the next generation of > medicines, for creating elder-care robots to make old age more > livable, for extending healthspan for those who want it, for aiding > the invention of new energy sources, for aiding in the fight against > physical and cyber terrorism, and so forth.... "Don't worry too > much, we'll be careful" is not a convincing counterargument -- a better > counterargument to the Musks, Hawkings, Bostroms and Yudkowskys of the > world is more like "Hey, I don't want your fear of science fiction bad > guys to > deny my grandma her life-extending, health-improving medicine and her > robot friend, to eliminate my future of virtually unlimited energy and > to put me at risk from terrorist attacks...." I.e. "DON'T LET THE > LUDDITES KILL YOUR GRANDMA AND TAKE YOUR TOYS AWAY!! EMBRACE AI AND > ROBOTS LIKE YOU'VE EMBRACED SMARTPHONES, AC POWER, THE INTERNET AND BIRTH > CONTROL PILLS -- AND YOUR LIFE WILL BE BETTER -- " .... > > OK I'm semi-joking ;) ;p ... but unfortunately I think it's a mistake to > overestimate the general public's appetite for rational, balanced > discussion and thinking ;p ... When careful nuanced thinking on difficult > issues it put out there, it tends to be vigorously ignored... > > -- Ben > > On Tue, Mar 17, 2015 at 8:01 PM, Benjamin Kapp via AGI <[email protected]> > wrote: > >> If you think of governments as an artificial man (as was done by >> Aristotle and Hobbes amongst others) which is composed of humans who are >> the muscles (military, police), the intelligence (spys, scientists), the >> judging and planning (judges, politicians), etc.. In a way the state is a >> leviathan (a thing which has power to overawe any individual or group of >> individuals). And in this way AGI (or a super intelligence) already >> exists. >> >> On Tue, Mar 17, 2015 at 6:29 AM, Calum Chace via AGI <[email protected]> >> wrote: >> >>> Thanks Basile. I agree with Pitrat, although I might dial up the >>> consideration of the downside possibility a touch. >>> >>> Hawking usually gets slightly mis-represented. He said that AGI could >>> be either the best or the worst thing ever to happen to humanity. The >>> "best" bit seems to get missed by both sides of the debate. >>> >>> So, my question is, what is the best way for people who think along >>> these lines to try and steer the public debate on AGI? Alarmism is >>> unhelpful, and hard to avoid. Secrecy won't work. Ben is tackling the >>> issue head-on (as in the video he posted just now), but it's a hard debate >>> to get right. >>> >>> Calum >>> >>> On 17 March 2015 at 11:17, Basile Starynkevitch < >>> [email protected]> wrote: >>> >>>> On Tue, Mar 17, 2015 at 09:33:22AM +0100, Calum Chace via AGI wrote: >>>> > Steve >>>> > >>>> > I sympathise with your very understandable preference not to be >>>> targeted by >>>> > anti-AI crazies! >>>> > >>>> > What do you think is the best way to try and shape the growing public >>>> > debate about AGI? Following Bostrom's book, and the comments by >>>> Hawking, >>>> > Musk and Gates, a fair proportion of the general public is now aware >>>> that >>>> > AGI might arrive in the medium term, and that it will have a very big >>>> > impact. >>>> > >>>> > Some AI researchers seem to be responding by saying, "Don't worry, it >>>> can't >>>> > happen for centuries, if ever". No doubt some of them genuinely >>>> believe >>>> > that, but I wonder whether some are saying it in the (forlorn?) hope >>>> the >>>> > debate will go away. It won't. In fact I suspect that the new >>>> Avengers >>>> > movie will kick it up a level. >>>> > >>>> > Others are saying, "Don't worry, AGI cannot and will not harm >>>> humans." To >>>> > my mind (and I realise I may be in a small minority here on this) >>>> that is >>>> > hard to be certain about - as Bostrom demonstrated. >>>> > >>>> > Yet others are saying, "AI researcher will solve the problem long >>>> before >>>> > AGI arrives, and it's best not to worry everyone else in the >>>> meantime." >>>> > That seems a dangerous approach to me. If the public ever feels >>>> (rightly >>>> > or wrongly) that things have been hidden from them, they will react >>>> badly. >>>> > >>>> > But I do definitely sympathise with the desire not to be targeted by >>>> > crazies, or to be vilified by journalists who have half-understood the >>>> > situation! >>>> > >>>> >>>> [...] >>>> >>>> > >> > ------------------------------------------- >>>> > >> > AGI >>>> > >> > Archives: https://www.listbox.com/member/archive/303/=now >>>> >>>> [....] >>>> >>>> >>>> I would suggest reading J.Pitrat's december 2014 blog entry on that >>>> subject. >>>> J.Pitrat is probably not subscribing to that list, i >>>> so I am blind-carbon-copying him. >>>> >>>> >>>> http://bootstrappingartificialintelligence.fr/WordPress3/2014/12/not-developing-an-advanced-artificial-intelligence-could-spell-the-end-of-the-human-race/ >>>> >>>> He is explaining that >>>> >>>> "Not developing an advanced artificial intelligence >>>> could spell the end of the human race" >>>> >>>> and I believe he has a point. Of course AGI researchers should be >>>> careful. >>>> >>>> Regards >>>> >>>> -- >>>> Basile Starynkevitch http://starynkevitch.net/Basile/ >>>> >>>> >>> >>> >>> -- >>> Regards >>> >>> Calum >>> >>> *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> >>> <https://www.listbox.com/member/archive/rss/303/26973278-698fd9ee> | >>> Modify <https://www.listbox.com/member/?&> Your Subscription >>> <http://www.listbox.com> >>> >> >> *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> >> <https://www.listbox.com/member/archive/rss/303/212726-deec6279> | Modify >> <https://www.listbox.com/member/?&> Your Subscription >> <http://www.listbox.com> >> > > > > -- > Ben Goertzel, PhD > http://goertzel.org > > "The reasonable man adapts himself to the world: the unreasonable one > persists in trying to adapt the world to himself. Therefore all progress > depends on the unreasonable man." -- George Bernard Shaw > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/26879140-5b8435c3> | > Modify > <https://www.listbox.com/member/?&> > Your Subscription <http://www.listbox.com> > -- Regards Calum ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657 Powered by Listbox: http://www.listbox.com
