The BIG problem that threatens the safety of everyone on this forum are crazies like some on this forum who publicly (this AGI forum is public and is Google searchable) claim that potentially dangerous AGIs will be up in running in 5 years or whenever. Regardless of the truth or lack thereof, movies (like the Terminator series) portraying the great social value of KILLING people who develop AGIs is enough to motivate other crazies out there to do really bad things to the crazies who make such statements.
Which of the crazies are more dangerous TO ME. That is easy - it is the crazies on this forum, who might guide other crazies to my door. There are LOTS of murderous crazies in our world, which I can live with, so long as no one is guiding them TO ME. It is really hard to believe that ANYONE who is SO stupid to be publicly making such obviously dangerous statements actually thinks they are smart enough to participate in making a genuine AGI. There is NO WAY that this could possibly happen. Stupid is forever. Personally, I have absolutely NO expectation of any sort of dangerous AGIs emerging anytime soon, most especially from people who are stupid enough to make such obviously dangerous public statements. The less mentally impaired participants on this forum are quietly working on their respective theories WITHOUT making such incendiary statements. If you the reader have made such public statements in the past, then GIVE UP working on AGI technology, because you have failed the intelligence test and are obviously NOT smart enough to ever succeed. Further, your potential future colleagues will see you as being too dangerous to have as an associate. If you are an assassin looking for someone to kill, then look elsewhere. These guys aren't worth your ammunition. *BEN*: As an act of social responsibility, you should purge ALL postings, including this posting, that mentions any sort of short time horizon for AGI development, and go all the way back to the beginning of this forum to do so. That demonstrations like the one in Texas can even come about shows the pushback that such statements can attract. Do we really need to be seen as dangerous social pariahs? Steve ================== On Mon, Mar 16, 2015 at 7:21 PM, Nanograte Knowledge Technologies via AGI < [email protected]> wrote: > If I may say something pelase? To my understanding, Google would and Elon > Musk would. > > However, AI is not the real threat. In my most-humble opinion, it is the > key to the solution to a real threat. The technology would still be > developed, regardless, and is probably being hastened as we speak. Perhaps > the following tenets should be applied to "their" rationale: "The cat's out > of the bag. What they don't know, won't harm them. Let's just go > underground and hurry it up some more." As such then, protests are > insignificant, red herrings. Protests could be staged to support public > statements, as a distraction and disabling strategy to detract from the > real issues at hand. Who knows? Who cares? Most people don't even read. I > agree more with the 5-year outlook, and it could even become 4, depending > on how quickly the key constraints to such progress could be resolved by > people like us. > > ------------------------------ > Date: Tue, 17 Mar 2015 08:39:22 +0800 > Subject: Re: [agi] AI Protest in Texas > From: [email protected] > To: [email protected] > CC: [email protected] > > > > yeah, that's more consistent with what I've heard from Demis in the past... > > > > On Tue, Mar 17, 2015 at 8:30 AM, Calum Chace <[email protected]> wrote: > > Sorry, Ben, it wasn't centuries for Hassabis. It was decades. Rather an > important difference! > > Last year, the American entrepreneur, Elon Musk, one of Deep Mind’s early > investors, described AI as humanity’s greatest existential threat. “Unless > you have direct exposure to groups like Deepmind, you have no idea how fast > [AI] is growing,” he said. “The risk of something seriously dangerous > happening is in the five year timeframe. Ten years at most.” > > However, the Google team played down the concerns. “We agree with him > there are risks that need to be borne in mind, but we’re decades away from > any sort of technology that we need to worry about,” Hassabis said. > > > http://www.theguardian.com/technology/2015/feb/25/google-develops-computer-program-capable-of-learning-tasks-independently > > Calum > > On 17 March 2015 at 01:20, Ben Goertzel <[email protected]> wrote: > > > Did Demis really say AGI is hundreds of years away? That surprises me.... > > I think Ng actually believes AGI is far off, he's conservative but I > believe he's a straight shooter. > > I don't know Yann and Christof F2F so I don't have a strong opinion on > their attitudes... > > -- Ben > > On Tue, Mar 17, 2015 at 8:13 AM, Calum Chace <[email protected]> wrote: > > Yes, but Austin, of all places. > > Ben, why do you think Yann LeCun, Andrew Ng, Christof Koch and Demis > Hassabis have all been lining up to say that AGI is hundreds of years > away? Are they worried about this sort of reaction? > > On 17 March 2015 at 01:10, Ben Goertzel via AGI <[email protected]> wrote: > > > And of course it has to be in Texas 8-D ... > > On Tue, Mar 17, 2015 at 5:17 AM, Piaget Modeler via AGI <[email protected]> > wrote: > > Straight out of Stephen Spielberg's film: A.I. > > ~PM > > > Date: Mon, 16 Mar 2015 12:59:25 -0700 > > Subject: Re: [agi] AI Protest in Texas > > From: [email protected] > > To: [email protected] > > > > > On 3/16/15, Aaron Hosford <[email protected]> wrote: > > > This sort of thing was predicted 50 years ago. > > > http://en.wikipedia.org/wiki/Butlerian_Jihad > > > > > > Nonetheless, yes, mind blowing. > > > > > > On Mon, Mar 16, 2015 at 11:41 AM, Mike Archbold via AGI < > [email protected]> > > > wrote: > > > > > > > Bultlerian, named after a guy from Stanwood, WA. I'm not far from > > there, actually, and there is a beautiful old Scandinavian farming > > community there, with falling down barns and images of tall blonde > > girls. > > > > A woman in the building I live in told me I have to find Jesus right > > away after she took a look at my book, presently at position about > > 5,000,000 on amazon. If I don't find Jesus right away it is all over. > > > > What a strange world. > > > > >> > > >> > http://en.yibada.com/articles/19837/20150316/humans-hold-anti-ai-robot-protest-sxsw-texas.htm > > >> > > >> I find this kind of mind blowing. Down with robots? Down with AI? > > >> > > >> > > >> ------------------------------------------- > > >> AGI > > >> Archives: https://www.listbox.com/member/archive/303/=now > > >> RSS Feed: > > >> https://www.listbox.com/member/archive/rss/303/23050605-2da819ff > > >> Modify Your Subscription: > > >> https://www.listbox.com/member/?& > > >> Powered by Listbox: http://www.listbox.com > > >> > > > > > > > > > ------------------------------------------- > > AGI > > Archives: https://www.listbox.com/member/archive/303/=now > > RSS Feed: > https://www.listbox.com/member/archive/rss/303/19999924-4a978ccc > > Modify Your Subscription: https://www.listbox.com/member/?& > > Powered by Listbox: http://www.listbox.com > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/212726-deec6279> | Modify > <https://www.listbox.com/member/?&> Your Subscription > <http://www.listbox.com> > > > > > -- > Ben Goertzel, PhD > http://goertzel.org > > "The reasonable man adapts himself to the world: the unreasonable one > persists in trying to adapt the world to himself. Therefore all progress > depends on the unreasonable man." -- George Bernard Shaw > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/26879140-5b8435c3> | > Modify <https://www.listbox.com/member/?&> Your Subscription > <http://www.listbox.com> > > > > > -- > Regards > > Calum > > > > > -- > Ben Goertzel, PhD > http://goertzel.org > > "The reasonable man adapts himself to the world: the unreasonable one > persists in trying to adapt the world to himself. Therefore all progress > depends on the unreasonable man." -- George Bernard Shaw > > > > > -- > Regards > > Calum > > > > > -- > Ben Goertzel, PhD > http://goertzel.org > > "The reasonable man adapts himself to the world: the unreasonable one > persists in trying to adapt the world to himself. Therefore all progress > depends on the unreasonable man." -- George Bernard Shaw > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/26941503-0abb15dc> | > Modify <https://www.listbox.com/member/?&> Your Subscription > <http://www.listbox.com> > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/10443978-6f4c28ac> | > Modify > <https://www.listbox.com/member/?&> > Your Subscription <http://www.listbox.com> > -- Full employment can be had with the stoke of a pen. Simply institute a six hour workday. That will easily create enough new jobs to bring back full employment. ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657 Powered by Listbox: http://www.listbox.com
