Steve

I sympathise with your very understandable preference not to be targeted by
anti-AI crazies!

What do you think is the best way to try and shape the growing public
debate about AGI?  Following Bostrom's book, and the comments by Hawking,
Musk and Gates, a fair proportion of the general public is now aware that
AGI might arrive in the medium term, and that it will have a very big
impact.

Some AI researchers seem to be responding by saying, "Don't worry, it can't
happen for centuries, if ever".  No doubt some of them genuinely believe
that, but I wonder whether some are saying it in the (forlorn?) hope the
debate will go away. It won't.  In fact I suspect that the new Avengers
movie will kick it up a level.

Others are saying, "Don't worry, AGI cannot and will not harm humans."  To
my mind (and I realise I may be in a small minority here on this) that is
hard to be certain about - as Bostrom demonstrated.

Yet others are saying, "AI researcher will solve the problem long before
AGI arrives, and it's best not to worry everyone else in the meantime."
 That seems a dangerous approach to me.  If the public ever feels (rightly
or wrongly) that things have been hidden from them, they will react badly.

But I do definitely sympathise with the desire not to be targeted by
crazies, or to be vilified by journalists who have half-understood the
situation!

I am not an AI researcher.  I am a retired businessman now resuming a
lapsed career as a writer, and my subject is AGI.  My first novel,
Pandora's Brain, is just out.  I chose the subject because I think it is
the most important one in the world, bar none.

The most likely outcome of my writing, of course, is zilch.  Half a million
new books are published in the US and the UK every year, and it would be
arrogant to think mine will stand out.  But I work hard on them, and one
can dream.

In the unlikely event that more than a handful of people will read Pandora,
I want to make a responsible contribution to the debate.  That is why I
chose the subject in the first place.

Having said all this, I'll understand if Ben just kicks me off the forum!

Regards
Calum



On 17 March 2015 at 08:36, Steve Richfield via AGI <[email protected]> wrote:

> The BIG problem that threatens the safety of everyone on this forum are
> crazies like some on this forum who publicly (this AGI forum is public and
> is Google searchable) claim that potentially dangerous AGIs will be up in
> running in 5 years or whenever. Regardless of the truth or lack thereof,
> movies (like the Terminator series) portraying the great social value of
> KILLING people who develop AGIs is enough to motivate other crazies out
> there to do really bad things to the crazies who make such statements.
>
> Which of the crazies are more dangerous TO ME. That is easy - it is the
> crazies on this forum, who might guide other crazies to my door. There are
> LOTS of murderous crazies in our world, which I can live with, so long as
> no one is guiding them TO ME.
>
> It is really hard to believe that ANYONE who is SO stupid to be publicly
> making such obviously dangerous statements actually thinks they are smart
> enough to participate in making a genuine AGI. There is NO WAY that this
> could possibly happen. Stupid is forever.
>
> Personally, I have absolutely NO expectation of any sort of dangerous AGIs
> emerging anytime soon, most especially from people who are stupid enough to
> make such obviously dangerous public statements.
>
> The less mentally impaired participants on this forum are quietly working
> on their respective theories WITHOUT making such incendiary statements.
>
> If you the reader have made such public statements in the past, then GIVE
> UP working on AGI technology, because you have failed the intelligence test
> and are obviously NOT smart enough to ever succeed. Further, your potential
> future colleagues will see you as being too dangerous to have as an
> associate.
>
> If you are an assassin looking for someone to kill, then look elsewhere.
> These guys aren't worth your ammunition.
>
> *BEN*: As an act of social responsibility, you should purge ALL postings,
> including this posting, that mentions any sort of short time horizon for
> AGI development, and go all the way back to the beginning of this forum to
> do so. That demonstrations like the one in Texas can even come about shows
> the pushback that such statements can attract. Do we really need to be seen
> as dangerous social pariahs?
>
> Steve
> ==================
>
>
> On Mon, Mar 16, 2015 at 7:21 PM, Nanograte Knowledge Technologies via AGI
> <[email protected]> wrote:
>
>> If I may say something pelase? To my understanding, Google would and Elon
>> Musk would.
>>
>> However, AI is not the real threat. In my most-humble opinion, it is the
>> key to the solution to a real threat. The technology would still be
>> developed, regardless, and is probably being hastened as we speak. Perhaps
>> the following tenets should be applied to "their" rationale: "The cat's out
>> of the bag. What they don't know, won't harm them. Let's just go
>> underground and hurry it up some more." As such then, protests are
>> insignificant, red herrings. Protests could be staged to support public
>> statements, as a distraction and disabling strategy to detract from the
>> real issues at hand. Who knows? Who cares? Most people don't even read. I
>> agree more with the 5-year outlook, and it could even become 4, depending
>> on how quickly the key constraints to such progress could be resolved by
>> people like us.
>>
>> ------------------------------
>> Date: Tue, 17 Mar 2015 08:39:22 +0800
>> Subject: Re: [agi] AI Protest in Texas
>> From: [email protected]
>> To: [email protected]
>> CC: [email protected]
>>
>>
>>
>> yeah, that's more consistent with what I've heard from Demis in the
>> past...
>>
>>
>>
>> On Tue, Mar 17, 2015 at 8:30 AM, Calum Chace <[email protected]> wrote:
>>
>> Sorry, Ben, it wasn't centuries for Hassabis.  It was decades.  Rather an
>> important difference!
>>
>> Last year, the American entrepreneur, Elon Musk, one of Deep Mind’s early
>> investors, described AI as humanity’s greatest existential threat. “Unless
>> you have direct exposure to groups like Deepmind, you have no idea how fast
>> [AI] is growing,” he said. “The risk of something seriously dangerous
>> happening is in the five year timeframe. Ten years at most.”
>>
>> However, the Google team played down the concerns. “We agree with him
>> there are risks that need to be borne in mind, but we’re decades away from
>> any sort of technology that we need to worry about,” Hassabis said.
>>
>>
>> http://www.theguardian.com/technology/2015/feb/25/google-develops-computer-program-capable-of-learning-tasks-independently
>>
>> Calum
>>
>> On 17 March 2015 at 01:20, Ben Goertzel <[email protected]> wrote:
>>
>>
>> Did Demis really say AGI is hundreds of years away?   That surprises
>> me....
>>
>> I think Ng actually believes AGI is far off, he's conservative but I
>> believe he's a straight shooter.
>>
>> I don't know Yann and Christof F2F so I don't have a strong opinion on
>> their attitudes...
>>
>> -- Ben
>>
>> On Tue, Mar 17, 2015 at 8:13 AM, Calum Chace <[email protected]> wrote:
>>
>> Yes, but Austin, of all places.
>>
>> Ben, why do you think Yann LeCun, Andrew Ng, Christof Koch and Demis
>> Hassabis have all been lining up to say that AGI is hundreds of years
>> away?  Are they worried about this sort of reaction?
>>
>> On 17 March 2015 at 01:10, Ben Goertzel via AGI <[email protected]> wrote:
>>
>>
>> And of course it has to be in Texas 8-D ...
>>
>> On Tue, Mar 17, 2015 at 5:17 AM, Piaget Modeler via AGI <[email protected]>
>> wrote:
>>
>> Straight out of Stephen Spielberg's film: A.I.
>>
>> ~PM
>>
>> > Date: Mon, 16 Mar 2015 12:59:25 -0700
>> > Subject: Re: [agi] AI Protest in Texas
>> > From: [email protected]
>> > To: [email protected]
>>
>> >
>> > On 3/16/15, Aaron Hosford <[email protected]> wrote:
>> > > This sort of thing was predicted 50 years ago.
>> > > http://en.wikipedia.org/wiki/Butlerian_Jihad
>> > >
>> > > Nonetheless, yes, mind blowing.
>> > >
>> > > On Mon, Mar 16, 2015 at 11:41 AM, Mike Archbold via AGI <
>> [email protected]>
>> > > wrote:
>> > >
>> >
>> > Bultlerian, named after a guy from Stanwood, WA. I'm not far from
>> > there, actually, and there is a beautiful old Scandinavian farming
>> > community there, with falling down barns and images of tall blonde
>> > girls.
>> >
>> > A woman in the building I live in told me I have to find Jesus right
>> > away after she took a look at my book, presently at position about
>> > 5,000,000 on amazon. If I don't find Jesus right away it is all over.
>> >
>> > What a strange world.
>> >
>> > >>
>> > >>
>> http://en.yibada.com/articles/19837/20150316/humans-hold-anti-ai-robot-protest-sxsw-texas.htm
>> > >>
>> > >> I find this kind of mind blowing. Down with robots? Down with AI?
>> > >>
>> > >>
>> > >> -------------------------------------------
>> > >> AGI
>> > >> Archives: https://www.listbox.com/member/archive/303/=now
>> > >> RSS Feed:
>> > >> https://www.listbox.com/member/archive/rss/303/23050605-2da819ff
>> > >> Modify Your Subscription:
>> > >> https://www.listbox.com/member/?&;
>> > >> Powered by Listbox: http://www.listbox.com
>> > >>
>> > >
>> >
>> >
>> > -------------------------------------------
>> > AGI
>> > Archives: https://www.listbox.com/member/archive/303/=now
>> > RSS Feed:
>> https://www.listbox.com/member/archive/rss/303/19999924-4a978ccc
>> > Modify Your Subscription: https://www.listbox.com/member/?&;
>> > Powered by Listbox: http://www.listbox.com
>>     *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
>> <https://www.listbox.com/member/archive/rss/303/212726-deec6279> | Modify
>> <https://www.listbox.com/member/?&;> Your Subscription
>> <http://www.listbox.com>
>>
>>
>>
>>
>> --
>> Ben Goertzel, PhD
>> http://goertzel.org
>>
>> "The reasonable man adapts himself to the world: the unreasonable one
>> persists in trying to adapt the world to himself. Therefore all progress
>> depends on the unreasonable man." -- George Bernard Shaw
>>    *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
>> <https://www.listbox.com/member/archive/rss/303/26879140-5b8435c3> |
>> Modify <https://www.listbox.com/member/?&;> Your Subscription
>> <http://www.listbox.com>
>>
>>
>>
>>
>> --
>> Regards
>>
>> Calum
>>
>>
>>
>>
>> --
>> Ben Goertzel, PhD
>> http://goertzel.org
>>
>> "The reasonable man adapts himself to the world: the unreasonable one
>> persists in trying to adapt the world to himself. Therefore all progress
>> depends on the unreasonable man." -- George Bernard Shaw
>>
>>
>>
>>
>> --
>> Regards
>>
>> Calum
>>
>>
>>
>>
>> --
>> Ben Goertzel, PhD
>> http://goertzel.org
>>
>> "The reasonable man adapts himself to the world: the unreasonable one
>> persists in trying to adapt the world to himself. Therefore all progress
>> depends on the unreasonable man." -- George Bernard Shaw
>>    *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
>> <https://www.listbox.com/member/archive/rss/303/26941503-0abb15dc> |
>> Modify <https://www.listbox.com/member/?&;> Your Subscription
>> <http://www.listbox.com>
>>    *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
>> <https://www.listbox.com/member/archive/rss/303/10443978-6f4c28ac> |
>> Modify <https://www.listbox.com/member/?&;> Your Subscription
>> <http://www.listbox.com>
>>
>
>
>
> --
> Full employment can be had with the stoke of a pen. Simply institute a six
> hour workday. That will easily create enough new jobs to bring back full
> employment.
>
>    *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
> <https://www.listbox.com/member/archive/rss/303/26879140-5b8435c3> |
> Modify
> <https://www.listbox.com/member/?&;>
> Your Subscription <http://www.listbox.com>
>



-- 
Regards

Calum



-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to