You're focusing on the "Tool AI" aspect of Holden's argument, which is only
one part of it.   Not all of his arguments rely on his point about Tool AI;
his other arguments about the utility of SI as an organization are largely
independent of his Tool AI point.

Regarding the potential power of Tool AI, I would say that's unclear.  The
"Very Powerful Optimization Process" that Eliezer has posited as part of
one version of CEV, is in fact a kind of Tool AI, it seems to me.   That
VPOP seems precisely a Tool AI designed to guide and possibly avert the
development of powerful Agent AI.

Personally, I fear powerful Tool AI more than I fear powerful Agent AI,
because it's humans who will be using the tools, and the propensity of a
certain percentage of humans to use powerful tools for harm against other
humans is well known.  That is: the dangers of Tool AI are obvious, whereas
the dangers of Agent AI are more confusing at present.

However, it's a fair point that SI is obsessing more about the potential
dangers of self-modifying Agent AI, and largely side-stepping (in its
public materials and discussions so far, anyway) the more obvious dangers
of Tool AI in the hands of power-hungry or malevolent humans.

-- Ben G

On Thu, Aug 23, 2012 at 7:52 AM, Michael Anissimov
<[email protected]>wrote:

> Discerning, but wrong...all based on the assumption that there can be
> sophisticated AI that behaves like Google Maps, which is flawed.
>
> On Tue, Aug 21, 2012 at 8:07 PM, Ben Goertzel <[email protected]> wrote:
>
>>
>> Hmmm... the reply on Tool AI is interesting, but Holden's original
>> critique of SI is also worth reading:
>>
>> http://lesswrong.com/lw/cbs/thoughts_on_the_singularity_institute_si/
>>
>> It's a good deal more discerning than Hugo's critique, I'd say...
>>
>>
>> ben
>>
>> On Tue, Aug 21, 2012 at 6:43 PM, Michael Anissimov <
>> [email protected]> wrote:
>>
>>> For those interested in current Singularity Institute research:
>>>
>>> http://singularity.org/research/
>>>
>>> Also possibly of interest, our executive director's recent reply to
>>> Holden Karnofsky on what he calls "tool AI":
>>>
>>> http://lesswrong.com/lw/cze/reply_to_holden_on_tool_ai/
>>>
>>> We have recently hired two research fellows, Alex Altair and Kaj Sotala.
>>> Both are exclusively focused on AI research.
>>>
>>> --
>>> Michael Anissimov
>>> Singularity Institute
>>> www.singularity.org
>>>
>>>    *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
>>> <https://www.listbox.com/member/archive/rss/303/212726-11ac2389> |
>>> Modify <https://www.listbox.com/member/?&;> Your Subscription
>>> <http://www.listbox.com>
>>>
>>
>>
>>
>> --
>> Ben Goertzel, PhD
>> http://goertzel.org
>>
>> "My humanity is a constant self-overcoming" -- Friedrich Nietzsche
>>
>>    *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
>> <https://www.listbox.com/member/archive/rss/303/3787904-e0bccf1e> |
>> Modify <https://www.listbox.com/member/?&;> Your Subscription
>> <http://www.listbox.com>
>>
>
>
>
> --
> Michael Anissimov
> Singularity Institute
> www.singularity.org
>
>    *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
> <https://www.listbox.com/member/archive/rss/303/212726-11ac2389> | 
> Modify<https://www.listbox.com/member/?&;>Your Subscription
> <http://www.listbox.com>
>



-- 
Ben Goertzel, PhD
http://goertzel.org

"My humanity is a constant self-overcoming" -- Friedrich Nietzsche



-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-c97d2393
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-2484a968
Powered by Listbox: http://www.listbox.com

Reply via email to