On Wed, Aug 27, 2014 at 3:49 PM, meekerdb <meeke...@verizon.net> wrote:

>  On 8/27/2014 4:53 AM, Terren Suydam wrote:
>
>     You're talking about an AI that arrives at novel solutions, which
>>> requires the ability to invent/simulate/act on new models in new domains
>>> (AGI).
>>>
>>
>>   Evolutionary computation already achieves novelty and invention, to a
>> degree. I concur that it is still not AGI. But it could already be a
>> threat, given enough computational resources.
>>
>
>  AGI is a threat because it's utility function would necessarily be
> sufficiently "meta" that it could create novel sub-goals. We would not
> necessarily be able to control whether it chose a goal that was compatible
> with ours.
>
>
> On the other hand we're not that good at choosing goals for ourselves -
> e.g. ISIS has chosen the goal of imposing a ruthless religious tyranny.
>
> Brent
>
>
The quality of the goal system is not what defines intelligence (though it
may suffice to define wisdom).

Terren

-- 
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to everything-list+unsubscr...@googlegroups.com.
To post to this group, send email to everything-list@googlegroups.com.
Visit this group at http://groups.google.com/group/everything-list.
For more options, visit https://groups.google.com/d/optout.

Reply via email to