On Wed, Aug 27, 2014 at 3:49 PM, meekerdb <[email protected]> wrote:

>  On 8/27/2014 4:53 AM, Terren Suydam wrote:
>
>     You're talking about an AI that arrives at novel solutions, which
>>> requires the ability to invent/simulate/act on new models in new domains
>>> (AGI).
>>>
>>
>>   Evolutionary computation already achieves novelty and invention, to a
>> degree. I concur that it is still not AGI. But it could already be a
>> threat, given enough computational resources.
>>
>
>  AGI is a threat because it's utility function would necessarily be
> sufficiently "meta" that it could create novel sub-goals. We would not
> necessarily be able to control whether it chose a goal that was compatible
> with ours.
>
>
> On the other hand we're not that good at choosing goals for ourselves -
> e.g. ISIS has chosen the goal of imposing a ruthless religious tyranny.
>
> Brent
>
>
The quality of the goal system is not what defines intelligence (though it
may suffice to define wisdom).

Terren

-- 
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To post to this group, send email to [email protected].
Visit this group at http://groups.google.com/group/everything-list.
For more options, visit https://groups.google.com/d/optout.

Reply via email to