On Wed, Aug 27, 2014 at 3:49 PM, meekerdb <[email protected]> wrote:
> On 8/27/2014 4:53 AM, Terren Suydam wrote: > > You're talking about an AI that arrives at novel solutions, which >>> requires the ability to invent/simulate/act on new models in new domains >>> (AGI). >>> >> >> Evolutionary computation already achieves novelty and invention, to a >> degree. I concur that it is still not AGI. But it could already be a >> threat, given enough computational resources. >> > > AGI is a threat because it's utility function would necessarily be > sufficiently "meta" that it could create novel sub-goals. We would not > necessarily be able to control whether it chose a goal that was compatible > with ours. > > > On the other hand we're not that good at choosing goals for ourselves - > e.g. ISIS has chosen the goal of imposing a ruthless religious tyranny. > > Brent > > The quality of the goal system is not what defines intelligence (though it may suffice to define wisdom). Terren -- You received this message because you are subscribed to the Google Groups "Everything List" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To post to this group, send email to [email protected]. Visit this group at http://groups.google.com/group/everything-list. For more options, visit https://groups.google.com/d/optout.

