On 8/27/2014 4:53 AM, Terren Suydam wrote:
You're talking about an AI that arrives at novel solutions, which
requires the
ability to invent/simulate/act on new models in new domains (AGI).
Evolutionary computation already achieves novelty and invention, to a
degree. I
concur that it is still not AGI. But it could already be a threat, given
enough
computational resources.
AGI is a threat because it's utility function would necessarily be sufficiently "meta"
that it could create novel sub-goals. We would not necessarily be able to control
whether it chose a goal that was compatible with ours.
On the other hand we're not that good at choosing goals for ourselves - e.g. ISIS has
chosen the goal of imposing a ruthless religious tyranny.
Brent
--
You received this message because you are subscribed to the Google Groups
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email
to [email protected].
To post to this group, send email to [email protected].
Visit this group at http://groups.google.com/group/everything-list.
For more options, visit https://groups.google.com/d/optout.