On Sat, Sep 17, 2016 at 11:32 PM, John Clark <[email protected]> wrote:
> On Fri, Sep 16, 2016  Telmo Menezes <[email protected]> wrote:
>
>>
>> >
>> it is not so clear to me that it has no intelligence. It
>> leads to better and better designs,
>
>
> If Evolution has intelligence it sure doesn't have much, despite the
> resources of the entire planet at its disposal Evolution was incredibly slow
> and sloppy in getting its work done; but it was the only way complex things
> could get made until it finally invented brains after 3 billion years of
> screwing around.
>
>>
>> >
>> my point is that what makes this a "bad idea" is our own
>> evolutionary context. Why is it a bad idea exactly? It becomes
>>
>> tautological. It is a bad idea because it hurts your ability to pursue
>> goals dictated by evolution.
>
>
> The only goal dictated by Evolution is a vague command to figure out a way
> to get your genes into the next generation, and just like the vague laws
> that humans make that leaves plenty of room for interpretation, unintended
> consequences, and loopholes.
>
>>
>> >
>> Outside of evolution, why would an entity
>> not choose the easy way out?
>
>
> Any intelligence regardless of if was produced by Evolution or by human
> engineers or even by God would have a constantly shifting goal structure
> depending on environmental circumstances. Even your iPhone has, in a sense,
> a goal to continue working, so personal survival would certainly be a goal
> for any brain even if it isn't guaranteed the permanent #1 position. Just
> like with people I expect that with a AI survival would usually, although
> not always, take the #1 slot.  But for some some AIs just like with some
> people, the mindless pursuit pleasure could be #1 , but I don't see why that
> would be more common in a AI than it is in a person unless the electronic
> version is far more powerful and addictive than the chemical we call crack.
> But could the electronic version really be that much more powerful and
> additive? I don't know, maybe it's possible. If it is then that explains the
> Fermi Paradox.
>
> Although we're using different language I think we may agree more than we
> disagree,

Indeed.

Telmo.

>
>  John K Clark
>
>
>
> --
> You received this message because you are subscribed to the Google Groups
> "Everything List" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to [email protected].
> To post to this group, send email to [email protected].
> Visit this group at https://groups.google.com/group/everything-list.
> For more options, visit https://groups.google.com/d/optout.

-- 
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To post to this group, send email to [email protected].
Visit this group at https://groups.google.com/group/everything-list.
For more options, visit https://groups.google.com/d/optout.

Reply via email to