On Wed, Feb 11, 2015 at 5:42 AM, meekerdb <[email protected]> wrote:

> On 2/10/2015 6:15 PM, Stathis Papaioannou wrote:
>
>> The implication is that if you believe in universal personhood then even
>> if you are selfish you will be motivated towards charity.
>>
>
> If humans are any indication, a super-intelligence will be incredibly good
> at rationalizing what it wants to do.  For example, if personhood is
> universal then what's good for me is good for the human race.
>

Not necessarily. If personhood is universal then your pleasure is my
pleasure, so the conclusion could be:
Do unto others as they want done to them.

Telmo


>
> Brent
>
>  But the selfishness itself, as a primary value, is not amenable to
>> rational analysis. There is no inconsistency in a superintelligent AI that
>> is selfish, or one that is charitable, or one that believes the single most
>> important thing in the world is to collect stamps.
>>
>
> --
> You received this message because you are subscribed to the Google Groups
> "Everything List" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to [email protected].
> To post to this group, send email to [email protected].
> Visit this group at http://groups.google.com/group/everything-list.
> For more options, visit https://groups.google.com/d/optout.
>

-- 
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To post to this group, send email to [email protected].
Visit this group at http://groups.google.com/group/everything-list.
For more options, visit https://groups.google.com/d/optout.

Reply via email to