On Wed, Feb 11, 2015 at 10:51 PM, meekerdb <[email protected]> wrote:

>  On 2/11/2015 7:50 AM, Jason Resch wrote:
>
>
>
> On Wed, Feb 11, 2015 at 4:25 AM, Stathis Papaioannou <[email protected]>
> wrote:
>
>>
>>
>> On Wednesday, February 11, 2015, Jason Resch <[email protected]>
>> wrote:
>>
>>>
>>>
>>> On Tue, Feb 10, 2015 at 8:15 PM, Stathis Papaioannou <[email protected]
>>> > wrote:
>>>
>>>>
>>>>
>>>> On Wednesday, February 11, 2015, Jason Resch <[email protected]>
>>>> wrote:
>>>>
>>>>>
>>>>>
>>>>> On Tue, Feb 10, 2015 at 3:30 PM, Stathis Papaioannou <
>>>>> [email protected]> wrote:
>>>>>
>>>>>>
>>>>>>
>>>>>> On Wednesday, February 11, 2015, Jason Resch <[email protected]>
>>>>>> wrote:
>>>>>>
>>>>>>> If you define increased intelligence as decreased probability of
>>>>>>> having a false belief on any randomly chosen proposition, then
>>>>>>> superintelligences will be wrong on almost nothing, and their beliefs 
>>>>>>> will
>>>>>>> converge as their intelligence rises. Therefore nearly all
>>>>>>> superintelligences will operate according to the same belief system. We
>>>>>>> should stop worrying about trying to ensure friendly AI, it will either 
>>>>>>> be
>>>>>>> friendly or it won't according to what is right.
>>>>>>>
>>>>>>>  I think chances are that it will be friendly, since I happen to
>>>>>>> believe in universal personhood, and if that belief is correct, then
>>>>>>> superintelligences will also come to believe it is correct. And with the
>>>>>>> belief in universal personhood it would know that harm to others is 
>>>>>>> harm to
>>>>>>> the self.
>>>>>>>
>>>>>>
>>>>>>  Having accurate beliefs about the world and having goals are two
>>>>>> unrelated things. If I like stamp collecting, being intelligent will help
>>>>>> me to collect stamps, it will help me see if stamp collecting clashes 
>>>>>> with
>>>>>> a higher priority goal, but it won't help me decide if my goals are 
>>>>>> worthy.
>>>>>>
>>>>>>
>>>>>>
>>>>>  Were all your goals set at birth and driven by biology, or are some
>>>>> of your goals based on what you've since learned about the world? Perhaps
>>>>> learning about universal personhood (for example), could lead one to
>>>>> believe that charity is a worthy goal, and perhaps deserving of more time
>>>>> than collecting stamps.
>>>>>
>>>>
>>>>   The implication is that if you believe in universal personhood then
>>>> even if you are selfish you will be motivated towards charity. But the
>>>> selfishness itself, as a primary value, is not amenable to rational
>>>> analysis. There is no inconsistency in a superintelligent AI that is
>>>> selfish, or one that is charitable, or one that believes the single most
>>>> important thing in the world is to collect stamps.
>>>>
>>>>
>>>>
>>>  But doing something well (regardless of what it is) is almost always
>>> improved by having greater knowledge, so would not gathering greater
>>> knowledge become a secondary sub goal for nearly any supintelligence that
>>> has goals? Is it impossible that it might discover and decide to pursue
>>> other goals during that time? After all, capacity to change one's mine
>>> seems to be a requirement for any intelligence process, or any process on
>>> the path towards superintelligence.
>>>
>>
>>   Sure, but the AI may still decide to do evil, perverse or self
>> destructive things. There is no contradiction in superintelligence behaving
>> this way.
>>
>>
>>
>  It's an assumption to say there is no contradiction. If it's beliefs are
> defined to be almost completely correct, why would its actions not follow
> its beliefs and also be almost completely correct?
>
>
> What does "correct" mean in this context?  Instrumentally correct, i.e.
> well chosen to achieve it's goals?  Or does it mean agreeing with Jason
> Resch's liberal humanist values?
>
>
Interesting description of my values.

By correct I mean in alignment with truth.

Jason

-- 
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To post to this group, send email to [email protected].
Visit this group at http://groups.google.com/group/everything-list.
For more options, visit https://groups.google.com/d/optout.

Reply via email to