On Tue, Feb 10, 2015 at 12:59 PM, Telmo Menezes <[email protected]>
wrote:

>
>
>
> On Tue, Feb 10, 2015 at 6:21 PM, Jason Resch <[email protected]> wrote:
>
>>
>>
>> On Tue, Feb 10, 2015 at 12:04 PM, Telmo Menezes <[email protected]>
>> wrote:
>>
>>>
>>>
>>> On Tue, Feb 10, 2015 at 4:47 PM, Jason Resch <[email protected]>
>>> wrote:
>>>
>>>> If you define increased intelligence as decreased probability of having
>>>> a false belief on any randomly chosen proposition, then superintelligences
>>>> will be wrong on almost nothing, and their beliefs will converge as their
>>>> intelligence rises. Therefore nearly all superintelligences will operate
>>>> according to the same belief system. We should stop worrying about trying
>>>> to ensure friendly AI, it will either be friendly or it won't according to
>>>> what is right.
>>>>
>>>
>>> I wonder if this isn't prevented by Gödel's incompleteness. Given that
>>> the superintelligence can never be certain of its own consistency, it must
>>> remain fundamentally agnostic. In this case, we might have different
>>> superintelligences working under different hypothesis, possibly occupying
>>> niches just like what happens with Darwinism.
>>>
>>
>> Interesting point. Yes a true super intelligence may never perform any
>> actions, as its trapped in never being certain (and knowing it never can be
>> certain) that its actions are right. Fitness for survival may play some
>> role in how intelligent active agents can be before they become inactive.
>>
>
> Yes, that's an interesting way to put it. I wonder.
>
>
>>
>>
>>>
>>>
>>>>
>>>> I think chances are that it will be friendly, since I happen to believe
>>>> in universal personhood, and if that belief is correct, then
>>>> superintelligences will also come to believe it is correct. And with the
>>>> belief in universal personhood it would know that harm to others is harm to
>>>> the self.
>>>>
>>>
>>> I agree with you, with the difference that I try to assume universal
>>> personhood without believing in it, to avoid becoming a religious
>>> fundamentalist.
>>>
>>>
>> Interesting. Why do you think having beliefs can lead to religious
>> fundamentalism. Would you not say you belief the Earth is round? Could such
>> a belief lead to religious fundamentalism and if not why not?
>>
>
> This leads us back to a recurring discussion on this mailing list. I would
> say that you can believe the Earth to be round in the informal sense of the
> word: your estimation of the probability that the earth is round is very
> close to one. I don't think you can believe the earth to be round with 100%
> certainty without falling into religious fundamentalism. This implies a
> total belief in your senses, for example. That is a strong position about
> the nature of reality that is not really backed up by anything. Just like
> believing literally in the Bible or the Quran or Atlas Shrugged.
>
>
I see. I did not mean it in the sense of absolute certitude, merely that
universal personhood is one of my current working hypotheses derived from
my consideration of various problems of personal identity.

Jason

-- 
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To post to this group, send email to [email protected].
Visit this group at http://groups.google.com/group/everything-list.
For more options, visit https://groups.google.com/d/optout.

Reply via email to