On Monday, October 15, 2018 at 6:45:11 PM UTC-5, Brent wrote:
>
>
>
> On 10/14/2018 11:13 PM, Philip Thrift wrote:
>
>
>
> On Sunday, October 14, 2018 at 9:53:07 PM UTC-5, Brent wrote: 
>>
>>
>>
>> On 10/14/2018 2:48 PM, John Clark wrote:
>>
>> >*And there are sound reasons for doubting the consciousness of 
>>> computers -*
>>>
>>  
>> Name one of them that could not also be used to doubt the consciousness 
>> of your fellow human beings.
>>
>>
>> The reason for not doubting that other human beings are conscious is that 
>> (1) I am conscious and (2) other human beings are made of the same stuff in 
>> approximately the same way that I am and (3) they behave the same way in 
>> relation to what I am conscious of, e.g. they jump at a sudden loud sound.
>>
>> Brent
>>
>
>
> The thought crossed my mind yesterday: I was helping a young man applying 
> for a Ph.D. program in chemical engineering with his application, and we 
> were talking about chemistry and consciousness*, and I mentioned a type of 
> zombie - a being that could converse (like an advanced Google Assistant or 
> Sophie Robot) but not be conscious - and I thought it was *possible* he was 
> a zombie.
>
> * cf. 
> *Experience processing*
> https://codicalist.wordpress.com/2018/10/14/experience-processing/
>
>
> I suppose you've read Scott Aaronson's take down of Tononi's theory.  So I 
> wonder why you would reference Tononi.
>
> One problem with the "experience is primary" theory is that there's no way 
> for it to evolve.  If it's a property of matter why are organized 
> information processing lumps of matter more capable of experience than 
> unorganized lumps of the same composition...the obvious answer is that the 
> former process information, and processing information is something natural 
> selection can work on.  Smart animals reproduce better.  Animals with 
> experiences...who cares?
>
> “Emotional-BDI agents are BDI agents whose behavior is guided not only by 
> beliefs, desires and intentions, but also by the role of emotions in 
> reasoning and decision-making."  This makes a false assumption that 
> emotions are something independent of beliefs, desire, intentions, 
> reasoning, and decision making.  But this says nothing about the 
> satisfaction and thwarting of desires and intentions.  Why are those enough 
> to explain emotions.  I agree that emotions are necessary for reasoning in 
> the sense that emotions are the value-weights given to events, including 
> those imagined by foresight, and that some values are primitive.  
>
> I think it is false that "Purely informational processing, which includes 
> intentional agent programming (learning from experience, self-modeling), 
> does not capture all true experiential processing (phenomenal 
> consciousness). "  It is a cheat to put in "purely".  In fact all learning 
> and intentional planning must include weighing alternatives and assigning 
> value/emotion to them.  I don't see any need for a further primitive 
> modality.  For example, a feeling of dizziness is a failure to maintain 
> personal spacial orientation which is a value at a very low (subconscious) 
> level.  Sure there are feelings and emotions...but I think they are all 
> derivative from more primitive values that are derivative from evolution.
>
> I think the reason you are attracted to this idea is that it is closed 
> within the computer/information/program frame.  And that is why I use the 
> example of the AI Mars Rover.  Sure, emotion cannot be derived within a 
> computer.  Emotion is something useful to a robot, an AI that works and 
> strives within an external world which also acts on it.
>
> Brent
>
>
>
>
I'll include the reference to

   *Why I Am Not An Integrated Information Theorist (or, The Unconscious 
Expander)*
   https://www.scottaaronson.com/blog/?p=1799

Since I say that IIT is still in the *information-oriented paradigm *and 
not in the *experience-oriented paradigm*, Aaronson's post helps my case.

I don't quite follow the rest. A person may feel pleasure (a modality) 
without reasoning "I need to feel pleasure now."

- pt



 

-- 
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To post to this group, send email to [email protected].
Visit this group at https://groups.google.com/group/everything-list.
For more options, visit https://groups.google.com/d/optout.

Reply via email to