On Monday, October 15, 2018 at 10:49:44 PM UTC-5, Brent wrote:
>
>
>
> On 10/15/2018 5:06 PM, Philip Thrift wrote:
>
>
>
> On Monday, October 15, 2018 at 6:45:11 PM UTC-5, Brent wrote: 
>>
>>
>>
>> On 10/14/2018 11:13 PM, Philip Thrift wrote:
>>
>>
>>
>> On Sunday, October 14, 2018 at 9:53:07 PM UTC-5, Brent wrote: 
>>>
>>>
>>>
>>> On 10/14/2018 2:48 PM, John Clark wrote:
>>>
>>> >*And there are sound reasons for doubting the consciousness of 
>>>> computers -*
>>>>
>>>  
>>> Name one of them that could not also be used to doubt the consciousness 
>>> of your fellow human beings.
>>>
>>>
>>> The reason for not doubting that other human beings are conscious is 
>>> that (1) I am conscious and (2) other human beings are made of the same 
>>> stuff in approximately the same way that I am and (3) they behave the same 
>>> way in relation to what I am conscious of, e.g. they jump at a sudden loud 
>>> sound.
>>>
>>> Brent
>>>
>>
>>
>> The thought crossed my mind yesterday: I was helping a young man applying 
>> for a Ph.D. program in chemical engineering with his application, and we 
>> were talking about chemistry and consciousness*, and I mentioned a type of 
>> zombie - a being that could converse (like an advanced Google Assistant or 
>> Sophie Robot) but not be conscious - and I thought it was *possible* he was 
>> a zombie.
>>
>> * cf. 
>> *Experience processing*
>> https://codicalist.wordpress.com/2018/10/14/experience-processing/
>>
>>
>> I suppose you've read Scott Aaronson's take down of Tononi's theory.  So 
>> I wonder why you would reference Tononi.
>>
>> One problem with the "experience is primary" theory is that there's no 
>> way for it to evolve.  If it's a property of matter why are organized 
>> information processing lumps of matter more capable of experience than 
>> unorganized lumps of the same composition...the obvious answer is that the 
>> former process information, and processing information is something natural 
>> selection can work on.  Smart animals reproduce better.  Animals with 
>> experiences...who cares?
>>
>> “Emotional-BDI agents are BDI agents whose behavior is guided not only by 
>> beliefs, desires and intentions, but also by the role of emotions in 
>> reasoning and decision-making."  This makes a false assumption that 
>> emotions are something independent of beliefs, desire, intentions, 
>> reasoning, and decision making.  But this says nothing about the 
>> satisfaction and thwarting of desires and intentions.  Why are those enough 
>> to explain emotions.  I agree that emotions are necessary for reasoning in 
>> the sense that emotions are the value-weights given to events, including 
>> those imagined by foresight, and that some values are primitive.  
>>
>> I think it is false that "Purely informational processing, which includes 
>> intentional agent programming (learning from experience, self-modeling), 
>> does not capture all true experiential processing (phenomenal 
>> consciousness). "  It is a cheat to put in "purely".  In fact all learning 
>> and intentional planning must include weighing alternatives and assigning 
>> value/emotion to them.  I don't see any need for a further primitive 
>> modality.  For example, a feeling of dizziness is a failure to maintain 
>> personal spacial orientation which is a value at a very low (subconscious) 
>> level.  Sure there are feelings and emotions...but I think they are all 
>> derivative from more primitive values that are derivative from evolution.
>>
>> I think the reason you are attracted to this idea is that it is closed 
>> within the computer/information/program frame.  And that is why I use the 
>> example of the AI Mars Rover.  Sure, emotion cannot be derived within a 
>> computer.  Emotion is something useful to a robot, an AI that works and 
>> strives within an external world which also acts on it.
>>
>> Brent
>>
>>
>>
>>
> I'll include the reference to
>
>    *Why I Am Not An Integrated Information Theorist (or, The Unconscious 
> Expander)*
>    https://www.scottaaronson.com/blog/?p=1799
>
> Since I say that IIT is still in the *information-oriented paradigm *and 
> not in the *experience-oriented paradigm*, Aaronson's post helps my case.
>
> I don't quite follow the rest. A person may feel pleasure (a modality) 
> without reasoning "I need to feel pleasure now."
>
>
> It's not a question of what it's possible to feel, but whether that it can 
> be accounted for by information processing.  "I feel pleasure now." may 
> well be a function of specific perceptions, values, and reasoning about 
> them.  I don't see any attempt to prove this cannot be the case.  It seems 
> that helping yourself to a primitive "experience" built into matter is just 
> baseless speculation unless you have some project to measure or 
> characterize this experience and show how it interacts with 
> information...because we certainly know that information can give pain or 
> pleasure.
>
> Brent
>

 

Information can give pain or pleasure, like watching Fox News vs. a cuddly 
nature show.

But here is the thesis I think of the experience-oriented (vs. 
information-oriented) paradigm*: *Experience cannot be represented: It does 
not exist outside of its material instantiation.*


* and why a biocomputer is a new kind of computer

- pt

-- 
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To post to this group, send email to [email protected].
Visit this group at https://groups.google.com/group/everything-list.
For more options, visit https://groups.google.com/d/optout.

Reply via email to