Stathis Papaioannou wrote:
> 2009/4/24 Brent Meeker <meeke...@dslextreme.com>:
>   
>> Stathis Papaioannou wrote:
>>     
>>> 2009/4/23 Brent Meeker <meeke...@dslextreme.com>:
>>>
>>>       
>>>>> Say a machine is in two separate parts M1 and M2, and the information
>>>>> on M1 in state A is written to a punchcard, walked over to M2, loaded,
>>>>> and M2 goes into state B. Then what you are suggesting is that this
>>>>> sequence could give rise to a few moments of consciousness, since A
>>>>> and B are causally connected; whereas if M1 and M2 simply went into
>>>>> the same respective states A and B at random, this would not give rise
>>>>> to the same consciousness, since the states would not have the right
>>>>> causal connection. Right?
>>>>>
>>>>>           
>>>> Maybe.  But I'm questioning more than the lack of causal connection.
>>>> I'm questioning the idea that a static thing like a state can be
>>>> conscious.  That consciousness goes through a set of states, each one
>>>> being an "instant", is an inference we make in analogy with how we would
>>>> write a program simulating a mind.  I'm saying I suspect something
>>>> essential is missing when we "digitize" it in this way.  Note that this
>>>> does not mean I'd say "No" to Burno's doctor - because the doctor is
>>>> proposing to replace part of my brain with a mechanism that instantiates
>>>> a process - not just discrete states.
>>>>
>>>>
>>>> Brent
>>>>         
>>> What is needed for the series of states to qualify as a process?
>>>       
>> I think that the states, by themselves, cannot qualify.  The has to be 
>> something
>> else, a rule of inference, a causal connection, that joins them into a 
>> process.
>>
>>     
>>> I
>>> assume that a causal connection between the states, as in my example
>>> above, would be enough, since it is what happens in normal brains and
>>> computers.
>>>       
>> Yes, I certainly agree that it would be sufficient.  But it may be more than 
>> is
>> necessary.  The idea of physical causality isn't that well defined and it 
>> hardly
>> even shows up in fundamental physics except to mean no-action-at-a-distance.
>>
>>     
>>> But what would you say about the examples I give below,
>>> where the causal connection is disrupted in various ways: is there a
>>> process or is there just an unfeeling sequence of states?
>>>
>>>       
>>>>> But then you could come up with variations on this experiment where
>>>>> the transfer of information doesn't happen in as straightforward a
>>>>> manner. For example, what if the operator who walks over the punchcard
>>>>> gets it mixed up in a filing cabinet full of all the possible
>>>>> punchcards variations, and either (a) loads one of the cards into M2
>>>>> because he gets a special vibe about it and it happens to be the right
>>>>> one, or (b) loads all of the punchcards into M2 in turn so as to be
>>>>> sure that the right one is among them? Would the machine be conscious
>>>>> if the operator loads the right card knowingly, but not if he is just
>>>>> lucky, and not if he is ignorant but systematic? If so, how could the
>>>>> computation know about the psychological state of the operator?
>>>>>           
>> So you are contemplating a process that consists of a sequence of states and 
>> a
>> rule that connects them thus constituting a process: punch cards (states) 
>> and a
>> machine which physically implements some rule producing a new punch card 
>> (state)
>> from a previous one.  And then you ask whether it is still a process if, 
>> instead
>> of the rule producing the next state it is produced in some other way.  I'd 
>> say
>> so long as the rule is followed (the operator loads the right card knowingly)
>> it's the same process.  Otherwise it is not the same process (the operator
>> selects the right card by chance or by a different rule).  If the process is 
>> a
>> conscious one, is the latter still conscious?  I'd say that it is.  If the
>> selection is by chance it's an instance of a Boltzmann brain.  But I don't 
>> worry
>> about Boltzmann brains; they're to improbable.
>>     
>
> Boltzmann brains are improbable, but the example of the punchcards is
> not. The operator could have two punchcards in his pocket, have a
> conversation with someone on the way from M1 to M2 and end up
> forgetting or almost forgetting which is the right one. That is, his
> certainty of picking the right card could vary between 0.5 and 1.
> Would you say that only if his certainty is 1 would the conscious
> process be implemented, and not if it is, say, 0.9?
>
>   

I said it would be implementing *the same* consciousness if he was 
following the rule.  If not he might be implementing a different 
consciousness by using a different rule.  Of course if it were different 
in only one "moment" that wouldn't really be much of a difference.  I 
don't think it depends on his certainty.  Even more difficult we might 
ask what it means for him to follow the rule - must he do it 
*consciously*; in which case do we have to know whether his brain is 
functioning according to the same rule?

You're asking a lot of questions, Stathis.  :-)  What do you think?

Brent

--~--~---------~--~----~------------~-------~--~----~
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To post to this group, send email to everything-list@googlegroups.com
To unsubscribe from this group, send email to 
everything-list+unsubscr...@googlegroups.com
For more options, visit this group at 
http://groups.google.com/group/everything-list?hl=en
-~----------~----~----~----~------~----~------~--~---

Reply via email to