Brent,
did your dog communicate to you (in dogese, of course) that she has - NO -
INNER NARRATIVE? or you are just ignorant to perceive such?
(Of course do not expect such at the complexity level of your 11b neurons)
John M

On 11/22/08, Brent Meeker <[EMAIL PROTECTED]> wrote:
>
> Günther Greindl wrote:
>>
>>
>> Kory Heath wrote:
>>> If Lucky Alice is conscious and Empty-Headed Alice is not conscious,
>>> then there are partial zombies halfway between them. Like you, I can't
>>> make any sense of these partial zombies. But
>>   also can't make any
>
> I don't see why partial "zombies" are problematic.  My dog is conscious of
> perceptions, of being an individual, of memories and even dreams, but he
> doesn't
> have an inner narrative - so is he a partial zombie?
>
> Brent
>
>>
>> I think a materialist would either have to argue that Lucky Alice is
>> conscious (if he focuses on physical states) and that removing neurons
>> would lead to fading qualia (the "partial zombies") or simply assume
>> that already Lucky Alice is a Zombie (because he focuses on causal
>> dynamics).
>>
>> (I would like to note that I have dropped MAT in the meantime and tend
>> to MECH. Just wanted to "simulate" a materialist argumentation :-) -
>> maybe I can convince myself of MAT and not MECH again *grin*)
>>
>> Could we say that MAT focuses on _physical states_ (exclusively) and
>> MECH on _dynamics_? And that MGA shows that one can't have both?
>>
>>
>> Cheers,
>> Günther
>>
>> >
>>
>
>
> >
>

--~--~---------~--~----~------------~-------~--~----~
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To post to this group, send email to [EMAIL PROTECTED]
To unsubscribe from this group, send email to [EMAIL PROTECTED]
For more options, visit this group at 
http://groups.google.com/group/everything-list?hl=en
-~----------~----~----~----~------~----~------~--~---

Reply via email to