On 3/11/2010 4:51 AM, Stathis Papaioannou wrote:
On 11 March 2010 13:57, Jack Mallah<jackmal...@yahoo.com>  wrote:
--- On Mon, 3/8/10, Stathis Papaioannou<stath...@gmail.com>  wrote:
In the original fading qualia thought experiment the artificial neurons could 
be considered black boxes, the consciousness status of which is unknown. The 
conclusion is that if the artificial neurons lack consciousness, then the brain 
would be partly zombified, which is absurd.
That's not the argument Chalmers made, and indeed he couldn't have, since he 
believes zombies are possible; he instead talks about fading qualia.

If you start out believing that computer zombies are NOT possible, the original 
thought experiment is moot; you already believe the conclusion.   His argument 
is aimed at dualists, who are NOT computationalists to start out.

Since partial consciousness is possible, which he didn't take into account, his 
argument _fails_; a dualist who does believe zombies are possible should have 
no problem believing that partial zombies are.  So dualists don't have to be 
computationalists after all.
A partial zombie is very different from a full zombie! The thing about
zombies is, although you can't tell if someone is a zombie, you know
with absolute certainty that *you* aren't a zombie (assuming that you
aren't a zombie, of course; if you are a zombie then you don't "know"
anything at all except in a mindless zombie way). If your visual
qualia were to fade but your behaviour remain unchanged, then that is
equivalent to partial zombification, and partial zombification is
absurd, impossible or meaningless (take your pick). The argument is
simply this: if zombie vision in an otherwise intact person is
possible, then I could have zombie vision right now. I behave as if I
can see normally, since I am typing this email, but that is consistent
with zombie vision. I am also absolutely convinced that I can see
normally, but that is also consistent with zombie vision. So it seems
that zombie vision is neither subjectively nor objectively different
from normal vision, which means it is not different from normal vision
in any way that matters. You might still say zombie vision is still
different in some metaphysical sense, a category neither objective nor
subjective, but now you are in the supernatural domain.

I think this holds *whatever* is in the black boxes: computers, biological 
tissue, a demon pulling strings or nothing.
Partial consciousness is possible and again ruins any such argument.  If you don't believe to start 
out that consciousness can be based on "whatever" (e.g. "nothing"), you don't 
have any reason to accept the conclusion.
It goes against the grain of functionalism to assume that
consciousness is due primarily to a physical process. The primary idea
is that if the black box replicates the function of a component in a
system, then any mental states that the system has will also be
replicated. Normally this is taken as implying a kind of materialism
since the black box won't be able to replicate behaviour of brain
components unless it contains a complex physical mechanism, but if
miraculously it could - if the black box were empty but the remaining
brain tissue behaves normally anyway - then the consciousness of the
system would remain intact. If a chunk were taken out of the CPU in a
computer but, miraculously, the remaining parts of the CPU behaved
exactly the same as if nothing had happened, then that magical CPU is
just as good as an intact one, and the computations it performs just
as valid.

whatever is going on inside the putative zombie's head, if it reproduces the 
I/O behaviour of a human, it will have the mind of a human.
That is behaviorism, not computationalism, and I certainly don't believe it.  I 
wouldn't say that a computer that uses a huge lookup table algorithm would be 
conscious.
Well, functionalism reduces to a type of behaviourism. Functionalism
is OK with replacing components of a system with functionally
identical analogues, regardless of the internal processes of the
functional analogues. If the internal processes don't matter then it
should not matter if a replaced neuron, for example, is driven by a
lookup table. In fact, a practical computational model of a neuron
would probably contain lookup tables as a matter of course, and it
would seem absurd to claim that its consciousness is inversely
proportion to the number of such devices used.

The requirement that a computer be able to handle the counterfactuals in order 
to be conscious seems to have been brought in to make computationalists feel 
better about computationalism.
Not at all.  It was always part of the notion of computation.  Would you buy a 
PC that only plays a movie?  It must handle all possible inputs in a reliable 
manner.
But if I did buy a PC that only did addition, for example, I don't see
how it would make sense to say that it isn't really doing that
computation, that it's not real addition. It might not qualify as a
"computer", but the one type of computation it does is still correct
and valid. But if there is some primitive consciousness associated
with addition, the idea is that the crippled computer would have it to
a lesser degree compared to a second general purpose computer that
goes through exactly the same sequence of physical states, simply
because the second machine might potentially do other computations as
well. It is this idea that seems to me to be ad hoc, but also quite
unnecessary: nothing is lost by dropping it.

Brains are all probabilistic in that disaster could at any point befall them 
causing them to deviate widely from normal behaviour
It is not a problem, it just seems like one at first glance.  Such cases include input to 
the formal system; for some inputs, the device halts or acts differently.  Hence my talk 
of "derailable computations" in my MCI paper.

or else prevent them from deviating at all from a rigidly determined pathway
If that were done, that would change what computation is being implemented.  
Depending on how it was done, it might or might not affect consciousness.  We 
can't do such an experimemt.
We can do a thought experiment. A brain is rigged to explode unless it
goes down one particular pathway. Does it change the computation being
implemented if it is given the right input so that it does go down
that pathway? Does it change the consciousness? Is it different to a
brain that lacks the connections to begin with so that it does not
explode but simply stops working unless it is provided with the right
input? What do you lose if you say both brains have exactly the same
conscious experience as a normal brain which goes down that pathway?

You might have diminished consciousness. If you identify consciousness with a computation, as in a digital computer, then any specific computation will leave some components unused. But 0's are as much a part of the computation as 1's. So just because the same causal chain of gates or neurons is used it is not the same computation unless it is relative to the same possible computations. Or at least that's one way to look at it. It's not magic, it's just that computation and consciousness maybe holistic properties of a system.

Brent

--
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To post to this group, send email to everything-l...@googlegroups.com.
To unsubscribe from this group, send email to 
everything-list+unsubscr...@googlegroups.com.
For more options, visit this group at 
http://groups.google.com/group/everything-list?hl=en.

Reply via email to