On Wed, 24 May 2023 at 15:37, Jason Resch <jasonre...@gmail.com> wrote:
On Wed, May 24, 2023, 1:15 AM Stathis Papaioannou
<stath...@gmail.com> wrote:
On Wed, 24 May 2023 at 04:03, Jason Resch
<jasonre...@gmail.com> wrote:
On Tue, May 23, 2023 at 7:15 AM Stathis Papaioannou
<stath...@gmail.com> wrote:
On Tue, 23 May 2023 at 21:09, Jason Resch
<jasonre...@gmail.com> wrote:
As I see this thread, Terren and Stathis are both
talking past each other. Please either of you
correct me if i am wrong, but in an effort to
clarify and perhaps resolve this situation:
I believe Stathis is saying the functional
substitution having the same fine-grained causal
organization *would* have the same phenomenology,
the same experience, and the same qualia as the
brain with the same fine-grained causal organization.
Therefore, there is no disagreement between your
positions with regards to symbols groundings,
mappings, etc.
When you both discuss the problem of symbology, or
bits, etc. I believe this is partly responsible
for why you are both talking past each other,
because there are many levels involved in brains
(and computational systems). I believe you were
discussing completely different levels in the
hierarchical organization.
There are high-level parts of minds, such as
ideas, thoughts, feelings, quale, etc. and there
are low-level, be they neurons, neurotransmitters,
atoms, quantum fields, and laws of physics as in
human brains, or circuits, logic gates, bits, and
instructions as in computers.
I think when Terren mentions a "symbol for the
smell of grandmother's kitchen" (GMK) the trouble
is we are crossing a myriad of levels. The quale
or idea or memory of the smell of GMK is a very
high-level feature of a mind. When Terren asks for
or discusses a symbol for it, a complete
answer/description for it can only be supplied in
terms of a vast amount of information concerning
low level structures, be they patterns of neuron
firings, or patterns of bits being processed. When
we consider things down at this low level,
however, we lose all context for what the meaning,
idea, and quale are or where or how they come in.
We cannot see or find the idea of GMK in any
neuron, no more than we can see or find it in any
neuron.
Of course then it should seem deeply mysterious,
if not impossible, how we get "it" (GMK or
otherwise) from "bit", but to me, this is no
greater a leap from how we get "it" from a bunch
of cells squirting ions back and forth. Trying to
understand a smartphone by looking at the flows of
electrons is a similar kind of problem, it would
seem just as difficult or impossible to explain
and understand the high-level features and
complexity out of the low-level simplicity.
This is why it's crucial to bear in mind and
explicitly discuss the level one is operation on
when one discusses symbols, substrates, or quale.
In summary, I think a chief reason you have been
talking past each other is because you are each
operating on different assumed levels.
Please correct me if you believe I am mistaken and
know I only offer my perspective in the hope it
might help the conversation.
I think you’ve captured my position. But in addition I
think replicating the fine-grained causal organisation
is not necessary in order to replicate higher level
phenomena such as GMK. By extension of Chalmers’
substitution experiment,
Note that Chalmers's argument is based on assuming the
functional substitution occurs at a certain level of
fine-grained-ness. If you lose this step, and look at only
the top-most input-output of the mind as black box, then
you can no longer distinguish a rock from a dreaming
person, nor a calculator computing 2+3 and a human
computing 2+3, and one also runs into the Blockhead
"lookup table" argument against functionalism.
Yes, those are perhaps problems with functionalism. But a
major point in Chalmers' argument is that if qualia were
substrate-specific (hence, functionalism false) it would be
possible to make a partial zombie or an entity whose
consciousness and behaviour diverged from the point the
substitution was made. And this argument works not just by
replacing the neurons with silicon chips, but by replacing any
part of the human with anything that reproduces the
interactions with the remaining parts.
How deeply do you have to go when you consider or define those
"other parts" though? That seems to be a critical but unstated
assumption, and something that depends on how finely grained you
consider the relevant/important parts of a brain to be.
For reference, this is what Chalmers says:
"In this paper I defend this view. Specifically, I defend a
principle of organizational invariance, holding that experience is
invariant across systems with the same fine-grained functional
organization. More precisely, the principle states that given any
system that has conscious experiences, then any system that has
the same functional organization at a fine enough grain will have
qualitatively identical conscious experiences. A full
specification of a system's fine-grained functional organization
will fully determine any conscious experiences that arise."
https://consc.net/papers/qualia.html
By substituting a fine-grained functional organization for a
coarse-grained one, you change the functional definition and can
no longer guarantee identical experiences, nor identical behaviors
in all possible situations. They're no longer"functional
isomorphs" as Chalmers's argument requires.
By substituting a recording of a computation for a computation,
you replace a conscious mind with a tape recording of the prior
behavior of a conscious mind. This is what happens in the
Blockhead thought experiment. The result is something that passes
a Turing test, but which is itself not conscious (though creating
such a recording requires prior invocation of a conscious mind or
extraordinary luck).
The replaced part must of course be functionally identical, otherwise
both the behaviour and the qualia could change. But this does not mean
that it must replicate the functional organisation at a particular
scale. If a volume of brain tissue is removed, in order to guarantee
identical behaviour the replacement part must interact at the cut
surfaces of the surrounding tissue in the same way as the original. It
is at these surfaces that the interactions must be sufficiently
fine-grained, but what goes inside the volume doesn't matter: it could
be conventional simulation of neurons, it could be a giant lookup
table. Also, the volume could be any size, and could comprise an
arbitrarily large proportion of the subject.