On Thu, 25 May 2023 at 04:08, Brent Meeker <[email protected]> wrote:

>
>
> On 5/24/2023 10:41 AM, Stathis Papaioannou wrote:
>
>
>
> On Thu, 25 May 2023 at 02:14, Brent Meeker <[email protected]> wrote:
>
>>
>>
>> On 5/24/2023 12:19 AM, Stathis Papaioannou wrote:
>>
>>
>>
>> On Wed, 24 May 2023 at 15:37, Jason Resch <[email protected]> wrote:
>>
>>>
>>>
>>> On Wed, May 24, 2023, 1:15 AM Stathis Papaioannou <[email protected]>
>>> wrote:
>>>
>>>>
>>>>
>>>> On Wed, 24 May 2023 at 04:03, Jason Resch <[email protected]> wrote:
>>>>
>>>>>
>>>>>
>>>>> On Tue, May 23, 2023 at 7:15 AM Stathis Papaioannou <
>>>>> [email protected]> wrote:
>>>>>
>>>>>>
>>>>>>
>>>>>> On Tue, 23 May 2023 at 21:09, Jason Resch <[email protected]>
>>>>>> wrote:
>>>>>>
>>>>>>> As I see this thread, Terren and Stathis are both talking past each
>>>>>>> other. Please either of you correct me if i am wrong, but in an effort 
>>>>>>> to
>>>>>>> clarify and perhaps resolve this situation:
>>>>>>>
>>>>>>> I believe Stathis is saying the functional substitution having the
>>>>>>> same fine-grained causal organization *would* have the same 
>>>>>>> phenomenology,
>>>>>>> the same experience, and the same qualia as the brain with the same
>>>>>>> fine-grained causal organization.
>>>>>>>
>>>>>>> Therefore, there is no disagreement between your positions with
>>>>>>> regards to symbols groundings, mappings, etc.
>>>>>>>
>>>>>>> When you both discuss the problem of symbology, or bits, etc. I
>>>>>>> believe this is partly responsible for why you are both talking past 
>>>>>>> each
>>>>>>> other, because there are many levels involved in brains (and 
>>>>>>> computational
>>>>>>> systems). I believe you were discussing completely different levels in 
>>>>>>> the
>>>>>>> hierarchical organization.
>>>>>>>
>>>>>>> There are high-level parts of minds, such as ideas, thoughts,
>>>>>>> feelings, quale, etc. and there are low-level, be they neurons,
>>>>>>> neurotransmitters, atoms, quantum fields, and laws of physics as in 
>>>>>>> human
>>>>>>> brains, or circuits, logic gates, bits, and instructions as in 
>>>>>>> computers.
>>>>>>>
>>>>>>> I think when Terren mentions a "symbol for the smell of
>>>>>>> grandmother's kitchen" (GMK) the trouble is we are crossing a myriad of
>>>>>>> levels. The quale or idea or memory of the smell of GMK is a very
>>>>>>> high-level feature of a mind. When Terren asks for or discusses a symbol
>>>>>>> for it, a complete answer/description for it can only be supplied in 
>>>>>>> terms
>>>>>>> of a vast amount of information concerning low level structures, be they
>>>>>>> patterns of neuron firings, or patterns of bits being processed. When we
>>>>>>> consider things down at this low level, however, we lose all context for
>>>>>>> what the meaning, idea, and quale are or where or how they come in. We
>>>>>>> cannot see or find the idea of GMK in any neuron, no more than we can 
>>>>>>> see
>>>>>>> or find it in any neuron.
>>>>>>>
>>>>>>> Of course then it should seem deeply mysterious, if not impossible,
>>>>>>> how we get "it" (GMK or otherwise) from "bit", but to me, this is no
>>>>>>> greater a leap from how we get "it" from a bunch of cells squirting ions
>>>>>>> back and forth. Trying to understand a smartphone by looking at the 
>>>>>>> flows
>>>>>>> of electrons is a similar kind of problem, it would seem just as 
>>>>>>> difficult
>>>>>>> or impossible to explain and understand the high-level features and
>>>>>>> complexity out of the low-level simplicity.
>>>>>>>
>>>>>>> This is why it's crucial to bear in mind and explicitly discuss the
>>>>>>> level one is operation on when one discusses symbols, substrates, or 
>>>>>>> quale.
>>>>>>> In summary, I think a chief reason you have been talking past each 
>>>>>>> other is
>>>>>>> because you are each operating on different assumed levels.
>>>>>>>
>>>>>>> Please correct me if you believe I am mistaken and know I only offer
>>>>>>> my perspective in the hope it might help the conversation.
>>>>>>>
>>>>>>
>>>>>> I think you’ve captured my position. But in addition I think
>>>>>> replicating the fine-grained causal organisation is not necessary in 
>>>>>> order
>>>>>> to replicate higher level phenomena such as GMK. By extension of 
>>>>>> Chalmers’
>>>>>> substitution experiment,
>>>>>>
>>>>>
>>>>> Note that Chalmers's argument is based on assuming the functional
>>>>> substitution occurs at a certain level of fine-grained-ness. If you lose
>>>>> this step, and look at only the top-most input-output of the mind as black
>>>>> box, then you can no longer distinguish a rock from a dreaming person, nor
>>>>> a calculator computing 2+3 and a human computing 2+3, and one also runs
>>>>> into the Blockhead "lookup table" argument against functionalism.
>>>>>
>>>>
>>>> Yes, those are perhaps problems with functionalism. But a major point
>>>> in Chalmers' argument is that if qualia were substrate-specific (hence,
>>>> functionalism false) it would be possible to make a partial zombie or an
>>>> entity whose consciousness and behaviour diverged from the point the
>>>> substitution was made. And this argument works not just by replacing the
>>>> neurons with silicon chips, but by replacing any part of the human with
>>>> anything that reproduces the interactions with the remaining parts.
>>>>
>>>
>>>
>>> How deeply do you have to go when you consider or define those "other
>>> parts" though? That seems to be a critical but unstated assumption, and
>>> something that depends on how finely grained you consider the
>>> relevant/important parts of a brain to be.
>>>
>>> For reference, this is what Chalmers says:
>>>
>>>
>>> "In this paper I defend this view. Specifically, I defend a principle of
>>> organizational invariance, holding that experience is invariant across
>>> systems with the same fine-grained functional organization. More precisely,
>>> the principle states that given any system that has conscious experiences,
>>> then any system that has the same functional organization at a fine enough
>>> grain will have qualitatively identical conscious experiences. A full
>>> specification of a system's fine-grained functional organization will fully
>>> determine any conscious experiences that arise."
>>> https://consc.net/papers/qualia.html
>>>
>>> By substituting a fine-grained functional organization for a
>>> coarse-grained one, you change the functional definition and can no longer
>>> guarantee identical experiences, nor identical behaviors in all possible
>>> situations. They're no longer"functional isomorphs" as Chalmers's argument
>>> requires.
>>>
>>> By substituting a recording of a computation for a computation, you
>>> replace a conscious mind with a tape recording of the prior behavior of a
>>> conscious mind. This is what happens in the Blockhead thought experiment.
>>> The result is something that passes a Turing test, but which is itself not
>>> conscious (though creating such a recording requires prior invocation of a
>>> conscious mind or extraordinary luck).
>>>
>>
>> The replaced part must of course be functionally identical, otherwise
>> both the behaviour and the qualia could change. But this does not mean that
>> it must replicate the functional organisation at a particular scale. If a
>> volume of brain tissue is removed, in order to guarantee identical
>> behaviour the replacement part must interact at the cut surfaces of the
>> surrounding tissue in the same way as the original. It is at these surfaces
>> that the interactions must be sufficiently fine-grained, but what goes
>> inside the volume doesn't matter: it could be conventional simulation of
>> neurons, it could be a giant lookup table. Also, the volume could be any
>> size, and could comprise an arbitrarily large proportion of the subject.
>>
>>
>> Doesn't it need to be able to change in order to have memory and to learn?
>>
>
> Yes, I meant change from what the original parts would do. If you get a
> neural implant you would want it to leave your brain functioning as it was
> originally, which means all the remaining neurons firing in the same way as
> they were originally. This would guarantee that your consciousness would
> also continue as it was originally.
>
> Except I couldn't learn anything or form any new memories; at least not if
> they depended on the implant.  Right?
>

It wouldn’t be functionally equivalent in that case.

> --
Stathis Papaioannou

-- 
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/everything-list/CAH%3D2ypWrypF1ejdKraf3pkXau9icDrjSSzApRmjZ5Q__ePiXhw%40mail.gmail.com.

Reply via email to