On Tue, 23 May 2023 at 15:58, Terren Suydam <[email protected]> wrote:

>
>
> On Tue, May 23, 2023 at 12:32 AM Stathis Papaioannou <[email protected]>
> wrote:
>
>>
>>
>> On Tue, 23 May 2023 at 14:23, Terren Suydam <[email protected]>
>> wrote:
>>
>>>
>>>
>>> On Tue, May 23, 2023 at 12:14 AM Stathis Papaioannou <[email protected]>
>>> wrote:
>>>
>>>>
>>>>
>>>> On Tue, 23 May 2023 at 13:37, Terren Suydam <[email protected]>
>>>> wrote:
>>>>
>>>>>
>>>>>
>>>>> On Mon, May 22, 2023 at 11:13 PM Stathis Papaioannou <
>>>>> [email protected]> wrote:
>>>>>
>>>>>>
>>>>>>
>>>>>> On Tue, 23 May 2023 at 10:48, Terren Suydam <[email protected]>
>>>>>> wrote:
>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Mon, May 22, 2023 at 8:42 PM Stathis Papaioannou <
>>>>>>> [email protected]> wrote:
>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> On Tue, 23 May 2023 at 10:03, Terren Suydam <
>>>>>>>> [email protected]> wrote:
>>>>>>>>
>>>>>>>>>
>>>>>>>>> it is true that my brain has been trained on a large amount of
>>>>>>>>> data - data that contains intelligence outside of my own. But when I
>>>>>>>>> introspect, I notice that my understanding of things is ultimately
>>>>>>>>> rooted/grounded in my phenomenal experience. Ultimately, everything we
>>>>>>>>> know, we know either by our experience, or by analogy to experiences 
>>>>>>>>> we've
>>>>>>>>> had. This is in opposition to how LLMs train on data, which is 
>>>>>>>>> strictly
>>>>>>>>> about how words/symbols relate to one another.
>>>>>>>>>
>>>>>>>>
>>>>>>>> The functionalist position is that phenomenal experience supervenes
>>>>>>>> on behaviour, such that if the behaviour is replicated (same output for
>>>>>>>> same input) the phenomenal experience will also be replicated. This is 
>>>>>>>> what
>>>>>>>> philosophers like Searle (and many laypeople) can’t stomach.
>>>>>>>>
>>>>>>>
>>>>>>> I think the kind of phenomenal supervenience you're talking about is
>>>>>>> typically asserted for behavior at the level of the neuron, not the 
>>>>>>> level
>>>>>>> of the whole agent. Is that what you're saying?  That chatGPT must be
>>>>>>> having a phenomenal experience if it talks like a human?   If so, that 
>>>>>>> is
>>>>>>> stretching the explanatory domain of functionalism past its breaking 
>>>>>>> point.
>>>>>>>
>>>>>>
>>>>>> The best justification for functionalism is David Chalmers' "Fading
>>>>>> Qualia" argument. The paper considers replacing neurons with functionally
>>>>>> equivalent silicon chips, but it could be generalised to replacing any 
>>>>>> part
>>>>>> of the brain with a functionally equivalent black box, the whole brain, 
>>>>>> the
>>>>>> whole person.
>>>>>>
>>>>>
>>>>> You're saying that an algorithm that provably does not have
>>>>> experiences of rabbits and lollipops - but can still talk about them in a
>>>>> way that's indistinguishable from a human - essentially has the same
>>>>> phenomenology as a human talking about rabbits and lollipops. That's just
>>>>> absurd on its face. You're essentially hand-waving away the grounding
>>>>> problem. Is that your position? That symbols don't need to be grounded in
>>>>> any sort of phenomenal experience?
>>>>>
>>>>
>>>> It's not just talking about them in a way that is indistinguishable
>>>> from a human, in order to have human-like consciousness the entire I/O
>>>> behaviour of the human would need to be replicated. But in principle, I
>>>> don't see why a LLM could not have some other type of phenomenal
>>>> experience. And I don't think the grounding problem is a problem: I was
>>>> never grounded in anything, I just grew up associating one symbol with
>>>> another symbol, it's symbols all the way down.
>>>>
>>>
>>> Is the smell of your grandmother's kitchen a symbol?
>>>
>>
>> Yes, I can't pull away the facade to check that there was a real
>> grandmother and a real kitchen against which I can check that the sense
>> data matches.
>>
>
> The ground problem is about associating symbols with a phenomenal
> experience, or the memory of one - which is not the same thing as the
> functional equivalent or the neural correlate. It's the feeling, what it's
> like to experience the thing the symbol stands for. The experience of
> redness. The shock of plunging into cold water. The smell of coffee. etc.
>
> Take a migraine headache - if that's just a symbol, then why does that
> symbol *feel* *bad* while others feel *good*?  Why does any symbol feel
> like anything? If you say evolution did it, that doesn't actually answer
> the question, because evolution doesn't do anything except select for
> traits, roughly speaking. So it just pushes the question to: how did the
> subjective feeling of pain or pleasure emerge from some genetic mutation,
> when it wasn't there before?
>
> Without a functionalist explanation of the *origin* of aesthetic valence,
> then I don't think you can "get it from bit".
>

That seems more like the hard problem of consciousness. There is no
solution to it.

-- 
Stathis Papaioannou

-- 
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/everything-list/CAH%3D2ypWZz4fP1nS_uSNRS6%3Drp63cCpWRLt0_Oeq77Yrfi8WS_w%40mail.gmail.com.

Reply via email to