Being a too-literal person, who never gets the joke, I have to say that these simple 
scalings, combinatorial or not, don't capture the interconnectionist point being made in 
the pain article. The absolute numbers of elements (neurons, synapses, signaling 
molecules, etc.) flatten it all out. But _ganglion_, that's a different thing. What we're 
looking for are loops and "integratory" structures. I think that's where we can 
start to find a scaling for smartness.

In that context, my guess is the heart is closer to ChatGPT in its smartness than either 
of those are to the human gut. But structure-based assessments like these merely 
complement behavior-based assessments. We could quantify the number of *jobs* done by the 
thing. The heart has fewer jobs to do than the gut. And the gut has fewer jobs to do than 
the dog. Etc. Of course, the lines between jobs aren't all that crisp, especially as the 
complexity of the thing grows. Behaviors in complex things are composable and 
polymorphic. In spite of our imagining what ChatGPT is doing, it's really only doing 1 
thing: choosing the most likely next token given the previous tokens. You *might* be able 
to serialize your dog and suggest she's really just choosing the most likely next 
behavior given the previous behaviors. But my guess is dog owners perceive (or impute) 
that dogs resolve contradictions that arise in parallel. (chase the ball? chew the bone? 
continue chewing the bone until you get to the ball?) Contradiction resolution is 
evidence of more than 1 task. You could gussy up the model by providing a single 
interface to an ensemble of models. Then it might look more like a dog, depending on the 
algorithm(s) used to resolve contradictions between models. But to get closer to 
dog-complexity, you'd have to wire the models together so that they could contradict each 
other but still feed off each other in some way. A model that changes its mind midway 
through its response would be good. I haven't had a dog in a long time. But I seem to 
remember they were easy to redirect, despite the old saying "like a dog with a 
bone".

On 1/17/23 12:51, Prof David West wrote:
Apropos of nothing:

The human heart has roughly 40,000 neurons and the human gut around 0.1 billion 
neurons (sensory neurons, neurotransmitters, ganglia, and motor neurons).

So the human gut is about 1/5 as smart as Marcus's dog??

davew


On Tue, Jan 17, 2023, at 1:08 PM, Marcus Daniels wrote:
Dogs have about 500 million neurons in their cortex.  Neurons have
about 7,000 synaptic connections, so I think my dog is a lot smarter
than a billion parameter LLM.  :-)

Sent from my iPhone

On Jan 17, 2023, at 11:35 AM, glen <[email protected]> wrote:


1) "I asked Chat GPT to write a song in the style of Nick Cave and this is what it 
produced. What do you think?"
https://www.theredhandfiles.com/chat-gpt-what-do-you-think/

2) "Is it pain if it does not hurt? On the unlikelihood of insect pain"
https://www.cambridge.org/core/journals/canadian-entomologist/article/is-it-pain-if-it-does-not-hurt-on-the-unlikelihood-of-insect-pain/9A60617352A45B15E25307F85FF2E8F2#

Taken separately, (1) and (2) are each interesting, if seemingly orthogonal. But what 
twines them, I think, is the concept of "mutual information". I read (2) before 
I read (1) because, for some bizarre reason, my day job involves trying to understand 
pain mechanisms. And (2) speaks directly (if only implicitly) to things like IIT. If you 
read (1) first, it's difficult to avoid snapping quickly into NickC's canal. Despite 
NickT's objection to an inner life, it seems clear that the nuance we see on the surface, 
at least longitudinally, *needs* an inner life. You simply can't get good stuff out of an 
entirely flat/transparent/reactive/Markovian object.

However, what NickC misses is that LLMs *have* some intertwined mutual information within them. Similar to asking whether an 
insect experiences pain, we can ask whether a X billion parameter LLM experiences something like "suffering". My guess 
is the answer is "yes". It may not be a good analog to what we call "suffering", though ... maybe 
"friction"? ... maybe "release"? My sense is that when you engage a LLM (embedded in a larger construct that 
handles the prompts and live learning, of course) in such a way that it assembles a response that nobody else has evoked, it 
might get something akin to a tingle ... or like the relief you feel when scratching an itch ... of course it would be primordial 
because the self-attention in such a system is hopelessly disabled compared to the rich self-attention loops we have in our meaty 
bodies. But it just *might* be there in some primitive sense.

As always, agnosticism is the only rational stance. And I won't trust the songs 
written by LLMs until I see a few of them commit suicide, overdose, or punch a 
TMZ cameraman in the face.

--
ꙮ Mɥǝu ǝlǝdɥɐuʇs ɟᴉƃɥʇ' ʇɥǝ ƃɹɐss snɟɟǝɹs˙ ꙮ
-. --- - / ...- .- .-.. .. -.. / -- --- .-. ... . / -.-. --- -.. .
FRIAM Applied Complexity Group listserv
Fridays 9a-12p Friday St. Johns Cafe   /   Thursdays 9a-12p Zoom 
https://bit.ly/virtualfriam
to (un)subscribe http://redfish.com/mailman/listinfo/friam_redfish.com
FRIAM-COMIC http://friam-comic.blogspot.com/
archives:  5/2017 thru present https://redfish.com/pipermail/friam_redfish.com/
 1/2003 thru 6/2021  http://friam.383.s1.nabble.com/

Reply via email to