Yeah, that's the germ of an actually profound observation about us vs them. "What it's 
like" to be some thing is nothing more than a strong analogy, both behavioral and structural. 
If you have all the same parts of another thing, and you have all the same behaviors as another 
thing, then you experience "what it's like" to be that other thing. True of computers. 
True of Trump.

An MD Psych therapist friend of mine divides people into 2 groups: 1) the 
splitters, who incessantly parse, and 2) the smooshers, who incessantly 
integrate. A splitter with a slightly different database of grievances would 
make great hay of their tiny differences with Trump (maybe Liz Cheney?). A 
smoosher would ignore any small differences between themselves and Trump (maybe 
Kevin McCarthy?).

On 6/20/22 09:35, Barry MacKichan wrote:
This prompts me to propose, with tongue slightly in cheek, the /Weak Turing 
Test/, which consists of a remote observer trying to distinguish between a 
computer and Donald Trump. It doesn’t require any deep analysis on the part of 
the computer, just a database of grievances.

—Barry

On 20 Jun 2022, at 11:45, glen wrote:

    "None of this has anything to do with artificial consciousness, of course. “I 
don’t anthropomorphize,” Chowdhery said bluntly. “We are simply predicting language.” 
Artificial consciousness is a remote dream that remains firmly entrenched in science 
fiction, because we have no idea what human consciousness is; there is no functioning 
falsifiable thesis of consciousness, just a bunch of vague notions. And if there is no 
way to test for consciousness, there is no way to program it. You can ask an algorithm to 
do only what you tell it to do. All that we can come up with to compare machines with 
humans are little games, such as Turing’s imitation game, that ultimately prove 
nothing."


    What amazes me is that few take the logical step of suggesting that there is no such thing as 
consciousness. We're all "simply predicting language." The only difference between an 
animal and a language predicting chatbot is that we *also* "simply predict" actions in 4D 
spacetime. Any disconnect between the serial prediction of tokens from the (not-so-serial) 
prediction of all actions is that dimension reduction.


    On 6/20/22 08:15, Marcus Daniels wrote:

        
https://www.theatlantic.com/technology/archive/2022/06/google-palm-ai-artificial-consciousness/661329/
 
<https://www.theatlantic.com/technology/archive/2022/06/google-palm-ai-artificial-consciousness/661329/>
 
<https://www.theatlantic.com/technology/archive/2022/06/google-palm-ai-artificial-consciousness/661329/
 
<https://www.theatlantic.com/technology/archive/2022/06/google-palm-ai-artificial-consciousness/661329/>>


--
ꙮ Mɥǝu ǝlǝdɥɐuʇs ɟᴉƃɥʇ' ʇɥǝ ƃɹɐss snɟɟǝɹs˙ ꙮ

-. --- - / ...- .- .-.. .. -.. / -- --- .-. ... . / -.-. --- -.. .
FRIAM Applied Complexity Group listserv
Fridays 9a-12p Friday St. Johns Cafe   /   Thursdays 9a-12p Zoom 
https://bit.ly/virtualfriam
to (un)subscribe http://redfish.com/mailman/listinfo/friam_redfish.com
FRIAM-COMIC http://friam-comic.blogspot.com/
archives:  5/2017 thru present https://redfish.com/pipermail/friam_redfish.com/
 1/2003 thru 6/2021  http://friam.383.s1.nabble.com/

Reply via email to