Clicking through the linked references: The OpenAI "research" reads
like a list of press releases promoting their agenda; Anthropic's list
reads much more like SciAm or similar level of popular publication
trying to actually communicate the issues:
https://transformer-circuits.pub/2022/toy_model/index.html#motivation
My "dive" into this is still very shallow (maybe someone can
de-metaphorize this for me?) but it aligns well with my own ad-hoc/naive
growing apprehensions. Lots here (for me) including what feels like a
strong parallel (apt for metaphorical/analogical domain transfer?) with
genotype/phenotype thinking.
On 7/17/2025 11:20 AM, Steve Smith wrote:
https://scitechdaily.com/researchers-decode-how-we-turn-thoughts-into-sentences/
I'm hoping/expecting some folks here are as fascinated with these
things as I am? LLM's, interperatability, natural vs are to me as
weather/vortices/entropy-intuition is to Nick?
As someone who spends way too much time composing sentences (in
writing) through this impedence-mismatched interface (keyboard) I have
a strong (if misleading, or at least ideosyncratic) apprehension of
how I might form sentences from thoughts, and perhaps even
forward/back propogate possible expressions and structures *all the
way* to where I imagine my interlocutors (often all y'all here)
reading and responding internally (mentally) and online. My
engagement with the LLMs in "casual conversation" includes a great
deal of this, albeit understanding that I'm talking to "a stochastic
parrot" or more aptly perhaps "making faces into a funhouse mirror"
(reminding me that I really want to compose a good-faith answer to
glen's very sincere and I think pivotal questions about metaphor).
I haven't parsed the linked article deeply yet and have not sought out
the actual paper itself yet, but find the ideas presented very
provocative or at least evocative? It triggers hopeful imaginings
about connections with the cortical column work of Hawkins/Numenta as
well as the never ending topics of FriAM: " Effing the inEffabl"e and
"Metaphors all the way Down?"
I don't expect this line of research to *answer* those questions, but
possibly shed some scattered light onto their periphery (oupsie, I
waxed up another metapho to shoot some curls)? For example, might
the electrocorticography during ideation-to-speech transmogrification
show us how strongly metaphorical constructions differ from more
concise or formal analogical versions (if they are a spectrum) or how
attempts to "eff the ineffable" might yield widely branching (bushy)
explorations, ending in some kind of truncation by fatigue or
(de)saturation?
https://www.nature.com/articles/s44271-025-00270-1
<https://www.nature.com/articles/s44271-025-00270-1>
And are attempts at Interpreting LLMs in some meaningful way colinear
or offer important parallax (to reference the
"steam-engine/thermodynamics" duality)?
And me, here, with obviously "way too much time" on my hands and a
fascination with LLMs and an urgency to try to keep traction on the
increasing slope of "the singularity" and a mild facility with visual
analytics and *I* haven't even begun to keep up... This list
(ironically) was formulated by GPT and I've not (and surely will not)
do much double-checking beyond (hopefully) diving deep(er) inoto the
work. I was mildly surprised there were no 2025 references... I'm
guessing the blogs are running commentary including current work.
I'll go click through as soon as I hit <send> here (imagine the
next-token prediction I am doing as I decide to try to stop typing and
hit <send>?)
*“A Survey of Explainability and Interpretability in Large
Language Models”* (ACM Computing Surveys, 2024)
Comprehensive classification of methods, with comparisons between
mechanistic and post‑hoc approaches.
Preprint link on arXiv: [arXiv:2310.01789]
<https://arxiv.org/abs/2310.01789>
*Anthropic’s Interpretability Research Pages* (2023–2024)
https://www.anthropic.com/research
*OpenAI’s Technical Blog: “Language Models and Interpretability”*
(2023)
Discussion of interpretability challenges, with examples from
GPT‑4-level models:
https://openai.com/research
*NeurIPS 2023 Workshop on XAI for Large Models*
Video talks & proceedings with up-to-date methods:
https://nips.cc/virtual/2023/workshop/66533
.- .-.. .-.. / ..-. --- --- - . .-. ... / .- .-. . / .-- .-. --- -. --. / ...
--- -- . / .- .-. . / ..- ... . ..-. ..- .-..
FRIAM Applied Complexity Group listserv
Fridays 9a-12p Friday St. Johns Cafe / Thursdays 9a-12p
Zoomhttps://bit.ly/virtualfriam
to (un)subscribehttp://redfish.com/mailman/listinfo/friam_redfish.com
FRIAM-COMIChttp://friam-comic.blogspot.com/
archives: 5/2017 thru presenthttps://redfish.com/pipermail/friam_redfish.com/
1/2003 thru 6/2021http://friam.383.s1.nabble.com/
.- .-.. .-.. / ..-. --- --- - . .-. ... / .- .-. . / .-- .-. --- -. --. / ...
--- -- . / .- .-. . / ..- ... . ..-. ..- .-..
FRIAM Applied Complexity Group listserv
Fridays 9a-12p Friday St. Johns Cafe / Thursdays 9a-12p Zoom
https://bit.ly/virtualfriam
to (un)subscribe http://redfish.com/mailman/listinfo/friam_redfish.com
FRIAM-COMIC http://friam-comic.blogspot.com/
archives: 5/2017 thru present https://redfish.com/pipermail/friam_redfish.com/
1/2003 thru 6/2021 http://friam.383.s1.nabble.com/