Jon, Gary F, List, For me, this has been a most valuable discussion. While I had earlier come to the conclusion that Artificial Intelligence is *not* intelligent, the comments and quotes included in this exchange strongly suggest to me that it will never be, *can* never be because it misses the necessary features that characterize intelligence.
As Jon concisely put it, "If genuine semiosis is truly continuous. . . then a digital computer, no matter how sophisticated, can only ever *simulate *it--just as the real numbers do not constitute a true continuum, but usefully *approximate *one for most practical purposes. After all, whenever we humans break up our own reasoning (arguments) into discrete steps--namely, "definitely formulated premisses" and conclusions (argumentations. . .) --we are always doing so artificially and retrospectively, *after *the real and continuous inferential process has already run its course." Yet, to the extent that AI may prove dangerous, I continue to think that it behooves us -- to the extent to which it is possible -- to move AI systems toward Peircean theoretical rhetoric within the communities of inquiry in which each of us may be engaged. Nevertheless, Gary F's warning shouldn't be ignored: "If present experience is any guide. . . , clearly AI systems are going to align with the values of the billionaire owners of those systems (and to a lesser extent the programmers who work for them), which is certainly no cause for optimism. Best, Gary R On Fri, Jan 9, 2026 at 12:30 PM Jon Alan Schmidt <[email protected]> wrote: > Gary R., Gary F., List: > > GF: Having read the fine print at the end of the paper, it’s clear that > Manheim’s article was co-written with several LLM chatbots, and I wonder if > some of the optimism comes from them (or some of them) rather than from the > human side. > > > I noticed that, too, with the result that it is more difficult for me to > take the article seriously. In a 1999 paper > <https://www.jstor.org/stable/40320779>, "Peirce's Inkstand as an > External Embodiment of Mind," Peter Skagestad quotes CP 7.366 (1902) and > points out that Peirce "is not *only *making the point that without ink > he would not be able to express his thoughts, but rather the point that > thoughts come to him in and through the act of writing, so that having > writing implements is a condition for having certain thoughts" (p. 551). I > know firsthand that the act of writing facilitates my own thinking, and I > cannot help wondering if Manheim's choice to delegate so much of the effort > for drafting his article to LLMs precluded him from carefully thinking > through everything that it ended up saying. > > GF: Successful "alignment" is supposed to be between a super "intelligent" > system and *human values*. One problem with this is that human values > vary widely between different groups of humans, so which values is future > AI supposed to align with? > > > If an artificial system were really intelligent, then it seems to me that > it would be capable of *choosing *its own values instead of having a > particular set of human values imposed on it. In a 2013 paper > <https://www.academia.edu/9898586/C_S_Peirce_and_Artificial_Intelligence_Historical_Heritage_and_New_Theoretical_Stakes>, > "C. S. Peirce and Artificial Intelligence: Historical Heritage and (New) > Theoretical Stakes," Pierre Steiner observes that according to Peirce ... > > PS: [H]uman reasoning is notably special (and, in that sense only, > *genuine*) in virtue of the *high *degrees of self-control and > self-correctiveness it can exercise on conduct: control on control, > self-criticism on control, and control on control on the basis of > (revisable and self-endorsed) norms and principles and, ultimately, > aesthetic and moral ideals. ... The fact that reasoning human agents have > *purposes > *is crucial here: it is on the basis of purposes that they are ready to > endorse, change or criticize specific methods of reasoning (inductive, > formal, empirical, ...), but also to revise and reject previous purposes. > Contrary to machines, humans do not only have *specified *purposes. Their > purposes are often vague and general. In other passages, Peirce suggests > that this ability for (higher-order and purposive) self-control is closely > related to the fact that human agents are living, and especially *growing*, > systems. (p. 272) > > > I suspect that much of the worry about "AI safety/alignment," as reflected > by common fictional storylines in popular culture, is a tacit admission of > this. What would prevent a sufficiently intelligent artificial system, > provided that such a thing is even possible, from *rejecting *human > values and instead adopting norms, principles, ideals, and purposes that we > would find objectionable, perhaps even abhorrent? More on the > living/growing aspect of intelligent systems below. > > GF: LLMs have to be artificially supplied with a giant database of > thousands or millions of symbolic texts, and it takes them months or years > to build up the level of language competence that a human toddler has; and > even then is is doubtful whether they *understand *any of it. > > > As with intelligence, I am unconvinced that it is accurate to ascribe > "language competence" to LLMs, especially given the well-founded doubt > about "whether they *understand *any of it." John Searle's famous > "Chinese room" thought experiment seems relevant here, e.g., as discussed > by John Fetzer in his online *Commens Encyclopedia* article > <http://www.commens.org/encyclopedia/article/fetzer-james-peirce-and-philosophy-artificial-intelligence>, > "Peirce and the Philosophy of Artificial Intelligence." Again, in my view, > LLMs do not actually *use *natural languages, they only *simulate *using > natural languages. > > GF: I can’t help thinking that all this has a bearing on the perennial > question of whether semiosis requires *life *or not. > > > In light of the following passage, Peirce's answer is evidently that *genuine > *semiosis requires life, given that it requires *genuine *triadic > relations; but he also seems to define "life" in this context much more > broadly than what we associate with the special science of biology. > > CSP: For forty years, that is, since the beginning of the year 1867, I > have been constantly on the alert to find a *genuine *triadic > relation--that is, one that does not consist in a mere collocation of > dyadic relations, or the negative of such, etc. (I prefer not to attempt a > perfectly definite definition)--which is not either an intellectual > relation or a relation concerned with the less comprehensible phenomena of > life. I have not met with one which could not reasonably be supposed to > belong to one or other of these two classes. ... In short, the problem of > how genuine triadic relationships first arose in the world is a better, > because more definite, formulation of the problem of how life first came > about; and no explanation has ever been offered except that of pure chance, > which we must suspect to be no explanation, owing to the suspicion that > pure chance may itself be a vital phenomenon. In that case, life in the > physiological sense would be due to life in the metaphysical sense. (CP > 6.322, 1907) > > > Elsewhere, Peirce maintains > <https://list.iu.edu/sympa/arc/peirce-l/2025-11/msg00044.html> that a > continuum is *defined* by a genuine triadic relation, so his remarks here > are consistent with my sense that what fundamentally precludes digital > computers from ever being truly intelligent is the *discreteness* of > their operations. As I said before, LLMs are surely quasi-minds whose > individual determinations are *dynamical *interpretants of sign *tokens*; > but those correlates are involved in *degenerate *triadic relations, > which are reducible to their constituent dyadic relations. In my view > <https://list.iu.edu/sympa/arc/peirce-l/2025-11/msg00056.html>, the *genuine > *triadic relation involves the *final *interpretant and the sign *itself*, > which is general > <https://list.iu.edu/sympa/arc/peirce-l/2025-11/msg00019.html> and > therefore a continuum of potential tokens that is *not *reducible to the > actual tokens that individually embody it. > > Regards, > > Jon Alan Schmidt - Olathe, Kansas, USA > Structural Engineer, Synechist Philosopher, Lutheran Christian > www.LinkedIn.com/in/JonAlanSchmidt / twitter.com/JonAlanSchmidt > > On Thu, Jan 8, 2026 at 11:17 AM <[email protected]> wrote: > >> List, I’d like to add a few comments to those already posted by Jon and >> Gary R about the Manheim paper — difficult as it is to focus on these >> issues given the awareness of what’s happening in Minnesota, Venezuela, >> Washington etc. (I may come back to that later.) >> >> Except for the odd usage of the term “interpretant” which Jon has already >> mentioned, I think Manheim’s simplified account of Peircean semiotics is >> cogent enough. But his paper seems to get increasingly muddled in the >> latter half of it. For instance, the “optimism” about future AI that Jon >> sees in it seems quite equivocal to me. Having read the fine print at the >> end of the paper, it’s clear that Manheim’s article was co-written with >> several LLM chatbots, and I wonder if some of the optimism comes from them >> (or some of them) rather than from the human side. >> >> Also, the paper makes a distinction between AI *safety* and the >> *alignment* problem, but then seems to gloss over the differences. >> Succesful “alignment” is supposed to be between a super”intelligent” system >> and *human values*. One problem with this is that human values vary >> widely between different groups of humans, so which values is future AI >> supposed to align with? If present experience is any guide (and it better >> be!), clearly AI systems are going to align with the values of the >> billionaire owners of those systems (and to a lesser extent the programmers >> who work for them), which is certainly no cause for optimism. >> >> I think Stanislas Dehaene’s 2020 book *How We Learn* deals with the >> deeper context of these issues better than Manheim and his chatbot >> co-authors. Its subtitle is *Why Brains Learn Better Than Any Machine … >> for Now.* Reducing this to simplest terms, it’s because brains learn >> from *experience* — “the total cognitive result of living,” as Peirce >> said* — and they do so by a scientific method (an algorithm, as Dehaene >> calls it) which is part of the *genetic* inheritance supplied by >> *biological* evolution. An absolute requirement of this method is what >> Peirce called *abduction* (or retroduction). >> >> For instance, human babies begin learning the language they are exposed >> to from birth, or even before — syntax, semantics, pragmatics and all — >> almost entirely without instruction, by a trial-and-error method. It >> enables them to pick up and remember the meaning and use of a new word *from >> one or two encounters with it*. LLMs have to be artificially supplied >> with a giant database of thousands or millions of symbolic texts, and it >> takes them months or years to build up the level of language competence >> that a human toddler has; and even then is is doubtful whether they >> *understand* any of it. LLM learning is entirely bottom-up and therefore >> works much slower than the holistic learning-from-experience of a living >> bodymind, even though the processing speed of a computer is much faster >> than a brain’s. (That’s why it is so much more energy-hungry than brains >> are.) >> >> I can’t help thinking that all this has a bearing on the perennial >> question of whether semiosis requires *life* or not. I can’t help >> thinking that *experience* requires life, and that is what a “scientific >> intelligence” has to learn from — including whatever *values* it learns. >> It has to be embodied, and providing it with sensors to gather data from >> the external world is not enough if that embodiment does not have a *whole >> world within* it in continuous dialogue with the world without — an >> internal *model*, as I (and Dehaene and others) call it. But I’d better >> stop there, as this is getting too long already. >> >> *The context of the Peirce quote above is here: Turning Signs 7: >> Experience and Experiment <https://gnusystems.ca/TS/xpt.htm#lgcsmtc> >> >> Love, gary f >> >> Coming from the ancestral lands of the Anishinaabeg >> >> >> >> *From:* Gary Richmond <[email protected]> >> *Sent:* 8-Jan-26 04:03 >> *To:* Peirce List <[email protected]>; Gary Fuhrman < >> [email protected]>; Jon Alan Schmidt <[email protected]> >> *Subject:* AI safety and semeiotic, was, Surdity, Feeling, and >> Consciousness, was, Truth and dyadic consciousnessg >> >> >> >> Gary F, Jon, List, >> >> In the discussion of Manheim's paper I think it's important to remember >> that his concern is primarily with* AI safety. Anything* that would >> contribute to that safely I would wholeheartedly support. In my view, >> Peircean semeiotic might prove to be of some value in the matter, but >> perhaps not exactly in the way that Manheim is thinking of it. >> >> Manheim remarks that his paper does not try to settle philosophical >> questions about whether LLMs genuinely reason or only simulate thought, and >> that resolving those debates isn’t necessary for building safer general AI. >> I won't take up that claim now, but suffice it to say that I don't >> *fully* agree with it, especially as I continue to agree with your >> argument, Jon, that AI is* not* 'intelligent'. Can it every be? >> >> What Mannheim claims is necessary re: AI safety is to move AI systems >> toward Peircean semiosis in the sense of their becoming 'participants' in >> interpretive processes. He holds that this is achievable through >> engineering and 'capability' advances rather than "philosophical >> breakthroughs;" though he also says that those advances remain insufficient >> on their own for safety. Remaining "insufficient on its own for full >> safety" sounds to me somewhat self-contradictory. But I think that more >> importantly, he is saying that* if* there are things -- including >> Peircean 'things' -- that we can begin to do now in consideration of AI >> safety, then we ought to consider them, do them! >> >> Manheim claims that AI safety depends on deliberately designing systems >> for what he calls 'grounded meaning', 'persistence across interactions' and >> 'shared semiotic communities' rather than 'isolated agents'. I would tend >> to strongly agree. In addition, AI safety requires goals that are *explicitly >> defined* but also open to ongoing discussion rather than quasi*-emerging >> implicitly* from methods like *Reinforcement Learning from Human >> Feedback *(RLHF) . Manheim seems to be saying that companies developing >> advanced AI should take steps in system design and goal setting -- >> including those mentioned above -- if safety is taken seriously. The >> choice, he says, is between ignoring the implications of Peircean semeiotic >> and continuing merely to refine current systems despite their deficiency >> vis-a-vis safety; OR to embrace Peircean semiosis (whatever that means) and >> intentionally build AI as genuine 'semiotic partners'. But,I haven't a >> clear notion of what he means by 'semeiotic partners', nor a method for >> implementing whatever he does have in mind. >> >> I think Manheim off-handedly and rather summarily unfortunately dismisses >> RLHF -- which is, falsely he argues, claimed as a way of 'aligning' models >> with human values. From what I've read it has not yet really been developed >> much in that direction. As far as I can tell, and this may relate to the >> reason why Manheim seems to reject RLHF in toto, it appears to be more a >> 'reward proxy' trained on human rankings of outputs which are then fed >> back through some kind of loop to strongly influence future responses. >> Human judgment enters only in the 'training'', not as something that a >> complex system can engage with and debate with or, possibly, revise >> understandings over time. In Manheim's view, RLHF is not 'bridging' human >> goals and machine behavior (as it claims) but merely facilitating machine >> outputs to fit *learned preferences*. >> >> Still, whatever else RLHF is doing that is geared specifically* toward >> AI safety*, it would likely be augmented by an understanding of Peircean >> cenoscopic science including semeiotic. I would suggest that the semeiotic >> ideas that it might most benefit from occur in the third branch of *Logic >> as Semeiotic*, namely methodology (methodeutic) , perhaps in the present >> context representing, almost to a T, Peirce's alternative title, *speculative >> rhetoric*. It's in this branch of semeiotic that pragmatism >> (pragmaticism) is analyzed. There is of course much more to be said on >> methodology and theoretical rhetoric. >> >> For now, I would tweak Manheim's idea a bit and would suggest that we >> might try to move AI systems toward Peircean semeiotic rhetoric within >> communities of inquiry. >> >> Best, >> >> Gary R >> > _ _ _ _ _ _ _ _ _ _ > ► PEIRCE-L subscribers: Click on "Reply List" or "Reply All" to REPLY ON > PEIRCE-L to this message. PEIRCE-L posts should go to [email protected] > . > ► <a href="mailto:[email protected]">UNSUBSCRIBE FROM > PEIRCE-L</a> . But, if your subscribed email account is not your default > email account, then go to > https://list.iu.edu/sympa/signoff/peirce-l . > ► PEIRCE-L is owned by THE PEIRCE GROUP; moderated by Gary Richmond; and > co-managed by him and Ben Udell.
_ _ _ _ _ _ _ _ _ _ ► PEIRCE-L subscribers: Click on "Reply List" or "Reply All" to REPLY ON PEIRCE-L to this message. PEIRCE-L posts should go to [email protected] . ► <a href="mailto:[email protected]">UNSUBSCRIBE FROM PEIRCE-L</a> . But, if your subscribed email account is not your default email account, then go to https://list.iu.edu/sympa/signoff/peirce-l . ► PEIRCE-L is owned by THE PEIRCE GROUP; moderated by Gary Richmond; and co-managed by him and Ben Udell.
