List, Gary, JAS I agree, I think, in gist, with what JAS has to say regarding the limitations of computation with respect to what one might term "genuine" semiosis (or genuine anything, to be honest).
The core of the misfit is that the simulation is a series of numbers — that is, whatever the front-end product of any given application/AI, the back-end is merely a series of numbers in one respect or another. That is, to take AI as genuine thought or semiosis, would be, metaphorically, to assume that a television can understand the human interaction with whatever is its "front-end" display when, if we make it AI-ish, it could only refer back to numbers regarding the output (the core is numerical/processual, and this is not true for people's experience of essentially everything). In other words, there is a poverty of stimulus, unbridgeable in AI which, though the models get processually better, as would any dynamic algorithm over time, especially ones which use people for training —they can never possibly understand what is never processual to begin with when no matter how sophisticated they get, they will always be processual in one degree or another: people need not, ever, be that way at all. Valuable tools, but reminds me of the pharmakon (Plato/Derrida). A double edged sword unless you can design it (I have some proto designs) in such a way that the person is in control and the AI (of whatever model) does two things: one, takes a seconday position to the person meanwhile also runs its own primary-secondary operations but which (even here) the person has final say over. Vast potential for this kind of thing across all sectors but ontologically, which is the debate here, really, I agree with much of what has been said and I think we see rare agreement with everyone (or many) on this list regarding this topic (increasingly, it is an agreement which is, by my discernment, also being understood in the sciences in different ways). Best, Jack ________________________________ From: [email protected] <[email protected]> on behalf of Jon Alan Schmidt <[email protected]> Sent: Wednesday, January 7, 2026 6:42 PM To: Peirce-L <[email protected]> Subject: Re: [PEIRCE-L] Surdity, Feeling, and Consciousness, was, Truth and dyadic consciousnessg Gary R., Gary F., List: The discussion of surdity echoes a List exchange<https://list.iu.edu/sympa/arc/peirce-l/2025-09/msg00024.html> that happened a few months ago, in which I noted that "every proposition involves at least one index or precept for finding an index to denote its dynamical object, not just words describing that object"; and the linked paper is reminiscent of the reservations that I expressed at that time about characterizing LLMs as "artificial intelligence." JAS: One of Peirce's definitions of "logic as semiotic" is the science that ascertains "what must be the characters of all signs used by a 'scientific' intelligence, that is to say, by an intelligence capable of learning by experience" (CP 2.227, c. 1897). LLMs do not qualify because they do not have that capability--they have no beliefs with corresponding habits of conduct that can be corroborated or confounded by the "outward clash" with reality itself. Just like formal systems of deductive logic, LLMs cannot establish truth, only preserve it once they have been "trained" by being fed a vast quantity of information; the principle of "garbage in, garbage out" still applies. LLMs cannot even verify the truth of any given sentence apart from comparing it with other sentences, so it is no wonder that they sometimes suffer from "hallucinations" that produce false or misleading responses to prompts. Manheim wonders whether LLMs themselves could someday be "genuine Peircean interpretants," but what he presumably means is that they could someday be quasi-minds whose determinations are "genuine Peircean interpretants." Even this is not quite right, because according to Peirce, any actual effect of a sign token is a dynamical interpretant of that sign token--a determination of an interpreting quasi-mind, which need not be an individual conscious mind like that of a human, such that LLMs presumably qualify already. My sense is that Manheim is really asking whether LLMs could someday go beyond only accessing and processing texts to have genuine Peircean experience of reality--the compulsive (2ns) aspect of cognition (3ns)--which would render them intrinsically self-correcting in the long run, if they consistently conform to the normative principles of inquiry (method of science). In any case, I still do not share his optimism, for the reasons expressed above. Moreover, although Manheim addresses the objection that LLMs are "finite-state systems" and situates human minds in that same category, he overlooks the fact that the former are always executing discrete operations while the latter (at least arguably) are performing continuous inferences. If genuine semiosis is truly continuous, as I maintain, then a digital computer, no matter how sophisticated, can only ever simulate it--just as the real numbers do not constitute a true continuum, but usefully approximate one for most practical purposes. After all, whenever we humans break up our own reasoning (arguments) into discrete steps--namely, "definitely formulated premisses" and conclusions (argumentations; see CP 6.456, EP 2:435, 1908)--we are always doing so artificially and retrospectively, after the real and continuous inferential process has already run its course (see CP 2.27, 1902; LF 3/1:234-5, 1906). Regards, Jon Alan Schmidt - Olathe, Kansas, USA Structural Engineer, Synechist Philosopher, Lutheran Christian www.LinkedIn.com/in/JonAlanSchmidt<http://www.LinkedIn.com/in/JonAlanSchmidt> / twitter.com/JonAlanSchmidt<http://twitter.com/JonAlanSchmidt> On Tue, Jan 6, 2026 at 8:55 AM <[email protected]<mailto:[email protected]>> wrote: Gary R, list, I’d like to pick up where your post ended, Gary: GR: Can we say that consciousness is, in its 'primitive' form, surd, while mind in its fullest sense is semiosic', that consciousness offers the brute 'given' of existence, while thought supplies the purposive, essentially semiotic structure that consciousness itself cannot provide? GF: Peirce’s remarks about surdity are mostly in the context of Secondness, i.e. dyadic relations. I would say it is the dyadic consciousness that “offers the brute ‘given’ of existence,” and that’s why “real relations” — genuine indexicality — are necessary to ground the ability of triadic or semiotic relations to convey information, or for a proposition to be true. Phaneroscopically, the “pure consciousness” of Firstness lacks that grounding. But this never happens in (what we call) reality! For the sense or feeling of reality to occur in perception, something must be “present to the mind” and other than the mind, external to it or independent of it, in order for words like “true” or “real” to have any meaning. So you can’t have genuine Thirdness without genuine Secondness. Or to put it physiologically, both the perceived object and the perceiver must be embodied, and the quality of the experience is determined by the real relation between the two. The percept is not just a representation of the object, nor is it just an artifact of internal activity within the perceiving mind or brain. It feel rather odd to even be writing this because it seems so obvious and yet feels like beating around the bush when I try to express it verbally. However I just came across an open access article which strikes me as an important application of it. It’s in Philosophy & Technology (2026) 39:9, https://doi.org/10.1007/s13347-025-00975-5, “Language Models’ Hall of Mirrors Problem: Why AI Alignment Requires Peircean Semiosis,” by David Manheim. Here is the abstract: [[ This paper examines some limitations of large language models (LLMs) through the framework of Peircean semiotics. We argue that basic LLMs exist within a “hall of mirrors,” reflecting only the linguistic surface of training data without indexical grounding in a shared external world, and manipulating symbols without participation in socially-mediated epistemology. We then argue that newer developments, including extended context windows, persistent memory, and mediated interactions with reality, are moving towards making newer Artificial Intelligence (AI) systems into genuine Peircean interpretants, and conclude that LLMs may be approaching this goal, and we identify no fundamental architectural barriers that would prevent this. This lens reframes a central challenge for AI alignment: without grounding in the semiotic process, a model’s linguistic encoding of goals may diverge from real-world values. By synthesizing Peirce’s pragmatic view of signs, contemporary discussions of AI alignment, and recent work on relational realism, we illustrate a fundamental epistemological and practical challenge to AI safety and point to part of a solution. ]] Love, gary f Coming from the ancestral lands of the Anishinaabeg From: [email protected]<mailto:[email protected]> <[email protected]<mailto:[email protected]>> On Behalf Of Gary Richmond Sent: 2-Jan-26 00:29 To: Peirce List <[email protected]<mailto:[email protected]>>; Gary Fuhrman <[email protected]<mailto:[email protected]>> Subject: [PEIRCE-L] Surdity, Feeling, and Consciousness, was, Truth and dyadic consciousnessg Gary F, List, The material you linked to in Turning Signs on 'consciousness' and 'feeling' is thought-provoking, especially your comments on 'surdity' in relation to 'feeling' and 'consciousness'. I have changed the subject line to show my emphasis on surdity. In TS you write: GF:The truth of a proposition depends on the dyadic or real relation (as opposed to a relation of reason) between that sign and its dynamic object. It must involve ‘action of brute force, physical or psychical,’ of the dynamic object upon the sign, so that the relation between the two is ‘real,’ i.e. surd – no sign can express or describe it. GR: I agree that the real relation between a sign and its object is surd, which concept I'd like to explore a bit in this post. You continued: GF: The Secondness of experience itself is a dyadic relation or dynamic action between two subjects; it is ‘brute,’ ‘surd,’ indicible, ineffable. But a subject capable of both attention and intention can become a host, as it were, of Thirdness, or semiosis, so that another subject can become an ‘object of thought’ (Peirce, CP 1.343, 1903). Now we have a triadic relation involving the object, the sign or ‘thought,’ and the experiencing subject, the system of interpretation or ‘mind.’ This is the essential structure of mental experience. The 1ns of experience would be the pure feeling that there is something other than feeling itself – a world appearing to the subject, and thus becoming an object of attention (emphasis added by GR). This reminded me that Peirce remarked -- and Joe Ransdell emphasized this point in several of his papers, on Peirce-L, and in private conversations -- that there can be no pure icon, which is to say that a qualisign must be embodied in a sinsign in order to be operative at all. So, there are no pure qualisigns in actual semiosis, only qualitative aspects of sinsigns: a qualisign is a 1ns that can signify only by being instantiated as a sinsign and interpreted via a legisign. At bottom consciousness is tied to what is immediate and irreducible, and this is linked to surdity. Consciousness is feeling: “. . . consciousness is nothing but Feeling, in general” (CP 7.365). Feeling is simple, immediate, yet qualitative. It can't infer, intend, or mean anything as it is non-relational and non-purposive. In this primitive (primary?) sense, then, consciousness lacks any internal rational structure: it simply is what it is. This primary consciousness is pure 1ns, a qualitative 'given' of unmediated feeling. This is to say that it is not a sign at all (which, btw, contradicts the pansemiotic views of some theorists). Peirce is careful not to conflate surdity with mentality, so he distinguishes consciousness from mind: consciousness belongs to 1ns as feeling, while mind belongs to 3ns in semiosis. “The mediate element of experience is the mental element, which is semiosic but not necessarily conscious” (CP 7.366). Mentality does not essentially require consciousness since semiosis can -- and often does -- proceed without awareness (the famous example of the growth of a crystal; but there are many others). What defines mind isn't consciousness itself but, rather, final causation, purpose, mind being a living complexus of signs, habits, and consequent effects which these produce. In an essential sense, surdity would seem to take on the form of brute reaction (2ns) which may or may not be conscious. For example, the pain of touching a hot burner is a conscious instance of 2ns, while the reflex of removing the hand from the source of the pain isn't. Can one say that surdity lies in the brute action/reaction itself, while consciousness accompanies surdity only when feeling is present? In contrast, 3ns interprets brute facts and, over time, can make experience intelligible. Yet, I firmly believe that 3ns cannot eliminate surdity without cutting thought off from brute actuality, the hic et nunc. So Peirce argues that thought requires resistance, and that signs require dynamic objects as a kind of check on thought. Consciousness, then, is the ground upon which both semiosis and thought operate. This would seem to resolve an apparent contradiction in Peirce’s use of the term 'consciousness.' In its primary sense, consciousness is feeling, so surd, immediate, so, categorially 1ns. In a looser sense, Peirce speaks of “three modes of consciousness” (CP 8.256): the awareness of feeling, action, law. I would say that this usage represents, at best, a secondary, mediated consciousness. Can we say that consciousness is, in its 'primitive' form, surd, while mind in its fullest sense is semiosic', that consciousness offers the brute 'given' of existence, while thought supplies the purposive, essentially semiotic structure that consciousness itself cannot provide? Best, Gary R
_ _ _ _ _ _ _ _ _ _ ► PEIRCE-L subscribers: Click on "Reply List" or "Reply All" to REPLY ON PEIRCE-L to this message. PEIRCE-L posts should go to [email protected] . ► <a href="mailto:[email protected]">UNSUBSCRIBE FROM PEIRCE-L</a> . But, if your subscribed email account is not your default email account, then go to https://list.iu.edu/sympa/signoff/peirce-l . ► PEIRCE-L is owned by THE PEIRCE GROUP; moderated by Gary Richmond; and co-managed by him and Ben Udell.
