A brief and non-technical description of the two types of semantics mentioned in the previous discussions:
(1) Model-Theoretic Semantics (MTS) (1.1) There is a world existing independently outside the intelligent system (human or machine). (1.2) In principle, there is an objective description of the world, in terms of objects, their properties, and relations among them. (1.3) Within the intelligent system, its knowledge is an approximation of the objective description of the world. (1.4) The meaning of a symbol within the system is the object it refers to in the world. (1.5) The truth-value of a statement within the system measures how close it approximates the fact in the world. (2) Experience-Grounded Semantics (EGS) (2.1) There is a world existing independently outside the intelligent system (human or machine). [same as (1.1), but the agreement stops here] (2.2) Even in principle, there is no objective description of the world. What the system has is its experience, the history of its interaction of the world. (2.3) Within the intelligent system, its knowledge is a summary of its experience. (2.4) The meaning of a symbol within the system is determined by its role in the experience. (2.5) The truth-value of a statement within the system measures how close it summarizes the relevant part of the experience. To further simplify the description, in the context of learning and reasoning: MTS takes "objective truth" of statements and "real meaning" of terms as aim of approximation, while EGS refuses them, but takes experience (input data) as the only thing to depend on. As usual, each theory has its strength and limitation. The issue is which one is more proper for AGI. MTS has been dominating in math, logic, and computer science, and therefore is accepted by the majority people. Even so, it has been attacked by other people (not only the EGS believers) for many reasons. A while ago I made a figure to illustrate this difference, which is at http://nars.wang.googlepages.com/wang.semantics-figure.pdf . A manifesto of EGS is at http://nars.wang.googlepages.com/wang.semantics.pdf Since the debate on the nature of "truth" and "meaning" has existed for thousands of years, I don't think we can settle down it here by some email exchanges. I just want to let the interested people know the theoretical background of the related discussions. Pei On Sat, Oct 11, 2008 at 8:34 PM, Ben Goertzel <[EMAIL PROTECTED]> wrote: > > > > Hi, > >> >> > What this highlights for me is the idea that NARS truth values attempt >> > to reflect the evidence so far, while probabilities attempt to reflect >> > the world > > I agree that probabilities attempt to reflect the world > >> >> . >> >> Well said. This is exactly the difference between an >> experience-grounded semantics and a model-theoretic semantics. > > I don't agree with this distinction ... unless you are construing "model > theoretic semantics" in a very restrictive way, which then does not apply to > PLN. > > If by model-theoretic semantics you mean something like what Wikipedia says > at http://en.wikipedia.org/wiki/Formal_semantics, > > *** > Model-theoretic semantics is the archetype of Alfred Tarski's semantic > theory of truth, based on his T-schema, and is one of the founding concepts > of model theory. This is the most widespread approach, and is based on the > idea that the meaning of the various parts of the propositions are given by > the possible ways we can give a recursively specified group of > interpretation functions from them to some predefined mathematical domains: > an interpretation of first-order predicate logic is given by a mapping from > terms to a universe of individuals, and a mapping from propositions to the > truth values "true" and "false". > *** > > then yes, PLN's semantics is based on a mapping from terms to a universe of > individuals, and a mapping from propositions to truth values. On the other > hand, these "individuals" may be for instance **elementary sensations or > actions**, rather than higher-level individuals like, say, a specific cat, > or the concept "cat". So there is nothing non-experience-based about > mapping terms into a "individuals" that are the system's direct experience > ... and then building up more abstract terms by grouping these > directly-experience-based terms. > > IMO, the dichotomy between experience-based and model-based semantics is a > misleading one. Model-based semantics has often been used in a > non-experience-based way, but that is not because it fundamentally **has** > to be used in that way. > > To say that PLN tries to model the world, is then just to say that it tries > to make probabilistic predictions about sensations and actions that have not > yet been experienced ... which is certainly the case. > >> >> Once >> again, the difference in truth-value functions is reduced to the >> difference in semantics, what is, what the "truth-value" attempts to >> measure. > > Agreed... > > Ben G > > ________________________________ > agi | Archives | Modify Your Subscription ------------------------------------------- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=117534816-b15a34 Powered by Listbox: http://www.listbox.com