When working on your new proposal, remember that in NARS all measurements must be based on what the system has --- limited evidence and resources. I don't allow any "objective probability" that only exists in a Platonic world or the infinite future.
Pei On Sun, Sep 21, 2008 at 1:53 PM, Abram Demski <[EMAIL PROTECTED]> wrote: > Hmm... I didn't mean infinite evidence, only infinite time and space > with which to compute the consequences of evidence. But that is > interesting too. > > The higher-order probabilities I'm talking about introducing do not > reflect inaccuracy at all. :) > This may seem odd, but it seems to me to follow from your development > of NARS... so the difficulty for me is to account for why you can > exclude it in your system. Of course, this need only arises from > interpreting your definitions probabilistically. > > I think I have come up with a more specific proposal. I will try to > write it up properly and see if it works. > > --Abram > > On Sat, Sep 20, 2008 at 11:28 PM, Pei Wang <[EMAIL PROTECTED]> wrote: >> On Sat, Sep 20, 2008 at 11:02 PM, Abram Demski <[EMAIL PROTECTED]> wrote: >>> You are right in what you say about (1). The truth is, my analysis is >>> meant to apply to NARS operating with unrestricted time and memory >>> resources (which of course is not the point of NARS!). So, the >>> question is whether NARS approaches a probability calculation as it is >>> given more time to use all its data. >> >> That is an interesting question. When the weight of evidence w goes to >> infinite, so does confidence, and frequency converge to the limit of >> positive evidence among all evidence, so it becomes probability, under >> a certain interpretation. Therefore, as far as a single truth value is >> concerned, probability theory is an extreme case of NARS. >> >> However, to take all truth values in the system into account, it is >> not necessarily true, because the two theories specify the relations >> among statements/propositions differently. For example, probability >> theory has conditional B|A, while NARS uses implication A==>B, which >> are similar, but not the same. Of course, there are some overlaps, >> such as disjunction and conjunction, where NARS converges to >> probability theory in the extreme case (infinite evidence). >> >>> As for higher values... NARS and PLN may be using them for the purpose >>> you mention, but that is not the purpose I am giving them in my >>> analysis! In my analysis, I am simply trying to justify the deductions >>> allowed in NARS in a probabilistic way. Higher-order probabilities are >>> potentially useful here because of the way you sum evidence. Simply >>> put, it is as if NARS purposefully ignores the distinction between >>> different probability levels, so that a NARS frequency is also a >>> frequency-of-frequencies and frequency-of-frequency-of frequencies and >>> so on, all the way up. >> >> I see what you mean, but as it is currently defined, in NARS there is >> no need to introduce higher-order probabilities --- frequency is not >> an estimation of a "true probability". It is uncertain because the >> influence of new evidence, not because it is inaccurate. >> >>> The simple way of dealing with this is to say that it is wrong, and >>> results from a confusion of similar-looking mathematical entities. >>> But, to some extent, it is intuitive: I should not care too much in >>> normal reasoning which "level" of inheritance I'm using when I say >>> that a truck is a type of vehicle. So the question is, can this be >>> justified probabilistically? I think I can give a very tentative >>> "yes". >> >> Hopefully we'll know better about that when you explore further. ;-) >> >> Pei >> >>> --Abram >>> >>> On Sat, Sep 20, 2008 at 9:38 PM, Pei Wang <[EMAIL PROTECTED]> wrote: >>>> On Sat, Sep 20, 2008 at 9:09 PM, Abram Demski <[EMAIL PROTECTED]> wrote: >>>>>> >>>>>> (1) In probability theory, an event E has a constant probability P(E) >>>>>> (which can be unknown). Given the assumption of insufficient knowledge >>>>>> and resources, in NARS P(A-->B) would change over time, when more and >>>>>> more evidence is taken into account. This process cannot be treated as >>>>>> conditioning, because, among other things, the system can neither >>>>>> explicitly list all evidence as condition, nor update the probability >>>>>> of all statements in the system for each piece of new evidence (so as >>>>>> to treat all background knowledge as a default condition). >>>>>> Consequently, at any moment P(A-->B) and P(B-->C) may be based on >>>>>> different, though unspecified, data, so it is invalid to use them in a >>>>>> rule to calculate the "probability" of A-->C --- probability theory >>>>>> does not allow cross-distribution probability calculation. >>>>> >>>>> This is not a problem the way I set things up. The likelihood of a >>>>> statement is welcome to change over time, as the evidence changes. >>>> >>>> If each of them is changed independently, you don't have a single >>>> probability distribution anymore, but a bunch of them. In the above >>>> case, you don't really have P(A-->B) and P(B-->C), but P_307(A-->B) >>>> and P_409(B-->C). How can you use two probability values together if >>>> they come from different distributions? >>>> >>>>>> (2) For the same reason, in NARS a statement might get different >>>>>> "probability" attached, when derived from different evidence. >>>>>> Probability theory does not have a general rule to handle >>>>>> inconsistency within a probability distribution. >>>>> >>>>> The same statement holds for PLN, right? >>>> >>>> Yes. Ben proposed a solution, which I won't comment until I see all >>>> the details in the PLN book. >>>> >>>>>> The first half is fine, but the second isn't. As the previous example >>>>>> shows, in NARS a high Confidence does implies that the Frequency value >>>>>> is a good summary of evidence, but a low Confidence does implies that >>>>>> the Frequency is bad, just that it is not very stable. >>>>> >>>>> But I'm not talking about confidence when I say "higher". I'm talking >>>>> about the system of levels I defined, for which it is perfectly OK. >>>> >>>> Yes, but the whole purpose of adding another value is to handle >>>> inconsistency and belief revision. Higher-order probability is >>>> mathematically sound, but won't do this work. >>>> >>>> Think about a concrete example: if from one source the system gets >>>> P(A-->B) = 0.9, and P(P(A-->B) = 0.9) = 0.5, while from another source >>>> P(A-->B) = 0.2, and P(P(A-->B) = 0.2) = 0.7, then what will be the >>>> conclusion when the two sources are considered together? >>>> >>>> Pei >>>> >>>> >>>> ------------------------------------------- >>>> agi >>>> Archives: https://www.listbox.com/member/archive/303/=now >>>> RSS Feed: https://www.listbox.com/member/archive/rss/303/ >>>> Modify Your Subscription: https://www.listbox.com/member/?& >>>> Powered by Listbox: http://www.listbox.com >>>> >>> >>> >>> ------------------------------------------- >>> agi >>> Archives: https://www.listbox.com/member/archive/303/=now >>> RSS Feed: https://www.listbox.com/member/archive/rss/303/ >>> Modify Your Subscription: https://www.listbox.com/member/?& >>> Powered by Listbox: http://www.listbox.com >>> >> >> >> ------------------------------------------- >> agi >> Archives: https://www.listbox.com/member/archive/303/=now >> RSS Feed: https://www.listbox.com/member/archive/rss/303/ >> Modify Your Subscription: https://www.listbox.com/member/?& >> Powered by Listbox: http://www.listbox.com >> > > > ------------------------------------------- > agi > Archives: https://www.listbox.com/member/archive/303/=now > RSS Feed: https://www.listbox.com/member/archive/rss/303/ > Modify Your Subscription: https://www.listbox.com/member/?& > Powered by Listbox: http://www.listbox.com > ------------------------------------------- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=114414975-3c8e69 Powered by Listbox: http://www.listbox.com