We all make conjectures all of the time, but we don't often don't have anyway to establish credibility for the claims that are made. So I wanted to examine one part of this field, and the idea that seemed most natural for me was Solomonoff Induction. I have reached a conclusion about the subject and that conclusion is that all of the claims that I have seen made about Solomonoff Induction are rationally unfounded including the one that you just made when you said: "We can produce upper bounds that get closer and closer w/o getting arbitrarily near, and we can produce numbers which do approach arbitrarily near to the correct number in the limit but sometimes dip below for a time; but we can't have both features."
Your inability to fully recognize or perhaps acknowledge that you cannot use Solomonoff Induction to make this claim is difficult for me to comprehend. While the fields of compression and probability have an impressive body of evidence supporting them, I simply have no reason to think the kind of claims that have been made about Solomonoff Induction have any merit. By natural induction I feel comfortable drawing the conclusion that this whole area related to algorithmic information theory is based on shallow methods of reasoning. It can be useful, as it was for me, just as means of exploring ideas that I would not have otherwise explored. But its usefulness comes in learning how to determine its lack of merit. I will write one more thing about my feelings about computability, but I will start a new thread and just mention the relation to this thread. Jim Bromer On Thu, Jul 15, 2010 at 2:45 PM, Abram Demski <[email protected]> wrote: > Jim, > > Yes this is true & provable: there is no way to compute a correct error > bound such that it converges to 0 as the computation of algorithmic > probability converges to the correct number. More specifically--- we can > approximate the algorithmic probability from below, computing better lower > bounds which converge to the correct number, but we cannot approximate it > from above, as there is no procedure (and can never be any procedure) which > creates closer and closer upper bounds converging to the correct number. > > (We can produce upper bounds that get closer and closer w/o getting > arbitrarily near, and we can produce numbers which do approach arbitrarily > near to the correct number in the limit but sometimes dip below for a time; > but we can't have both features.) > > The question of whether the function would be useful for the "sorts of > things we keep talking about" ... well, I think the best argument that I can > give is that MDL is strongly supported by both theory and practice for many > *subsets* of the full program space. The concern might be that, so far, it > is only supported by *theory* for the full program space-- and since > approximations have very bad error-bound properties, it may never be > supported in practice. My reply to this would be that it still appears > useful to approximate Solomonoff induction, since most successful predictors > can be viewed as approximations to Solomonoff induction. "It approximates > solomonoff induction" appears to be a good _explanation_ for the success of > many systems. > > What sort of alternatives do you have in mind, by the way? > > --Abram > > On Thu, Jul 15, 2010 at 11:50 AM, Jim Bromer <[email protected]>wrote: > >> I think that Solomonoff Induction includes a computational method that >> produces probabilities of some sort and whenever those probabilities were >> computed (in a way that would make the function computable) they would sum >> up to 1. But the issue that I am pointing out is that there is no way that >> you can determine the margin of error in what is being computed for what has >> been repeatedly claimed that the function is capable of computing. Since >> you are not able to rely on something like the theory of limits, you are not >> able to determine the degree of error in what is being computed. And in >> fact, there is no way to determine that what the function would compute >> would be in any way useful for the sort of things that you guys keep talking >> about. >> >> Jim Bromer >> >> >> >> On Thu, Jul 15, 2010 at 8:18 AM, Jim Bromer <[email protected]> wrote: >> >>> On Wed, Jul 14, 2010 at 7:46 PM, Abram Demski <[email protected]>wrote: >>> >>>> Jim, >>>> >>>> There is a simple proof of convergence for the sum involved in defining >>>> the probability of a given string in the Solomonoff distribution: >>>> >>>> At its greatest, a particular string would be output by *all* programs. >>>> In this case, its sum would come to 1. This puts an upper bound on the sum. >>>> Since there is no subtraction, there is a lower bound at 0 and the sum >>>> monotonically increases as we take the limit. Knowing these facts, suppose >>>> it *didn't* converge. It must then increase without bound, since it cannot >>>> fluctuate back and forth (it can only go up). But this contradicts the >>>> upper >>>> bound of 1. So, the sum must stop at 1 or below (and in fact we can prove >>>> it >>>> stops below 1, though we can't say where precisely without the infinite >>>> computing power required to compute the limit). >>>> >>>> --Abram >>> >>> >>> I believe that Solomonoff Induction would be computable given infinite >>> time and infinite resources (the Godel Theorem fits into this category) but >>> some people disagree for reasons I do not understand. >>> >>> If it is not computable then it is not a mathematical theorem and the >>> question of whether the sum of probabilities equals 1 is pure fantasy. >>> >>> If it is computable then the central issue is whether it could (given >>> infinite time and infinite resources) be used to determine the probability >>> of a particular string being produced from all possible programs. The >>> question about the sum of all the probabilities is certainly an interesting >>> question. However, the problem of making sure that the function was actually >>> computable would interfere with this process of determining the probability >>> of each particular string that can be produced. For example, since some >>> strings would be infinite, the computability problem makes it imperative >>> that the infinite strings be partially computed at an iteration (or else the >>> function would be hung up at some particular iteration and the infinite >>> other calculations could not be considered computable). >>> >>> My criticism is that even though I believe the function may be >>> theoretically computable, the fact that each particular probability (of each >>> particular string that is produced) cannot be proven to approach a limit >>> through mathematical analysis, and since the individual probabilities will >>> fluctuate with each new string that is produced, one would have to know how >>> to reorder the production of the probabilities in order to demonstrate that >>> the individual probabilities do approach a limit. If they don't, then the >>> claim that this function could be used to define the probabilities of a >>> particular string from all possible program is unprovable. (Some >>> infinite calculations fluctuate infinitely.) Since you do not have any way >>> to determine how to reorder the infinite probabilities a priori, your >>> algorithm would have to be able to compute all possible reorderings to find >>> the ordering and filtering of the computations that would produce evaluable >>> limits. Since there are trans infinite rearrangements of an infinite list >>> (I am not sure that I am using the term 'trans infinite' properly) this >>> shows that the conclusion that the theorem can be used to derive the desired >>> probabilities is unprovable through a variation of Cantor's Diagonal >>> Argument, and that you can't use Solomonoff Induction the way you have been >>> talking about using it. >>> >>> Since you cannot fully compute every string that may be produced at a >>> certain iteration, you cannot make the claim that you even know the >>> probabilities of any possible string before infinity and therefore your >>> claim that the sum of the probabilities can be computed is not provable. >>> >>> But I could be wrong. >>> Jim Bromer >>> >> >> *agi* | Archives <https://www.listbox.com/member/archive/303/=now> >> <https://www.listbox.com/member/archive/rss/303/> | >> Modify<https://www.listbox.com/member/?&>Your Subscription >> <http://www.listbox.com/> >> > > > > -- > Abram Demski > http://lo-tho.blogspot.com/ > http://groups.google.com/group/one-logic > *agi* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/> | > Modify<https://www.listbox.com/member/?&>Your Subscription > <http://www.listbox.com/> > ------------------------------------------- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c Powered by Listbox: http://www.listbox.com
