At 2:39 PM -0800 3/28/02, Hal Finney wrote:
>Bill Jefferys, <[EMAIL PROTECTED]>, writes:
>>  >>  Ockham's razor is a consequence of probability theory, if you look at
>>  >  > things from a Bayesian POV, as I do.
>>
>>  This is well known in Bayesian circles as the Bayesian Ockham's
>>  Razor. A simple discussion is found in the paper that Jim Berger and
>>  I wrote:
>>
>>      http://bayesrules.net/papers/ockham.pdf
>
>This is an interesting paper, however it uses a slightly unusual
>interpretation of Ockham's Razor.  Usually this is stated as that the
>simpler theory is preferred, or as your paper says, "an explanation of
>the facts should be no more complicated than is necessary."  However the
>bulk of your paper seems to use a different definition, which is that
>the simpler theory is the one which is more easily falsified and which
>makes sharper predictions.
>
>I think most people have an intuitive sense of what "simpler" means, and
>while being more easily falsified frequently means being simpler, they
>aren't exactly the same.  It is true that a theory with many parameters
>is both more complex and often less easily falsified, because it has more
>knobs to tweak to try to match the facts.  So the two concepts often do
>go together.
>
>But not always.  You give the example of a strongly biased coin being
>a simpler hypothesis than a fair coin.  I don't think that is what
>most people mean by "simpler".  If anything, the fair coin seems like
>a simpler hypothesis (by the common meaning) since a biased coin has a
>parameter to tweak, the degree of bias.

Depends on whether you know the degree of bias. If you are choosing 
between a two-headed coin and a fair coin, the two-headed coin is 
simpler since it can explain only one outcome, whereas a fair coin 
would be consistent with any outcome. On the other hand, if you don't 
know the bias, then between a fair coin and a coin with unknown bias, 
the fair coin is simpler. This automatically pops out when you do the 
analysis.

>By equating "simpler" with "more easily falsified" you are able to tie it
>into the Bayesian paradigm, which essentially deals with falsifiability.
>A more easily falsified theory gets a Bayesian boost when it happens to
>be correct, because that was a priori unlikely.  But I don't think you
>can legitimately say that this is a Bayesian version of Ockham's Razor,
>because you have to use this rather specialized definition of simple,
>which is more restricted than what people usually mean when they are
>discussing Ockham.

Regardless, it is called the Bayesian Ockham's razor in the 
literature; I will agree that there are some differences between it 
and the "philosopher's" Ockham's razor, and Jim and I (and other 
Bayesians) don't claim otherwise. The interesting thing is that a 
Bayesian approach automatically penalizes models with more parameters 
relative to those with fewer parameters. It does not rely on _ad 
hockery_.

Bill

Reply via email to