Pei,

I sympathize with your care in wording, because I'm very aware of the
strange meaning that the word "model" takes on in formal accounts of
semantics. While a cognitive scientist might talk about a person's
"model of the world", a logician would say that the world is "a model
of a first-order theory". I do want to avoid the second meaning. But,
I don't think I could fare well by saying "system" instead, because
the models are only a part of the larger system... so I'm not sure
there is a word that is both neutral and sufficiently meaningful.

Do you think it is impossible to apply probability to open
models/theories/systems, or merely undesirable?

On Thu, Sep 4, 2008 at 8:10 PM, Pei Wang <[EMAIL PROTECTED]> wrote:
> Abram,
>
> I agree with the spirit of your post, and I even go further to include
> "being open" in my working definition of intelligence --- see
> http://nars.wang.googlepages.com/wang.logic_intelligence.pdf
>
> I also agree with your comment on Solomonoff induction and Bayesian prior.
>
> However, I talk about "open system", not "open model", because I think
> model-theoretic semantics is the wrong theory to be used here --- see
> http://nars.wang.googlepages.com/wang.semantics.pdf
>
> Pei
>
> On Thu, Sep 4, 2008 at 2:19 PM, Abram Demski <[EMAIL PROTECTED]> wrote:
>> A closed model is one that is interpreted as representing all truths
>> about that which is modeled. An open model is instead interpreted as
>> making a specific set of assertions, and leaving the rest undecided.
>> Formally, we might say that a closed model is interpreted to include
>> all of the truths, so that any other statements are false. This is
>> also known as the closed-world assumption.
>>
>> A typical example of an open model is a set of statements in predicate
>> logic. This could be changed to a closed model simply by applying the
>> closed-world assumption. A possibly more typical example of a
>> closed-world model is a computer program that outputs the data so far
>> (and predicts specific future output), as in Solomonoff induction.
>>
>> These two types of model are very different! One important difference
>> is that we can simply *add* to an open model if we need to account for
>> new data, while we must always *modify* a closed model if we want to
>> account for more information.
>>
>> The key difference I want to ask about here is: a length-based
>> bayesian prior seems to apply well to closed models, but not so well
>> to open models.
>>
>> First, such priors are generally supposed to apply to entire joint
>> states; in other words, probability theory itself (and in particular
>> bayesian learning) is built with an assumption of an underlying space
>> of closed models, not open ones.
>>
>> Second, an open model always has room for additional stuff somewhere
>> else in the universe, unobserved by the agent. This suggests that,
>> made probabilistic, open models would generally predict universes with
>> infinite description length. Whatever information was known, there
>> would be an infinite number of chances for other unknown things to be
>> out there; so it seems as if the probability of *something* more being
>> there would converge to 1. (This is not, however, mathematically
>> necessary.) If so, then taking that other thing into account, the same
>> argument would still suggest something *else* was out there, and so
>> on; in other words, a probabilistic open-model-learner would seem to
>> predict a universe with an infinite description length. This does not
>> make it easy to apply the description length principle.
>>
>> I am not arguing that open models are a necessity for AI, but I am
>> curious if anyone has ideas of how to handle this. I know that Pei
>> Wang suggests abandoning standard probability in order to learn open
>> models, for example.
>>
>> --Abram Demski
>>
>>
>> -------------------------------------------
>> agi
>> Archives: https://www.listbox.com/member/archive/303/=now
>> RSS Feed: https://www.listbox.com/member/archive/rss/303/
>> Modify Your Subscription: https://www.listbox.com/member/?&;
>> Powered by Listbox: http://www.listbox.com
>>
>
>
> -------------------------------------------
> agi
> Archives: https://www.listbox.com/member/archive/303/=now
> RSS Feed: https://www.listbox.com/member/archive/rss/303/
> Modify Your Subscription: https://www.listbox.com/member/?&;
> Powered by Listbox: http://www.listbox.com
>


-------------------------------------------
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=111637683-c8fa51
Powered by Listbox: http://www.listbox.com

Reply via email to