I'm considering nonmonotonic reasoning using Bayes net, and got stuck.
There is an example on p483 of J Pearl's 1988 book PRIIS:
Given:
"birds can fly"
"penguins are birds"
"penguins cannot fly"
The desiderata is to conclude that "penguins are birds, but penguins
cannot fly".
Pearl translates the KB to:
P(f | b) = high
P(f | p) = low
P(b | p) = high
where high and low means arbitrarily close to 1 and 0, respectively.
If you draw this on paper you'll see a triangular loop.
Then Pearl continues to deduce:
Conditioning P(f | p) on both b and ~b,
P(f | p) = P(f | p,b) P(b | p) + P(f | p,~b) [1-P(b | p)]
> P(f | p,b) P(b | p)
Thus
P(f | p,b) < P(f | p) / P(b | p) which is close to 0.
Thus Pearl concludes that "given penguin and bird, fly is not true".
But I found something wrong here. It seems that the Bayes net is
loopy and we can conclude that "fly" given "penguin" and "bird" can be
either 0 or 1. (The loop is somewhat symmetric).
Ben, do you have a similar problem dealing with nonmonotonicity using
probabilistic networks?
YKY
-------------------------------------------
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription:
http://www.listbox.com/member/?member_id=8660244&id_secret=106510220-47b225
Powered by Listbox: http://www.listbox.com