David, Yes, this makes sense to me.
To go back to your original query, I still think you will find a rich community relevant to your work if you look into the MDL literature (which additionally does not rely on probability theory, though as I said I'd say it's equivalent). Perhaps this book might be helpful: http://www.amazon.com/Description-Principle-Adaptive-Computation-Learning/dp/0262072815/ref=sr_1_1?ie=UTF8&s=books&qid=1279036776&sr=8-1 It includes a (short-ish?) section comparing the pros/cons of MDL and Bayesianism, and examining some of the mathematical linkings between them-- with the aim of showing that MDL is a broader principle. I disagree there, of course. :) --Abram On Tue, Jul 13, 2010 at 10:01 AM, David Jones <[email protected]> wrote: > Abram, > > Thanks for the clarification Abram. I don't have a single way to deal with > uncertainty. I try not to decide on a method ahead of time because what I > really want to do is analyze the problems and find a solution. But, at the > same time. I have looked at the probabilistic approaches and they don't seem > to be sufficient to solve the problem as they are now. So, my inclination is > to use methods that don't gloss over important details. For me, the most > important way of dealing with uncertainty is through explanatory-type > reasoning. But, explanatory reasoning has not been well defined yet. So, the > implementation is not yet clear. That's where I am now. > > I've begun to approach problems as follows. I try to break the problem down > and answer the following questions: > 1) How do we come up with or construct possible hypotheses. > 2) How do we compare hypotheses to determine which is better. > 3) How do we lower the uncertainty of hypotheses. > 4) How do we determine the likelihood or strength of a single hypothesis > all by itself. Is it sufficient on its own? > > With those questions in mind, the solution seems to be to break possible > hypotheses down into pieces that are generally applicable. For example, in > image analysis, a particular type of hypothesis might be related to 1) > motion or 2) attachment relationships or 3) change or movement behavior of > an object, etc. > > By breaking the possible hypotheses into very general pieces, you can apply > them to just about any problem. With that as a tool, you can then develop > general methods for resolving uncertainty of such hypotheses using > explanatory scoring, consistency, and even statistical analysis. > > Does that make sense to you? > > Dave > > > On Tue, Jul 13, 2010 at 2:29 AM, Abram Demski <[email protected]>wrote: > >> PS-- I am not denying that statistics is applied probability theory. :) >> When I say they are different, what I mean is that saying "I'm going to use >> probability theory" and "I'm going to use statistics" tend to indicate very >> different approaches. Probability is a set of axioms, whereas statistics is >> a set of methods. The probability theory camp tends to be bayesian, whereas >> the stats camp tends to be frequentist. >> >> Your complaint that probability theory doesn't try to figure out why it >> was wrong in the 30% (or whatever) it misses is a common objection. >> Probability theory glosses over important detail, it encourages lazy >> thinking, etc. However, this all depends on the space of hypotheses being >> examined. Statistical methods will be prone to this objection because they >> are essentially narrow-AI methods: they don't *try* to search in the space >> of all hypotheses a human might consider. An AGI setup can and should have >> such a large hypothesis space. Note that AIXI is typically formulated as >> using a space of crisp (non-probabilistic) hypotheses, though probability >> theory is used to reason about them. This means no theory it considers will >> gloss over detail in this way: every theory completely explains the data. (I >> use AIXI as a convenient example, not because I agree with it.) >> >> --Abram >> > > *agi* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/> | > Modify<https://www.listbox.com/member/?&>Your Subscription > <http://www.listbox.com> > -- Abram Demski http://lo-tho.blogspot.com/ http://groups.google.com/group/one-logic ------------------------------------------- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c Powered by Listbox: http://www.listbox.com
