:) You say that as if bayesian explanatory reasoning is the only way.

There is much debate over bayesian explanatory reasoning and non-bayesian.
There are pros and cons to bayesian methods. Likewise, there is the problem
with non-bayesian methods because few have figured out how to do it
effectively. I'm still going to pursue a non-bayesian approach because I
believe there is likely more merit to it and that the short-comings can be
overcome.

Dave

On Thu, Jul 15, 2010 at 10:54 AM, Matt Mahoney <[email protected]> wrote:

> Hypotheses are scored using Bayes law. Let D be your observed data and H be
> your hypothesis. Then p(H|D) = p(D|H)p(H)/p(D). Since p(D) is constant, you
> can remove it and rank hypotheses by p(D|H)p(H).
>
> p(H) can be estimated using the minimum description length principle or
> Solomonoff induction. Ideally, p(H) = 2^-|H| where |H| is the length (in
> bits) of the description of the hypothesis. The value is language dependent,
> so this method is not perfect.
>
>
> -- Matt Mahoney, [email protected]
>
>
> ------------------------------
> *From:* David Jones <[email protected]>
> *To:* agi <[email protected]>
> *Sent:* Thu, July 15, 2010 10:22:44 AM
> *Subject:* Re: [agi] How do we Score Hypotheses?
>
> It is no wonder that I'm having a hard time finding documentation on
> hypothesis scoring. Few can agree on how to do it and there is much debate
> about it.
>
> I noticed though that a big reason for the problems is that explanatory
> reasoning is being applied to many diverse problems. I think, like I
> mentioned before, that people should not try to come up with a single
> universal rule set for applying explanatory reasoning to every possible
> problem. So, maybe that's where the hold up is.
>
> I've been testing my ideas out on complex examples. But now I'm going to go
> back to simplified model testing (although not as simple as black squares :)
> ) and work my way up again.
>
> Dave
>
> On Wed, Jul 14, 2010 at 12:59 PM, David Jones <[email protected]>wrote:
>
>> Actually, I just realized that there is a way to included inductive
>> knowledge and experience into this algorithm. Inductive knowledge and
>> experience about a specific object or object type can be exploited to know
>> which hypotheses in the past were successful, and therefore which hypothesis
>> is most likely. By choosing the most likely hypothesis first, we skip a lot
>> of messy hypothesis comparison processing and analysis. If we choose the
>> right hypothesis first, all we really have to do is verify that this
>> hypothesis reveals in the data what we expect to be there. If we confirm
>> what we expect, that is reason enough not to look for other hypotheses
>> because the data is explained by what we originally believed to be likely.
>> We only look for additional hypotheses when we find something unexplained.
>> And even then, we don't look at the whole problem. We only look at what we
>> have to to explain the unexplained data. In fact, we could even ignore the
>> unexplained data if we believe, from experience, that it isn't pertinent.
>>
>> I discovered this because I'm analyzing how a series of hypotheses are
>> navigated when analyzing images. It seems to me that it is done very
>> similarly to way we do it. We sort of confirm what we expect and try to
>> explain what we don't expect. We try out hypotheses in a sort of trial and
>> error manor and see how each hypothesis affects what we find in the image.
>> If we confirm things because of the hypothesis, we are likely to keep it. We
>> keep going, navigating the tree of hypotheses, conflicts and unexpected
>> observations until we find a good hypothesis. Something like that. I'm
>> attempting to construct an algorithm for doing this as I analyze specific
>> problems.
>>
>> Dave
>>
>>
>> On Wed, Jul 14, 2010 at 10:22 AM, David Jones <[email protected]>wrote:
>>
>>> What do you mean by definitive events?
>>>
>>> I guess the first problem I see with my approach is that the movement of
>>> the window is also a hypothesis. I need to analyze it in more detail and see
>>> how the tree of hypotheses affects the hypotheses regarding the "e"s on the
>>> windows.
>>>
>>> What I believe is that these problems can be broken down into types of
>>> hypotheses,  types of events and types of relationships. then those types
>>> can be reasoned about in a general way. If possible, then you have a method
>>> for reasoning about any object that is covered by the types of hypotheses,
>>> events and relationships that you have defined.
>>>
>>> How to reason about specific objects should not be preprogrammed. But, I
>>> think the solution to this part of AGI is to find general ways to reason
>>> about a small set of concepts that can be combined to describe specific
>>> objects and situations.
>>>
>>> There are other parts to AGI that I am not considering yet. I believe the
>>> problem has to be broken down into separate pieces and understood before
>>> putting it back together into a complete system. I have not covered
>>> inductive learning for example, which would be an important part of AGI. I
>>> have also not yet incorporated learned experience into the algorithm, which
>>> is also important.
>>>
>>> The general AI problem is way too complicated to consider all at once. I
>>> simply can't solve hypothesis generation, comparison and disambiguation
>>> while at the same time solving induction and experience-based reasoning. It
>>> becomes unwieldly. So, I'm starting where I can and I'll work my way up to
>>> the full complexity of the problem.
>>>
>>> I don't really understand what you mean here: "The central unsolved
>>> problem, in my view, is: How can hypotheses be conceptually integrated along
>>> with the observable definitive events of the problem to form good
>>> explanatory connections that can mesh well with other knowledge about the
>>> problem that is considered to be reliable.  The second problem is finding
>>> efficient ways to represent this complexity of knowledge so that the program
>>> can utilize it efficiently."
>>>
>>> You also might want to include concrete problems to analyze for your
>>> central problem suggestions. That would help define the problem a bit better
>>> for analysis.
>>>
>>> Dave
>>>
>>>
>>> On Wed, Jul 14, 2010 at 8:30 AM, Jim Bromer <[email protected]> wrote:
>>>
>>>>
>>>>
>>>> On Tue, Jul 13, 2010 at 9:05 PM, Jim Bromer <[email protected]>wrote:
>>>> Even if you refined your model until it was just right, you would have
>>>> only caught up to everyone else with a solution to a narrow AI problem.
>>>>
>>>>
>>>> I did not mean that you would just have a solution to a narrow AI
>>>> problem, but that your solution, if put in the form of scoring of points on
>>>> the basis of the observation *of definitive* events, would constitute a
>>>> narrow AI method.  The central unsolved problem, in my view, is: How can
>>>> hypotheses be conceptually integrated along with the observable definitive
>>>> events of the problem to form good explanatory connections that can mesh
>>>> well with other knowledge about the problem that is considered to be
>>>> reliable.  The second problem is finding efficient ways to represent this
>>>> complexity of knowledge so that the program can utilize it efficiently.
>>>>
>>>>    *agi* | Archives <https://www.listbox.com/member/archive/303/=now>
>>>> <https://www.listbox.com/member/archive/rss/303/> | 
>>>> Modify<https://www.listbox.com/member/?&;>Your Subscription
>>>> <http://www.listbox.com>
>>>>
>>>
>>>
>>
>    *agi* | Archives <https://www.listbox.com/member/archive/303/=now>
> <https://www.listbox.com/member/archive/rss/303/> | 
> Modify<https://www.listbox.com/member/?&;>Your Subscription
> <http://www.listbox.com>
>    *agi* | Archives <https://www.listbox.com/member/archive/303/=now>
> <https://www.listbox.com/member/archive/rss/303/> | 
> Modify<https://www.listbox.com/member/?&;>Your Subscription
> <http://www.listbox.com>
>



-------------------------------------------
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com

Reply via email to