Because simpler is not better if it is less predictive.

On Thu, Jul 22, 2010 at 1:21 PM, Abram Demski <abramdem...@gmail.com> wrote:

> Jim,
>
> Why more predictive *and then* simpler?
>
> --Abram
>
> On Thu, Jul 22, 2010 at 11:49 AM, David Jones <davidher...@gmail.com>wrote:
>
>> An Update....
>>
>> I think the following gets to the heart of general AI and what it takes to
>> achieve it. It also provides us with evidence as to why general AI is so
>> difficult. With this new knowledge in mind, I think I will be much more
>> capable now of solving the problems and making it work.
>>
>> I've come to the conclusion lately that the best hypothesis is better
>> because it is more predictive and then simpler than other hypotheses (in
>> that order.... more predictive... then simpler). But, I am amazed at how
>> difficult it is to quantitatively define more predictive and simpler for
>> specific problems. This is why I have sometimes doubted the truth of the
>> statement.
>>
>> In addition, the observations that the AI gets are not representative of
>> all observations! This means that if your measure of "predictiveness"
>> depends on the number of certain observations, it could make mistakes! So,
>> the specific observations you are aware of may be unrepresentative of the
>> predictiveness of a hypothesis relative to the truth. If you try to
>> calculate which hypothesis is more predictive and you don't have the
>> critical observations that would give you the right answer, you may get the
>> wrong answer! This all depends of course on your method of calculation,
>> which is quite elusive to define.
>>
>> Visual input from screenshots, for example, can be somewhat malicious.
>> Things can move, appear, disappear or occlude each other suddenly. So,
>> without sufficient knowledge it is hard to decide whether matches you find
>> between such large changes are because it is the same object or a different
>> object. This may indicate that bias and preprogrammed experience should be
>> introduced to the AI before training. Either that or the training inputs
>> should be carefully chosen to avoid malicious input and to make them nice
>> for learning.
>>
>> This is the "correspondence problem" that is typical of computer vision
>> and has never been properly solved. Such malicious input also makes it
>> difficult to learn automatically because the AI doesn't have sufficient
>> experience to know which changes or transformations are acceptable and which
>> are not. It is immediately bombarded with malicious inputs.
>>
>> I've also realized that if a hypothesis is more "explanatory", it may be
>> better. But quantitatively defining explanatory is also elusive and truly
>> depends on the specific problems you are applying it to because it is a
>> heuristic. It is not a true measure of correctness. It is not loyal to the
>> truth. "More explanatory" is really a heuristic that helps us find
>> hypothesis that are more predictive. The true measure of whether a
>> hypothesis is better is simply the most accurate and predictive hypothesis.
>> That is the ultimate and true measure of correctness.
>>
>> Also, since we can't measure every possible prediction or every last
>> prediction (and we certainly can't predict everything), our measure of
>> predictiveness can't possibly be right all the time! We have no choice but
>> to use a heuristic of some kind.
>>
>> So, its clear to me that the right hypothesis is "more predictive and then
>> simpler". But, it is also clear that there will never be a single measure of
>> this that can be applied to all problems. I hope to eventually find a nice
>> model for how to apply it to different problems though. This may be the
>> reason that so many people have tried and failed to develop general AI. Yes,
>> there is a solution. But there is no silver bullet that can be applied to
>> all problems. Some methods are better than others. But I think another major
>> reason of the failures is that people think they can predict things without
>> sufficient information. By approaching the problem this way, we compound the
>> need for heuristics and the errors they produce because we simply don't have
>> sufficient information to make a good decision with limited evidence. If
>> approached correctly, the right solution would solve many more problems with
>> the same efforts than a poor solution would. It would also eliminate some of
>> the difficulties we currently face if sufficient data is available to learn
>> from.
>>
>> In addition to all this theory about better hypotheses, you have to add on
>> the need to solve problems in reasonable time. This also compounds the
>> difficulty of the problem and the complexity of solutions.
>>
>> I am always fascinated by the extraordinary difficulty and complexity of
>> this problem. The more I learn about it, the more I appreciate it.
>>
>> Dave
>>    *agi* | Archives <https://www.listbox.com/member/archive/303/=now>
>> <https://www.listbox.com/member/archive/rss/303/> | 
>> Modify<https://www.listbox.com/member/?&;>Your Subscription
>> <http://www.listbox.com>
>>
>
>
>
> --
> Abram Demski
> http://lo-tho.blogspot.com/
> http://groups.google.com/group/one-logic
>    *agi* | Archives <https://www.listbox.com/member/archive/303/=now>
> <https://www.listbox.com/member/archive/rss/303/> | 
> Modify<https://www.listbox.com/member/?&;>Your Subscription
> <http://www.listbox.com>
>



-------------------------------------------
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com

Reply via email to