David, That's why, imho, the rules need to be *learned* (and, when need be, unlearned). IE, what we need to work on is general learning algorithms, not general visual processing algorithms.
As you say, there's not even such a thing as a general visual processing algorithm. Learning algorithms suffer similar environment-dependence, but (by their nature) not as severe... --Abram On Thu, Jul 8, 2010 at 3:17 PM, David Jones <[email protected]> wrote: > I've learned something really interesting today. I realized that general > rules of inference probably don't really exists. There is no such thing as > complete generality for these problems. The rules of inference that work for > one environment would fail in alien environments. > > So, I have to modify my approach to solving these problems. As I studied > over simplified problems, I realized that there are probably an infinite > number of environments with their own behaviors that are not representative > of the environments we want to put a general AI in. > > So, it is not ok to just come up with any case study and solve it. The case > study has to actually be representative of a problem we want to solve in an > environment we want to apply AI. Otherwise the solution required will take > too long to develop because of it tries to accommodate too much > "generality". As I mentioned, such a general solution is likely impossible. > So, someone could easily get stuck trying to solve an impossible task of > creating one general solution to too many problems that don't allow a > general solution. > > The best course is a balance between the time required to write a very > general solution and the time required to write less general solutions for > multiple problem types and environments. The best way to do this is to > choose representative case studies to solve and make sure the solutions are > truth-tropic and justified for the environments they are to be applied. > > Dave > > > On Sun, Jun 27, 2010 at 1:31 AM, David Jones <[email protected]>wrote: > >> A method for comparing hypotheses in explanatory-based reasoning: * >> >> We prefer the hypothesis or explanation that ***expects* more >> observations. If both explanations expect the same observations, then the >> simpler of the two is preferred (because the unnecessary terms of the more >> complicated explanation do not add to the predictive power).* >> >> *Why are expected events so important?* They are a measure of 1) >> explanatory power and 2) predictive power. The more predictive and the more >> explanatory a hypothesis is, the more likely the hypothesis is when compared >> to a competing hypothesis. >> >> Here are two case studies I've been analyzing from sensory perception of >> simplified visual input: >> The goal of the case studies is to answer the following: How do you >> generate the most likely motion hypothesis in a way that is general and >> applicable to AGI? >> *Case Study 1)* Here is a link to an example: animated gif of two black >> squares move from left to >> right<http://practicalai.org/images/CaseStudy1.gif>. >> *Description: *Two black squares are moving in unison from left to right >> across a white screen. In each frame the black squares shift to the right so >> that square 1 steals square 2's original position and square two moves an >> equal distance to the right. >> *Case Study 2) *Here is a link to an example: the interrupted >> square<http://practicalai.org/images/CaseStudy2.gif>. >> *Description:* A single square is moving from left to right. Suddenly in >> the third frame, a single black square is added in the middle of the >> expected path of the original black square. This second square just stays >> there. So, what happened? Did the square moving from left to right keep >> moving? Or did it stop and then another square suddenly appeared and moved >> from left to right? >> >> *Here is a simplified version of how we solve case study 1: >> *The important hypotheses to consider are: >> 1) the square from frame 1 of the video that has a very close position to >> the square from frame 2 should be matched (we hypothesize that they are the >> same square and that any difference in position is motion). So, what >> happens is that in each two frames of the video, we only match one square. >> The other square goes unmatched. >> 2) We do the same thing as in hypothesis #1, but this time we also match >> the remaining squares and hypothesize motion as follows: the first square >> jumps over the second square from left to right. We hypothesize that this >> happens over and over in each frame of the video. Square 2 stops and square >> 1 jumps over it.... over and over again. >> 3) We hypothesize that both squares move to the right in unison. This is >> the correct hypothesis. >> >> So, why should we prefer the correct hypothesis, #3 over the other two? >> >> Well, first of all, #3 is correct because it has the most explanatory >> power of the three and is the simplest of the three. Simpler is better >> because, with the given evidence and information, there is no reason to >> desire a more complicated hypothesis such as #2. >> >> So, the answer to the question is because explanation #3 expects the most >> observations, such as: >> 1) the consistent relative positions of the squares in each frame are >> expected. >> 2) It also expects their new positions in each from based on velocity >> calculations. >> 3) It expects both squares to occur in each frame. >> >> Explanation 1 ignores 1 square from each frame of the video, because it >> can't match it. Hypothesis #1 doesn't have a reason for why the a new square >> appears in each frame and why one disappears. It doesn't expect these >> observations. In fact, explanation 1 doesn't expect anything that happens >> because something new happens in each frame, which doesn't give it a chance >> to confirm its hypotheses in subsequent frames. >> >> The power of this method is immediately clear. It is general and it solves >> the problem very cleanly. >> >> *Here is a simplified version of how we solve case study 2:* >> We expect the original square to move at a similar velocity from left to >> right because we hypothesized that it did move from left to right and we >> calculated its velocity. If this expectation is confirmed, then it is more >> likely than saying that the square suddenly stopped and another started >> moving. Such a change would be unexpected and such a conclusion would be >> unjustifiable. >> >> I also believe that explanations which generate fewer incorrect >> expectations should be preferred over those that more incorrect >> expectations. >> >> The idea I came up with earlier this month regarding high frame rates to >> reduce uncertainty is still applicable. It is important that all generated >> hypotheses have as low uncertainty as possible given our constraints and >> resources available. >> >> I thought I'd share my progress with you all. I'll be testing the ideas on >> test cases such as the ones I mentioned in the coming days and weeks. >> >> Dave >> > > *agi* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/> | > Modify<https://www.listbox.com/member/?&>Your Subscription > <http://www.listbox.com> > -- Abram Demski http://lo-tho.blogspot.com/ http://groups.google.com/group/one-logic ------------------------------------------- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c Powered by Listbox: http://www.listbox.com
