There's a general fallacy that misleads many AGI people, of the following form ...
" -- Capability or method X, if you could do it incredibly (i.e. unrealistically) well, would enable arbitrarily great general intelligence -- Simple versions of X, seem to lead to interesting "narrow AI" behaviors THEREFORE... -- By pursuing more and more complex versions of X, we can get high levels (e.g. human-level) of real-world general intelligence " In the case we're discussing here X = Prediction .. In other cases, X = logical reasoning, or pattern recognition, or automated program learning, or simulation, etc. etc. Unfortunately, things just don't work that way ;/ ... ben On Tue, Jun 19, 2012 at 9:21 PM, Steve Richfield <[email protected]>wrote: > Jim, > > I think we are in agreement here. Computing optimal action without a > guiding prediction is NOT easy. I mentioned high speed trading because they > appear to be doing just that, albeit within a narrow domain. I suspect that > failure to grok this area is just one of many areas where AGI is going to > have to make progress before it can become "serious". > > Steve > ============= > On Tue, Jun 19, 2012 at 6:00 PM, Jim Bromer <[email protected]> wrote: > >> Steve, >> High speed trading is something that people are not good at but narrow AI >> can be. It has to be narrow to keep it efficient. >> Your idea of optimal action in the absence of prediction is a pretty wild >> abstraction, and it would be difficult to implement. Just using >> correlation for example would tell you a lot about the relations of what >> was obvious and previously identifiable but little about the relations of >> causation and (ironically) co-occurrence. Correlation can identify perfect >> co-occurrence but it cannot be relied on -in itself- to identify imperfect >> or conditional co-occurrence. >> So for something like correlation to actually work to reliably identify >> conditional co-occurrences it has to have some way to identify or speculate >> on interactions that may be hidden. (I think Abram was telling me that >> hidden Markov processes were capable of doing this but here the guiding >> form of a Markov process is a conditional premise of the efficacy of the >> method.) >> >> In order then to generalize this example of effectively using correlation >> just to identify interactive co-occurrence I feel the system would have to >> be capable of a dealing with a great many possible complications just >> to identify the few interactive co-occurrences that might have impact on a >> subsystem that is being observed. First, how do you simultaneously watch >> multiple subsystems? This is the easy part because our ideas about simple >> observations are inherently absurd. A simple system can easily be a >> complicated system. This relativistic view is the first clue then to >> discovering new ideas about simultaneously watching multiple complex >> systems. Many of the systems that can be observed easily are complex >> systems. So the real problem is how do we discriminate or recognize subtle >> relations that are hidden in the seeming simplicity of an observed event. >> For example if a cognitive system was deriving insight from a camera, the >> video of an observation event would contain all the complexity that could >> be inferred from what the camera could capture. So while a cognitive >> system might make jump to some real time conclusions, reanalysis of the >> recording of the event might provide more insight from a more sophisticated >> cognitive basis. >> >> Although this sounds like nothing new, it is still new just because no >> one has made much progress in identifying how representations of complexity >> work. >> >> Jim Bromer >> >> >> On Tue, Jun 19, 2012 at 5:08 PM, Steve Richfield < >> [email protected]> wrote: >> >>> Just a comment to inject into this discussion: >>> >>> Prediction is great when it is possible, but that is rare in our world >>> of imperfect information. >>> >>> Modern economics has brought us the concept of optimal action in the >>> face if imperfect information. This leans on concepts like volatility, >>> where tiny (and hence unpredictable) contributions can have huge effects. >>> >>> I think the emphasis should NOT be prediction, but rather on the >>> computation of optimal action in the ABSENCE of prediction. Of course that >>> is a more complex concept, so it has so far evaded deep discussion here. >>> >>> Note that this is the stock and trade of high speed trading software, so >>> people ARE already making this work in the real world. >>> >>> Any thoughts? >>> >>> Steve Richfield >>> ================== >>> >>> On Tue, Jun 19, 2012 at 1:50 PM, Jim Bromer <[email protected]> wrote: >>> >>>> Logan Streondj <[email protected]> wrote: >>>> Infering is a form of carrying knowledge from one place and applying it >>>> in a different place or time. >>>> >>>> >>>> This is the real question but it does not provide us with an answer. >>>> All the narrow forms of AI do offer solutions to certain kinds of problems, >>>> but is there a general way to work from uncertainty (about most every >>>> basis to make the determination) toward greater certainty that would allow >>>> us to say that a particular kind of knowledge that worked in another >>>> situation could work in this situation? If you base inference on >>>> similarities then the problem is how do you use automation (in other words >>>> a program) to detect similarities without some absolute method that some of >>>> the aspects of the two similar events are of a kind? >>>> >>>> Jim Bromer >>>> >>>> >>>> >>>> >>>> On Tue, Jun 19, 2012 at 3:42 PM, Jim Bromer <[email protected]>wrote: >>>> >>>>> Logan, >>>>> I also cannot predict the kinds of replies that I will get. >>>>> Of course, prediction is something that human beings can do. I >>>>> mentioned that in my previous message. However, I do not see this group >>>>> as >>>>> swirling around the question of what kinds of things human beings can do >>>>> but around the question of what can we do to make our computer programs >>>>> act >>>>> smarter. In that sense, intelligence is not a product of prediction, >>>>> prediction is a product of intelligence. >>>>> If anyone has proof that AGI is a product of prediction then he has >>>>> solved the problems that are constantly being discussed in this group. >>>>> >>>>> I think there was another author who popularized probability and >>>>> prediction in the 1970s. >>>>> >>>>> Your definition of inference is interesting and it is a more >>>>> sophisticated way of understanding the problem then the usual refrain of >>>>> the false promise of probability and prediction. >>>>> Jim Bromer >>>>> >>>>> >>>>> >>>>> On Tue, Jun 19, 2012 at 3:33 PM, Logan Streondj <[email protected]>wrote: >>>>> >>>>>> I agree with Jim that there is too much focus on "prediction". >>>>>> >>>>>> Though here the obsession might be related to how Ray Kurzweil, a man >>>>>> famous for his accurate predictions, chooses to define intelligence as >>>>>> the >>>>>> ability to predict. >>>>>> >>>>>> If prediction was a normal thing, then Ray wouldn't have gotten >>>>>> famous for it. >>>>>> Amongst us ordinary humans, at least around here in Canada, and on >>>>>> the internet, people rarely if ever mention the word "prediction". And >>>>>> very >>>>>> few if anyone claims to make accurate predictions. >>>>>> >>>>>> Kurzweil mentions prediction as necessary for things like hunting or >>>>>> gathering food. >>>>>> For instance projecting the vector of an animal as is runs away, >>>>>> or infering that if fruit was gathered in an area at a time of year, >>>>>> it may be available at same place next year at similar time. >>>>>> >>>>>> However projecting, infering, and predicting are 3 different words, >>>>>> with different meanings. >>>>>> >>>>>> Projecting is a form of planning ahead. >>>>>> Infering is a form of carrying knowledge from one place and applying >>>>>> it in a different place or time. >>>>>> Predicting is a form of prophecy or foretelling based on special >>>>>> knowledge. >>>>>> >>>>>> Most people infer and project, but very few predict. >>>>>> >>>>>> For instance in New Age cultures, many humans have had their careers >>>>>> broken by making false predictions. Also no one (but Ray) claims to make >>>>>> predictions based on intelligence, but usually based on things like >>>>>> intuition, and telepathic communications. >>>>>> >>>>>> Maybe I'm committing the fallacy of making a distinction without a >>>>>> difference. Perhaps all those words mean the same thing. >>>>>> >>>>>> Though I'd like to state that there are other ways of getting food, >>>>>> that don't relate to predictive ability. >>>>>> For instance how does an animal get food, they find a signature of >>>>>> the food-item, for instance a smell, sound or shape, and then move >>>>>> towards >>>>>> it, until their jaws are clenching it. >>>>>> I guess you could say they "predicted" the food was there, based on >>>>>> the signature. >>>>>> As it certainly is possible that following a signature, will lead to >>>>>> a non-food item, >>>>>> for instance if the signature is lost, or is being produced by >>>>>> something else, i.e. carnivorous flower. >>>>>> >>>>>> >>>>>> >>>>>> It is also possible that everything can be classified as prediction, >>>>>> for instance when saying the alphabet, we predict what the next >>>>>> letter is, before saying it. >>>>>> This would mean that every hello world program uses prediction, as it >>>>>> loads the hello world string, before printing it to screen. >>>>>> *shrugs* >>>>>> >>>>>> it depends, do we want prediction to lose all meaning, by applying it >>>>>> just about anything. >>>>>> Or do we want to be specific with what we are saying, and only use >>>>>> words like "prediction" in the way that non AGI mailing list people use >>>>>> it, as a shorter-term version of prophesy, based on special knowledge. >>>>>> >>>>>> For instance even though I consider myself intelligence, I can't >>>>>> predict the content of any reply or if there even will be replies, >>>>>> though >>>>>> I can infer that there may be replies, as this is a mailing list, where >>>>>> people often reply. >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> On Tue, Jun 19, 2012 at 1:48 PM, Piaget Modeler < >>>>>> [email protected]> wrote: >>>>>> >>>>>>> Jim, >>>>>>> >>>>>>> Is this your actual belief or is this disinformation? >>>>>>> >>>>>>> ------------------------------ >>>>>>> Date: Tue, 19 Jun 2012 10:40:22 -0400 >>>>>>> Subject: [agi] Prediction Did Not Work (except in narrow ai.) >>>>>>> From: [email protected] >>>>>>> To: [email protected] >>>>>>> >>>>>>> >>>>>>> The original idea behind the use of "prediction" in AI was that the >>>>>>> prediction could be compared against the actuality and that comparison >>>>>>> could be used to test the theory that produced the prediction. (Some >>>>>>> author popularized that model for AI but it was proposed by academic >>>>>>> researchers before he did so. Karl Popper used the concept as part of >>>>>>> his >>>>>>> model of scientific discovery in the 1930s, but his principles, which >>>>>>> were >>>>>>> based on logical positivism, have become more dubious because logical >>>>>>> certainty has become a more dubious principle of knowledge. And, oh, by >>>>>>> the >>>>>>> way, Popper did not believe that AI was possible.) So, continuing with >>>>>>> the >>>>>>> march of the use of "prediction" in AI, AI people could see that our >>>>>>> expectations were like "predictions" and so it did seem that the human >>>>>>> mind did indeed use a method of prediction. Of course the principle >>>>>>> that a >>>>>>> prediction could be compared against an actuality in order to evaluate >>>>>>> the >>>>>>> accuracy of a theory only works in narrow AI, and as narrow AI failed to >>>>>>> produce simple AGI that part of the cherished notion of "prediction" has >>>>>>> been gradually eroded. >>>>>>> >>>>>>> This group uses the term prediction to simply refer to something >>>>>>> that is "known" and as such it is a concept which is pretty shallow >>>>>>> since >>>>>>> its verification as a mental product is thereby based on the experience >>>>>>> that when we know something we act as if we were confident that it would >>>>>>> happen. The problem with such concepts like "knowing" or "prediction" >>>>>>> is >>>>>>> that they -do not- confirm the efficacy of theories that an AGI program >>>>>>> might produce, except in those circumstances which would be considered >>>>>>> narrow AI by this group. >>>>>>> >>>>>>> Let me repeat that. >>>>>>> The problem with such concepts like "knowing" or "prediction" is >>>>>>> that they -DO NOT- confirm the effaciacy of theories that an AGI program >>>>>>> might produce, except in those circumstances which would be considered >>>>>>> narrow AI by this group. >>>>>>> >>>>>>> So sure, when someone points out that the human mind uses >>>>>>> "expectation" and expectation is a little like "prediction" I do agree. >>>>>>> But here the word prediction is just being used to describe "knowing >>>>>>> something." There is no principle of confirmation or disconfirmation of >>>>>>> the use of "prediction" that can be used to produce AGI, except for >>>>>>> special >>>>>>> cases. After years and years of the repetition of the word in these >>>>>>> types >>>>>>> of discussions there is still no AGI so that should give you a hint >>>>>>> about >>>>>>> how good an idea it was. >>>>>>> >>>>>>> If the use of prediction as a confirming method can only be used >>>>>>> in a limited set of circumstances then its power in these discussions >>>>>>> has >>>>>>> been so diminished that it should not be used as if it were a magical >>>>>>> concept. Without some efficacy the word should not be used as a special >>>>>>> technical term. The word should be used in the way it is usually used. >>>>>>> >>>>>>> As I implied, Popper originally used the word the concept in a >>>>>>> logical model of scientific theory. If a theory could be used to >>>>>>> predict a >>>>>>> confirming or disconfirming observable event then the theory could be >>>>>>> disconfirmed by the failure of the event to occur. (If the event >>>>>>> occurred >>>>>>> it still might be caused by a coincidence.) >>>>>>> >>>>>>> It is coming back to me. (Or else my creative memory is kicking >>>>>>> in.) The author who popularized the theory of confirmation through >>>>>>> prediction had a model of probability and confirmation by prediction. >>>>>>> That >>>>>>> model is inherently contradictory. >>>>>>> >>>>>>> It amazes me that you guys don't get this. >>>>>>> >>>>>>> Jim Bromer >>>>>>> *AGI* | Archives<https://www.listbox.com/member/archive/303/=now> >>>>>>> <https://www.listbox.com/member/archive/rss/303/19999924-5cfde295>| >>>>>>> Modify <https://www.listbox.com/member/?&> Your Subscription >>>>>>> <http://www.listbox.com> >>>>>>> *AGI* | Archives<https://www.listbox.com/member/archive/303/=now> >>>>>>> <https://www.listbox.com/member/archive/rss/303/5037279-6ef01b0b> | >>>>>>> Modify <https://www.listbox.com/member/?&> Your Subscription >>>>>>> <http://www.listbox.com> >>>>>>> >>>>>> >>>>>> *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> >>>>>> <https://www.listbox.com/member/archive/rss/303/10561250-164650b2> | >>>>>> Modify <https://www.listbox.com/member/?&> Your Subscription >>>>>> <http://www.listbox.com> >>>>>> >>>>> >>>>> >>>> *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> >>>> <https://www.listbox.com/member/archive/rss/303/10443978-6f4c28ac> | >>>> Modify <https://www.listbox.com/member/?&> Your Subscription >>>> <http://www.listbox.com> >>>> >>> >>> >>> >>> -- >>> Full employment can be had with the stoke of a pen. Simply institute a >>> six hour workday. That will easily create enough new jobs to bring back >>> full employment. >>> >>> >>> *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> >>> <https://www.listbox.com/member/archive/rss/303/10561250-164650b2> | >>> Modify <https://www.listbox.com/member/?&> Your Subscription >>> <http://www.listbox.com> >>> >> >> *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> >> <https://www.listbox.com/member/archive/rss/303/10443978-6f4c28ac> | >> Modify <https://www.listbox.com/member/?&> Your Subscription >> <http://www.listbox.com> >> > > > > -- > Full employment can be had with the stoke of a pen. Simply institute a six > hour workday. That will easily create enough new jobs to bring back full > employment. > > > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/212726-11ac2389> | > Modify<https://www.listbox.com/member/?&>Your Subscription > <http://www.listbox.com> > -- Ben Goertzel, PhD http://goertzel.org "My humanity is a constant self-overcoming" -- Friedrich Nietzsche ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-c97d2393 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-2484a968 Powered by Listbox: http://www.listbox.com
