Jim, I think we are in agreement here. Computing optimal action without a guiding prediction is NOT easy. I mentioned high speed trading because they appear to be doing just that, albeit within a narrow domain. I suspect that failure to grok this area is just one of many areas where AGI is going to have to make progress before it can become "serious".
Steve ============= On Tue, Jun 19, 2012 at 6:00 PM, Jim Bromer <[email protected]> wrote: > Steve, > High speed trading is something that people are not good at but narrow AI > can be. It has to be narrow to keep it efficient. > Your idea of optimal action in the absence of prediction is a pretty wild > abstraction, and it would be difficult to implement. Just using > correlation for example would tell you a lot about the relations of what > was obvious and previously identifiable but little about the relations of > causation and (ironically) co-occurrence. Correlation can identify perfect > co-occurrence but it cannot be relied on -in itself- to identify imperfect > or conditional co-occurrence. > So for something like correlation to actually work to reliably identify > conditional co-occurrences it has to have some way to identify or speculate > on interactions that may be hidden. (I think Abram was telling me that > hidden Markov processes were capable of doing this but here the guiding > form of a Markov process is a conditional premise of the efficacy of the > method.) > > In order then to generalize this example of effectively using correlation > just to identify interactive co-occurrence I feel the system would have to > be capable of a dealing with a great many possible complications just > to identify the few interactive co-occurrences that might have impact on a > subsystem that is being observed. First, how do you simultaneously watch > multiple subsystems? This is the easy part because our ideas about simple > observations are inherently absurd. A simple system can easily be a > complicated system. This relativistic view is the first clue then to > discovering new ideas about simultaneously watching multiple complex > systems. Many of the systems that can be observed easily are complex > systems. So the real problem is how do we discriminate or recognize subtle > relations that are hidden in the seeming simplicity of an observed event. > For example if a cognitive system was deriving insight from a camera, the > video of an observation event would contain all the complexity that could > be inferred from what the camera could capture. So while a cognitive > system might make jump to some real time conclusions, reanalysis of the > recording of the event might provide more insight from a more sophisticated > cognitive basis. > > Although this sounds like nothing new, it is still new just because no one > has made much progress in identifying how representations of complexity > work. > > Jim Bromer > > > On Tue, Jun 19, 2012 at 5:08 PM, Steve Richfield < > [email protected]> wrote: > >> Just a comment to inject into this discussion: >> >> Prediction is great when it is possible, but that is rare in our world of >> imperfect information. >> >> Modern economics has brought us the concept of optimal action in the face >> if imperfect information. This leans on concepts like volatility, where >> tiny (and hence unpredictable) contributions can have huge effects. >> >> I think the emphasis should NOT be prediction, but rather on the >> computation of optimal action in the ABSENCE of prediction. Of course that >> is a more complex concept, so it has so far evaded deep discussion here. >> >> Note that this is the stock and trade of high speed trading software, so >> people ARE already making this work in the real world. >> >> Any thoughts? >> >> Steve Richfield >> ================== >> >> On Tue, Jun 19, 2012 at 1:50 PM, Jim Bromer <[email protected]> wrote: >> >>> Logan Streondj <[email protected]> wrote: >>> Infering is a form of carrying knowledge from one place and applying it >>> in a different place or time. >>> >>> >>> This is the real question but it does not provide us with an answer. >>> All the narrow forms of AI do offer solutions to certain kinds of problems, >>> but is there a general way to work from uncertainty (about most every >>> basis to make the determination) toward greater certainty that would allow >>> us to say that a particular kind of knowledge that worked in another >>> situation could work in this situation? If you base inference on >>> similarities then the problem is how do you use automation (in other words >>> a program) to detect similarities without some absolute method that some of >>> the aspects of the two similar events are of a kind? >>> >>> Jim Bromer >>> >>> >>> >>> >>> On Tue, Jun 19, 2012 at 3:42 PM, Jim Bromer <[email protected]> wrote: >>> >>>> Logan, >>>> I also cannot predict the kinds of replies that I will get. >>>> Of course, prediction is something that human beings can do. I >>>> mentioned that in my previous message. However, I do not see this group as >>>> swirling around the question of what kinds of things human beings can do >>>> but around the question of what can we do to make our computer programs act >>>> smarter. In that sense, intelligence is not a product of prediction, >>>> prediction is a product of intelligence. >>>> If anyone has proof that AGI is a product of prediction then he has >>>> solved the problems that are constantly being discussed in this group. >>>> >>>> I think there was another author who popularized probability and >>>> prediction in the 1970s. >>>> >>>> Your definition of inference is interesting and it is a more >>>> sophisticated way of understanding the problem then the usual refrain of >>>> the false promise of probability and prediction. >>>> Jim Bromer >>>> >>>> >>>> >>>> On Tue, Jun 19, 2012 at 3:33 PM, Logan Streondj <[email protected]>wrote: >>>> >>>>> I agree with Jim that there is too much focus on "prediction". >>>>> >>>>> Though here the obsession might be related to how Ray Kurzweil, a man >>>>> famous for his accurate predictions, chooses to define intelligence as the >>>>> ability to predict. >>>>> >>>>> If prediction was a normal thing, then Ray wouldn't have gotten famous >>>>> for it. >>>>> Amongst us ordinary humans, at least around here in Canada, and on the >>>>> internet, people rarely if ever mention the word "prediction". And very >>>>> few >>>>> if anyone claims to make accurate predictions. >>>>> >>>>> Kurzweil mentions prediction as necessary for things like hunting or >>>>> gathering food. >>>>> For instance projecting the vector of an animal as is runs away, >>>>> or infering that if fruit was gathered in an area at a time of year, >>>>> it may be available at same place next year at similar time. >>>>> >>>>> However projecting, infering, and predicting are 3 different words, >>>>> with different meanings. >>>>> >>>>> Projecting is a form of planning ahead. >>>>> Infering is a form of carrying knowledge from one place and applying >>>>> it in a different place or time. >>>>> Predicting is a form of prophecy or foretelling based on special >>>>> knowledge. >>>>> >>>>> Most people infer and project, but very few predict. >>>>> >>>>> For instance in New Age cultures, many humans have had their careers >>>>> broken by making false predictions. Also no one (but Ray) claims to make >>>>> predictions based on intelligence, but usually based on things like >>>>> intuition, and telepathic communications. >>>>> >>>>> Maybe I'm committing the fallacy of making a distinction without a >>>>> difference. Perhaps all those words mean the same thing. >>>>> >>>>> Though I'd like to state that there are other ways of getting food, >>>>> that don't relate to predictive ability. >>>>> For instance how does an animal get food, they find a signature of >>>>> the food-item, for instance a smell, sound or shape, and then move >>>>> towards >>>>> it, until their jaws are clenching it. >>>>> I guess you could say they "predicted" the food was there, based on >>>>> the signature. >>>>> As it certainly is possible that following a signature, will lead to a >>>>> non-food item, >>>>> for instance if the signature is lost, or is being produced by >>>>> something else, i.e. carnivorous flower. >>>>> >>>>> >>>>> >>>>> It is also possible that everything can be classified as prediction, >>>>> for instance when saying the alphabet, we predict what the next letter >>>>> is, before saying it. >>>>> This would mean that every hello world program uses prediction, as it >>>>> loads the hello world string, before printing it to screen. >>>>> *shrugs* >>>>> >>>>> it depends, do we want prediction to lose all meaning, by applying it >>>>> just about anything. >>>>> Or do we want to be specific with what we are saying, and only use >>>>> words like "prediction" in the way that non AGI mailing list people use >>>>> it, as a shorter-term version of prophesy, based on special knowledge. >>>>> >>>>> For instance even though I consider myself intelligence, I can't >>>>> predict the content of any reply or if there even will be replies, though >>>>> I can infer that there may be replies, as this is a mailing list, where >>>>> people often reply. >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> On Tue, Jun 19, 2012 at 1:48 PM, Piaget Modeler < >>>>> [email protected]> wrote: >>>>> >>>>>> Jim, >>>>>> >>>>>> Is this your actual belief or is this disinformation? >>>>>> >>>>>> ------------------------------ >>>>>> Date: Tue, 19 Jun 2012 10:40:22 -0400 >>>>>> Subject: [agi] Prediction Did Not Work (except in narrow ai.) >>>>>> From: [email protected] >>>>>> To: [email protected] >>>>>> >>>>>> >>>>>> The original idea behind the use of "prediction" in AI was that the >>>>>> prediction could be compared against the actuality and that comparison >>>>>> could be used to test the theory that produced the prediction. (Some >>>>>> author popularized that model for AI but it was proposed by academic >>>>>> researchers before he did so. Karl Popper used the concept as part of his >>>>>> model of scientific discovery in the 1930s, but his principles, which >>>>>> were >>>>>> based on logical positivism, have become more dubious because logical >>>>>> certainty has become a more dubious principle of knowledge. And, oh, by >>>>>> the >>>>>> way, Popper did not believe that AI was possible.) So, continuing with >>>>>> the >>>>>> march of the use of "prediction" in AI, AI people could see that our >>>>>> expectations were like "predictions" and so it did seem that the human >>>>>> mind did indeed use a method of prediction. Of course the principle >>>>>> that a >>>>>> prediction could be compared against an actuality in order to evaluate >>>>>> the >>>>>> accuracy of a theory only works in narrow AI, and as narrow AI failed to >>>>>> produce simple AGI that part of the cherished notion of "prediction" has >>>>>> been gradually eroded. >>>>>> >>>>>> This group uses the term prediction to simply refer to something that >>>>>> is "known" and as such it is a concept which is pretty shallow since >>>>>> its verification as a mental product is thereby based on the experience >>>>>> that when we know something we act as if we were confident that it would >>>>>> happen. The problem with such concepts like "knowing" or "prediction" is >>>>>> that they -do not- confirm the efficacy of theories that an AGI program >>>>>> might produce, except in those circumstances which would be considered >>>>>> narrow AI by this group. >>>>>> >>>>>> Let me repeat that. >>>>>> The problem with such concepts like "knowing" or "prediction" is that >>>>>> they -DO NOT- confirm the effaciacy of theories that an AGI program might >>>>>> produce, except in those circumstances which would be considered narrow >>>>>> AI by this group. >>>>>> >>>>>> So sure, when someone points out that the human mind uses >>>>>> "expectation" and expectation is a little like "prediction" I do agree. >>>>>> But here the word prediction is just being used to describe "knowing >>>>>> something." There is no principle of confirmation or disconfirmation of >>>>>> the use of "prediction" that can be used to produce AGI, except for >>>>>> special >>>>>> cases. After years and years of the repetition of the word in these >>>>>> types >>>>>> of discussions there is still no AGI so that should give you a hint about >>>>>> how good an idea it was. >>>>>> >>>>>> If the use of prediction as a confirming method can only be used in a >>>>>> limited set of circumstances then its power in these discussions has been >>>>>> so diminished that it should not be used as if it were a magical concept. >>>>>> Without some efficacy the word should not be used as a special technical >>>>>> term. The word should be used in the way it is usually used. >>>>>> >>>>>> As I implied, Popper originally used the word the concept in a >>>>>> logical model of scientific theory. If a theory could be used to >>>>>> predict a >>>>>> confirming or disconfirming observable event then the theory could be >>>>>> disconfirmed by the failure of the event to occur. (If the event >>>>>> occurred >>>>>> it still might be caused by a coincidence.) >>>>>> >>>>>> It is coming back to me. (Or else my creative memory is kicking in.) >>>>>> The author who popularized the theory of confirmation through prediction >>>>>> had a model of probability and confirmation by prediction. That model is >>>>>> inherently contradictory. >>>>>> >>>>>> It amazes me that you guys don't get this. >>>>>> >>>>>> Jim Bromer >>>>>> *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> >>>>>> <https://www.listbox.com/member/archive/rss/303/19999924-5cfde295> | >>>>>> Modify <https://www.listbox.com/member/?&> Your Subscription >>>>>> <http://www.listbox.com> >>>>>> *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> >>>>>> <https://www.listbox.com/member/archive/rss/303/5037279-6ef01b0b> | >>>>>> Modify <https://www.listbox.com/member/?&> Your Subscription >>>>>> <http://www.listbox.com> >>>>>> >>>>> >>>>> *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> >>>>> <https://www.listbox.com/member/archive/rss/303/10561250-164650b2> | >>>>> Modify <https://www.listbox.com/member/?&> Your Subscription >>>>> <http://www.listbox.com> >>>>> >>>> >>>> >>> *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> >>> <https://www.listbox.com/member/archive/rss/303/10443978-6f4c28ac> | >>> Modify <https://www.listbox.com/member/?&> Your Subscription >>> <http://www.listbox.com> >>> >> >> >> >> -- >> Full employment can be had with the stoke of a pen. Simply institute a >> six hour workday. That will easily create enough new jobs to bring back >> full employment. >> >> >> *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> >> <https://www.listbox.com/member/archive/rss/303/10561250-164650b2> | >> Modify <https://www.listbox.com/member/?&> Your Subscription >> <http://www.listbox.com> >> > > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/10443978-6f4c28ac> | > Modify<https://www.listbox.com/member/?&>Your Subscription > <http://www.listbox.com> > -- Full employment can be had with the stoke of a pen. Simply institute a six hour workday. That will easily create enough new jobs to bring back full employment. ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-c97d2393 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-2484a968 Powered by Listbox: http://www.listbox.com
