Ben,

I am in absolute agreement with your very eloquent posting. Just one
question: Doesn't this also apply equally well to your OpenCog approach to
AGI? Indeed, THIS (as described in your posting) has been my primary
objection to the past AGI-related efforts that I have seen.

Again, very well said.

Steve
===============
On Tue, Jun 19, 2012 at 6:31 PM, Ben Goertzel <[email protected]> wrote:

>
> There's a general fallacy that misleads many AGI people, of the following
> form ...
>
> "
> -- Capability or method X, if you could do it incredibly (i.e.
> unrealistically) well, would enable arbitrarily great general intelligence
> -- Simple versions of X, seem to lead to interesting "narrow AI" behaviors
> THEREFORE...
> -- By pursuing  more and more complex versions of X, we can get high
> levels (e.g. human-level) of real-world general intelligence
> "
>
> In the case we're discussing here X = Prediction ..
>
> In other cases, X = logical reasoning, or pattern recognition, or
> automated program learning, or simulation, etc. etc.
>
> Unfortunately, things just don't work that way ;/ ...
>
> ben
>
>
> On Tue, Jun 19, 2012 at 9:21 PM, Steve Richfield <
> [email protected]> wrote:
>
>> Jim,
>>
>> I think we are in agreement here. Computing optimal action without a
>> guiding prediction is NOT easy. I mentioned high speed trading because they
>> appear to be doing just that, albeit within a narrow domain. I suspect that
>> failure to grok this area is just one of many areas where AGI is going to
>> have to make progress before it can become "serious".
>>
>> Steve
>> =============
>> On Tue, Jun 19, 2012 at 6:00 PM, Jim Bromer <[email protected]> wrote:
>>
>>> Steve,
>>> High speed trading is something that people are not good at but narrow
>>> AI can be.  It has to be narrow to keep it efficient.
>>> Your idea of optimal action in the absence of prediction is a pretty
>>> wild abstraction, and it would be difficult to implement.  Just using
>>> correlation for example would tell you a lot about the relations of what
>>> was obvious and previously identifiable but little about the relations of
>>> causation and (ironically) co-occurrence.  Correlation can identify perfect
>>> co-occurrence but it cannot be relied on -in itself- to identify imperfect
>>> or conditional co-occurrence.
>>> So for something like correlation to actually work to reliably identify
>>> conditional co-occurrences it has to have some way to identify or speculate
>>> on interactions that may be hidden.  (I think Abram was telling me that
>>> hidden Markov processes were capable of doing this but here the guiding
>>> form of a Markov process is a conditional premise of the efficacy of the
>>> method.)
>>>
>>> In order then to generalize this example of effectively using
>>> correlation just to identify interactive co-occurrence I feel the system
>>> would have to be capable of a dealing with a great many possible
>>> complications just to identify the few interactive co-occurrences that
>>> might have impact on a subsystem that is being observed.  First, how do you
>>> simultaneously watch multiple subsystems?  This is the easy part because
>>> our ideas about simple observations are inherently absurd.  A simple system
>>> can easily be a complicated system.  This relativistic view is the first
>>> clue then to discovering new ideas about simultaneously watching multiple
>>> complex systems.  Many of the systems that can be observed easily are
>>> complex systems. So the real problem is how do we discriminate or recognize
>>> subtle relations that are hidden in the seeming simplicity of an observed
>>> event.  For example if a cognitive system was deriving insight from a
>>> camera, the video of an observation event would contain all the complexity
>>> that could be inferred from what the camera could capture.  So while a
>>> cognitive system might make jump to some real time conclusions, reanalysis
>>> of the recording of the event might provide more insight from a more
>>> sophisticated cognitive basis.
>>>
>>> Although this sounds like nothing new, it is still new just because no
>>> one has made much progress in identifying how representations of complexity
>>> work.
>>>
>>> Jim Bromer
>>>
>>>
>>> On Tue, Jun 19, 2012 at 5:08 PM, Steve Richfield <
>>> [email protected]> wrote:
>>>
>>>> Just a comment to inject into this discussion:
>>>>
>>>> Prediction is great when it is possible, but that is rare in our world
>>>> of imperfect information.
>>>>
>>>> Modern economics has brought us the concept of optimal action in the
>>>> face if imperfect information. This leans on concepts like volatility,
>>>> where tiny (and hence unpredictable) contributions can have huge effects.
>>>>
>>>> I think the emphasis should NOT be prediction, but rather on the
>>>> computation of optimal action in the ABSENCE of prediction. Of course that
>>>> is a more complex concept, so it has so far evaded deep discussion here.
>>>>
>>>> Note that this is the stock and trade of high speed trading software,
>>>> so people ARE already making this work in the real world.
>>>>
>>>> Any thoughts?
>>>>
>>>> Steve Richfield
>>>> ==================
>>>>
>>>> On Tue, Jun 19, 2012 at 1:50 PM, Jim Bromer <[email protected]>wrote:
>>>>
>>>>> Logan Streondj <[email protected]> wrote:
>>>>> Infering is a form of carrying knowledge from one place and applying
>>>>> it in a different place or time.
>>>>>
>>>>>
>>>>> This is the real question but it does not provide us with an answer.
>>>>> All the narrow forms of AI do offer solutions to certain kinds of 
>>>>> problems,
>>>>> but is there a general way to work from uncertainty (about most every
>>>>> basis to make the determination) toward greater certainty that would allow
>>>>> us to say that a particular kind of knowledge that worked in another
>>>>> situation could work in this situation?  If you base inference on
>>>>> similarities then the problem is how do you use automation (in other words
>>>>> a program) to detect similarities without some absolute method that some 
>>>>> of
>>>>> the aspects of the two similar events are of a kind?
>>>>>
>>>>> Jim Bromer
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> On Tue, Jun 19, 2012 at 3:42 PM, Jim Bromer <[email protected]>wrote:
>>>>>
>>>>>> Logan,
>>>>>> I also cannot predict the kinds of replies that I will get.
>>>>>> Of course, prediction is something that human beings can do.  I
>>>>>> mentioned that in my previous message.  However, I do not see this group 
>>>>>> as
>>>>>> swirling around the question of what kinds of things human beings can do
>>>>>> but around the question of what can we do to make our computer programs 
>>>>>> act
>>>>>> smarter.  In that sense, intelligence is not a product of prediction,
>>>>>> prediction is a product of intelligence.
>>>>>> If anyone has proof that AGI is a product of prediction then he has
>>>>>> solved the problems that are constantly being discussed in this group.
>>>>>>
>>>>>> I think there was another author who popularized probability and
>>>>>> prediction in the 1970s.
>>>>>>
>>>>>> Your definition of inference is interesting and it is a more
>>>>>> sophisticated way of understanding the problem then the usual refrain of
>>>>>> the false promise of probability and prediction.
>>>>>> Jim Bromer
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Tue, Jun 19, 2012 at 3:33 PM, Logan Streondj 
>>>>>> <[email protected]>wrote:
>>>>>>
>>>>>>> I agree with Jim that there is too much focus on "prediction".
>>>>>>>
>>>>>>> Though here the obsession might be related to how Ray Kurzweil, a
>>>>>>> man famous for his accurate predictions, chooses to define intelligence 
>>>>>>> as
>>>>>>> the ability to predict.
>>>>>>>
>>>>>>> If prediction was a normal thing, then Ray wouldn't have gotten
>>>>>>> famous for it.
>>>>>>> Amongst us ordinary humans, at least around here in Canada, and on
>>>>>>> the internet, people rarely if ever mention the word "prediction". And 
>>>>>>> very
>>>>>>> few if anyone claims to make accurate predictions.
>>>>>>>
>>>>>>> Kurzweil mentions prediction as necessary for things like hunting or
>>>>>>> gathering food.
>>>>>>> For instance projecting the vector of an animal as is runs away,
>>>>>>> or infering that if fruit was gathered in an area at a time of year,
>>>>>>> it may be available at same place next year at similar time.
>>>>>>>
>>>>>>> However projecting, infering, and predicting are 3 different words,
>>>>>>> with different meanings.
>>>>>>>
>>>>>>> Projecting is  a form of planning ahead.
>>>>>>> Infering  is a form of carrying knowledge from one place and
>>>>>>> applying it in a different place or time.
>>>>>>> Predicting is a form of prophecy or foretelling based on special
>>>>>>> knowledge.
>>>>>>>
>>>>>>> Most people infer and project, but very few predict.
>>>>>>>
>>>>>>> For instance in New Age cultures, many humans have had their careers
>>>>>>> broken by making false predictions.  Also no one (but Ray) claims to 
>>>>>>> make
>>>>>>> predictions based on intelligence, but usually based on things like
>>>>>>> intuition, and telepathic communications.
>>>>>>>
>>>>>>> Maybe I'm committing the fallacy of making a distinction without a
>>>>>>> difference. Perhaps all those words mean the same thing.
>>>>>>>
>>>>>>> Though I'd like to state that there are other ways of getting food,
>>>>>>> that don't relate to predictive ability.
>>>>>>>  For instance how does an animal get food, they find a signature of
>>>>>>> the food-item, for instance a smell, sound or shape,  and then move 
>>>>>>> towards
>>>>>>> it, until their jaws are clenching it.
>>>>>>> I guess you could say they "predicted" the food was there, based on
>>>>>>> the signature.
>>>>>>> As it certainly is possible that following a signature, will lead to
>>>>>>> a non-food item,
>>>>>>> for instance if the signature is lost, or is being produced by
>>>>>>> something else, i.e. carnivorous flower.
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> It is also possible that everything can be classified as prediction,
>>>>>>> for instance when saying the alphabet, we predict what the next
>>>>>>> letter is, before saying it.
>>>>>>> This would mean that every hello world program uses prediction, as
>>>>>>> it loads the hello world string, before printing it to screen.
>>>>>>> *shrugs*
>>>>>>>
>>>>>>> it depends, do we want prediction to lose all meaning, by applying
>>>>>>> it just about anything.
>>>>>>> Or do we want to be specific with what we are saying, and only use
>>>>>>> words like "prediction" in the way that non AGI mailing list people use
>>>>>>> it,   as a shorter-term version of prophesy, based on special knowledge.
>>>>>>>
>>>>>>> For instance even though I consider myself intelligence, I can't
>>>>>>> predict the content of any reply or if there even will be replies,  
>>>>>>> though
>>>>>>> I can infer that there may be replies, as this is a mailing list, where
>>>>>>> people often reply.
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Tue, Jun 19, 2012 at 1:48 PM, Piaget Modeler <
>>>>>>> [email protected]> wrote:
>>>>>>>
>>>>>>>>  Jim,
>>>>>>>>
>>>>>>>> Is this your actual belief or is this disinformation?
>>>>>>>>
>>>>>>>> ------------------------------
>>>>>>>> Date: Tue, 19 Jun 2012 10:40:22 -0400
>>>>>>>> Subject: [agi] Prediction Did Not Work (except in narrow ai.)
>>>>>>>> From: [email protected]
>>>>>>>> To: [email protected]
>>>>>>>>
>>>>>>>>
>>>>>>>> The original idea behind the use of "prediction" in AI was that the
>>>>>>>> prediction could be compared against the actuality and that comparison
>>>>>>>> could be used to test the theory that produced the prediction.  (Some
>>>>>>>> author popularized that model for AI but it was proposed by academic
>>>>>>>> researchers before he did so. Karl Popper used the concept as part of 
>>>>>>>> his
>>>>>>>> model of scientific discovery in the 1930s, but his principles, which 
>>>>>>>> were
>>>>>>>> based on logical positivism, have become more dubious because logical
>>>>>>>> certainty has become a more dubious principle of knowledge. And, oh, 
>>>>>>>> by the
>>>>>>>> way, Popper did not believe that AI was possible.)  So, continuing 
>>>>>>>> with the
>>>>>>>> march of the use of "prediction" in AI, AI people could see that our
>>>>>>>> expectations were like "predictions" and so it did seem that the human
>>>>>>>> mind did indeed use a method of prediction.  Of course the principle 
>>>>>>>> that a
>>>>>>>> prediction could be compared against an actuality in order to evaluate 
>>>>>>>> the
>>>>>>>> accuracy of a theory only works in narrow AI, and as narrow AI failed 
>>>>>>>> to
>>>>>>>> produce simple AGI that part of the cherished notion of "prediction" 
>>>>>>>> has
>>>>>>>> been gradually eroded.
>>>>>>>>
>>>>>>>> This group uses the term prediction to simply refer to something
>>>>>>>> that is "known" and as such it is a concept which is pretty shallow 
>>>>>>>> since
>>>>>>>> its verification as a mental product is thereby based on the experience
>>>>>>>> that when we know something we act as if we were confident that it 
>>>>>>>> would
>>>>>>>> happen.  The problem with such concepts like "knowing" or "prediction" 
>>>>>>>> is
>>>>>>>> that they -do not- confirm the efficacy of theories that an AGI program
>>>>>>>> might produce, except in those circumstances which would be considered
>>>>>>>> narrow AI by this group.
>>>>>>>>
>>>>>>>> Let me repeat that.
>>>>>>>> The problem with such concepts like "knowing" or "prediction" is
>>>>>>>> that they -DO NOT- confirm the effaciacy of theories that an AGI 
>>>>>>>> program
>>>>>>>> might produce, except in those circumstances which would be considered
>>>>>>>> narrow AI by this group.
>>>>>>>>
>>>>>>>> So sure, when someone points out that the human mind uses
>>>>>>>> "expectation" and expectation is a little like "prediction" I do agree.
>>>>>>>> But here the word prediction is just being used to describe "knowing
>>>>>>>> something."  There is no principle of confirmation or disconfirmation 
>>>>>>>> of
>>>>>>>> the use of "prediction" that can be used to produce AGI, except for 
>>>>>>>> special
>>>>>>>> cases.  After years and years of the repetition of the word in these 
>>>>>>>> types
>>>>>>>> of discussions there is still no AGI so that should give you a hint 
>>>>>>>> about
>>>>>>>> how good an idea it was.
>>>>>>>>
>>>>>>>> If the use of prediction as a confirming method can only be used
>>>>>>>> in a limited set of circumstances then its power in these discussions 
>>>>>>>> has
>>>>>>>> been so diminished that it should not be used as if it were a magical
>>>>>>>> concept.  Without some efficacy the word should not be used as a 
>>>>>>>> special
>>>>>>>> technical term.  The word should be used in the way it is usually used.
>>>>>>>>
>>>>>>>> As I implied, Popper originally used the word the concept in a
>>>>>>>> logical model of scientific theory.  If a theory could be used to 
>>>>>>>> predict a
>>>>>>>> confirming or disconfirming observable event then the theory could be
>>>>>>>> disconfirmed by the failure of the event to occur.  (If the event 
>>>>>>>> occurred
>>>>>>>> it still might be caused by a coincidence.)
>>>>>>>>
>>>>>>>> It is coming back to me. (Or else my creative memory is kicking
>>>>>>>> in.)  The author who popularized the theory of confirmation through
>>>>>>>> prediction had a model of probability and confirmation by prediction.  
>>>>>>>> That
>>>>>>>> model is inherently contradictory.
>>>>>>>>
>>>>>>>> It amazes me that you guys don't get this.
>>>>>>>>
>>>>>>>> Jim Bromer
>>>>>>>>    *AGI* | Archives<https://www.listbox.com/member/archive/303/=now>
>>>>>>>> <https://www.listbox.com/member/archive/rss/303/19999924-5cfde295>|
>>>>>>>> Modify <https://www.listbox.com/member/?&;> Your Subscription
>>>>>>>> <http://www.listbox.com>
>>>>>>>>    *AGI* | Archives<https://www.listbox.com/member/archive/303/=now>
>>>>>>>> <https://www.listbox.com/member/archive/rss/303/5037279-6ef01b0b>|
>>>>>>>> Modify <https://www.listbox.com/member/?&;> Your Subscription
>>>>>>>> <http://www.listbox.com>
>>>>>>>>
>>>>>>>
>>>>>>>    *AGI* | Archives<https://www.listbox.com/member/archive/303/=now>
>>>>>>> <https://www.listbox.com/member/archive/rss/303/10561250-164650b2>|
>>>>>>> Modify <https://www.listbox.com/member/?&;> Your Subscription
>>>>>>> <http://www.listbox.com>
>>>>>>>
>>>>>>
>>>>>>
>>>>>    *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
>>>>> <https://www.listbox.com/member/archive/rss/303/10443978-6f4c28ac> |
>>>>> Modify <https://www.listbox.com/member/?&;> Your Subscription
>>>>> <http://www.listbox.com>
>>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> Full employment can be had with the stoke of a pen. Simply institute a
>>>> six hour workday. That will easily create enough new jobs to bring back
>>>> full employment.
>>>>
>>>>
>>>>    *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
>>>> <https://www.listbox.com/member/archive/rss/303/10561250-164650b2> |
>>>> Modify <https://www.listbox.com/member/?&;> Your Subscription
>>>> <http://www.listbox.com>
>>>>
>>>
>>>    *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
>>> <https://www.listbox.com/member/archive/rss/303/10443978-6f4c28ac> |
>>> Modify <https://www.listbox.com/member/?&;> Your Subscription
>>> <http://www.listbox.com>
>>>
>>
>>
>>
>> --
>> Full employment can be had with the stoke of a pen. Simply institute a
>> six hour workday. That will easily create enough new jobs to bring back
>> full employment.
>>
>>
>>    *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
>> <https://www.listbox.com/member/archive/rss/303/212726-11ac2389> | 
>> Modify<https://www.listbox.com/member/?&;>Your Subscription
>> <http://www.listbox.com>
>>
>
>
>
> --
> Ben Goertzel, PhD
> http://goertzel.org
>
> "My humanity is a constant self-overcoming" -- Friedrich Nietzsche
>
>    *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
> <https://www.listbox.com/member/archive/rss/303/10443978-6f4c28ac> |
> Modify<https://www.listbox.com/member/?&;>Your Subscription
> <http://www.listbox.com>
>



-- 
Full employment can be had with the stoke of a pen. Simply institute a six
hour workday. That will easily create enough new jobs to bring back full
employment.



-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-c97d2393
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-2484a968
Powered by Listbox: http://www.listbox.com

Reply via email to