RICHARD LOOSEMORE>>>> I cannot even begin to do justice, here, to the issues
involved in solving "the high dimensional problem of seeking to understand
the meaning of text, which often involve multiple levels of implication,
which would normally be accomplished by some sort of search of a large
semantic space"

You talk as if an extension of some current strategy will solve this ... but
it is not at all clear that any current strategy for solving this 
problem actually does scale up to a full solution to the problem.  I don't
care how many toy examples you come up with, you have to show a 
strategy for dealing with some of the core issues, AND reasons to believe
that those strategies really will work (other than "I find them 
quite promising").

Not only that, but there at least some people (to wit, myself) who believe
there are positive reasons to believe that the current 
strategies *will* not scale up.

ED PORTER>>>>  I don't know if you read the Shastri paper I linked to or
not, but it shows we do know how to do many of the types of implication
which are used in NL.  What he shows needs some extensions, so it is more
generalized, but it and other known inference schemes explain a lot of how
text understanding could be done.  

With regard to the scaling issue, it is a real issue.  But there are
multiple reasons to believe the scaling problems can be overcome.  Not
proofs, Richard, so you are entitled to your doubts.  But open your mind to
the possibilities they present.  They include:

-1---------the likely availability of roughly brain level representational,
computational, and interconnect capacities within the several hundred
thousand to 1 million dollar range in seven to ten years.

-2---------the fact that human experience and representation does not
explode combinatorially.  Instead it is quite finite.  It fits insides our
heads.  

Thus, although you are dealing with extremely high dimensional spaces, most
of that space is empty.  There are know ways to deal with extremely high
dimensional spaces while avoiding the exponential explosion made possible by
such high dimensionality.  

Take the well know Growing Neural Gas (GNG) algorithm.  It automatically
creates a relative compact representation of a possibly infinite dimensional
space, by allocated nodes to only those parts of the high dimensional space
where there is stuff, or, if resource are more limited, where the most stuff
is.

Or take indexing, it takes one only to places in the hyperspace where
something actually occurred or was thought about.  One can have
probabilitistically selected hierarchical indexing (something like John Rose
suggested) which make indexing much more efficient.

-3---------experiential computers focus most learning, most models, and most
search on things that actually have happened in the past or on things that
in many ways are similar to what has happened in the past.  This tends to
greatly reduce representational and search spaces.

When such a system synthesizes or perceives new patterns that have never
happened before the system will normally have to explore large search
spaces, but because of the capacity of brain level hardware it will have
considerable capability to do so.  The type of hardware that will be
available for human-level agi in the next decade will probably have
sustainable cross sectional bandwidths of 10G to 1T messages/sec with 64Byte
payloads/msg.  With branching tree activations and the fact that many
messages will be regional, the total amount of messaging could well be 100G
to 100T such msg/sec.

Lets assume our hardware has 10T msg/sec and that we want to read 10 words a
second.  That would allow 1T msg/word.  With a dumb spreading activation
rule that would allow you to: active the 30K most probably implications; and
for each of them the 3K most probable implications; and for each of them the
300 most probable implications; and for each of them the 30 most probable
implications.  As dumb as this method of inferencing would be, it actually
would make a high percent of the appropriate multi-step inferences,
particularly when you consider that the probability of activation at the
successive stages would be guided by probabilities from other activations in
the current context.

Of course there are much more intelligent ways to guide activation that
this.

Also it is important to understand that at every level in many of the
searches or explorations in such a system there will be guidance and
limitations provided by similar models from past experience, greatly
reducing the amount of or the number of explorations that are required to
produce reasonable results.

-4---------Michael Collins a few years ago had was many AI researches
considered to be the best grammatical parser, which used the kernel trick to
effectively match parse trees in, I think it was, 500K dimensions.  By use
of the Kernel trick the actual computation usually was performed in a small
subset of these dimensions and the parser was relatively efficient. 

-5---------Hecht-Nielsen's sentence completion program (produced by his
"confabulation" see http://r.ucsd.edu), just by appropriately tying together
probabilistic implications learned from sequences of words, automatically
creates grammatically correct sentences that are related to a prior
sentense, allegedly without any knowledge of grammar, using millions of
probability activations per word, without any un-computable combinatorial
explosion.  The search space that is being explored at any one time
theoretically is considering more possibilities than there are particles in
the known universe -- yet it works.  At any given time several, lets, say 6
to 12 word or phrase slots can be under computation, in which each of
approximately 100K or so words or phrases is receiving scores.  One could
consider the search space to include each of the possible words or phrase
being considered in each of those say 10 ordered slots as the possible
permutation of 10 slot fillers each chosen from a set of about 10^5 words or
phrases, a permuation that has (10^5)!/(10^4)! possibilities.  This is a
very large search space  -- just 100!/10! is over 10^151¸and (10^5)!/(10^4)!
is much, much, much larger space than that -- and yet it all compute with
somewhere within several orders of magnitude of a billion opps.  This very
large search space is actually handled with a superposition of probabilities
(somewhat as in quantum computing) which are collapsed in a sequential
manner, in a rippling propagation of decisions and ensuing probability
propagations. 

So Richard there are ways to do searches efficiently in very high
dimensional spaces, including in the case of confabulation spaces that are
in some ways trillions and trillions of times larger than the known universe
-- all on relatively small computers.  

So lift thine eyes up unto Hecht-Nielsen -- (and his cat with whom he
generously shares credit for Confabulation) -- and believe!

Ed Porter



-----Original Message-----
From: Richard Loosemore [mailto:[EMAIL PROTECTED] 
Sent: Monday, December 03, 2007 12:49 PM
To: agi@v2.listbox.com
Subject: Re: Hacker intelligence level [WAS Re: [agi] Funding AGI research]

Ed Porter wrote:
> Richard,
> 
> It is false to imply that knowledge of how to draw implications from a
> series of statements by some sort of search mechanism is equally unknown
as
> that of how to make an anti-gravity drive -- if by "anti-gravity drive"
you
> mean some totally unknown form of physics, rather than just anything, such
> as human legs, that can push against gravity.  
> 
> It is unfair because there is a fair amount of knowledge about how to draw
> implications from sequences of statements.  For example view Shastri's
> www.icsi.berkeley.edu/~shastri/psfiles/cogsci00.ps.  Also Ben Goertzel has
> demonstrated a program that draws implications from statements contained
in
> different medical texts.
> 
> Ed Porter 
> 
> P.S., I have enclosed an inexact, but, at least to me, useful drawing I
made
> of the type of "search" involved in understanding the multiple
implications
> contained in the series of statements contained in Shastri's "John fell in
> the Hallway. Tom had cleaned it.  He was hurt" example.  Of course, what
is
> most missing from this drawing are all the other, dead end, implications
> which do not provide a likely implication.  Only one of such dead end is
> shown (the implication between fall and trip).  As a result you don't
sense
> how many dead ends have to be searched to find the implications which best
> explain the statements.   EWP

Well, bear in mind that I was not meaning the analogy to be *that* 
exact, or I would have given up on AGI long ago - I'm sure you know that 
I don't believe that getting an understanding system working is as 
impossible as getting an AG drive built.

The purpose of my comment was to point to a huge gap in understanding, 
and the mistaken strategy of dealing with all the peripheral issues 
before having a clear idea how to solve the central problem.

I cannot even begin to do justice, here, to the issues involved in 
solving "the high dimensional problem of seeking to understand the 
meaning of text, which often involve multiple levels of implication, 
which would normally be accomplished by some sort of search of a large 
semantic space"

You talk as if an extension of some current strategy will solve this ... 
but it is not at all clear that any current strategy for solving this 
problem actually does scale up to a full solution to the problem.  I 
don't care how many toy examples you come up with, you have to show a 
strategy for dealing with some of the core issues, AND reasons to 
believe that those strategies really will work (other than "I find them 
quite promising").

Not only that, but there at least some people (to wit, myself) who 
believe there are positive reasons to believe that the current 
strategies *will* not scale up.



Richard Loosemore



> -----Original Message-----
> From: Richard Loosemore [mailto:[EMAIL PROTECTED] 
> Sent: Monday, December 03, 2007 10:07 AM
> To: agi@v2.listbox.com
> Subject: Re: Hacker intelligence level [WAS Re: [agi] Funding AGI
research]
> 
> Ed Porter wrote:
>> Once you build up good models for parsing and word sense, then you read
>> large amounts of text and start building up model of the realities
> described
>> and generalizations from them.
>>
>> Assuming this is a continuation of the discussion of an AGI-at-home P2P
>> system, you are going to be very limited by the lack of bandwidth,
>> particularly for attacking the high dimensional problem of seeking to
>> understand the meaning of text, which often involve multiple levels of
>> implication, which would normally be accomplished by some sort of search
> of
>> a large semantic space, which is going to be difficult with limited
>> bandwidth.
>>
>> But a large amount of text with appropriate parsing and word sense
> labeling
>> would still provide a valuable aid for web and text search and for many
>> forms of automatic learning.  And the level of understanding that such a
> P2P
>> system could derive from reading huge amounts of text could be a valuable
>> initial source of one component of world knowledge for use by AGI.
> 
> I know you always find it teious when I express scepticism, so I will 
> preface my remarks with:  take this advice or ignore it, your choice.
> 
> This description of how to get AGI done reminds me of my childhood 
> project to build a Mars-bound spacecraft modeled after James Blish's 
> Book "Welcome to Mars".  I Knew that I could build it in time for the 
> next conjunction of Mars, but I hadn't quite gotten the anti-gravity 
> drive sorted out, so instead I collected all the other materials 
> described in the book, so everything would be ready when the AG drive 
> started working...
> 
> The reason it reminds me of this episode is that you are calmly talking 
> here about "the high dimensional problem of seeking to understand the 
> meaning of text, which often involve multiple levels of implication, 
> which would normally be accomplished by some sort of search of a large 
> semantic space" ......... this is your equivalent of the anti-gravity 
> drive.  This is the part that needs extremely detailed knowledge of AI 
> and psychology, just to be understand the nature of the problem (never 
> mind to solve it).  If you had any idea bout how to solve this part of 
> the problem, everything else would drop into your lap.  You wouldn't 
> need a P2P AGI-at-home system, because with this solution in hand you 
> would have people beating down your door to give you a supercomputer.
> 
> Menawhile, unfortunately, solving all those other issues like making 
> parsers and trying to do word-sense disambiguation would not help one 
> whit to get the real theoretical task done.
> 
> I am not being negative, I am just relaying the standard understanding 
> of priorities in the AGI field as a whole.  Send complaints addressed to 
> "AGI Community", not to me, please.
> 
> 
> 
> Richard Loosemore
> 
> 
> -----
> This list is sponsored by AGIRI: http://www.agiri.org/email
> To unsubscribe or change your options, please go to:
> http://v2.listbox.com/member/?&;
> 
> -----
> This list is sponsored by AGIRI: http://www.agiri.org/email
> To unsubscribe or change your options, please go to:
> http://v2.listbox.com/member/?&;
> 
> 
> ------------------------------------------------------------------------
> 

-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?&;

-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?member_id=8660244&id_secret=71673866-9b97ad

<<attachment: winmail.dat>>

Reply via email to