Boris,
It is going to take some time for me to understand what you wrote. But I do
have a quick reaction to something easy to understand.

You said:
This reflects on our basic disagreement, - you (& most logicians,
mathematicians, & programmers) start from deduction / pattern projection,
which is based on direct operations. And I think real GI must start from
induction / pattern discovery, which is intrinsically an inverse operation.
It's pretty dumb to generate / project patterns at random, vs. first
discovering them in the real world & projecting accordingly.

Well, no and yes.  No I don't disagree that GI must start from induction
and pattern discovery, but I do believe that projection, first produced
from primal or a programmed methods and then later from learned reactions
is obviously a necessary part of general intelligence.  Since they are
obviously necessary, I don't see any reason not to amp them up so long as
they do not interfere with the ability of an AGI program to do some genuine
learning from induction-like methods.

It will take me some time to figure out what you are saying in the rest of
the message.
Jim


On Wed, Aug 8, 2012 at 10:21 AM, Boris Kazachenko <bori...@verizon.net>wrote:

> **
> Jim,
>
> I agree with your focus on binary computational compression, but, as you
> said, that efficiency depends on specific operands. Even though low-power
> operations (addition) are more efficient for most data, it's the exceptions
> that matter. Most data is noise, what we care about is patterns. So, to
> improve both representational & computational compression, we need
> to quantify it for each operand ) operation. And the atomic operation that
> quantifies compression is what I call comparison, which starts with an
> inverse, vs. direct arithmetic operation. This reflects on our basic
> disagreement, - you (& most logicians, mathematicians, & programmers) start
> from deduction / pattern projection, which is based on direct operations.
> And I think real GI must start from induction / pattern discovery, which is
> intrinsically an inverse operation.  It's pretty dumb to generate / project
> patterns at random, vs. first discovering them in the real world &
> projecting accordingly.
>
> This is how I proposed to quantify compression (pattern strength) in my
> intro, part 2:
>
>  "AIT quantifies compression for sequences of inputs, while I define
> match for comparisons among individual inputs. On this level, a match is a
> lossless compression by replacing a larger comparand with its derivative
> (miss), relative to the smaller comparand. In other words, a match a
> complementary of a miss. That’s a deeper level of analysis, which I think
> can enable a far more incremental (thus potentially scalable) approach.
>
> Given incremental complexity of representation, initial inputs should have
> binary resolution. However, average binary match won’t justify the cost of
> comparison, which adds a syntactic overhead of newly differentiated match &
> miss to positionally distinct inputs. Rather, these binary inputs are
> compressed by digitization: a selective carry, aggregated & then forwarded
> up the hierarchy of digits. This is analogous to hierarchical search,
> explained in the next chapter, where selected templates are compared &
> conditionally forwarded up the hierarchy of expansion levels: a “digital
> hierarchy” of a corresponding coordinate. Digitization is done on inputs
> within a shared coordinate, the resolution of which is adjusted by
> feedback. This resolution must form average integers that are large enough
> for an average match between them (a subset of their magnitude) to merit
> the above-mentioned costs of comparison.
>
> Hence, the next order of compression is comparison across coordinates
> (initially defined with binary resolution as before | after input). Any
> comparison is an inverse arithmetic operation of incremental power: Boolean
> AND, subtraction, division, logarithm, & so on. Binary match is a sum of
> AND: partial identity of uncompressed bit strings, & miss is !AND. Binary
> comparison is useful for digitization, but it won’t further compress the
> integers produced thereby. In general, the products of a given-power
> comparison are further compressed only by a higher-power comparison between
> them, where match is the *additive* compression.
>
> Thus, initial comparison between digitized integers is done by
> subtraction, which increases match by compressing miss from !AND to
> difference, in which opposite-sign bits cancel each other via carry |
> borrow. The match is increased because it is a complimentary of difference,
> equal to the smaller of the comparands.
>
> All-to-all comparison across 1D queue of pixels forms signed derivatives,
> complemented by which new inputs can losslessly & compressively replace
> older templates. At the same time, current input match determines whether
> individual derivatives are also compared (vs. aggregated), forming
> successively higher derivatives. “Atomic” comparison is between a
> single-variable input & a template (older input):
> Comparison: match= min (input, template), miss= dif (i-t): aggregated over
> the span of constant sign.
> Evaluation: match - average_match_per_average_difference_match, formed on
> the next search level.
> This evaluation is for comparing higher derivatives, vs. evaluation for
> higher-level inputs explained in part 3. It can also be increasingly
> complex, but I will need a meaningful feedback to elaborate.
>
> Division further reduces difference to a ratio, which can then be reduced
> to a logarithm, & so on. Thus, complimentary match is increased with the
> power of comparison. But the costs may grow even faster, for both
> operations & incremental syntax to record incidental sign, fraction,
> irrational fraction. The power of comparison is increased if current match
> plus miss predict an improvement, as indicated by higher-order comparison
> between the results from different powers of comparison. This
> meta-comparison can discover algorithms, or meta-patterns..."
>
> *http://www.cognitivealgorithm.info/2012/01/cognitive-algorithm.html*<http://www.cognitivealgorithm.info/2012/01/cognitive-algorithm.html>
> **
> *From:* Jim Bromer <jimbro...@gmail.com>
> *Sent:* Wednesday, August 08, 2012 9:08 AM
> *To:* AGI <a...@listbox.com>
> *Subject:* [agi] Re: Addition is the Engine of Computation
>
> Rereading my original message I realized that what I said may not have
> been very easy to read.  It is, however, interesting and serious
> programmers should be aware of it.
>
> An algorithm can be (and usually is) like a procedural compression
> method.  (I have sometimes called it a transformational compression
> method.)  That is not to say that algorithms typically compress data, but
> that they are extremely efficient. For instance, multiplication is defined
> as the repeated addition of a particular number.  However, the
> standard multiplier algorithm is much more efficient than doing the
> repeated additions because it is, what I am calling, a procedural
> compression method.  Furthermore, both addition and multiplication can use
> binary numbers which, as I explained, are extremely efficient compressions
> of the representations of values.
>
> It is difficult to compare the complexity (that is the efficiency) of
> these compressed procedural methods, so one way to do so is to expand the
> algorithm into a true formula of Boolean Logic where only AND, OR, Negation
> and parentheses are used with atomic Boolean variables.   If you can find
> the shortest Boolean Formula then we can say (if we agree to do so) that
> this is a measure of the complexity of the algorithm.  Most programmers
> (including myself) do not know how you go about it and the problem of
> efficiently finding the most efficient Boolean Formula may be a problem
> that has probably not yet been solved.  However, we can expand some
> algorithms into Boolean formulas and at least get an intuitive sense of
> just how efficient these algorithms are.
>
> What I said in this thread is that I was surprised to find that the real
> engine of efficiency in arithmetic procedures is found in additions of
> multiple addends. Multiplication is more efficient than repeated addition
> for one class of multiple addends, but even there it is my opinion that the
> real power of multiplication comes from the addition of the multiple
> partial products.
>
> Most examples of compressed data cannot be used without decompressing the
> data. The benefits of addition and multiplication is that it is not
> necessary to 'decompress' the binary representations of values in order to
> use them in arithmetic operations.  So addition and multiplication (and of
> the computational-numerical methods which are derived from addition) are
> not only procedural compressions but they also use data in it's compressed
> form without first decompressing it!  This is an amazing power, and I
> believe that it is exactly where the power of computation comes from.
>
> What does this have to do with you?
> It is my opinion, based on this analysis, that if you can -effectively-
> use multiple addend addition in your programs you would be well advised to
> consider doing so.  The problem with many supposed numerical 'solutions' to
> AGI is that no one has been able to find an effective method to use numbers
> to represent or model the problem space of AGI.
>
> Jim Bromer
>
>
>
>
>   *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
> <https://www.listbox.com/member/archive/rss/303/18407320-d9907b69> |
> Modify <https://www.listbox.com/member/?&;> Your Subscription
> <http://www.listbox.com>
>    *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
> <https://www.listbox.com/member/archive/rss/303/10561250-164650b2> |
> Modify<https://www.listbox.com/member/?&;>Your Subscription
> <http://www.listbox.com>
>



-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-c97d2393
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-2484a968
Powered by Listbox: http://www.listbox.com

Reply via email to