The positional notation system is a powerful compression of a representation of a number. The decimal number system and the binary number system are both examples of this method. If you were to use a single mark to represent 1, two marks to represent 2, three marks to represent 3 and so on, you would find that the positional notation system has a logarithmic compression value. Decompressed, a value of n (significant)bits expressed in the binary system would decompress into 2^n sticks (in the unary system that I just mentioned.) This is a major representational compression method and it represents one of greatest of historical achievements. I feel that the system fails badly and leaves a lot of artifacts when used to express fractional amounts but I don't want to get into that can of worms (which has already been opened for centuries).
The methods of addition and multiplication of numbers expressed in the positional notation system represent major procedural compression methods because they can be used to compute their functions so efficiently. I always felt that since multiplication was more efficient than addition (when used to do what they do) that it had to be the more powerful system. Cross multiplication (where you multiply the values in the ones column of the two multiplicands, then go to multiply the ones column of one multiplicand by the twos column of the other, and so on) seemed to be the most amazing procedural compression method. However, I was wrong. The addition of multiple numbers is the most powerful procedural engine. When you try to express the addition of multiple binary numbers in the terms of a true Boolean formula (on the input bit fields of the addends) it expands into a formula that has to be expressed at a roughly exponential degree of complexity. (That is the formula is itself exponentially long relative to some function of the number of bits in the addends and the number of addends.) If you look at cross multiplication alone, ignoring the subsequent step when you add the partial products of the multiplication, you would find that it expands into a Boolean formula that is roughly n^2 steps (based on the input fields of multiplicands which were each n bits long). But when you realize that the addition of the partial products has a complexity which is at least exponential relative to the input bit fields you can see that this step of the addition is main engine of efficiency. So it is true that multiplication is more complex than addition, but it is the part where the additions are being done which is most complex (that is to say the part which is most efficient.) (Of course there are some cases where this analysis of the complexity may not hold, one can find short cuts for some situations, as when one of the multiplicands is zero or one, but in general, the best case of procedural efficiency, as found through an expansion into a true Boolean formula, can be found in binary addition. You can say that multiplication is more powerful, but without binary addition - you got squat. (I have not worked the details out but I am confident that this analysis is in the ballpark.) Jim Bromer ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-c97d2393 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-2484a968 Powered by Listbox: http://www.listbox.com