> The reason for the extra reduction is that there is some redundancy
> between files.  For instance, they probably have similar headers and
> footers.

        Another option, albeit slower (but still available, and GPL) is
rzip. It is based on the same checksum routines that rsync uses (weak
32-bit and stronger 128-bit algorithms), which can find redundant parts,
and not re-compress them.

        For example, if you have 10 home directories, and 5 users have the
same Linux kernel source unpacked in their home directory, rzip can detect
that and will just not compress the "redundant" data over again. Very
slick stuff.

        If you're plucking 5,000 pages, and the header, footer, and
graphics are all the same around the page ornaments, rzip can omit those
redundant parts, and only compress the ones which "differ".

d.

_______________________________________________
plucker-dev mailing list
[EMAIL PROTECTED]
http://lists.rubberchicken.org/mailman/listinfo/plucker-dev

Reply via email to