On 4/20/05, Linus Torvalds <[EMAIL PROTECTED]> wrote:
> It really _shouldn't_ be faster. It still does the compression, and throws
> the end result away.
Am I misunderstanding or is the proglem that doing:
<file with unknown status> -> compress -> sha1 -> compare with existing hash
What about doing:
<file it's supposed to be equal to> -> uncompress -> compare with
unknown status file
It's more file I/O, but the uncompress is much cheaper than the compress.
On a second issue, what's the format of the main 'index' file? Is it:
If so, that's not going to compress well. A file like:
Will compress better.
Stop me if I'm way off base--I'm just following the mailing list, I
haven't tried out the code.
To unsubscribe from this list: send the line "unsubscribe git" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at http://vger.kernel.org/majordomo-info.html