On Tue, May 11, 2010 at 7:13 AM, Adam GROSZER <agros...@gmail.com> wrote:
> Hello Jim,
>
> Tuesday, May 11, 2010, 12:33:04 PM, you wrote:
>
> JF> On Tue, May 11, 2010 at 3:16 AM, Adam GROSZER <agros...@gmail.com> wrote:
>>> Hello Jim,
>>>
>>> Monday, May 10, 2010, 1:27:00 PM, you wrote:
>>>
>>> JF> On Sun, May 9, 2010 at 4:59 PM, Roel Bruggink <r...@fourdigits.nl> 
>>> wrote:
>>>>> That's really interesting! Did you notice any issues performance wise, or
>>>>> didn't you check that yet?
>>>
>>> JF> I didn't check performance. I just iterated over a file storage file,
>>> JF> checking compressed and uncompressed pickle sizes.
>>>
>>> I'd say some checksum is then also needed to detect bit failures that
>>> mess up the compressed data.
>
> JF> Why?
>
> I think the gzip algo compresses to a bit-stream, where even one bit
> has an error the rest of the uncompressed data might be a total mess.
> If that one bit is relatively early in the stream it's fatal.
> Salvaging the data is not a joy either.
> I know at this level we should expect that the OS and any underlying
> infrastructure should provide error-free data or fail.
> Tho I've seen some magic situations where the file copied without
> error through a network, but at the end CRC check failed on it :-O

How would a checksum help?  All it would do is tell you your hosed.
It wouldn't make you any less hosed.

Jim

-- 
Jim Fulton
_______________________________________________
For more information about ZODB, see the ZODB Wiki:
http://www.zope.org/Wikis/ZODB/

ZODB-Dev mailing list  -  ZODB-Dev@zope.org
https://mail.zope.org/mailman/listinfo/zodb-dev

Reply via email to