Whenever somebody asks the question, "How much memory do I need to dedup X
terabytes filesystem," the standard answer is "as much as you can afford to
buy."  This is true and correct, but I don't believe it's the best we can
do.  Because "as much as you can buy" is a true assessment for memory in
*any* situation.

 

To improve knowledge in this area, I think the question just needs to be
asked differently.  "How much *extra* memory is required for X terabytes,
with dedup enabled versus disabled?"

 

I hope somebody knows more about this than me.  I expect the answer will be
something like ...

 

The default ZFS block size is 128K.  If you have a filesystem with 128G
used, that means you are consuming 1,048,576 blocks, each of which must be
checksummed.  ZFS uses adler32 and sha256, which means 4bytes and 32bytes
...  36 bytes * 1M blocks = an extra 36 Mbytes and some fluff consumed by
enabling dedup.

 

I suspect my numbers are off, because 36Mbytes seems impossibly small.  But
I hope some sort of similar (and more correct) logic will apply.  ;-)

_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to