On Oct 4, 2011, at 4:14 PM, Daniel Carosone wrote:
> I sent a zvol from host a, to host b, twice. Host b has two pools,
> one ashift=9, one ashift=12. I sent the zvol to each of the pools on
> b. The original source pool is ashift=9, and an old revision (2009_06
> because it's still running xen).
> I sent it twice, because something strange happened on the first send,
> to the ashift=12 pool. "zfs list -o space" showed figures at least
> twice those on the source, maybe roughly 2.5 times.
Can you share the output?
"15% of nothin' is nothin'!'"
> I suspected this may be related to ashift, so tried the second send to
> the ahsift=9 pool; these received snapshots line up with the same
> space consumption as the source.
> What is going on? Is there really that much metadata overhead? How
> many metadata blocks are needed for each 8k vol block, and are they
> each really only holding 512 bytes of metadata in a 4k allocation?
> Can they not be packed appropriately for the ashift?
Doesn't matter how small metadata compresses, the minimum size you can write
> Longer term, if zfs were to pack metadata into full blocks by ashift,
> is it likely that this could be introduced via a zpool upgrade, with
> space recovered as metadata is rewritten - or would it need the pool
> to be recreated? Or is there some other solution in the works?
I think we'd need to see the exact layout of the internal data. This can be
achieved with the zfs_blkstats macro in mdb. Perhaps we can take this offline
and report back?
ZFS and performance consulting
VMworld Copenhagen, October 17-20
OpenStorage Summit, San Jose, CA, October 24-27
LISA '11, Boston, MA, December 4-9
zfs-discuss mailing list