Robert Milkowski writes:
Hello Selim,
Wednesday, March 28, 2007, 5:45:42 AM, you wrote:
SD talking of which,
SD what's the effort and consequences to increase the max allowed block
SD size in zfs to highr figures like 1M...
Think what would happen then if you try to read 100KB of data - due to
chekcsumming ZFS would have to read entire MB.
However it should be possible to batch several IOs together and issue
one larger with ZFS - at least I hope it's possible.
As you note The max coherency unit (blocksize) in ZFS is
128K. It's also the max I/O size. And smaller I/O are
already aggregated or batched up to that size.
At 128K size the control to data ratio on the wire is already
quite reasonable. So I don't see much benefit to increasing this
(there maybe some but the context needs to be well defined).
The issue subject to debate because traditionally, one I/O
came with an implied overhead of a full head seek. In that
case, the larger the I/O the better. So at 60MB/s throughput
and 5ms head seek time, we need I/Os 300K to make the data
transfer time larger than the seek time and ~ 3MB I/O
sizes to reach the point of diminishing return.
But with a write allocate scheme we are not hit with the
head seek for every I/O and common I/O size wisdom needs to
be reconsidered.
-r
--
Best regards,
Robertmailto:[EMAIL PROTECTED]
http://milek.blogspot.com
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss