> From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
> boun...@opensolaris.org] On Behalf Of Nathan Kroenert
> That reminds me of something I have been wondering about... Why only 12x
> faster? If we are effectively reading from memory - as compared to a
> disk reading at approximately 100MB/s (which is about an average PC HDD
> reading sequentially), I'd have thought it should be a lot faster than
> Can we really only pull stuff from cache at only a little over one
> gigabyte per second if it's dedup data?
Actually, cpu's and memory aren't as fast as you might think. In a system
with 12 disks, I've had to write my own "dd" replacement, because "dd
if=/dev/zero bs=1024k" wasn't fast enough to keep the disks busy. Later, I
wanted to do something similar, using unique data, and it was simply
impossible to generate random data fast enough. I had to tweak my "dd"
replacement to write serial numbers, which still wasn't fast enough, so I
had to tweak my "dd" replacement to write a big block of static data,
followed by a serial number, followed by another big block (always smaller
than the disk block, so it would be treated as unique when hitting the
1 typical disk sustains 1Gbit/sec. In theory, 12 should be able to sustain
12 Gbit/sec. According to Nathan's email, the memory bandwidth might be 25
Gbit, of which, you probably need to both read & write, thus making it
effectively 12.5 Gbit... I'm sure the actual bandwidth available varies by
system and memory type.
zfs-discuss mailing list