Can you share your 'zpool status' output for both pools? Also you may want to run the following a few times in a loop and provide the output:
# echo "::walk spa | ::print spa_t spa_name spa_last_io spa_scrub_inflight" | mdb -k Thanks, George On Sat, May 14, 2011 at 8:29 AM, Donald Stahl <d...@blacksun.org> wrote: >> The scrub I/O has lower priority than other I/O. >> >> In later ZFS releases, scrub I/O is also throttled. When the throttle >> kicks in, the scrub can drop to 5-10 IOPS. This shouldn't be much of >> an issue, scrubs do not need to be, and are not intended to be, run >> very often -- perhaps once a quarter or so. > I understand the lower priority I/O and such but what confuses me is this: > On my primary head: > scan: scrub in progress since Fri May 13 14:04:46 2011 > 24.5G scanned out of 14.2T at 340K/s, (scan is slow, no estimated time) > 0 repaired, 0.17% done > > I have a second NAS head, also running OI 147 on the same type of > server, with the same SAS card, connected to the same type of disk > shelf- and a zpool scrub over there is showing : > scan: scrub in progress since Sat May 14 11:10:51 2011 > 29.0G scanned out of 670G at 162M/s, 1h7m to go > 0 repaired, 4.33% done > > Obviously there is less data on the second server- but the first > server has 88 x SAS drives and the second one has 10 x 7200 SATA > drives. I would expect those 88 SAS drives to be able to outperform 10 > SATA drives- but they aren't. > > On the first server iostat -Xn is showing 30-40 IOPS max per drive, > while on the second server iostat -Xn is showing 400 IOPS per drive. > > On the first server the disk busy numbers never climb higher than 30% > while on the secondary they will spike to 96%. > > This performance problem isn't just related to scrubbing either. I see > mediocre performance when trying to write to the array as well. If I > were seeing hardware errors, high service times, high load, or other > errors, then that might make sense. Unfortunately I seem to have > mostly idle disks that don't get used. It's almost as if ZFS is just > sitting around twiddling its thumbs instead of writing data. > > I'm happy to provide real numbers, suffice it to say none of these > numbers make any sense to me. > > The array actually has 88 disks + 4 hot spares (1 each of two sizes > per controller channel) + 4 Intel X-25E 32GB SSD's (2 x 2 way mirror > split across controller channels). > > Any ideas or things I should test and I will gladly look into them. > > -Don > _______________________________________________ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > -- George Wilson M: +1.770.853.8523 F: +1.650.494.1676 275 Middlefield Road, Suite 50 Menlo Park, CA 94025 http://www.delphix.com _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss