Can you share your 'zpool status' output for both pools?

Also you may want to run the following a few times in a loop and
provide the output:

# echo "::walk spa | ::print spa_t spa_name spa_last_io
spa_scrub_inflight" | mdb -k

Thanks,
George

On Sat, May 14, 2011 at 8:29 AM, Donald Stahl <d...@blacksun.org> wrote:
>> The scrub I/O has lower priority than other I/O.
>>
>> In later ZFS releases, scrub I/O is also throttled. When the throttle
>> kicks in, the scrub can drop to 5-10 IOPS. This shouldn't be much of
>> an issue, scrubs do not need to be, and are not intended to be, run
>> very often -- perhaps once a quarter or so.
> I understand the lower priority I/O and such but what confuses me is this:
> On my primary head:
>  scan: scrub in progress since Fri May 13 14:04:46 2011
>    24.5G scanned out of 14.2T at 340K/s, (scan is slow, no estimated time)
>    0 repaired, 0.17% done
>
> I have a second NAS head, also running OI 147 on the same type of
> server, with the same SAS card, connected to the same type of disk
> shelf- and a zpool scrub over there is showing :
>  scan: scrub in progress since Sat May 14 11:10:51 2011
>    29.0G scanned out of 670G at 162M/s, 1h7m to go
>    0 repaired, 4.33% done
>
> Obviously there is less data on the second server- but the first
> server has 88 x SAS drives and the second one has 10 x 7200 SATA
> drives. I would expect those 88 SAS drives to be able to outperform 10
> SATA drives- but they aren't.
>
> On the first server iostat -Xn is showing 30-40 IOPS max per drive,
> while on the second server iostat -Xn is showing 400 IOPS per drive.
>
> On the first server the disk busy numbers never climb higher than 30%
> while on the secondary they will spike to 96%.
>
> This performance problem isn't just related to scrubbing either. I see
> mediocre performance when trying to write to the array as well. If I
> were seeing hardware errors, high service times, high load, or other
> errors, then that might make sense. Unfortunately I seem to have
> mostly idle disks that don't get used. It's almost as if ZFS is just
> sitting around twiddling its thumbs instead of writing data.
>
> I'm happy to provide real numbers, suffice it to say none of these
> numbers make any sense to me.
>
> The array actually has 88 disks + 4 hot spares (1 each of two sizes
> per controller channel) + 4 Intel X-25E 32GB SSD's (2 x 2 way mirror
> split across controller channels).
>
> Any ideas or things I should test and I will gladly look into them.
>
> -Don
> _______________________________________________
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>



-- 
George Wilson



M: +1.770.853.8523
F: +1.650.494.1676
275 Middlefield Road, Suite 50
Menlo Park, CA 94025
http://www.delphix.com
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to