The problem appears to be memory modules that are not quite up-to-spec.
After removing the originally flagged module, others still were
reporting errors and fmd was getting in the way.
I ended up with 24G that worked under S10U8 on one of the two identical
boxes. Under this configuration, the pool was able to give 800M/s write
as you would expect.
The i/o hardware is LSI-1068E cards to Samsung 1T enterprise drives.
johan...@sun.com wrote:
Paul:
I'm interested in getting a box like this. I hope this problem has a
simple fix.
When I create a 8x mirrored vdev pool (1T samsung enterise drives) and
do a simple dd test it maxes out at 100M/s, where I'd normally expect
500M/s+ at least.
Will you include the output from the following commands?
$ zpool status
$ zpool list
lo...@dev-storage-01:/storage# dd if=/dev/zero of=test.zeros bs=1M
count=64000
64000+0 records in
64000+0 records out
67108864000 bytes (67 GB) copied, 630.583 s, 106 MB/s
ZFS's default block size is 128k. Does setting bs=128k improve
performance at all?
lo...@dev-storage-01:~$ mpstat 15
...
mpstat(1) showed a lot of system time. You can run the following DTrace
as root to get more detail about what the OS is doing:
# dtrace -n 'profile-1997hz {...@a[stack(50)] = count();} END {trunc(@a, 20);}'
This will show you the 20 most frequent kernel stacks, sampled at 1997 Hz.
I agree with Richard's advice about resolving your FMA problems.
Whatever is going on, it's certainly not helping your performance.
I'd also be curious to see the iostat output for these transactions
while the dd is running. I typically use the following.
$ iostat -xn 1
Disk drivers can often have an adverse affect on performance. I've seen
some poorly supported SATA cards/drivers cause ZFS performance to look
really bad, when in fact the problem was in the I/O hardware/software.
It might also be useful to know what kind of I/O hardware you're using
in this system.
Thanks,
-j
_______________________________________________
perf-discuss mailing list
perf-discuss@opensolaris.org