Hi, Today I'm doing new benchmarks with RA, NORA, WB and WT in the controller:
With NORA ----------------- dd if=/vol02/bonnie/DD of=/dev/null bs=8M count=16384 16384+0 records in 16384+0 records out 137438953472 bytes (137 GB) copied, 318,306 s, 432 MB/s With RA ------------ dd if=/vol02/bonnie/DD of=/dev/null bs=8M count=16384 16384+0 records in 16384+0 records out 137438953472 bytes (137 GB) copied, 179,712 s, 765 MB/s dd if=/vol02/bonnie/DD of=/dev/null bs=8M count=16384 16384+0 records in 16384+0 records out 137438953472 bytes (137 GB) copied, 202,948 s, 677 MB/s dd if=/vol02/bonnie/DD of=/dev/null bs=8M count=16384 16384+0 records in 16384+0 records out 137438953472 bytes (137 GB) copied, 213,157 s, 645 MB/s With Adaptative RA ----------------- [root@cltbbdd01 ~]# dd if=/vol02/bonnie/DD of=/dev/null bs=8M count=16384 16384+0 records in 16384+0 records out 137438953472 bytes (137 GB) copied, 169,533 s, 811 MB/s [root@cltbbdd01 ~]# dd if=/vol02/bonnie/DD of=/dev/null bs=8M count=16384 16384+0 records in 16384+0 records out 137438953472 bytes (137 GB) copied, 207,223 s, 663 MB/s It's very strange the differences between the same test under same conditions... It seems thah adaptative read ahead is the best solution. For write test, I apply tuned-adm throughput-performance, that change IO elevator to deadline and grow up vm.dirty_ratio to 40.... ?¿?¿? With WB ------------- dd if=/dev/zero of=/vol02/bonnie/DD bs=8M count=16384 16384+0 records in 16384+0 records out 137438953472 bytes (137 GB) copied, 539,041 s, 255 MB/s dd if=/dev/zero of=/vol02/bonnie/DD bs=8M count=16384 16384+0 records in 16384+0 records out 137438953472 bytes (137 GB) copied, 505,695 s, 272 MB/s Enforce WB ----------------- dd if=/dev/zero of=/vol02/bonnie/DD bs=8M count=16384 16384+0 records in 16384+0 records out 137438953472 bytes (137 GB) copied, 662,538 s, 207 MB/s With WT -------------- dd if=/dev/zero of=/vol02/bonnie/DD bs=8M count=16384 16384+0 records in 16384+0 records out 137438953472 bytes (137 GB) copied, 750,615 s, 183 MB/s I think that this results are more logical... WT results in bad performance and differences, inside the same test, are minimum. Later I have put pair of dd at same time: dd if=/dev/zero of=/vol02/bonnie/DD2 bs=8M count=16384 16384+0 records in 16384+0 records out 137438953472 bytes (137 GB) copied, 633,613 s, 217 MB/s dd if=/dev/zero of=/vol02/bonnie/DD bs=8M count=16384 16384+0 records in 16384+0 records out 137438953472 bytes (137 GB) copied, 732,759 s, 188 MB/s Is very strange, that with parallel DD I take 400MBps. It's like if Centos have limit in IO throughput of a process... El 5 de abril de 2012 22:06, Tomas Vondra <t...@fuzzy.cz> escribió: > On 5.4.2012 20:43, Merlin Moncure wrote: > > The original problem is read based performance issue though and this > > will not have any affect on that whatsoever (although it's still > > excellent advice). Also dd should bypass the o/s buffer cache. I > > still pretty much convinced that there is a fundamental performance > > issue with the raid card dell needs to explain. > > Well, there are two issues IMHO. > > 1) Read performance that's not exactly as good as one'd expect from a > 12 x 15k SAS RAID10 array. Given that the 15k Cheetah drives usually > give like 170 MB/s for sequential reads/writes. I'd definitely > expect more than 533 MB/s when reading the data. At least something > near 1GB/s (equal to 6 drives). > > Hmm, the dd read performance seems to grow over time - I wonder if > this is the issue with adaptive read policy, as mentioned in the > xbitlabs report. > > Cesar, can you set the read policy to a 'read ahead' > > megacli -LDSetProp RA -LALL -aALL > > or maybe 'no read-ahead' > > megacli -LDSetProp NORA -LALL -aALL > > It's worth a try, maybe it somehow conflicts with the way kernel > handles read-ahead or something. I find these adaptive heuristics > a bit unpredictable ... > > Another thing - I see the patrol reads are enabled. Can you disable > that and try how that affects the performance? > > 2) Write performance behaviour, that's much more suspicious ... > > Not sure if it's related to the read performance issues. > > Tomas > > -- > Sent via pgsql-performance mailing list (pgsql-performance@postgresql.org) > To make changes to your subscription: > http://www.postgresql.org/mailpref/pgsql-performance > -- César Martín Pérez cmart...@gmail.com