On a quad xeon with 3.75 gb ram there was a big
regression in dbench and tiobench with many threads
between 2.4.20-pre1 and 2.4.20-pre2. Below is from
5 dbench runs.
There was a 23% regression at 192 processes
reiserfs
dbench 192 processes Average High Low
2.4.20-pre1 55.94 58.17 54.26
2.4.20-pre2 42.98 44.73 42.30
And 13% regression at 64 processes.
dbench 64 processes Average High Low
2.4.20-pre1 70.98 72.53 69.47
2.4.20-pre2 61.93 64.01 57.31
ext2 and ext3 did not change between 2.4.20-pre1 and pre2.
ext3
dbench 192 processes Average High Low
2.4.20-pre1 60.50 61.47 59.01
2.4.20-pre2 60.61 61.63 58.80
dbench 64 processes Average High Low
2.4.20-pre1 82.73 89.63 64.67
2.4.20-pre2 83.12 91.28 65.11
ext2
dbench 192 processes Average High Low
2.4.20-pre1 116.32 124.44 110.95
2.4.20-pre2 116.45 119.89 109.89
dbench 64 processes Average High Low
2.4.20-pre1 145.61 158.67 102.33
2.4.20-pre2 145.77 158.96 103.89
CPU utilization on reiserfs went down in the high thread count
tiobench tests, but it also had a throughput regression when there was
more than one thread most tests. ext2 and ext3 didn't have a regression.
tiobench-0.3.3
Unit information
================
File size = 12288 megabytes
Blk Size = 4096 bytes
Rate = megabytes per second
CPU% = percentage of CPU used during the test
Latency = milliseconds
Lat% = percent of requests that took longer than X seconds
CPU Eff = Rate divided by CPU% - throughput per cpu load
Sequential Reads reiserfs had the biggest drop in throughput.
Num Avg Maximum Lat% Lat% CPU
Kernel Thr Rate (CPU%) Latency Latency >2s >10s Eff
------------- --- ----------------------------------------------------------
2.4.20-pre1 1 35.11 23.02% 0.331 107.07 0.00000 0.00000 152
2.4.20-pre2 1 34.68 22.83% 0.335 119.57 0.00000 0.00000 152
2.4.20-pre1 32 27.27 20.12% 11.536 258436.68 0.01275 0.01135 136
2.4.20-pre2 32 8.00 6.05% 43.881 7568.44 0.00096 0.00000 132
2.4.20-pre1 64 26.50 19.63% 21.328 361756.08 0.02183 0.02053 135
2.4.20-pre2 64 8.42 6.46% 79.039 15721.84 0.18813 0.00000 130
2.4.20-pre1 128 27.72 21.17% 35.485 814716.92 0.02671 0.02549 131
2.4.20-pre2 128 9.30 7.10% 135.056 32520.09 3.92532 0.00035 131
2.4.20-pre1 256 26.55 19.95% 68.425 825224.03 0.04636 0.04515 133
2.4.20-pre2 256 8.47 6.49% 285.430 62013.50 4.68931 0.11676 131
Random Reads reiserfs went down when there is more than 1 thread.
Num Avg Maximum Lat% Lat% CPU
Kernel Thr Rate (CPU%) Latency Latency >2s >10s Eff
------------- --- ----------------------------------------------------------
2.4.20-pre1 1 0.68 0.84% 17.320 51.54 0.00000 0.00000 81
2.4.20-pre2 1 0.70 0.76% 16.787 53.99 0.00000 0.00000 92
2.4.20-pre1 32 3.09 6.80% 102.783 535.46 0.00000 0.00000 46
2.4.20-pre2 32 1.98 3.84% 147.780 16015.44 0.55000 0.00000 52
2.4.20-pre1 64 3.10 5.99% 178.641 10722.38 0.50403 0.00000 52
2.4.20-pre2 64 2.18 3.70% 229.511 16850.53 2.77218 0.00000 59
2.4.20-pre1 128 3.35 5.27% 226.876 9018.16 1.91532 0.00000 64
2.4.20-pre2 128 2.31 3.98% 378.511 17932.69 5.77117 0.00000 58
2.4.20-pre1 256 3.26 6.44% 249.264 5751.33 0.57292 0.00000 51
2.4.20-pre2 256 2.20 5.53% 724.340 19268.40 13.09896 0.00000 40
Sequential Writes reiserfs cpu utilization went down, but throughput dropped.
Num Avg Maximum Lat% Lat% CPU
Kernel Thr Rate (CPU%) Latency Latency >2s >10s Eff
------------- --- ----------------------------------------------------------
2.4.20-pre1 1 39.31 71.71% 0.282 6301.59 0.00012 0.00000 55
2.4.20-pre2 1 38.28 73.39% 0.290 8058.03 0.00019 0.00000 52
2.4.20-pre1 32 29.64 119.5% 11.171 16619.06 0.18094 0.00000 25
2.4.20-pre2 32 23.52 103.0% 10.242 17721.55 0.13629 0.00000 23
2.4.20-pre1 64 29.02 115.6% 22.065 38792.62 0.38675 0.00426 25
2.4.20-pre2 64 20.92 84.06% 19.449 41273.31 0.33824 0.00480 25
2.4.20-pre1 128 31.52 111.6% 36.264 104743.20 0.40728 0.11428 28
2.4.20-pre2 128 22.31 87.18% 36.251 78916.66 0.46791 0.10154 26
2.4.20-pre1 256 28.95 132.7% 75.959 163159.61 0.79301 0.24067 22
2.4.20-pre2 256 20.08 93.50% 74.113 114883.50 0.83399 0.23864 21
Random Writes reiserfs - slightly improved.
Num Avg Maximum Lat% Lat% CPU
Kernel Thr Rate (CPU%) Latency Latency >2s >10s Eff
------------- --- ----------------------------------------------------------
2.4.20-pre1 1 3.57 3.80% 0.091 0.57 0.00000 0.00000 94
2.4.20-pre2 1 3.63 3.79% 0.089 0.68 0.00000 0.00000 96
2.4.20-pre1 32 3.52 9.15% 0.654 268.43 0.00000 0.00000 38
2.4.20-pre2 32 3.57 8.15% 0.377 258.42 0.00000 0.00000 44
2.4.20-pre1 64 3.68 9.50% 0.567 196.77 0.00000 0.00000 39
2.4.20-pre2 64 3.78 9.76% 0.450 319.24 0.00000 0.00000 39
2.4.20-pre1 128 3.48 8.38% 2.923 626.60 0.00000 0.00000 42
2.4.20-pre2 128 3.54 8.98% 0.669 735.87 0.00000 0.00000 39
2.4.20-pre1 256 3.45 9.11% 3.416 572.87 0.00000 0.00000 38
2.4.20-pre2 256 3.59 10.12% 1.350 824.15 0.00000 0.00000 35
Bonnie++ did not have these regressions, but I run that as a single thread.
More on the tests at:
http://home.earthlink.net/~rwhron/kernel/bigbox.html
--
Randy Hron