Hi,

On 23/04/15 17:45, Bob Peterson wrote:
----- Original Message -----
Hi,

That looks better. Do you get better results with it compared with the
previous version?

Steve.

On 22/04/15 18:00, Bob Peterson wrote:
Hi,

This patch changes function gfs2_rgrp_congested so that it only factors
in non-zero values into its average round trip time. If the round-trip
time is zero for a particular cpu, that cpu has obviously never dealt
with bouncing the resource group in question, so factoring in a zero
value will only skew the numbers. It also fixes a compile error on
some arches related to division.

Regards,

Bob Peterson
Signed-off-by: Bob Peterson <rpete...@redhat.com>
It's not straightforward because the "preferred rgrps" patch skews the
performance results greatly. However, based on my performance tests, yes,
the numbers do look better. I temporarily commented out that "preferred
rgrps" patch to get a clearer picture. Here are numbers from my testing:
This test consists of 5 nodes all simultaneously doing this 'dd' command:
dd if=/dev/zero of=/mnt/gfs2/`hostname` bs=1M count=100
(after drop_caches)

Stock (nothing disabled):
             Run 1
          --------
pats     716 MB/s
jets     615 MB/s
bills    669 MB/s
dolphins 605 MB/s
ravens   735 MB/s
          --------
Average: 668 MB/s

Stock (plus preferred rgrps disabled): AKA 'c3'
             Run 1     Run 2
          --------  --------
pats     766 MB/s  675 MB/s
jets     649 MB/s  716 MB/s
bills    790 MB/s  735 MB/s
dolphins 761 MB/s  727 MB/s
ravens   712 MB/s  728 MB/s
          --------  --------
Average: 736 MB/s  716 MB/s

Without latest patch (plus preferred rgrps disabled): AKA 'c2'
(In other words, this is the previous patch which was called
"GFS2: Use average srttb value in congestion calculations")
             Run 1     Run 2     Run 3
          --------  --------  ---------
pats     830 MB/s  688 MB/s  697 MB/s
jets     833 MB/s  622 MB/s  645 MB/s
bills    831 MB/s  796 MB/s  637 MB/s
dolphins 834 MB/s  597 MB/s  690 MB/s
ravens   815 MB/s  731 MB/s  734 MB/s
          --------  --------  ---------
Average: 829 MB/s  687 MB/s  681 MB/s

Latest patch (plus preferred rgrps disabled):
             Run 1     Run 2     Run 3
          --------  --------  ---------
pats     811 MB/s  829 MB/s  652 MB/s
jets     825 MB/s  863 MB/s  702 MB/s
bills    846 MB/s  825 MB/s  710 MB/s
dolphins 845 MB/s  845 MB/s  683 MB/s
ravens   820 MB/s  818 MB/s  682 MB/s
          --------  --------  ---------
Average: 829 MB/s  836 MB/s  686 MB/s

Latest patch (nothing disabled):
             Run 1     Run 2     Run 3
          --------  --------  ---------
pats     834 MB/s  817 MB/s  819 MB/s
jets     837 MB/s  835 MB/s  836 MB/s
bills    841 MB/s  837 MB/s  834 MB/s
dolphins 838 MB/s  851 MB/s  842 MB/s
ravens   795 MB/s  808 MB/s  815 MB/s
          --------  --------  ---------
Average: 829 MB/s  830 MB/s  829 MB/s

This test (simultaneous dd) is known to be a worst case scenario,
so I expect it to show the most improvement. For ordinary block
allocations, I don't expect that big of an improvement.

Regards,

Bob Peterson
Red Hat File Systems

That looks very encouraging I think, definitely a good step forward.
Acked-by: Steven Whitehouse <swhit...@redhat.com>

Steve.

Reply via email to