----- Original Message -----
> Hi,
> 
> That looks better. Do you get better results with it compared with the
> previous version?
> 
> Steve.
> 
> On 22/04/15 18:00, Bob Peterson wrote:
> > Hi,
> >
> > This patch changes function gfs2_rgrp_congested so that it only factors
> > in non-zero values into its average round trip time. If the round-trip
> > time is zero for a particular cpu, that cpu has obviously never dealt
> > with bouncing the resource group in question, so factoring in a zero
> > value will only skew the numbers. It also fixes a compile error on
> > some arches related to division.
> >
> > Regards,
> >
> > Bob Peterson
> > Signed-off-by: Bob Peterson <[email protected]>

It's not straightforward because the "preferred rgrps" patch skews the
performance results greatly. However, based on my performance tests, yes,
the numbers do look better. I temporarily commented out that "preferred
rgrps" patch to get a clearer picture. Here are numbers from my testing:
This test consists of 5 nodes all simultaneously doing this 'dd' command:
dd if=/dev/zero of=/mnt/gfs2/`hostname` bs=1M count=100
(after drop_caches)

Stock (nothing disabled):
            Run 1
         --------
pats     716 MB/s
jets     615 MB/s
bills    669 MB/s
dolphins 605 MB/s
ravens   735 MB/s
         --------
Average: 668 MB/s

Stock (plus preferred rgrps disabled): AKA 'c3'
            Run 1     Run 2
         --------  --------
pats     766 MB/s  675 MB/s  
jets     649 MB/s  716 MB/s  
bills    790 MB/s  735 MB/s  
dolphins 761 MB/s  727 MB/s  
ravens   712 MB/s  728 MB/s  
         --------  --------
Average: 736 MB/s  716 MB/s

Without latest patch (plus preferred rgrps disabled): AKA 'c2'
(In other words, this is the previous patch which was called
"GFS2: Use average srttb value in congestion calculations")
            Run 1     Run 2     Run 3 
         --------  --------  ---------
pats     830 MB/s  688 MB/s  697 MB/s
jets     833 MB/s  622 MB/s  645 MB/s
bills    831 MB/s  796 MB/s  637 MB/s
dolphins 834 MB/s  597 MB/s  690 MB/s
ravens   815 MB/s  731 MB/s  734 MB/s
         --------  --------  ---------
Average: 829 MB/s  687 MB/s  681 MB/s

Latest patch (plus preferred rgrps disabled):
            Run 1     Run 2     Run 3 
         --------  --------  ---------
pats     811 MB/s  829 MB/s  652 MB/s
jets     825 MB/s  863 MB/s  702 MB/s
bills    846 MB/s  825 MB/s  710 MB/s
dolphins 845 MB/s  845 MB/s  683 MB/s
ravens   820 MB/s  818 MB/s  682 MB/s
         --------  --------  ---------
Average: 829 MB/s  836 MB/s  686 MB/s

Latest patch (nothing disabled):
            Run 1     Run 2     Run 3 
         --------  --------  ---------
pats     834 MB/s  817 MB/s  819 MB/s
jets     837 MB/s  835 MB/s  836 MB/s
bills    841 MB/s  837 MB/s  834 MB/s
dolphins 838 MB/s  851 MB/s  842 MB/s
ravens   795 MB/s  808 MB/s  815 MB/s
         --------  --------  ---------
Average: 829 MB/s  830 MB/s  829 MB/s

This test (simultaneous dd) is known to be a worst case scenario,
so I expect it to show the most improvement. For ordinary block
allocations, I don't expect that big of an improvement.

Regards,

Bob Peterson
Red Hat File Systems

Reply via email to