On Thu, May 18, 2017 at 6:39 PM, Joe Julian <[email protected]> wrote:
> On the other hand, tracking that stat between versions with a known test > sequence may be valuable for watching for performance issues or > improvements. > +1 Once we have nightly build setup, we should take this up. Can do these after branch out to allow time for fixing major issues if any. -Amar > > On May 17, 2017 10:03:28 PM PDT, Ravishankar N <[email protected]> > wrote: >> >> On 05/17/2017 11:07 PM, Pranith Kumar Karampuri wrote: >> >> + gluster-devel >> >> On Wed, May 17, 2017 at 10:50 PM, mabi <[email protected]> wrote: >> >>> I don't know exactly what kind of context-switches it was but what I >>> know is that it is the "cs" number under "system" when you run vmstat. >>> >> Okay, that could be due to the syscalls themselves or pre-emptive >> multitasking in case there aren't enough cpu cores. I think the spike in >> numbers is due to more users accessing the files at the same time like you >> observed, translating into more syscalls. You can try capturing the >> gluster volume profile info the next time it occurs and co-relate with the >> cs count. If you don't see any negative performance impact, I think you >> don't need to be bothered much by the numbers. >> >> HTH, >> Ravi >> >> >>> Also I use the percona linux monitoring template for cacti ( >>> https://www.percona.com/doc/percona-monitoring-plugins/LATE >>> ST/cacti/linux-templates.html) which monitors context switches too. If >>> that's of any use interrupts where also quite high during that time with >>> peaks up to 50k interrupts. >>> >>> >>> >>> -------- Original Message -------- >>> Subject: Re: [Gluster-users] 120k context switches on GlsuterFS nodes >>> Local Time: May 17, 2017 2:37 AM >>> UTC Time: May 17, 2017 12:37 AM >>> From: [email protected] >>> To: mabi <[email protected]>, Gluster Users <[email protected]> >>> >>> >>> On 05/16/2017 11:13 PM, mabi wrote: >>> >>> Today I even saw up to 400k context switches for around 30 minutes on my >>> two nodes replica... Does anyone else have so high context switches on >>> their GlusterFS nodes? >>> >>> I am wondering what is "normal" and if I should be worried... >>> >>> >>> >>> >>> -------- Original Message -------- >>> Subject: 120k context switches on GlsuterFS nodes >>> Local Time: May 11, 2017 9:18 PM >>> UTC Time: May 11, 2017 7:18 PM >>> From: [email protected] >>> To: Gluster Users <[email protected]> >>> <[email protected]> >>> >>> Hi, >>> >>> Today I noticed that for around 50 minutes my two GlusterFS 3.8.11 nodes >>> had a very high amount of context switches, around 120k. Usually the >>> average is more around 1k-2k. So I checked what was happening and there >>> where just more users accessing (downloading) their files at the same time. >>> These are directories with typical cloud files, which means files of any >>> sizes ranging from a few kB to MB and a lot of course. >>> >>> Now I never saw such a high number in context switches in my entire life >>> so I wanted to ask if this is normal or to be expected? I do not find any >>> signs of errors or warnings in any log files. >>> >>> >>> What context switch are you referring to (syscalls context-switch on the >>> bricks?) ? How did you measure this? >>> -Ravi >>> >>> My volume is a replicated volume on two nodes with ZFS as filesystem >>> behind and the volume is mounted using FUSE on the client (the cloud >>> server). On that cloud server the glusterfs process was using quite a lot >>> of system CPU but that server (VM) only has 2 vCPUs so maybe I should >>> increase the number of vCPUs... >>> >>> Any ideas or recommendations? >>> >>> >>> >>> Regards, >>> M. >>> >>> >>> >>> >>> _______________________________________________ >>> Gluster-users mailing >>> [email protected]http://lists.gluster.org/mailman/listinfo/gluster-users >>> >>> _______________________________________________ Gluster-users mailing >>> list [email protected] http://lists.gluster.org/mailm >>> an/listinfo/gluster-users >> >> -- >> Pranith >> >> > -- > Sent from my Android device with K-9 Mail. Please excuse my brevity. > > _______________________________________________ > Gluster-devel mailing list > [email protected] > http://lists.gluster.org/mailman/listinfo/gluster-devel > -- Amar Tumballi (amarts)
_______________________________________________ Gluster-devel mailing list [email protected] http://lists.gluster.org/mailman/listinfo/gluster-devel
