Stale super_blocks in 2.2

2001-02-13 Thread Phil Auld
Hello, It appears that the umount path in the 2.2 series kernels does not do anything to invalidate the buffers associated with the unmounted device. We then rely on disk change detection on a subsequent mount to prevent us from seeing the old super_block. Since deja was gobbled by goog

Re: Stale super_blocks in 2.2

2001-02-13 Thread Phil Auld
Alan Cox wrote: > > > does not do anything to invalidate the buffers associated with the > > unmounted device. We then rely on disk change detection on a > > subsequent mount to prevent us from seeing the old super_block. > > 2.2 yes, 2.4 no That can be a problem for fiber channel devices. I sa

Re: Stale super_blocks in 2.2

2001-02-14 Thread Phil Auld
Alan Cox wrote: > > > That can be a problem for fiber channel devices. I saw some issues with > > invalidate_buffers and page caching discussed in 2.4 space. Any reasons > > come to mind why I shouldn't call invalidate on the the way down instead > > (or in addition)? > > The I/O completed a few

Re: Configure.help is complete

2001-06-01 Thread Phil Auld
Alexander Viro wrote: ...snip... > > We should start removing the crap from procfs in 2.5. Documenting shit is > a good step, but taking it out would be better. > Not to open a what may be can of worms but ... What's wrong with procfs? It allows a general interface to the kernel that does

Re: [Patch] sched/fair: Avoid throttle_list starvation with low cfs quota

2018-10-09 Thread Phil Auld
I believe that's a different issue, though. The kernel allows this setting and should handle it better than it currently does. The proposed patch fixes it so that all the tasks make progress (even if not much progress) rather than having some starve at the back of the list. Cheers, Ph

Re: [Patch] sched/fair: Avoid throttle_list starvation with low cfs quota

2018-10-10 Thread Phil Auld
eeds to be fixed - and at first sight the quota > > of 1000 looks very > > low - could we improve the arithmetics perhaps? > > > > A low quota of 1000 is used because there's many VMs or containers > > provisioned on the system > > that is trig

[Patch] sched/fair: Avoid throttle_list starvation with low cfs quota

2018-10-08 Thread Phil Auld
From: "Phil Auld" sched/fair: Avoid throttle_list starvation with low cfs quota With a very low cpu.cfs_quota_us setting, such as the minimum of 1000, distribute_cfs_runtime may not empty the throttled_list before it runs out of runtime to distribute. In that case, due to the c

Re: [PATCH] sched/rt: Clean up usage of rt_task()

2024-05-14 Thread Phil Auld
Hi Qais, On Wed, May 15, 2024 at 12:41:12AM +0100 Qais Yousef wrote: > rt_task() checks if a task has RT priority. But depends on your > dictionary, this could mean it belongs to RT class, or is a 'realtime' > task, which includes RT and DL classes. > > Since this has caused some confusion alre

Re: [PATCH] sched/rt: Clean up usage of rt_task()

2024-05-15 Thread Phil Auld
On Wed, May 15, 2024 at 10:32:38AM +0200 Peter Zijlstra wrote: > On Tue, May 14, 2024 at 07:58:51PM -0400, Phil Auld wrote: > > > > Hi Qais, > > > > On Wed, May 15, 2024 at 12:41:12AM +0100 Qais Yousef wrote: > > > rt_task() checks if a task has RT priority.

Re: [PATCH] sched/rt: Clean up usage of rt_task()

2024-05-15 Thread Phil Auld
On Wed, May 15, 2024 at 01:06:13PM +0100 Qais Yousef wrote: > On 05/15/24 07:20, Phil Auld wrote: > > On Wed, May 15, 2024 at 10:32:38AM +0200 Peter Zijlstra wrote: > > > On Tue, May 14, 2024 at 07:58:51PM -0400, Phil Auld wrote: > > > > > > > > Hi Qais

[RFC] sched/fair: hard lockup in sched_cfs_period_timer

2019-03-01 Thread Phil Auld
Hi, I have a reproducible case of this: [ 217.264946] NMI watchdog: Watchdog detected hard LOCKUP on cpu 24 [ 217.264948] Modules linked in: sunrpc iTCO_wdt gpio_ich iTCO_vendor_support intel_powerclamp coretemp kvm_intel kvm ipmi_ssif irqbypass crct10dif_pclmul crc32_pclmul ghash_clmulni_in

Re: [tip:sched/core] sched/fair: Limit sched_cfs_period_timer loop to avoid hard lockup

2019-04-16 Thread Phil Auld
On Tue, Apr 09, 2019 at 03:05:27PM +0200 Peter Zijlstra wrote: > On Tue, Apr 09, 2019 at 08:48:16AM -0400, Phil Auld wrote: > > Hi Ingo, Peter, > > > > On Wed, Apr 03, 2019 at 01:38:39AM -0700 tip-bot for Phil Auld wrote: > > > Commit-ID: 06ec5d30e8d57b820d44df6

Re: [tip:sched/core] sched/fair: Limit sched_cfs_period_timer loop to avoid hard lockup

2019-04-16 Thread Phil Auld
On Tue, Apr 09, 2019 at 03:05:27PM +0200 Peter Zijlstra wrote: > On Tue, Apr 09, 2019 at 08:48:16AM -0400, Phil Auld wrote: > > Hi Ingo, Peter, > > > > On Wed, Apr 03, 2019 at 01:38:39AM -0700 tip-bot for Phil Auld wrote: > > > Commit-ID: 06ec5d30e8d57b820d44df6

Re: [tip:sched/urgent] sched/fair: Limit sched_cfs_period_timer() loop to avoid hard lockup

2019-04-16 Thread Phil Auld
Hi Sasha, On Tue, Apr 16, 2019 at 08:32:09AM -0700 tip-bot for Phil Auld wrote: > Commit-ID: 2e8e19226398db8265a8e675fcc0118b9e80c9e8 > Gitweb: > https://git.kernel.org/tip/2e8e19226398db8265a8e675fcc0118b9e80c9e8 > Author: Phil Auld > AuthorDate: Tue, 19 Mar 2019

Re: [RFC PATCH v2 12/17] sched: A quick and dirty cgroup tagging interface

2019-04-25 Thread Phil Auld
On Tue, Apr 23, 2019 at 04:18:17PM + Vineeth Remanan Pillai wrote: > From: Peter Zijlstra (Intel) > > Marks all tasks in a cgroup as matching for core-scheduling. > > Signed-off-by: Peter Zijlstra (Intel) > --- > kernel/sched/core.c | 62 > ker

Re: [RFC PATCH v2 11/17] sched: Basic tracking of matching tasks

2019-04-25 Thread Phil Auld
On Wed, Apr 24, 2019 at 08:43:36PM + Vineeth Remanan Pillai wrote: > > A minor nitpick. I find keeping the vruntime base readjustment in > > core_prio_less probably is more straight forward rather than pass a > > core_cmp bool around. > > The reason I moved the vruntime base adjustment to __p

Re: [RFC] sched/fair: hard lockup in sched_cfs_period_timer

2019-03-12 Thread Phil Auld
On Mon, Mar 11, 2019 at 04:25:36PM -0400 Phil Auld wrote: > On Mon, Mar 11, 2019 at 10:44:25AM -0700 bseg...@google.com wrote: > > Letting it spin for 100ms and then only increasing by 6% seems extremely > > generous. If we went this route I'd probably say "after loo

[PATCH] sched/fair: Limit sched_cfs_period_timer loop to avoid hard lockup

2019-03-13 Thread Phil Auld
as suggested by Ben Segall . Signed-off-by: Phil Auld Cc: Ben Segall Cc: Ingo Molnar Cc: Peter Zijlstra (Intel) --- Note: This is against v5.0 as suggested by the documentation. It won't apply to 5.0+ due to the change to raw_spin_lock_irqsave. I can respin as needed. kernel/

Re: [RFC] sched/fair: hard lockup in sched_cfs_period_timer

2019-03-13 Thread Phil Auld
On Wed, Mar 13, 2019 at 10:44:09AM -0700 bseg...@google.com wrote: > Phil Auld writes: > > > On Mon, Mar 11, 2019 at 04:25:36PM -0400 Phil Auld wrote: > >> On Mon, Mar 11, 2019 at 10:44:25AM -0700 bseg...@google.com wrote: > >> > Letting it spin for 100ms and

Re: [RFC] sched/fair: hard lockup in sched_cfs_period_timer

2019-03-13 Thread Phil Auld
On Wed, Mar 13, 2019 at 01:26:51PM -0700 bseg...@google.com wrote: > Phil Auld writes: > > > On Wed, Mar 13, 2019 at 10:44:09AM -0700 bseg...@google.com wrote: > >> Phil Auld writes: > >> > >> > On Mon, Mar 11, 2019 at 04:25:36PM -0400 Phil Auld wr

Re: [PATCH] sched/fair: Limit sched_cfs_period_timer loop to avoid hard lockup

2019-03-15 Thread Phil Auld
On Fri, Mar 15, 2019 at 11:11:50AM +0100 Peter Zijlstra wrote: > On Wed, Mar 13, 2019 at 11:08:26AM -0400, Phil Auld wrote: > > diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c > > index 310d0637fe4b..90cc67bbf592 100644 > > --- a/kernel/sched/fair.c > >

Re: [PATCH] sched/fair: Limit sched_cfs_period_timer loop to avoid hard lockup

2019-03-15 Thread Phil Auld
On Fri, Mar 15, 2019 at 11:33:57AM +0100 Peter Zijlstra wrote: > On Fri, Mar 15, 2019 at 11:11:50AM +0100, Peter Zijlstra wrote: > > diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c > > index ea74d43924b2..b71557be6b42 100644 > > --- a/kernel/sched/fair.c > > +++ b/kernel/sched/fair.c > > @@

Re: [PATCH] sched/fair: Limit sched_cfs_period_timer loop to avoid hard lockup

2019-03-15 Thread Phil Auld
On Fri, Mar 15, 2019 at 11:11:50AM +0100 Peter Zijlstra wrote: > On Wed, Mar 13, 2019 at 11:08:26AM -0400, Phil Auld wrote: ... > Computers _suck_ at /100. And since you're free to pick the constant, > pick a power of two, computers love those. > > > + > > +

Re: [PATCH] sched/fair: Limit sched_cfs_period_timer loop to avoid hard lockup

2019-03-15 Thread Phil Auld
On Fri, Mar 15, 2019 at 05:03:47PM +0100 Peter Zijlstra wrote: > On Fri, Mar 15, 2019 at 11:30:42AM -0400, Phil Auld wrote: > > > In my defense here, all the fair.c imbalance pct code also uses 100 :) > > Yes, I know, I hate on that too ;-) Just never got around to fixing >

Re: [PATCH] sched/fair: Limit sched_cfs_period_timer loop to avoid hard lockup

2019-03-15 Thread Phil Auld
On Fri, Mar 15, 2019 at 04:59:33PM +0100 Peter Zijlstra wrote: > On Fri, Mar 15, 2019 at 09:51:25AM -0400, Phil Auld wrote: > > On Fri, Mar 15, 2019 at 11:33:57AM +0100 Peter Zijlstra wrote: > > > On Fri, Mar 15, 2019 at 11:11:50AM +0100, Peter Zijlstra wrote: > > >

Re: [tip:sched/core] sched/fair: Limit sched_cfs_period_timer loop to avoid hard lockup

2019-04-09 Thread Phil Auld
Hi Ingo, Peter, On Wed, Apr 03, 2019 at 01:38:39AM -0700 tip-bot for Phil Auld wrote: > Commit-ID: 06ec5d30e8d57b820d44df6340dcb25010d6d0fa > Gitweb: > https://git.kernel.org/tip/06ec5d30e8d57b820d44df6340dcb25010d6d0fa > Author: Phil Auld > AuthorDate: Tue, 19 Mar 2019

Re: [tip:sched/core] sched/fair: Limit sched_cfs_period_timer loop to avoid hard lockup

2019-04-09 Thread Phil Auld
On Tue, Apr 09, 2019 at 03:05:27PM +0200 Peter Zijlstra wrote: > On Tue, Apr 09, 2019 at 08:48:16AM -0400, Phil Auld wrote: > > Hi Ingo, Peter, > > > > On Wed, Apr 03, 2019 at 01:38:39AM -0700 tip-bot for Phil Auld wrote: > > > Commit-ID: 06ec5d30e8d57b820d44df6

Re: [PATCH v2] cpuset: restore sanity to cpuset_cpus_allowed_fallback()

2019-04-10 Thread Phil Auld
sed if > + * _every_ other avenue has been traveled. > + **/ > + > void cpuset_cpus_allowed_fallback(struct task_struct *tsk) > { > rcu_read_lock(); > - do_set_cpus_allowed(tsk, task_cs(tsk)->effective_cpus); > + do_set_cpus_allowed(tsk, is_in_v2_mode() ? > + task_cs(tsk)->cpus_allowed : cpu_possible_mask); > rcu_read_unlock(); > > /* > -- > 2.18.1 > Fwiw, Acked-by: Phil Auld --

[PATCH] sched: Fix use of count for nr_running tracepoint

2020-08-05 Thread Phil Auld
The count field is meant to tell if an update to nr_running is an add or a subtract. Make it do so by adding the missing minus sign. Fixes: 9d246053a691 ("sched: Add a tracepoint to track rq->nr_running") Signed-off-by: Phil Auld --- kernel/sched/sched.h | 2 +- 1 file changed

Re: [PATCH 2/2] sched/fair: Relax task_hot() for misfit tasks

2021-04-19 Thread Phil Auld
On Fri, Apr 16, 2021 at 10:43:38AM +0100 Valentin Schneider wrote: > On 15/04/21 16:39, Rik van Riel wrote: > > On Thu, 2021-04-15 at 18:58 +0100, Valentin Schneider wrote: > >> Consider the following topology: > >> > >> Long story short, preempted misfit tasks are affected by task_hot(), > >> whil

Re: [PATCH 2/2] sched/fair: Relax task_hot() for misfit tasks

2021-04-19 Thread Phil Auld
On Mon, Apr 19, 2021 at 06:17:47PM +0100 Valentin Schneider wrote: > On 19/04/21 08:59, Phil Auld wrote: > > On Fri, Apr 16, 2021 at 10:43:38AM +0100 Valentin Schneider wrote: > >> On 15/04/21 16:39, Rik van Riel wrote: > >> > On Thu, 2021-04-15 at 18:58 +

Re: [PATCH v4 1/4] sched/fair: Introduce primitives for CFS bandwidth burst

2021-03-18 Thread Phil Auld
On Thu, Mar 18, 2021 at 09:26:58AM +0800 changhuaixin wrote: > > > > On Mar 17, 2021, at 4:06 PM, Peter Zijlstra wrote: > > > > On Wed, Mar 17, 2021 at 03:16:18PM +0800, changhuaixin wrote: > > > >>> Why do you allow such a large burst? I would expect something like: > >>> > >>> if (burst >

Re: [PATCH] sched/fair: Limit sched_cfs_period_timer loop to avoid hard lockup

2019-03-18 Thread Phil Auld
On Fri, Mar 15, 2019 at 05:03:47PM +0100 Peter Zijlstra wrote: > On Fri, Mar 15, 2019 at 11:30:42AM -0400, Phil Auld wrote: > > >> I'll rework the maths in the averaged version and post v2 if that makes > >> sense. > > > > It may have the extra timer f

Re: [PATCH] sched/fair: Limit sched_cfs_period_timer loop to avoid hard lockup

2019-03-18 Thread Phil Auld
On Mon, Mar 18, 2019 at 10:14:22AM -0700 bseg...@google.com wrote: > Phil Auld writes: > > > On Fri, Mar 15, 2019 at 05:03:47PM +0100 Peter Zijlstra wrote: > >> On Fri, Mar 15, 2019 at 11:30:42AM -0400, Phil Auld wrote: > >> > >> >> I'll rewor

[PATCH v2] sched/fair: Limit sched_cfs_period_timer loop to avoid hard lockup

2019-03-19 Thread Phil Auld
is state and the new values. v2: Math reworked/simplified by Peter Zijlstra. Signed-off-by: Phil Auld Cc: Ben Segall Cc: Ingo Molnar Cc: Peter Zijlstra (Intel) Cc: Anton Blanchard --- kernel/sched/fair.c | 25 + 1 file changed, 25 insertions(+) diff --git a/kernel/sc

Re: [RFC PATCH v3 00/16] Core scheduling v3

2019-08-29 Thread Phil Auld
On Wed, Aug 28, 2019 at 06:01:14PM +0200 Peter Zijlstra wrote: > On Wed, Aug 28, 2019 at 11:30:34AM -0400, Phil Auld wrote: > > On Tue, Aug 27, 2019 at 11:50:35PM +0200 Peter Zijlstra wrote: > > > > And given MDS, I'm still not entirely convinced it all makes sense.

Re: [PATCH v2 0/8] sched/fair: rework the CFS load balance

2019-08-29 Thread Phil Auld
wrong group in find_busiest_group due to using the average load. The second was in fix_small_imbalance(). The "load" of the lu.C tasks was so low it often failed to move anything even when it did find a group that was overloaded (nr_running > width). I have two small patches which fix this but since Vincent was > embarking on a re-work which also addressed this I dropped them. We've also run a series of performance tests we use to check for regressions and did not find any bad results on our workloads and systems. So... Tested-by: Phil Auld Cheers, Phil --

Re: [PATCH] sched: use rq_lock/unlock in online_fair_sched_group

2019-08-15 Thread Phil Auld
On Fri, Aug 09, 2019 at 06:43:09PM +0100 Valentin Schneider wrote: > On 09/08/2019 14:33, Phil Auld wrote: > > On Tue, Aug 06, 2019 at 03:03:34PM +0200 Peter Zijlstra wrote: > >> On Thu, Aug 01, 2019 at 09:37:49AM -0400, Phil Auld wrote: > >>> Enabling WARN_DOU

[PATCH] sched/rt: silence double clock update warning by using rq_lock wrappers

2019-08-15 Thread Phil Auld
er does: raw_spin_lock(&rq->lock); update_rq_clock(rq); which triggers the warning because of not using the rq_lock wrappers. So, use the wrappers. Signed-off-by: Phil Auld Cc: Peter Zijlstra (Intel) Cc: Ingo Molnar Cc: Valentin Schneider Cc: Dietmar Eggemann ---

Re: [RFC PATCH v3 00/16] Core scheduling v3

2019-08-28 Thread Phil Auld
On Tue, Aug 27, 2019 at 11:50:35PM +0200 Peter Zijlstra wrote: > On Tue, Aug 27, 2019 at 10:14:17PM +0100, Matthew Garrett wrote: > > Apple have provided a sysctl that allows applications to indicate that > > specific threads should make use of core isolation while allowing > > the rest of the sy

Re: [PATCH -next v2] sched/fair: fix -Wunused-but-set-variable warnings

2019-08-23 Thread Phil Auld
On Fri, Aug 23, 2019 at 10:28:02AM -0700 bseg...@google.com wrote: > Dave Chiluk writes: > > > On Wed, Aug 21, 2019 at 12:36 PM wrote: > >> > >> Qian Cai writes: > >> > >> > The linux-next commit "sched/fair: Fix low cpu usage with high > >> > throttling by removing expiration of cpu-local slic

Re: [PATCH] sched/fair: scale quota and period without losing quota/period ratio precision

2019-10-03 Thread Phil Auld
Hi, On Thu, Oct 03, 2019 at 05:12:43PM -0700 Xuewei Zhang wrote: > quota/period ratio is used to ensure a child task group won't get more > bandwidth than the parent task group, and is calculated as: > normalized_cfs_quota() = [(quota_us << 20) / period_us] > > If the quota/period ratio was chang

Re: [PATCH v3 0/8] sched/fair: rework the CFS load balance

2019-10-08 Thread Phil Auld
Hi Vincent, On Thu, Sep 19, 2019 at 09:33:31AM +0200 Vincent Guittot wrote: > Several wrong task placement have been raised with the current load > balance algorithm but their fixes are not always straight forward and > end up with using biased values to force migrations. A cleanup and rework > of

Re: [PATCH v3 0/8] sched/fair: rework the CFS load balance

2019-10-09 Thread Phil Auld
On Tue, Oct 08, 2019 at 05:53:11PM +0200 Vincent Guittot wrote: > Hi Phil, > ... > While preparing v4, I have noticed that I have probably oversimplified > the end of find_idlest_group() in patch "sched/fair: optimize > find_idlest_group" when it compares local vs the idlest other group. > Espe

Re: [PATCH] sched/fair: scale quota and period without losing quota/period ratio precision

2019-10-04 Thread Phil Auld
On Thu, Oct 03, 2019 at 07:05:56PM -0700 Xuewei Zhang wrote: > +cc neeln...@google.com and hao...@google.com, they helped a lot > for this issue. Sorry I forgot to include them when sending out the patch. > > On Thu, Oct 3, 2019 at 5:55 PM Phil Auld wrote: > > > > Hi

Re: [PATCH] sched/fair: scale quota and period without losing quota/period ratio precision

2019-10-07 Thread Phil Auld
Hi Xuewei, On Fri, Oct 04, 2019 at 05:28:15PM -0700 Xuewei Zhang wrote: > On Fri, Oct 4, 2019 at 6:14 AM Phil Auld wrote: > > > > On Thu, Oct 03, 2019 at 07:05:56PM -0700 Xuewei Zhang wrote: > > > +cc neeln...@google.com and hao...@google.com, they helped a lot >

Re: [PATCH] sched/fair: scale quota and period without losing quota/period ratio precision

2019-10-07 Thread Phil Auld
20, cfs_quota_us = 3200) [ 1393.965140] cfs_period_timer[cpu11]: period too short, but cannot scale up without losing precision (cfs_period_us = 20, cfs_quota_us = 3200) I suspect going higher could cause the original lockup, but that'd be the case with the old code as well. An

Re: [PATCH v4 00/10] sched/fair: rework the CFS load balance

2019-10-21 Thread Phil Auld
On Mon, Oct 21, 2019 at 10:44:20AM +0200 Vincent Guittot wrote: > On Mon, 21 Oct 2019 at 09:50, Ingo Molnar wrote: > > > > > > * Vincent Guittot wrote: > > > > > Several wrong task placement have been raised with the current load > > > balance algorithm but their fixes are not always straight for

Re: [RFC][PATCH 03/16] sched: Wrap rq::lock access

2019-02-19 Thread Phil Auld
On Mon, Feb 18, 2019 at 05:56:23PM +0100 Peter Zijlstra wrote: > In preparation of playing games with rq->lock, abstract the thing > using an accessor. > > Signed-off-by: Peter Zijlstra (Intel) Hi Peter, Sorry... what tree are these for? They don't apply to mainline. Some branch on tip, I gue

Re: [RFC][PATCH 03/16] sched: Wrap rq::lock access

2019-02-19 Thread Phil Auld
On Tue, Feb 19, 2019 at 05:22:50PM +0100 Peter Zijlstra wrote: > On Tue, Feb 19, 2019 at 11:13:43AM -0500, Phil Auld wrote: > > On Mon, Feb 18, 2019 at 05:56:23PM +0100 Peter Zijlstra wrote: > > > In preparation of playing games with rq->lock, abstract the thing &g

Re: [RFC PATCH v2 13/17] sched: Add core wide task selection and scheduling.

2019-05-20 Thread Phil Auld
On Sat, May 18, 2019 at 11:37:56PM +0800 Aubrey Li wrote: > On Wed, Apr 24, 2019 at 12:18 AM Vineeth Remanan Pillai > wrote: > > > > From: Peter Zijlstra (Intel) > > > > Instead of only selecting a local task, select a task for all SMT > > siblings for every reschedule on the core (irrespective w

Re: [PATCH v2 1/1] sched/fair: Fix low cpu usage with high throttling by removing expiration of cpu-local slices

2019-05-24 Thread Phil Auld
On Fri, May 24, 2019 at 10:14:36AM -0500 Dave Chiluk wrote: > On Fri, May 24, 2019 at 9:32 AM Phil Auld wrote: > > On Thu, May 23, 2019 at 02:01:58PM -0700 Peter Oskolkov wrote: > > > > If the machine runs at/close to capacity, won't the overallocation > >

Re: [RFC PATCH v2 00/17] Core scheduling v2

2019-04-23 Thread Phil Auld
Hi, On Tue, Apr 23, 2019 at 04:18:05PM + Vineeth Remanan Pillai wrote: > Second iteration of the core-scheduling feature. Thanks for spinning V2 of this. > > This version fixes apparent bugs and performance issues in v1. This > doesn't fully address the issue of core sharing between process

Re: [RFC PATCH v2 00/17] Core scheduling v2

2019-04-26 Thread Phil Auld
On Thu, Apr 25, 2019 at 08:53:43PM +0200 Ingo Molnar wrote: > Interesting. This strongly suggests sub-optimal SMT-scheduling in the > non-saturated HT case, i.e. a scheduler balancing bug. > > As long as loads are clearly below the physical cores count (which they > are in the early phases of yo

Re: [RFC PATCH v2 12/17] sched: A quick and dirty cgroup tagging interface

2019-04-26 Thread Phil Auld
On Fri, Apr 26, 2019 at 04:13:07PM +0200 Peter Zijlstra wrote: > On Thu, Apr 25, 2019 at 10:26:53AM -0400, Phil Auld wrote: > > diff --git a/kernel/sched/core.c b/kernel/sched/core.c > > index e8e5f26db052..b312ea1e28a4 100644 > > --- a/kernel/sched/core.c > >

Re: [RFC PATCH v2 00/17] Core scheduling v2

2019-04-29 Thread Phil Auld
On Mon, Apr 29, 2019 at 09:25:35PM +0800 Li, Aubrey wrote: > On 2019/4/29 14:14, Ingo Molnar wrote: > > > > * Li, Aubrey wrote: > > > >>> I suspect it's pretty low, below 1% for all rows? > >> > >> Hope my this mail box works for this... > >> > >> .---

Re: [tip:sched/core] sched/fair: Use rq_lock/unlock in online_fair_sched_group

2019-08-12 Thread Phil Auld
On Mon, Aug 12, 2019 at 05:52:04AM -0700 tip-bot for Phil Auld wrote: > Commit-ID: a46d14eca7b75fffe35603aa8b81df654353d80f > Gitweb: > https://git.kernel.org/tip/a46d14eca7b75fffe35603aa8b81df654353d80f > Author: Phil Auld > AuthorDate: Thu, 1 Aug 2019 09:37:49 -0

Re: [PATCH] sched: use rq_lock/unlock in online_fair_sched_group

2019-08-09 Thread Phil Auld
On Tue, Aug 06, 2019 at 03:03:34PM +0200 Peter Zijlstra wrote: > On Thu, Aug 01, 2019 at 09:37:49AM -0400, Phil Auld wrote: > > Enabling WARN_DOUBLE_CLOCK in /sys/kernel/debug/sched_features causes > > ISTR there were more issues; but it sure is good to start picking them > off

Re: [tip:sched/core] sched/fair: Use rq_lock/unlock in online_fair_sched_group

2019-08-09 Thread Phil Auld
On Fri, Aug 09, 2019 at 06:21:22PM +0200 Dietmar Eggemann wrote: > On 8/8/19 1:01 PM, tip-bot for Phil Auld wrote: > > [...] > > > diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c > > index 19c58599e967..d9407517dae9 100644 > > --- a/kernel/sched/fair.c

[PATCH] sched: use rq_lock/unlock in online_fair_sched_group

2019-08-01 Thread Phil Auld
e raw locking removes this warning. Signed-off-by: Phil Auld Cc: Peter Zijlstra Cc: Ingo Molnar Cc: Vincent Guittot --- Resend with PATCH instead of CHANGE in subject, and more recent upstream x86 backtrace. kernel/sched/fair.c | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-)

[CHANGE] sched: use rq_lock/unlock in online_fair_sched_group

2019-07-26 Thread Phil Auld
0/0x130 [ 612.546585] online_fair_sched_group+0x70/0x140 [ 612.551092] sched_online_group+0xd0/0xf0 [ 612.555082] sched_autogroup_create_attach+0xd0/0x198 [ 612.560108] sys_setsid+0x140/0x160 [ 612.563579] el0_svc_naked+0x44/0x48 Signed-off-by: Phil Auld Cc: Peter Zijlstra Cc: Ingo Molnar Cc:

Re: [PATCH] sched: use rq_lock/unlock in online_fair_sched_group

2019-08-05 Thread Phil Auld
On Fri, Aug 02, 2019 at 05:20:38PM +0800 Hillf Danton wrote: > > On Thu, 1 Aug 2019 09:37:49 -0400 Phil Auld wrote: > > > > Enabling WARN_DOUBLE_CLOCK in /sys/kernel/debug/sched_features causes > > warning to fire in update_rq_clock. This seems to be caused by onlining &g

Re: [RFC PATCH v3 00/16] Core scheduling v3

2019-08-05 Thread Phil Auld
Hi, On Fri, Aug 02, 2019 at 11:37:15AM -0400 Julien Desfossez wrote: > We tested both Aaron's and Tim's patches and here are our results. > > Test setup: > - 2 1-thread sysbench, one running the cpu benchmark, the other one the > mem benchmark > - both started at the same time > - both are pinn

Re: [PATCH] sched: use rq_lock/unlock in online_fair_sched_group

2019-08-06 Thread Phil Auld
On Tue, Aug 06, 2019 at 02:04:16PM +0800 Hillf Danton wrote: > > On Mon, 5 Aug 2019 22:07:05 +0800 Phil Auld wrote: > > > > If we're to clear that flag right there, outside of the lock pinning code, > > then I think we might as well just remove the flag and all as

Re: [RFC][PATCH 02/13] stop_machine: Fix stop_cpus_in_progress ordering

2019-07-30 Thread Phil Auld
On Fri, Jul 26, 2019 at 04:54:11PM +0200 Peter Zijlstra wrote: > Make sure the entire for loop has stop_cpus_in_progress set. > > Cc: Valentin Schneider > Cc: Aaron Lu > Cc: keesc...@chromium.org > Cc: mi...@kernel.org > Cc: Pawan Gupta > Cc: Phil Auld > Cc: torva..

Re: [PATCH] sched: use rq_lock/unlock in online_fair_sched_group

2019-08-06 Thread Phil Auld
On Tue, Aug 06, 2019 at 03:03:34PM +0200 Peter Zijlstra wrote: > On Thu, Aug 01, 2019 at 09:37:49AM -0400, Phil Auld wrote: > > Enabling WARN_DOUBLE_CLOCK in /sys/kernel/debug/sched_features causes > > ISTR there were more issues; but it sure is good to start picking them > o

Re: [RFC PATCH v3 00/16] Core scheduling v3

2019-08-06 Thread Phil Auld
On Tue, Aug 06, 2019 at 09:54:01PM +0800 Aaron Lu wrote: > On Mon, Aug 05, 2019 at 04:09:15PM -0400, Phil Auld wrote: > > Hi, > > > > On Fri, Aug 02, 2019 at 11:37:15AM -0400 Julien Desfossez wrote: > > > We tested both Aaron's and Tim's patches and here

Re: [RFC PATCH v3 00/16] Core scheduling v3

2019-08-06 Thread Phil Auld
On Tue, Aug 06, 2019 at 10:41:25PM +0800 Aaron Lu wrote: > On 2019/8/6 22:17, Phil Auld wrote: > > On Tue, Aug 06, 2019 at 09:54:01PM +0800 Aaron Lu wrote: > >> On Mon, Aug 05, 2019 at 04:09:15PM -0400, Phil Auld wrote: > >>> Hi, > >>> > >&g

Re: [PATCH v2] sched/fair: don't push cfs_bandwith slack timers forward

2019-06-11 Thread Phil Auld
> > booldistribute_running; > + boolslack_started; > #endif > }; > > -- > 2.22.0.rc1.257.g3120a18244-goog > I think this looks good. I like not delaying that further even if it does not fix Dave's use case. It does make it glaring that I should have used false/true for setting distribute_running though :) Acked-by: Phil Auld --

Re: [PATCH v2] sched/fair: don't push cfs_bandwith slack timers forward

2019-06-11 Thread Phil Auld
On Tue, Jun 11, 2019 at 03:53:25PM +0200 Peter Zijlstra wrote: > On Thu, Jun 06, 2019 at 10:21:01AM -0700, bseg...@google.com wrote: > > diff --git a/kernel/sched/sched.h b/kernel/sched/sched.h > > index efa686eeff26..60219acda94b 100644 > > --- a/kernel/sched/sched.h > > +++ b/kernel/sched/sched.h

Re: [PATCH v2] sched/fair: don't push cfs_bandwith slack timers forward

2019-06-11 Thread Phil Auld
On Tue, Jun 11, 2019 at 04:24:43PM +0200 Peter Zijlstra wrote: > On Tue, Jun 11, 2019 at 10:12:19AM -0400, Phil Auld wrote: > > > That looks reasonable to me. > > > > Out of curiosity, why not bool? Is sizeof bool architecture dependent? > > Yeah, sizeof(_Bool)

Re: [RFC] sched/fair: hard lockup in sched_cfs_period_timer

2019-03-09 Thread Phil Auld
On Wed, Mar 06, 2019 at 11:25:02AM -0800 bseg...@google.com wrote: > Phil Auld writes: > > > On Tue, Mar 05, 2019 at 12:45:34PM -0800 bseg...@google.com wrote: > >> Phil Auld writes: > >> > >> > Interestingly, if I limit the number of child cgroups

Re: [RFC] sched/fair: hard lockup in sched_cfs_period_timer

2019-03-11 Thread Phil Auld
On Mon, Mar 11, 2019 at 10:44:25AM -0700 bseg...@google.com wrote: > Phil Auld writes: > > > On Wed, Mar 06, 2019 at 11:25:02AM -0800 bseg...@google.com wrote: > >> Phil Auld writes: > >> > >> > On Tue, Mar 05, 2019 at 12:45:34PM -0800 bseg...@g

Re: [PATCH v2] sched/fair: Limit sched_cfs_period_timer loop to avoid hard lockup

2019-03-21 Thread Phil Auld
On Thu, Mar 21, 2019 at 07:01:37PM +0100 Peter Zijlstra wrote: > On Tue, Mar 19, 2019 at 09:00:05AM -0400, Phil Auld wrote: > > sched/fair: Limit sched_cfs_period_timer loop to avoid hard lockup > > > > With extremely short cfs_period_us setting on a parent task group with a

Re: [RFC] sched/fair: hard lockup in sched_cfs_period_timer

2019-03-04 Thread Phil Auld
On Mon, Mar 04, 2019 at 10:13:49AM -0800 bseg...@google.com wrote: > Phil Auld writes: > > > Hi, > > > > I have a reproducible case of this: > > > > [ 217.264946] NMI watchdog: Watchdog detected hard LOCKUP on cpu 24 > > [ 217.264948] M

Re: [RFC] sched/fair: hard lockup in sched_cfs_period_timer

2019-03-05 Thread Phil Auld
On Tue, Mar 05, 2019 at 10:49:01AM -0800 bseg...@google.com wrote: > Phil Auld writes: > > >> > > >> > raw_spin_lock(&cfs_b->lock); > >> > for (;;) { > >> > overrun = hrtimer_forward_now(timer, cfs_b->peri

Re: [RFC] sched/fair: hard lockup in sched_cfs_period_timer

2019-03-06 Thread Phil Auld
On Tue, Mar 05, 2019 at 12:45:34PM -0800 bseg...@google.com wrote: > Phil Auld writes: > > > Interestingly, if I limit the number of child cgroups to the number of > > them I'm actually putting processes into (16 down from 2500) the problem > > does not r

Re: [RFC][PATCH] sched: Better document ttwu()

2020-07-02 Thread Phil Auld
Hi Peter, On Thu, Jul 02, 2020 at 02:52:11PM +0200 Peter Zijlstra wrote: > > Dave hit the problem fixed by commit: > > b6e13e85829f ("sched/core: Fix ttwu() race") > > and failed to understand much of the code involved. Per his request a > few comments to (hopefully) clarify things. > > Re

Re: [RFC PATCH 00/13] Core scheduling v5

2020-06-30 Thread Phil Auld
On Fri, Jun 26, 2020 at 11:10:28AM -0400 Joel Fernandes wrote: > On Fri, Jun 26, 2020 at 10:36:01AM -0400, Vineeth Remanan Pillai wrote: > > On Thu, Jun 25, 2020 at 9:47 PM Joel Fernandes > > wrote: > > > > > > On Thu, Jun 25, 2020 at 4:12 PM Vineeth Remanan Pillai > > > wrote: > > > [...] > > >

[PATCH] Sched: Add a tracepoint to track rq->nr_running

2020-06-19 Thread Phil Auld
acepoints are added to add_nr_running() and sub_nr_running() which are in kernel/sched/sched.h. Since sched.h includes trace/events/tlb.h via mmu_context.h we had to limit when CREATE_TRACE_POINTS is defined. Signed-off-by: Phil Auld CC: Qais Yousef CC: Ingo Molnar CC: Peter Zijlstra CC: Vincen

Re: [PATCH] sched/fair: Remove the force parameter of update_tg_load_avg()

2020-09-25 Thread Phil Auld
_load_avg(cfs_rq, false); > + update_tg_load_avg(cfs_rq); > propagate_entity_cfs_rq(se); > } > > @@ -10805,7 +10804,7 @@ static void attach_entity_cfs_rq(struct sched_entity > *se) > /* Synchronize entity with its cfs_rq */ > update_load_avg(cfs_rq, se, sched_feat(ATTACH_AGE_LOAD) ? 0 : > SKIP_AGE_LOAD); > attach_entity_load_avg(cfs_rq, se); > - update_tg_load_avg(cfs_rq, false); > + update_tg_load_avg(cfs_rq); > propagate_entity_cfs_rq(se); > } > > -- > 2.17.1 > LGTM, Reviewed-by: Phil Auld --

Re: [PATCH RFC] sched: Add a per-thread core scheduling interface

2020-05-28 Thread Phil Auld
On Thu, May 28, 2020 at 07:01:28PM +0200 Peter Zijlstra wrote: > On Sun, May 24, 2020 at 10:00:46AM -0400, Phil Auld wrote: > > On Fri, May 22, 2020 at 05:35:24PM -0400 Joel Fernandes wrote: > > > On Fri, May 22, 2020 at 02:59:05PM +0200, Peter Zijlstra wrote: > > >

Re: [PATCH RFC] sched: Add a per-thread core scheduling interface

2020-05-28 Thread Phil Auld
On Thu, May 28, 2020 at 02:17:19PM -0400 Phil Auld wrote: > On Thu, May 28, 2020 at 07:01:28PM +0200 Peter Zijlstra wrote: > > On Sun, May 24, 2020 at 10:00:46AM -0400, Phil Auld wrote: > > > On Fri, May 22, 2020 at 05:35:24PM -0400 Joel Fernandes wrote: > > > > On F

Re: [PATCH v2] sched/fair: enqueue_task_fair optimization

2020-05-13 Thread Phil Auld
it doesn't jump to the label then se must be NULL for the loop to terminate. The final loop is a NOP if se is NULL. The check wasn't protecting that. Otherwise still > Reviewed-by: Phil Auld Cheers, Phil > Signed-off-by: Vincent Guittot > --- > > v2 changes: > -

Re: [PATCH v2] sched/fair: fix unthrottle_cfs_rq for leaf_cfs_rq list

2020-05-13 Thread Phil Auld
the same pattern as > enqueue_task_fair(). This fixes a problem already faced with the latter and > add an optimization in the last for_each_sched_entity loop. > > Reported-by Tao Zhou > Reviewed-by: Phil Auld > Signed-off-by: Vincent Guittot > --- > > v2 changes: > - R

Re: [PATCH v2] sched/fair: enqueue_task_fair optimization

2020-05-13 Thread Phil Auld
On Wed, May 13, 2020 at 03:10:28PM +0200 Vincent Guittot wrote: > On Wed, 13 May 2020 at 14:45, Phil Auld wrote: > > > > Hi Vincent, > > > > On Wed, May 13, 2020 at 02:33:35PM +0200 Vincent Guittot wrote: > > > enqueue_task_fair jumps to enqueue_

Re: [PATCH v2] sched/fair: enqueue_task_fair optimization

2020-05-13 Thread Phil Auld
On Wed, May 13, 2020 at 03:15:53PM +0200 Vincent Guittot wrote: > On Wed, 13 May 2020 at 15:13, Phil Auld wrote: > > > > On Wed, May 13, 2020 at 03:10:28PM +0200 Vincent Guittot wrote: > > > On Wed, 13 May 2020 at 14:45, Phil Auld wrote: > > > > > > >

Re: [PATCH v2] sched/fair: enqueue_task_fair optimization

2020-05-13 Thread Phil Auld
On Wed, May 13, 2020 at 03:25:29PM +0200 Vincent Guittot wrote: > On Wed, 13 May 2020 at 15:18, Phil Auld wrote: > > > > On Wed, May 13, 2020 at 03:15:53PM +0200 Vincent Guittot wrote: > > > On Wed, 13 May 2020 at 15:13, Phil Auld wrote: > > > > > > &g

Re: [PATCH RFC] sched: Add a per-thread core scheduling interface

2020-05-24 Thread Phil Auld
On Fri, May 22, 2020 at 05:35:24PM -0400 Joel Fernandes wrote: > On Fri, May 22, 2020 at 02:59:05PM +0200, Peter Zijlstra wrote: > [..] > > > > It doens't allow tasks for form their own groups (by for example setting > > > > the key to that of another task). > > > > > > So for this, I was thinking

Re: [tip: sched/core] sched/fair: Remove distribute_running from CFS bandwidth

2020-06-08 Thread Phil Auld
> > don't start a distribution while one is already running. However, even > > in the event that this race occurs, it is fine to have two distributions > > running (especially now that distribute grabs the cfs_b->lock to > > determine remaining quota before assigning). > &g

Re: Requirements to control kernel isolation/nohz_full at runtime

2020-09-03 Thread Phil Auld
On Thu, Sep 03, 2020 at 03:30:15PM -0300 Marcelo Tosatti wrote: > On Thu, Sep 03, 2020 at 03:23:59PM -0300, Marcelo Tosatti wrote: > > On Tue, Sep 01, 2020 at 12:46:41PM +0200, Frederic Weisbecker wrote: > > > Hi, > > > > Hi Frederic, > > > > Thanks for the summary! Looking forward to your commen

Re: [PATCH v2] sched/debug: Add new tracepoint to track cpu_capacity

2020-09-02 Thread Phil Auld
On Wed, Sep 02, 2020 at 12:44:42PM +0200 Dietmar Eggemann wrote: > + Phil Auld > Thanks Dietmar. > On 28/08/2020 19:26, Qais Yousef wrote: > > On 08/28/20 19:10, Dietmar Eggemann wrote: > >> On 28/08/2020 12:27, Qais Yousef wrote: > >>> On 08/28/20 10

Re: [PATCH v2] sched/debug: Add new tracepoint to track cpu_capacity

2020-09-08 Thread Phil Auld
Hi Quais, On Mon, Sep 07, 2020 at 12:02:24PM +0100 Qais Yousef wrote: > On 09/02/20 09:54, Phil Auld wrote: > > > > > > I think this decoupling is not necessary. The natural place for those > > > scheduler trace_event based on trace_points extension files is &g

Re: [PATCH 0/8] Style and small fixes for core-scheduling

2020-10-28 Thread Phil Auld
Hi John, On Wed, Oct 28, 2020 at 05:19:09AM -0700 John B. Wyatt IV wrote: > Patchset of style and small fixes for the 8th iteration of the > Core-Scheduling feature. > > Style fixes include changing spaces to tabs, inserting new lines before > declarations, removing unused braces, and spelling. >

Re: default cpufreq gov, was: [PATCH] sched/fair: check for idle core

2020-10-22 Thread Phil Auld
On Thu, Oct 22, 2020 at 03:58:13PM +0100 Colin Ian King wrote: > On 22/10/2020 15:52, Mel Gorman wrote: > > On Thu, Oct 22, 2020 at 02:29:49PM +0200, Peter Zijlstra wrote: > >> On Thu, Oct 22, 2020 at 02:19:29PM +0200, Rafael J. Wysocki wrote: > However I do want to retire ondemand, conservati

Re: default cpufreq gov, was: [PATCH] sched/fair: check for idle core

2020-10-22 Thread Phil Auld
On Thu, Oct 22, 2020 at 09:32:55PM +0100 Mel Gorman wrote: > On Thu, Oct 22, 2020 at 07:59:43PM +0200, Rafael J. Wysocki wrote: > > > > Agreed. I'd like the option to switch back if we make the default > > > > change. > > > > It's on the table and I'd like to be able to go that way. > > > > > > >

Re: [PATCH] sched/fair: remove the spin_lock operations

2020-10-30 Thread Phil Auld
> @@ -5105,9 +5105,6 @@ static void do_sched_cfs_slack_timer(struct > cfs_bandwidth *cfs_b) > return; > > distribute_cfs_runtime(cfs_b); > - > - raw_spin_lock_irqsave(&cfs_b->lock, flags); > - raw_spin_unlock_irqrestore(&cfs_b->lock, flags); > } > > /* > -- > 2.29.0 > > Nice :) Reviewed-by: Phil Auld --

Re: [PATCH v1] sched/fair: update_pick_idlest() Select group with lowest group_util when idle_cpus are equal

2020-11-09 Thread Phil Auld
Hi, On Fri, Nov 06, 2020 at 04:00:10PM + Mel Gorman wrote: > On Fri, Nov 06, 2020 at 02:33:56PM +0100, Vincent Guittot wrote: > > On Fri, 6 Nov 2020 at 13:03, Mel Gorman wrote: > > > > > > On Wed, Nov 04, 2020 at 09:42:05AM +, Mel Gorman wrote: > > > > While it's possible that some other

Re: [PATCH v1] sched/fair: update_pick_idlest() Select group with lowest group_util when idle_cpus are equal

2020-11-09 Thread Phil Auld
On Mon, Nov 09, 2020 at 03:38:15PM + Mel Gorman wrote: > On Mon, Nov 09, 2020 at 10:24:11AM -0500, Phil Auld wrote: > > Hi, > > > > On Fri, Nov 06, 2020 at 04:00:10PM + Mel Gorman wrote: > > > On Fri, Nov 06, 2020 at 02:33:56PM +0100, Vincent Guittot wrote:

Re: [PATCH] sched/fair: remove the spin_lock operations

2020-11-02 Thread Phil Auld
On Fri, Oct 30, 2020 at 10:16:29PM + David Laight wrote: > From: Benjamin Segall > > Sent: 30 October 2020 18:48 > > > > Hui Su writes: > > > > > Since 'ab93a4bc955b ("sched/fair: Remove > > > distribute_running fromCFS bandwidth")',there is > > > nothing to protect between raw_spin_lock_irq

  1   2   >