On 3/18/2026 6:15 PM, Boqun Feng wrote:
> On Wed, Mar 18, 2026 at 02:55:48PM -0700, Boqun Feng wrote:
>> On Wed, Mar 18, 2026 at 02:52:48PM -0700, Boqun Feng wrote:
>> [...]
>>>> Ah so it is an ABBA deadlock, not a ABA self-deadlock. I guess this is a
>>>> different issue, from the NMI issue? It is more of an issue of calling
>>>> call_srcu API with scheduler locks held.
>>>>
>>>> Something like below I think:
>>>>
>>>> CPU A (BPF tracepoint) CPU B (concurrent call_srcu)
>>>> ---------------------------- ------------------------------------
>>>> [1] holds &rq->__lock
>>>> [2]
>>>> -> call_srcu
>>>> -> srcu_gp_start_if_needed
>>>> -> srcu_funnel_gp_start
>>>> -> spin_lock_irqsave_ssp_content...
>>>> -> holds srcu locks
>>>>
>>>> [4] calls call_rcu_tasks_trace() [5] srcu_funnel_gp_start (cont..)
>>>> -> queue_delayed_work
>>>> -> call_srcu() -> __queue_work()
>>>> -> srcu_gp_start_if_needed() -> wake_up_worker()
>>>> -> srcu_funnel_gp_start() -> try_to_wake_up()
>>>> -> spin_lock_irqsave_ssp_contention() [6] WANTS rq->__lock
>>>> -> WANTS srcu locks
>>>
>>> I see, we can also have a self deadlock even without CPU B, when CPU A
>>> is going to try_to_wake_up() the a worker on the same CPU.
>>>
>>> An interesting observation is that the deadlock can be avoided in
>>> queue_delayed_work() uses a non-zero delay, that means a timer will be
>>> armed instead of acquiring the rq lock.
>>>
>
> If my observation is correct, then this can probably fix the deadlock
> issue with runqueue lock (untested though), but it won't work if BPF
> tracepoint can happen with timer base lock held.
>
> Regards,
> Boqun
>
> ------>
> diff --git a/kernel/rcu/srcutree.c b/kernel/rcu/srcutree.c
> index 2328827f8775..a5d67264acb5 100644
> --- a/kernel/rcu/srcutree.c
> +++ b/kernel/rcu/srcutree.c
> @@ -1061,6 +1061,7 @@ static void srcu_funnel_gp_start(struct srcu_struct
> *ssp, struct srcu_data *sdp,
> struct srcu_node *snp_leaf;
> unsigned long snp_seq;
> struct srcu_usage *sup = ssp->srcu_sup;
> + bool irqs_were_disabled;
>
> /* Ensure that snp node tree is fully initialized before traversing
> it */
> if (smp_load_acquire(&sup->srcu_size_state) < SRCU_SIZE_WAIT_BARRIER)
> @@ -1098,6 +1099,7 @@ static void srcu_funnel_gp_start(struct srcu_struct
> *ssp, struct srcu_data *sdp,
>
> /* Top of tree, must ensure the grace period will be started. */
> raw_spin_lock_irqsave_ssp_contention(ssp, &flags);
> + irqs_were_disabled = irqs_disabled_flags(flags);
> if (ULONG_CMP_LT(sup->srcu_gp_seq_needed, s)) {
> /*
> * Record need for grace period s. Pair with load
> @@ -1118,9 +1120,16 @@ static void srcu_funnel_gp_start(struct srcu_struct
> *ssp, struct srcu_data *sdp,
> // it isn't. And it does not have to be. After all, it
> // can only be executed during early boot when there is only
> // the one boot CPU running with interrupts still disabled.
> + //
> + // If irq was disabled when call_srcu() is called, then we
> + // could be in the scheduler path with a runqueue lock held,
> + // delay the process_srcu() work 1 more jiffies so we don't go
> + // through the kick_pool() -> wake_up_process() path below,
> and
> + // we could avoid deadlock with runqueue lock.
> if (likely(srcu_init_done))
> queue_delayed_work(rcu_gp_wq, &sup->work,
> - !!srcu_get_delay(ssp));
> + !!srcu_get_delay(ssp) +
> + !!irqs_were_disabled);
Nice, I wonder if it is better to do this in __queue_delayed_work() itself.
Do we have queue_delayed_work() with zero delays that are in irq-disabled
regions, and they depend on that zero-delay for correctness? Even with
delay of 0 though, the work item doesn't execute right away anyway, the
worker thread has to also be scheduler right?
Also if IRQ is disabled, I'd think this is a critical path that is not
wanting to run the work item right-away anyway since workqueue is more a
bottom-half mechanism, than "run this immediately".
IOW, would be good to make the workqueue-layer more resilient to waking up
the scheduler when a delay would have been totally ok. But maybe +Tejun can
yell if that sounds insane.
thanks,
--
Joel Fernandes