On Wed, Dec 24, 2025 at 05:00:26PM +0100, Marco Crivellari wrote:
> This patch continues the effort to refactor workqueue APIs, which has begun
> with the changes introducing new workqueues and a new alloc_workqueue flag:
> 
>    commit 128ea9f6ccfb ("workqueue: Add system_percpu_wq and system_dfl_wq")
>    commit 930c2ea566af ("workqueue: Add new WQ_PERCPU flag")
> 
> The point of the refactoring is to eventually alter the default behavior of
> workqueues to become unbound by default so that their workload placement is
> optimized by the scheduler.
> 
> Before that to happen after a careful review and conversion of each individual
> case, workqueue users must be converted to the better named new workqueues 
> with
> no intended behaviour changes:
> 
>    system_wq -> system_percpu_wq
>    system_unbound_wq -> system_dfl_wq
> 
> This way the old obsolete workqueues (system_wq, system_unbound_wq) can be
> removed in the future.
> 
> Suggested-by: Tejun Heo <[email protected]>
> Signed-off-by: Marco Crivellari <[email protected]>
> ---
>  drivers/gpu/drm/xe/xe_tlb_inval.c | 6 +++---
>  1 file changed, 3 insertions(+), 3 deletions(-)
> 
> diff --git a/drivers/gpu/drm/xe/xe_tlb_inval.c 
> b/drivers/gpu/drm/xe/xe_tlb_inval.c
> index 918a59e686ea..b2cf6e17fbc5 100644
> --- a/drivers/gpu/drm/xe/xe_tlb_inval.c
> +++ b/drivers/gpu/drm/xe/xe_tlb_inval.c
> @@ -94,7 +94,7 @@ static void xe_tlb_inval_fence_timeout(struct work_struct 
> *work)
>               xe_tlb_inval_fence_signal(fence);
>       }
>       if (!list_empty(&tlb_inval->pending_fences))
> -             queue_delayed_work(system_wq, &tlb_inval->fence_tdr,
> +             queue_delayed_work(system_percpu_wq, &tlb_inval->fence_tdr,

Actually system_wq or system_percpu_wq doesn't work here as this is the
fence signaling path. We should use one Xe's ordered work queues which
is properly setup to be reclaim same.

Matt

>                                  timeout_delay);
>       spin_unlock_irq(&tlb_inval->pending_lock);
>  }
> @@ -226,7 +226,7 @@ static void xe_tlb_inval_fence_prep(struct 
> xe_tlb_inval_fence *fence)
>       list_add_tail(&fence->link, &tlb_inval->pending_fences);
>  
>       if (list_is_singular(&tlb_inval->pending_fences))
> -             queue_delayed_work(system_wq, &tlb_inval->fence_tdr,
> +             queue_delayed_work(system_percpu_wq, &tlb_inval->fence_tdr,
>                                  tlb_inval->ops->timeout_delay(tlb_inval));
>       spin_unlock_irq(&tlb_inval->pending_lock);
>  
> @@ -378,7 +378,7 @@ void xe_tlb_inval_done_handler(struct xe_tlb_inval 
> *tlb_inval, int seqno)
>       }
>  
>       if (!list_empty(&tlb_inval->pending_fences))
> -             mod_delayed_work(system_wq,
> +             mod_delayed_work(system_percpu_wq,
>                                &tlb_inval->fence_tdr,
>                                tlb_inval->ops->timeout_delay(tlb_inval));
>       else
> -- 
> 2.52.0
> 

Reply via email to