On Tue, Mar 3, 2026 at 10:52 PM Yu, Lang <[email protected]> wrote:
>
> [AMD Official Use Only - AMD Internal Distribution Only]
>
> >-----Original Message-----
> >From: Alex Deucher <[email protected]>
> >Sent: Wednesday, March 4, 2026 10:48 AM
> >To: Yu, Lang <[email protected]>
> >Cc: Deucher, Alexander <[email protected]>; amd-
> >[email protected]
> >Subject: Re: [PATCH] drm/amdgpu/gfx12.1: add support for disable_kq
> >
> >On Tue, Mar 3, 2026 at 9:02 PM Yu, Lang <[email protected]> wrote:
> >>
> >> [AMD Official Use Only - AMD Internal Distribution Only]
> >>
> >> >-----Original Message-----
> >> >From: amd-gfx <[email protected]> On Behalf Of
> >> >Alex Deucher
> >> >Sent: Tuesday, March 3, 2026 11:23 PM
> >> >To: Deucher, Alexander <[email protected]>
> >> >Cc: [email protected]
> >> >Subject: Re: [PATCH] drm/amdgpu/gfx12.1: add support for disable_kq
> >> >
> >> >Ping?
> >> >
> >> >Alex
> >> >
> >> >On Wed, Jan 28, 2026 at 3:12 PM Alex Deucher
> >> ><[email protected]>
> >> >wrote:
> >> >>
> >> >> Plumb in support for disabling kernel queues and make it the default.
> >> >> For testing, kernel queues can be re-enabled by setting
> >> >> amdgpu.user_queue=0
> >> >>
> >> >> Signed-off-by: Alex Deucher <[email protected]>
> >> >> ---
> >> >>  drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c | 109
> >> >> +++++++++++++++++++------
> >> >>  1 file changed, 82 insertions(+), 27 deletions(-)
> >> >>
> >> >> diff --git a/drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
> >> >> b/drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
> >> >> index 08ae50a6313f3..f93ee275ce398 100644
> >> >> --- a/drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
> >> >> +++ b/drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
> >> >> @@ -1155,11 +1155,13 @@ static int gfx_v12_1_sw_init(struct
> >> >> amdgpu_ip_block
> >> >*ip_block)
> >> >>                 break;
> >> >>         }
> >> >>
> >> >> -       /* recalculate compute rings to use based on hardware 
> >> >> configuration */
> >> >> -       num_compute_rings = (adev->gfx.mec.num_pipe_per_mec *
> >> >> -                            adev->gfx.mec.num_queue_per_pipe) / 2;
> >> >> -       adev->gfx.num_compute_rings = min(adev->gfx.num_compute_rings,
> >> >> -                                         num_compute_rings);
> >> >> +       if (adev->gfx.num_compute_rings) {
> >> >> +               /* recalculate compute rings to use based on hardware 
> >> >> configuration
> >*/
> >> >> +               num_compute_rings = (adev->gfx.mec.num_pipe_per_mec *
> >> >> +                                    adev->gfx.mec.num_queue_per_pipe) 
> >> >> / 2;
> >> >> +               adev->gfx.num_compute_rings = 
> >> >> min(adev->gfx.num_compute_rings,
> >> >> +                                                 num_compute_rings);
> >> >> +       }
> >> >>
> >> >>         num_xcc = NUM_XCC(adev->gfx.xcc_mask);
> >> >>
> >> >> @@ -2794,6 +2796,36 @@ static void gfx_v12_1_xcc_fini(struct
> >> >> amdgpu_device
> >> >*adev,
> >> >>         gfx_v12_1_xcc_enable_gui_idle_interrupt(adev, false,
> >> >> xcc_id); }
> >> >>
> >> >> +static int gfx_v12_1_set_userq_eop_interrupts(struct amdgpu_device 
> >> >> *adev,
> >> >> +                                             bool enable) {
> >> >> +       unsigned int irq_type;
> >> >> +       int m, p, r, x, num_xcc;
> >> >> +
> >> >> +       if (adev->gfx.disable_kq) {
> >> >> +               num_xcc = NUM_XCC(adev->gfx.xcc_mask);
> >> >> +               for (x = 0; x < num_xcc; x++) {
> >> >> +                       for (m = 0; m < adev->gfx.mec.num_mec; ++m) {
> >> >> +                               for (p = 0; p < 
> >> >> adev->gfx.mec.num_pipe_per_mec; p++) {
> >> >> +                                       irq_type =
> >> >AMDGPU_CP_IRQ_COMPUTE_MEC1_PIPE0_EOP
> >> >> +                                               + (m * 
> >> >> adev->gfx.mec.num_pipe_per_mec)
> >> >> +                                               + p;
> >> >> +                                       if (enable)
> >> >> +                                               r = 
> >> >> amdgpu_irq_get(adev, &adev->gfx.eop_irq,
> >> >> +                                                                  
> >> >> irq_type);
> >> >> +                                       else
> >> >> +                                               r = 
> >> >> amdgpu_irq_put(adev, &adev->gfx.eop_irq,
> >> >> +                                                                  
> >> >> irq_type);
> >> >> +                                       if (r)
> >> >> +                                               return r;
> >> >> +                               }
> >> >> +                       }
> >> >> +               }
> >> >> +       }
> >> >> +
> >> >> +       return 0;
> >> >> +}
> >> >> +
> >> >>  static int gfx_v12_1_hw_fini(struct amdgpu_ip_block *ip_block)  {
> >> >>         struct amdgpu_device *adev = ip_block->adev; @@ -2801,6
> >> >> +2833,7 @@ static int gfx_v12_1_hw_fini(struct amdgpu_ip_block
> >> >> *ip_block)
> >> >>
> >> >>         amdgpu_irq_put(adev, &adev->gfx.priv_reg_irq, 0);
> >> >>         amdgpu_irq_put(adev, &adev->gfx.priv_inst_irq, 0);
> >> >> +       gfx_v12_1_set_userq_eop_interrupts(adev, false);
> >> >>
> >> >>         num_xcc = NUM_XCC(adev->gfx.xcc_mask);
> >> >>         for (i = 0; i < num_xcc; i++) { @@ -2868,10 +2901,26 @@
> >> >> static int gfx_v12_1_early_init(struct amdgpu_ip_block *ip_block)  {
> >> >>         struct amdgpu_device *adev = ip_block->adev;
> >> >>
> >> >> +
> >> >> +       switch (amdgpu_user_queue) {
> >> >> +       case -1:
> >> >> +       default:
> >> >> +               adev->gfx.disable_kq = true;
> >> >> +               adev->gfx.disable_uq = true;
> >> >> +               break;
> >>
> >> This disables both user queue and kernel queue by default.
> >> Is it expected?
> >
> >Yes.  there's not really a need for kernel queues or KGD user queues in this 
> >case.
>
> Thanks.  This also disables the sanity test (ring/ib) test in kernel by 
> default one of kernel queue's functionality.

Yes, but we don't want to enable kernel queues for this IP.

Alex

>
> Regards
> Lang
>
> >Alex
> >
> >>
> >> Regards
> >> Lang
> >>
> >> >> +       case 0:
> >> >> +               adev->gfx.disable_kq = false;
> >> >> +               adev->gfx.disable_uq = true;
> >> >> +               break;
> >> >> +       }
> >> >> +
> >> >>         adev->gfx.funcs = &gfx_v12_1_gfx_funcs;
> >> >>
> >> >> -       adev->gfx.num_compute_rings = min(amdgpu_gfx_get_num_kcq(adev),
> >> >> -                                         AMDGPU_MAX_COMPUTE_RINGS);
> >> >> +       if (adev->gfx.disable_kq)
> >> >> +               adev->gfx.num_compute_rings = 0;
> >> >> +       else
> >> >> +               adev->gfx.num_compute_rings =
> >> >> + min(amdgpu_gfx_get_num_kcq(adev),
> >> >> +
> >> >> + AMDGPU_MAX_COMPUTE_RINGS);
> >> >>
> >> >>         gfx_v12_1_set_kiq_pm4_funcs(adev);
> >> >>         gfx_v12_1_set_ring_funcs(adev); @@ -2898,6 +2947,10 @@
> >> >> static int gfx_v12_1_late_init(struct amdgpu_ip_block *ip_block)
> >> >>         if (r)
> >> >>                 return r;
> >> >>
> >> >> +       r = gfx_v12_1_set_userq_eop_interrupts(adev, true);
> >> >> +       if (r)
> >> >> +               return r;
> >> >> +
> >> >>         return 0;
> >> >>  }
> >> >>
> >> >> @@ -3719,29 +3772,31 @@ static void
> >> >> gfx_v12_1_handle_priv_fault(struct
> >> >amdgpu_device *adev,
> >> >>         if (xcc_id == -EINVAL)
> >> >>                 return;
> >> >>
> >> >> -       switch (me_id) {
> >> >> -       case 0:
> >> >> -               for (i = 0; i < adev->gfx.num_gfx_rings; i++) {
> >> >> -                       ring = &adev->gfx.gfx_ring[i];
> >> >> -                       /* we only enabled 1 gfx queue per pipe for now 
> >> >> */
> >> >> -                       if (ring->me == me_id && ring->pipe == pipe_id)
> >> >> -                               drm_sched_fault(&ring->sched);
> >> >> -               }
> >> >> -               break;
> >> >> -       case 1:
> >> >> -       case 2:
> >> >> -               for (i = 0; i < adev->gfx.num_compute_rings; i++) {
> >> >> -                       ring = &adev->gfx.compute_ring
> >> >> +       if (!adev->gfx.disable_kq) {
> >> >> +               switch (me_id) {
> >> >> +               case 0:
> >> >> +                       for (i = 0; i < adev->gfx.num_gfx_rings; i++) {
> >> >> +                               ring = &adev->gfx.gfx_ring[i];
> >> >> +                               /* we only enabled 1 gfx queue per pipe 
> >> >> for now */
> >> >> +                               if (ring->me == me_id && ring->pipe == 
> >> >> pipe_id)
> >> >> +                                       drm_sched_fault(&ring->sched);
> >> >> +                       }
> >> >> +                       break;
> >> >> +               case 1:
> >> >> +               case 2:
> >> >> +                       for (i = 0; i < adev->gfx.num_compute_rings; 
> >> >> i++) {
> >> >> +                               ring = &adev->gfx.compute_ring
> >> >>                                         [i +
> >> >>                                          xcc_id * 
> >> >> adev->gfx.num_compute_rings];
> >> >> -                       if (ring->me == me_id && ring->pipe == pipe_id 
> >> >> &&
> >> >> -                           ring->queue == queue_id)
> >> >> -                               drm_sched_fault(&ring->sched);
> >> >> +                               if (ring->me == me_id && ring->pipe == 
> >> >> pipe_id &&
> >> >> +                                   ring->queue == queue_id)
> >> >> +                                       drm_sched_fault(&ring->sched);
> >> >> +                       }
> >> >> +                       break;
> >> >> +               default:
> >> >> +                       BUG();
> >> >> +                       break;
> >> >>                 }
> >> >> -               break;
> >> >> -       default:
> >> >> -               BUG();
> >> >> -               break;
> >> >>         }
> >> >>  }
> >> >>
> >> >> --
> >> >> 2.52.0
> >> >>

Reply via email to