Le 05/02/2026 à 15:20, Alex Deucher a écrit :
On Thu, Feb 5, 2026 at 8:52 AM Pierre-Eric Pelloux-Prayer
<[email protected]> wrote:



Le 30/01/2026 à 18:30, Alex Deucher a écrit :
Switch to using a job structure for IBs.

Acked-by: Christian König <[email protected]>
Signed-off-by: Alex Deucher <[email protected]>
---
   drivers/gpu/drm/amd/amdgpu/amdgpu_vpe.c  |  37 +++---
   drivers/gpu/drm/amd/amdgpu/cik_sdma.c    |  31 ++---
   drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c   |  29 ++---
   drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c   |  29 ++---
   drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c   |  29 ++---
   drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c   |  29 ++---
   drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c    |  24 ++--
   drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c    |  25 ++--
   drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c    | 139 ++++++++++++-----------
   drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c    | 138 +++++++++++-----------
   drivers/gpu/drm/amd/amdgpu/gfx_v9_4_2.c  |  26 ++---
   drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c  |  29 ++---
   drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c   |  38 ++++---
   drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c   |  38 ++++---
   drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c   |  38 ++++---
   drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c |  38 ++++---
   drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c   |  37 +++---
   drivers/gpu/drm/amd/amdgpu/sdma_v5_2.c   |  36 +++---
   drivers/gpu/drm/amd/amdgpu/sdma_v6_0.c   |  36 +++---
   drivers/gpu/drm/amd/amdgpu/sdma_v7_0.c   |  36 +++---
   drivers/gpu/drm/amd/amdgpu/sdma_v7_1.c   |  36 +++---
   drivers/gpu/drm/amd/amdgpu/si_dma.c      |  29 +++--
   22 files changed, 500 insertions(+), 427 deletions(-)

diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vpe.c 
b/drivers/gpu/drm/amd/amdgpu/amdgpu_vpe.c
index fd881388d6125..9fb1946be1ba2 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vpe.c
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vpe.c
@@ -817,7 +817,8 @@ static int vpe_ring_test_ib(struct amdgpu_ring *ring, long 
timeout)
   {
       struct amdgpu_device *adev = ring->adev;
       const uint32_t test_pattern = 0xdeadbeef;
-     struct amdgpu_ib ib = {};
+     struct amdgpu_job *job;
+     struct amdgpu_ib *ib;
       struct dma_fence *f = NULL;
       uint32_t index;
       uint64_t wb_addr;
@@ -832,23 +833,28 @@ static int vpe_ring_test_ib(struct amdgpu_ring *ring, 
long timeout)
       adev->wb.wb[index] = 0;
       wb_addr = adev->wb.gpu_addr + (index * 4);

-     ret = amdgpu_ib_get(adev, NULL, 256, AMDGPU_IB_POOL_DIRECT, &ib);
+     ret = amdgpu_job_alloc_with_ib(ring->adev, NULL, NULL, 256,
+                                    AMDGPU_IB_POOL_DIRECT, &job,
+                                    AMDGPU_KERNEL_JOB_ID_VPE_RING_TEST);
       if (ret)
               goto err0;
-
-     ib.ptr[0] = VPE_CMD_HEADER(VPE_CMD_OPCODE_FENCE, 0);
-     ib.ptr[1] = lower_32_bits(wb_addr);
-     ib.ptr[2] = upper_32_bits(wb_addr);
-     ib.ptr[3] = test_pattern;
-     ib.ptr[4] = VPE_CMD_HEADER(VPE_CMD_OPCODE_NOP, 0);
-     ib.ptr[5] = VPE_CMD_HEADER(VPE_CMD_OPCODE_NOP, 0);
-     ib.ptr[6] = VPE_CMD_HEADER(VPE_CMD_OPCODE_NOP, 0);
-     ib.ptr[7] = VPE_CMD_HEADER(VPE_CMD_OPCODE_NOP, 0);
-     ib.length_dw = 8;
-
-     ret = amdgpu_ib_schedule(ring, 1, &ib, NULL, &f);
-     if (ret)
+     ib = &job->ibs[0];
+
+     ib->ptr[0] = VPE_CMD_HEADER(VPE_CMD_OPCODE_FENCE, 0);
+     ib->ptr[1] = lower_32_bits(wb_addr);
+     ib->ptr[2] = upper_32_bits(wb_addr);
+     ib->ptr[3] = test_pattern;
+     ib->ptr[4] = VPE_CMD_HEADER(VPE_CMD_OPCODE_NOP, 0);
+     ib->ptr[5] = VPE_CMD_HEADER(VPE_CMD_OPCODE_NOP, 0);
+     ib->ptr[6] = VPE_CMD_HEADER(VPE_CMD_OPCODE_NOP, 0);
+     ib->ptr[7] = VPE_CMD_HEADER(VPE_CMD_OPCODE_NOP, 0);
+     ib->length_dw = 8;
+
+     ret = amdgpu_job_submit_direct(job, ring, &f);
+     if (ret) {
+             amdgpu_job_free(job);
               goto err1;
+     }

This is a nice cleanup. What about going one step further and provide
a helper like:

     u32 test_dws[] = {
        VPE_CMD_HEADER(VPE_CMD_OPCODE_FENCE, 0),
        lower_32_bits(wb_addr),
        ...
     };
     ret = amdgpu_ring_run_dwords(ring, ARRAY_SIZE(test_dws), test_dws,
timeout);

Because that's what most _ring_test_ib function do.

I can look into that as a follow up series.

Sure.
In any case this patch is Acked-by me.

Pierre-Eric


Alex


Thanks,
Pierre-Eric


       ret = dma_fence_wait_timeout(f, false, timeout);
       if (ret <= 0) {
@@ -859,7 +865,6 @@ static int vpe_ring_test_ib(struct amdgpu_ring *ring, long 
timeout)
       ret = (le32_to_cpu(adev->wb.wb[index]) == test_pattern) ? 0 : -EINVAL;

   err1:
-     amdgpu_ib_free(&ib, NULL);
       dma_fence_put(f);
   err0:
       amdgpu_device_wb_free(adev, index);
diff --git a/drivers/gpu/drm/amd/amdgpu/cik_sdma.c 
b/drivers/gpu/drm/amd/amdgpu/cik_sdma.c
index 9e8715b4739da..e2ca96f5a7cfb 100644
--- a/drivers/gpu/drm/amd/amdgpu/cik_sdma.c
+++ b/drivers/gpu/drm/amd/amdgpu/cik_sdma.c
@@ -652,7 +652,8 @@ static int cik_sdma_ring_test_ring(struct amdgpu_ring *ring)
   static int cik_sdma_ring_test_ib(struct amdgpu_ring *ring, long timeout)
   {
       struct amdgpu_device *adev = ring->adev;
-     struct amdgpu_ib ib;
+     struct amdgpu_job *job;
+     struct amdgpu_ib *ib;
       struct dma_fence *f = NULL;
       unsigned index;
       u32 tmp = 0;
@@ -666,22 +667,27 @@ static int cik_sdma_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
       gpu_addr = adev->wb.gpu_addr + (index * 4);
       tmp = 0xCAFEDEAD;
       adev->wb.wb[index] = cpu_to_le32(tmp);
-     memset(&ib, 0, sizeof(ib));
-     r = amdgpu_ib_get(adev, NULL, 256,
-                                     AMDGPU_IB_POOL_DIRECT, &ib);
+
+     r = amdgpu_job_alloc_with_ib(ring->adev, NULL, NULL, 256,
+                                  AMDGPU_IB_POOL_DIRECT, &job,
+                                  AMDGPU_KERNEL_JOB_ID_SDMA_RING_TEST);
       if (r)
               goto err0;
+     ib = &job->ibs[0];

-     ib.ptr[0] = SDMA_PACKET(SDMA_OPCODE_WRITE,
+     ib->ptr[0] = SDMA_PACKET(SDMA_OPCODE_WRITE,
                               SDMA_WRITE_SUB_OPCODE_LINEAR, 0);
-     ib.ptr[1] = lower_32_bits(gpu_addr);
-     ib.ptr[2] = upper_32_bits(gpu_addr);
-     ib.ptr[3] = 1;
-     ib.ptr[4] = 0xDEADBEEF;
-     ib.length_dw = 5;
-     r = amdgpu_ib_schedule(ring, 1, &ib, NULL, &f);
-     if (r)
+     ib->ptr[1] = lower_32_bits(gpu_addr);
+     ib->ptr[2] = upper_32_bits(gpu_addr);
+     ib->ptr[3] = 1;
+     ib->ptr[4] = 0xDEADBEEF;
+     ib->length_dw = 5;
+
+     r = amdgpu_job_submit_direct(job, ring, &f);
+     if (r) {
+             amdgpu_job_free(job);
               goto err1;
+     }

       r = dma_fence_wait_timeout(f, false, timeout);
       if (r == 0) {
@@ -697,7 +703,6 @@ static int cik_sdma_ring_test_ib(struct amdgpu_ring *ring, 
long timeout)
               r = -EINVAL;

   err1:
-     amdgpu_ib_free(&ib, NULL);
       dma_fence_put(f);
   err0:
       amdgpu_device_wb_free(adev, index);
diff --git a/drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c 
b/drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
index 1893ceeeb26c8..50954b60e49df 100644
--- a/drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
+++ b/drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
@@ -4071,15 +4071,14 @@ static int gfx_v10_0_ring_test_ring(struct amdgpu_ring 
*ring)
   static int gfx_v10_0_ring_test_ib(struct amdgpu_ring *ring, long timeout)
   {
       struct amdgpu_device *adev = ring->adev;
-     struct amdgpu_ib ib;
+     struct amdgpu_job *job;
+     struct amdgpu_ib *ib;
       struct dma_fence *f = NULL;
       unsigned int index;
       uint64_t gpu_addr;
       uint32_t *cpu_ptr;
       long r;

-     memset(&ib, 0, sizeof(ib));
-
       r = amdgpu_device_wb_get(adev, &index);
       if (r)
               return r;
@@ -4088,22 +4087,27 @@ static int gfx_v10_0_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
       adev->wb.wb[index] = cpu_to_le32(0xCAFEDEAD);
       cpu_ptr = &adev->wb.wb[index];

-     r = amdgpu_ib_get(adev, NULL, 20, AMDGPU_IB_POOL_DIRECT, &ib);
+     r = amdgpu_job_alloc_with_ib(ring->adev, NULL, NULL, 20,
+                                  AMDGPU_IB_POOL_DIRECT, &job,
+                                  AMDGPU_KERNEL_JOB_ID_GFX_RING_TEST);
       if (r) {
               drm_err(adev_to_drm(adev), "failed to get ib (%ld).\n", r);
               goto err1;
       }
+     ib = &job->ibs[0];

-     ib.ptr[0] = PACKET3(PACKET3_WRITE_DATA, 3);
-     ib.ptr[1] = WRITE_DATA_DST_SEL(5) | WR_CONFIRM;
-     ib.ptr[2] = lower_32_bits(gpu_addr);
-     ib.ptr[3] = upper_32_bits(gpu_addr);
-     ib.ptr[4] = 0xDEADBEEF;
-     ib.length_dw = 5;
+     ib->ptr[0] = PACKET3(PACKET3_WRITE_DATA, 3);
+     ib->ptr[1] = WRITE_DATA_DST_SEL(5) | WR_CONFIRM;
+     ib->ptr[2] = lower_32_bits(gpu_addr);
+     ib->ptr[3] = upper_32_bits(gpu_addr);
+     ib->ptr[4] = 0xDEADBEEF;
+     ib->length_dw = 5;

-     r = amdgpu_ib_schedule(ring, 1, &ib, NULL, &f);
-     if (r)
+     r = amdgpu_job_submit_direct(job, ring, &f);
+     if (r) {
+             amdgpu_job_free(job);
               goto err2;
+     }

       r = dma_fence_wait_timeout(f, false, timeout);
       if (r == 0) {
@@ -4118,7 +4122,6 @@ static int gfx_v10_0_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
       else
               r = -EINVAL;
   err2:
-     amdgpu_ib_free(&ib, NULL);
       dma_fence_put(f);
   err1:
       amdgpu_device_wb_free(adev, index);
diff --git a/drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c 
b/drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
index 427975b5a1d97..f7c73ca34e025 100644
--- a/drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
+++ b/drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
@@ -604,7 +604,8 @@ static int gfx_v11_0_ring_test_ring(struct amdgpu_ring 
*ring)
   static int gfx_v11_0_ring_test_ib(struct amdgpu_ring *ring, long timeout)
   {
       struct amdgpu_device *adev = ring->adev;
-     struct amdgpu_ib ib;
+     struct amdgpu_job *job;
+     struct amdgpu_ib *ib;
       struct dma_fence *f = NULL;
       unsigned index;
       uint64_t gpu_addr;
@@ -616,8 +617,6 @@ static int gfx_v11_0_ring_test_ib(struct amdgpu_ring *ring, 
long timeout)
           ring->funcs->type == AMDGPU_RING_TYPE_KIQ)
               return 0;

-     memset(&ib, 0, sizeof(ib));
-
       r = amdgpu_device_wb_get(adev, &index);
       if (r)
               return r;
@@ -626,22 +625,27 @@ static int gfx_v11_0_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
       adev->wb.wb[index] = cpu_to_le32(0xCAFEDEAD);
       cpu_ptr = &adev->wb.wb[index];

-     r = amdgpu_ib_get(adev, NULL, 20, AMDGPU_IB_POOL_DIRECT, &ib);
+     r = amdgpu_job_alloc_with_ib(ring->adev, NULL, NULL, 20,
+                                  AMDGPU_IB_POOL_DIRECT, &job,
+                                  AMDGPU_KERNEL_JOB_ID_GFX_RING_TEST);
       if (r) {
               drm_err(adev_to_drm(adev), "failed to get ib (%ld).\n", r);
               goto err1;
       }
+     ib = &job->ibs[0];

-     ib.ptr[0] = PACKET3(PACKET3_WRITE_DATA, 3);
-     ib.ptr[1] = WRITE_DATA_DST_SEL(5) | WR_CONFIRM;
-     ib.ptr[2] = lower_32_bits(gpu_addr);
-     ib.ptr[3] = upper_32_bits(gpu_addr);
-     ib.ptr[4] = 0xDEADBEEF;
-     ib.length_dw = 5;
+     ib->ptr[0] = PACKET3(PACKET3_WRITE_DATA, 3);
+     ib->ptr[1] = WRITE_DATA_DST_SEL(5) | WR_CONFIRM;
+     ib->ptr[2] = lower_32_bits(gpu_addr);
+     ib->ptr[3] = upper_32_bits(gpu_addr);
+     ib->ptr[4] = 0xDEADBEEF;
+     ib->length_dw = 5;

-     r = amdgpu_ib_schedule(ring, 1, &ib, NULL, &f);
-     if (r)
+     r = amdgpu_job_submit_direct(job, ring, &f);
+     if (r) {
+             amdgpu_job_free(job);
               goto err2;
+     }

       r = dma_fence_wait_timeout(f, false, timeout);
       if (r == 0) {
@@ -656,7 +660,6 @@ static int gfx_v11_0_ring_test_ib(struct amdgpu_ring *ring, 
long timeout)
       else
               r = -EINVAL;
   err2:
-     amdgpu_ib_free(&ib, NULL);
       dma_fence_put(f);
   err1:
       amdgpu_device_wb_free(adev, index);
diff --git a/drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c 
b/drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
index 79ea1af363a53..e1eb6a3bfb899 100644
--- a/drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
+++ b/drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
@@ -493,7 +493,8 @@ static int gfx_v12_0_ring_test_ring(struct amdgpu_ring 
*ring)
   static int gfx_v12_0_ring_test_ib(struct amdgpu_ring *ring, long timeout)
   {
       struct amdgpu_device *adev = ring->adev;
-     struct amdgpu_ib ib;
+     struct amdgpu_job *job;
+     struct amdgpu_ib *ib;
       struct dma_fence *f = NULL;
       unsigned index;
       uint64_t gpu_addr;
@@ -505,8 +506,6 @@ static int gfx_v12_0_ring_test_ib(struct amdgpu_ring *ring, 
long timeout)
           ring->funcs->type == AMDGPU_RING_TYPE_KIQ)
               return 0;

-     memset(&ib, 0, sizeof(ib));
-
       r = amdgpu_device_wb_get(adev, &index);
       if (r)
               return r;
@@ -515,22 +514,27 @@ static int gfx_v12_0_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
       adev->wb.wb[index] = cpu_to_le32(0xCAFEDEAD);
       cpu_ptr = &adev->wb.wb[index];

-     r = amdgpu_ib_get(adev, NULL, 16, AMDGPU_IB_POOL_DIRECT, &ib);
+     r = amdgpu_job_alloc_with_ib(ring->adev, NULL, NULL, 16,
+                                  AMDGPU_IB_POOL_DIRECT, &job,
+                                  AMDGPU_KERNEL_JOB_ID_GFX_RING_TEST);
       if (r) {
               drm_err(adev_to_drm(adev), "failed to get ib (%ld).\n", r);
               goto err1;
       }
+     ib = &job->ibs[0];

-     ib.ptr[0] = PACKET3(PACKET3_WRITE_DATA, 3);
-     ib.ptr[1] = WRITE_DATA_DST_SEL(5) | WR_CONFIRM;
-     ib.ptr[2] = lower_32_bits(gpu_addr);
-     ib.ptr[3] = upper_32_bits(gpu_addr);
-     ib.ptr[4] = 0xDEADBEEF;
-     ib.length_dw = 5;
+     ib->ptr[0] = PACKET3(PACKET3_WRITE_DATA, 3);
+     ib->ptr[1] = WRITE_DATA_DST_SEL(5) | WR_CONFIRM;
+     ib->ptr[2] = lower_32_bits(gpu_addr);
+     ib->ptr[3] = upper_32_bits(gpu_addr);
+     ib->ptr[4] = 0xDEADBEEF;
+     ib->length_dw = 5;

-     r = amdgpu_ib_schedule(ring, 1, &ib, NULL, &f);
-     if (r)
+     r = amdgpu_job_submit_direct(job, ring, &f);
+     if (r) {
+             amdgpu_job_free(job);
               goto err2;
+     }

       r = dma_fence_wait_timeout(f, false, timeout);
       if (r == 0) {
@@ -545,7 +549,6 @@ static int gfx_v12_0_ring_test_ib(struct amdgpu_ring *ring, 
long timeout)
       else
               r = -EINVAL;
   err2:
-     amdgpu_ib_free(&ib, NULL);
       dma_fence_put(f);
   err1:
       amdgpu_device_wb_free(adev, index);
diff --git a/drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c 
b/drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
index eb9725ae1607a..cc5422a398a40 100644
--- a/drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
+++ b/drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
@@ -307,7 +307,8 @@ static int gfx_v12_1_ring_test_ring(struct amdgpu_ring 
*ring)
   static int gfx_v12_1_ring_test_ib(struct amdgpu_ring *ring, long timeout)
   {
       struct amdgpu_device *adev = ring->adev;
-     struct amdgpu_ib ib;
+     struct amdgpu_job *job;
+     struct amdgpu_ib *ib;
       struct dma_fence *f = NULL;
       unsigned index;
       uint64_t gpu_addr;
@@ -319,8 +320,6 @@ static int gfx_v12_1_ring_test_ib(struct amdgpu_ring *ring, 
long timeout)
           ring->funcs->type == AMDGPU_RING_TYPE_KIQ)
               return 0;

-     memset(&ib, 0, sizeof(ib));
-
       r = amdgpu_device_wb_get(adev, &index);
       if (r)
               return r;
@@ -329,22 +328,27 @@ static int gfx_v12_1_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
       adev->wb.wb[index] = cpu_to_le32(0xCAFEDEAD);
       cpu_ptr = &adev->wb.wb[index];

-     r = amdgpu_ib_get(adev, NULL, 16, AMDGPU_IB_POOL_DIRECT, &ib);
+     r = amdgpu_job_alloc_with_ib(ring->adev, NULL, NULL, 16,
+                                  AMDGPU_IB_POOL_DIRECT, &job,
+                                  AMDGPU_KERNEL_JOB_ID_GFX_RING_TEST);
       if (r) {
               dev_err(adev->dev, "amdgpu: failed to get ib (%ld).\n", r);
               goto err1;
       }
+     ib = &job->ibs[0];

-     ib.ptr[0] = PACKET3(PACKET3_WRITE_DATA, 3);
-     ib.ptr[1] = WRITE_DATA_DST_SEL(5) | WR_CONFIRM;
-     ib.ptr[2] = lower_32_bits(gpu_addr);
-     ib.ptr[3] = upper_32_bits(gpu_addr);
-     ib.ptr[4] = 0xDEADBEEF;
-     ib.length_dw = 5;
+     ib->ptr[0] = PACKET3(PACKET3_WRITE_DATA, 3);
+     ib->ptr[1] = WRITE_DATA_DST_SEL(5) | WR_CONFIRM;
+     ib->ptr[2] = lower_32_bits(gpu_addr);
+     ib->ptr[3] = upper_32_bits(gpu_addr);
+     ib->ptr[4] = 0xDEADBEEF;
+     ib->length_dw = 5;

-     r = amdgpu_ib_schedule(ring, 1, &ib, NULL, &f);
-     if (r)
+     r = amdgpu_job_submit_direct(job, ring, &f);
+     if (r) {
+             amdgpu_job_free(job);
               goto err2;
+     }

       r = dma_fence_wait_timeout(f, false, timeout);
       if (r == 0) {
@@ -359,7 +363,6 @@ static int gfx_v12_1_ring_test_ib(struct amdgpu_ring *ring, 
long timeout)
       else
               r = -EINVAL;
   err2:
-     amdgpu_ib_free(&ib, NULL);
       dma_fence_put(f);
   err1:
       amdgpu_device_wb_free(adev, index);
diff --git a/drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c 
b/drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c
index 73223d97a87f5..2f8aa99f17480 100644
--- a/drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c
+++ b/drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c
@@ -1895,24 +1895,29 @@ static int gfx_v6_0_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
   {
       struct amdgpu_device *adev = ring->adev;
       struct dma_fence *f = NULL;
-     struct amdgpu_ib ib;
+     struct amdgpu_job *job;
+     struct amdgpu_ib *ib;
       uint32_t tmp = 0;
       long r;

       WREG32(mmSCRATCH_REG0, 0xCAFEDEAD);
-     memset(&ib, 0, sizeof(ib));
-     r = amdgpu_ib_get(adev, NULL, 256, AMDGPU_IB_POOL_DIRECT, &ib);
+     r = amdgpu_job_alloc_with_ib(ring->adev, NULL, NULL, 256,
+                                  AMDGPU_IB_POOL_DIRECT, &job,
+                                  AMDGPU_KERNEL_JOB_ID_GFX_RING_TEST);
       if (r)
               return r;

-     ib.ptr[0] = PACKET3(PACKET3_SET_CONFIG_REG, 1);
-     ib.ptr[1] = mmSCRATCH_REG0 - PACKET3_SET_CONFIG_REG_START;
-     ib.ptr[2] = 0xDEADBEEF;
-     ib.length_dw = 3;
+     ib = &job->ibs[0];
+     ib->ptr[0] = PACKET3(PACKET3_SET_CONFIG_REG, 1);
+     ib->ptr[1] = mmSCRATCH_REG0 - PACKET3_SET_CONFIG_REG_START;
+     ib->ptr[2] = 0xDEADBEEF;
+     ib->length_dw = 3;

-     r = amdgpu_ib_schedule(ring, 1, &ib, NULL, &f);
-     if (r)
+     r = amdgpu_job_submit_direct(job, ring, &f);
+     if (r) {
+             amdgpu_job_free(job);
               goto error;
+     }

       r = dma_fence_wait_timeout(f, false, timeout);
       if (r == 0) {
@@ -1928,7 +1933,6 @@ static int gfx_v6_0_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
               r = -EINVAL;

   error:
-     amdgpu_ib_free(&ib, NULL);
       dma_fence_put(f);
       return r;
   }
diff --git a/drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c 
b/drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c
index 2b691452775bc..fa235b981c2e9 100644
--- a/drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c
+++ b/drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c
@@ -2291,25 +2291,31 @@ static void gfx_v7_ring_emit_cntxcntl(struct 
amdgpu_ring *ring, uint32_t flags)
   static int gfx_v7_0_ring_test_ib(struct amdgpu_ring *ring, long timeout)
   {
       struct amdgpu_device *adev = ring->adev;
-     struct amdgpu_ib ib;
+     struct amdgpu_job *job;
+     struct amdgpu_ib *ib;
       struct dma_fence *f = NULL;
       uint32_t tmp = 0;
       long r;

       WREG32(mmSCRATCH_REG0, 0xCAFEDEAD);
-     memset(&ib, 0, sizeof(ib));
-     r = amdgpu_ib_get(adev, NULL, 256, AMDGPU_IB_POOL_DIRECT, &ib);
+
+     r = amdgpu_job_alloc_with_ib(ring->adev, NULL, NULL, 256,
+                                  AMDGPU_IB_POOL_DIRECT, &job,
+                                  AMDGPU_KERNEL_JOB_ID_GFX_RING_TEST);
       if (r)
               return r;

-     ib.ptr[0] = PACKET3(PACKET3_SET_UCONFIG_REG, 1);
-     ib.ptr[1] = mmSCRATCH_REG0 - PACKET3_SET_UCONFIG_REG_START;
-     ib.ptr[2] = 0xDEADBEEF;
-     ib.length_dw = 3;
+     ib = &job->ibs[0];
+     ib->ptr[0] = PACKET3(PACKET3_SET_UCONFIG_REG, 1);
+     ib->ptr[1] = mmSCRATCH_REG0 - PACKET3_SET_UCONFIG_REG_START;
+     ib->ptr[2] = 0xDEADBEEF;
+     ib->length_dw = 3;

-     r = amdgpu_ib_schedule(ring, 1, &ib, NULL, &f);
-     if (r)
+     r = amdgpu_job_submit_direct(job, ring, &f);
+     if (r) {
+             amdgpu_job_free(job);
               goto error;
+     }

       r = dma_fence_wait_timeout(f, false, timeout);
       if (r == 0) {
@@ -2325,7 +2331,6 @@ static int gfx_v7_0_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
               r = -EINVAL;

   error:
-     amdgpu_ib_free(&ib, NULL);
       dma_fence_put(f);
       return r;
   }
diff --git a/drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c 
b/drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c
index a6b4c8f41dc11..4736216cd0211 100644
--- a/drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c
+++ b/drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c
@@ -868,9 +868,9 @@ static int gfx_v8_0_ring_test_ring(struct amdgpu_ring *ring)
   static int gfx_v8_0_ring_test_ib(struct amdgpu_ring *ring, long timeout)
   {
       struct amdgpu_device *adev = ring->adev;
-     struct amdgpu_ib ib;
+     struct amdgpu_job *job;
+     struct amdgpu_ib *ib;
       struct dma_fence *f = NULL;
-
       unsigned int index;
       uint64_t gpu_addr;
       uint32_t tmp;
@@ -882,22 +882,26 @@ static int gfx_v8_0_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)

       gpu_addr = adev->wb.gpu_addr + (index * 4);
       adev->wb.wb[index] = cpu_to_le32(0xCAFEDEAD);
-     memset(&ib, 0, sizeof(ib));

-     r = amdgpu_ib_get(adev, NULL, 20, AMDGPU_IB_POOL_DIRECT, &ib);
+     r = amdgpu_job_alloc_with_ib(ring->adev, NULL, NULL, 20,
+                                  AMDGPU_IB_POOL_DIRECT, &job,
+                                  AMDGPU_KERNEL_JOB_ID_GFX_RING_TEST);
       if (r)
               goto err1;

-     ib.ptr[0] = PACKET3(PACKET3_WRITE_DATA, 3);
-     ib.ptr[1] = WRITE_DATA_DST_SEL(5) | WR_CONFIRM;
-     ib.ptr[2] = lower_32_bits(gpu_addr);
-     ib.ptr[3] = upper_32_bits(gpu_addr);
-     ib.ptr[4] = 0xDEADBEEF;
-     ib.length_dw = 5;
+     ib = &job->ibs[0];
+     ib->ptr[0] = PACKET3(PACKET3_WRITE_DATA, 3);
+     ib->ptr[1] = WRITE_DATA_DST_SEL(5) | WR_CONFIRM;
+     ib->ptr[2] = lower_32_bits(gpu_addr);
+     ib->ptr[3] = upper_32_bits(gpu_addr);
+     ib->ptr[4] = 0xDEADBEEF;
+     ib->length_dw = 5;

-     r = amdgpu_ib_schedule(ring, 1, &ib, NULL, &f);
-     if (r)
+     r = amdgpu_job_submit_direct(job, ring, &f);
+     if (r) {
+             amdgpu_job_free(job);
               goto err2;
+     }

       r = dma_fence_wait_timeout(f, false, timeout);
       if (r == 0) {
@@ -914,7 +918,6 @@ static int gfx_v8_0_ring_test_ib(struct amdgpu_ring *ring, 
long timeout)
               r = -EINVAL;

   err2:
-     amdgpu_ib_free(&ib, NULL);
       dma_fence_put(f);
   err1:
       amdgpu_device_wb_free(adev, index);
@@ -1474,7 +1477,8 @@ static const u32 sec_ded_counter_registers[] =
   static int gfx_v8_0_do_edc_gpr_workarounds(struct amdgpu_device *adev)
   {
       struct amdgpu_ring *ring = &adev->gfx.compute_ring[0];
-     struct amdgpu_ib ib;
+     struct amdgpu_job *job;
+     struct amdgpu_ib *ib;
       struct dma_fence *f = NULL;
       int r, i;
       u32 tmp;
@@ -1505,106 +1509,108 @@ static int gfx_v8_0_do_edc_gpr_workarounds(struct 
amdgpu_device *adev)
       total_size += sizeof(sgpr_init_compute_shader);

       /* allocate an indirect buffer to put the commands in */
-     memset(&ib, 0, sizeof(ib));
-     r = amdgpu_ib_get(adev, NULL, total_size,
-                                     AMDGPU_IB_POOL_DIRECT, &ib);
+     r = amdgpu_job_alloc_with_ib(ring->adev, NULL, NULL, total_size,
+                                  AMDGPU_IB_POOL_DIRECT, &job,
+                                  AMDGPU_KERNEL_JOB_ID_RUN_SHADER);
       if (r) {
               drm_err(adev_to_drm(adev), "failed to get ib (%d).\n", r);
               return r;
       }
+     ib = &job->ibs[0];

       /* load the compute shaders */
       for (i = 0; i < ARRAY_SIZE(vgpr_init_compute_shader); i++)
-             ib.ptr[i + (vgpr_offset / 4)] = vgpr_init_compute_shader[i];
+             ib->ptr[i + (vgpr_offset / 4)] = vgpr_init_compute_shader[i];

       for (i = 0; i < ARRAY_SIZE(sgpr_init_compute_shader); i++)
-             ib.ptr[i + (sgpr_offset / 4)] = sgpr_init_compute_shader[i];
+             ib->ptr[i + (sgpr_offset / 4)] = sgpr_init_compute_shader[i];

       /* init the ib length to 0 */
-     ib.length_dw = 0;
+     ib->length_dw = 0;

       /* VGPR */
       /* write the register state for the compute dispatch */
       for (i = 0; i < ARRAY_SIZE(vgpr_init_regs); i += 2) {
-             ib.ptr[ib.length_dw++] = PACKET3(PACKET3_SET_SH_REG, 1);
-             ib.ptr[ib.length_dw++] = vgpr_init_regs[i] - 
PACKET3_SET_SH_REG_START;
-             ib.ptr[ib.length_dw++] = vgpr_init_regs[i + 1];
+             ib->ptr[ib->length_dw++] = PACKET3(PACKET3_SET_SH_REG, 1);
+             ib->ptr[ib->length_dw++] = vgpr_init_regs[i] - 
PACKET3_SET_SH_REG_START;
+             ib->ptr[ib->length_dw++] = vgpr_init_regs[i + 1];
       }
       /* write the shader start address: mmCOMPUTE_PGM_LO, mmCOMPUTE_PGM_HI */
-     gpu_addr = (ib.gpu_addr + (u64)vgpr_offset) >> 8;
-     ib.ptr[ib.length_dw++] = PACKET3(PACKET3_SET_SH_REG, 2);
-     ib.ptr[ib.length_dw++] = mmCOMPUTE_PGM_LO - PACKET3_SET_SH_REG_START;
-     ib.ptr[ib.length_dw++] = lower_32_bits(gpu_addr);
-     ib.ptr[ib.length_dw++] = upper_32_bits(gpu_addr);
+     gpu_addr = (ib->gpu_addr + (u64)vgpr_offset) >> 8;
+     ib->ptr[ib->length_dw++] = PACKET3(PACKET3_SET_SH_REG, 2);
+     ib->ptr[ib->length_dw++] = mmCOMPUTE_PGM_LO - PACKET3_SET_SH_REG_START;
+     ib->ptr[ib->length_dw++] = lower_32_bits(gpu_addr);
+     ib->ptr[ib->length_dw++] = upper_32_bits(gpu_addr);

       /* write dispatch packet */
-     ib.ptr[ib.length_dw++] = PACKET3(PACKET3_DISPATCH_DIRECT, 3);
-     ib.ptr[ib.length_dw++] = 8; /* x */
-     ib.ptr[ib.length_dw++] = 1; /* y */
-     ib.ptr[ib.length_dw++] = 1; /* z */
-     ib.ptr[ib.length_dw++] =
+     ib->ptr[ib->length_dw++] = PACKET3(PACKET3_DISPATCH_DIRECT, 3);
+     ib->ptr[ib->length_dw++] = 8; /* x */
+     ib->ptr[ib->length_dw++] = 1; /* y */
+     ib->ptr[ib->length_dw++] = 1; /* z */
+     ib->ptr[ib->length_dw++] =
               REG_SET_FIELD(0, COMPUTE_DISPATCH_INITIATOR, COMPUTE_SHADER_EN, 
1);

       /* write CS partial flush packet */
-     ib.ptr[ib.length_dw++] = PACKET3(PACKET3_EVENT_WRITE, 0);
-     ib.ptr[ib.length_dw++] = EVENT_TYPE(7) | EVENT_INDEX(4);
+     ib->ptr[ib->length_dw++] = PACKET3(PACKET3_EVENT_WRITE, 0);
+     ib->ptr[ib->length_dw++] = EVENT_TYPE(7) | EVENT_INDEX(4);

       /* SGPR1 */
       /* write the register state for the compute dispatch */
       for (i = 0; i < ARRAY_SIZE(sgpr1_init_regs); i += 2) {
-             ib.ptr[ib.length_dw++] = PACKET3(PACKET3_SET_SH_REG, 1);
-             ib.ptr[ib.length_dw++] = sgpr1_init_regs[i] - 
PACKET3_SET_SH_REG_START;
-             ib.ptr[ib.length_dw++] = sgpr1_init_regs[i + 1];
+             ib->ptr[ib->length_dw++] = PACKET3(PACKET3_SET_SH_REG, 1);
+             ib->ptr[ib->length_dw++] = sgpr1_init_regs[i] - 
PACKET3_SET_SH_REG_START;
+             ib->ptr[ib->length_dw++] = sgpr1_init_regs[i + 1];
       }
       /* write the shader start address: mmCOMPUTE_PGM_LO, mmCOMPUTE_PGM_HI */
-     gpu_addr = (ib.gpu_addr + (u64)sgpr_offset) >> 8;
-     ib.ptr[ib.length_dw++] = PACKET3(PACKET3_SET_SH_REG, 2);
-     ib.ptr[ib.length_dw++] = mmCOMPUTE_PGM_LO - PACKET3_SET_SH_REG_START;
-     ib.ptr[ib.length_dw++] = lower_32_bits(gpu_addr);
-     ib.ptr[ib.length_dw++] = upper_32_bits(gpu_addr);
+     gpu_addr = (ib->gpu_addr + (u64)sgpr_offset) >> 8;
+     ib->ptr[ib->length_dw++] = PACKET3(PACKET3_SET_SH_REG, 2);
+     ib->ptr[ib->length_dw++] = mmCOMPUTE_PGM_LO - PACKET3_SET_SH_REG_START;
+     ib->ptr[ib->length_dw++] = lower_32_bits(gpu_addr);
+     ib->ptr[ib->length_dw++] = upper_32_bits(gpu_addr);

       /* write dispatch packet */
-     ib.ptr[ib.length_dw++] = PACKET3(PACKET3_DISPATCH_DIRECT, 3);
-     ib.ptr[ib.length_dw++] = 8; /* x */
-     ib.ptr[ib.length_dw++] = 1; /* y */
-     ib.ptr[ib.length_dw++] = 1; /* z */
-     ib.ptr[ib.length_dw++] =
+     ib->ptr[ib->length_dw++] = PACKET3(PACKET3_DISPATCH_DIRECT, 3);
+     ib->ptr[ib->length_dw++] = 8; /* x */
+     ib->ptr[ib->length_dw++] = 1; /* y */
+     ib->ptr[ib->length_dw++] = 1; /* z */
+     ib->ptr[ib->length_dw++] =
               REG_SET_FIELD(0, COMPUTE_DISPATCH_INITIATOR, COMPUTE_SHADER_EN, 
1);

       /* write CS partial flush packet */
-     ib.ptr[ib.length_dw++] = PACKET3(PACKET3_EVENT_WRITE, 0);
-     ib.ptr[ib.length_dw++] = EVENT_TYPE(7) | EVENT_INDEX(4);
+     ib->ptr[ib->length_dw++] = PACKET3(PACKET3_EVENT_WRITE, 0);
+     ib->ptr[ib->length_dw++] = EVENT_TYPE(7) | EVENT_INDEX(4);

       /* SGPR2 */
       /* write the register state for the compute dispatch */
       for (i = 0; i < ARRAY_SIZE(sgpr2_init_regs); i += 2) {
-             ib.ptr[ib.length_dw++] = PACKET3(PACKET3_SET_SH_REG, 1);
-             ib.ptr[ib.length_dw++] = sgpr2_init_regs[i] - 
PACKET3_SET_SH_REG_START;
-             ib.ptr[ib.length_dw++] = sgpr2_init_regs[i + 1];
+             ib->ptr[ib->length_dw++] = PACKET3(PACKET3_SET_SH_REG, 1);
+             ib->ptr[ib->length_dw++] = sgpr2_init_regs[i] - 
PACKET3_SET_SH_REG_START;
+             ib->ptr[ib->length_dw++] = sgpr2_init_regs[i + 1];
       }
       /* write the shader start address: mmCOMPUTE_PGM_LO, mmCOMPUTE_PGM_HI */
-     gpu_addr = (ib.gpu_addr + (u64)sgpr_offset) >> 8;
-     ib.ptr[ib.length_dw++] = PACKET3(PACKET3_SET_SH_REG, 2);
-     ib.ptr[ib.length_dw++] = mmCOMPUTE_PGM_LO - PACKET3_SET_SH_REG_START;
-     ib.ptr[ib.length_dw++] = lower_32_bits(gpu_addr);
-     ib.ptr[ib.length_dw++] = upper_32_bits(gpu_addr);
+     gpu_addr = (ib->gpu_addr + (u64)sgpr_offset) >> 8;
+     ib->ptr[ib->length_dw++] = PACKET3(PACKET3_SET_SH_REG, 2);
+     ib->ptr[ib->length_dw++] = mmCOMPUTE_PGM_LO - PACKET3_SET_SH_REG_START;
+     ib->ptr[ib->length_dw++] = lower_32_bits(gpu_addr);
+     ib->ptr[ib->length_dw++] = upper_32_bits(gpu_addr);

       /* write dispatch packet */
-     ib.ptr[ib.length_dw++] = PACKET3(PACKET3_DISPATCH_DIRECT, 3);
-     ib.ptr[ib.length_dw++] = 8; /* x */
-     ib.ptr[ib.length_dw++] = 1; /* y */
-     ib.ptr[ib.length_dw++] = 1; /* z */
-     ib.ptr[ib.length_dw++] =
+     ib->ptr[ib->length_dw++] = PACKET3(PACKET3_DISPATCH_DIRECT, 3);
+     ib->ptr[ib->length_dw++] = 8; /* x */
+     ib->ptr[ib->length_dw++] = 1; /* y */
+     ib->ptr[ib->length_dw++] = 1; /* z */
+     ib->ptr[ib->length_dw++] =
               REG_SET_FIELD(0, COMPUTE_DISPATCH_INITIATOR, COMPUTE_SHADER_EN, 
1);

       /* write CS partial flush packet */
-     ib.ptr[ib.length_dw++] = PACKET3(PACKET3_EVENT_WRITE, 0);
-     ib.ptr[ib.length_dw++] = EVENT_TYPE(7) | EVENT_INDEX(4);
+     ib->ptr[ib->length_dw++] = PACKET3(PACKET3_EVENT_WRITE, 0);
+     ib->ptr[ib->length_dw++] = EVENT_TYPE(7) | EVENT_INDEX(4);

       /* shedule the ib on the ring */
-     r = amdgpu_ib_schedule(ring, 1, &ib, NULL, &f);
+     r = amdgpu_job_submit_direct(job, ring, &f);
       if (r) {
               drm_err(adev_to_drm(adev), "ib submit failed (%d).\n", r);
+             amdgpu_job_free(job);
               goto fail;
       }

@@ -1629,7 +1635,6 @@ static int gfx_v8_0_do_edc_gpr_workarounds(struct 
amdgpu_device *adev)
               RREG32(sec_ded_counter_registers[i]);

   fail:
-     amdgpu_ib_free(&ib, NULL);
       dma_fence_put(f);

       return r;
diff --git a/drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c 
b/drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
index 7e9d753f4a808..36f0300a21bfa 100644
--- a/drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
+++ b/drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
@@ -1224,9 +1224,9 @@ static int gfx_v9_0_ring_test_ring(struct amdgpu_ring 
*ring)
   static int gfx_v9_0_ring_test_ib(struct amdgpu_ring *ring, long timeout)
   {
       struct amdgpu_device *adev = ring->adev;
-     struct amdgpu_ib ib;
+     struct amdgpu_job *job;
+     struct amdgpu_ib *ib;
       struct dma_fence *f = NULL;
-
       unsigned index;
       uint64_t gpu_addr;
       uint32_t tmp;
@@ -1238,22 +1238,26 @@ static int gfx_v9_0_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)

       gpu_addr = adev->wb.gpu_addr + (index * 4);
       adev->wb.wb[index] = cpu_to_le32(0xCAFEDEAD);
-     memset(&ib, 0, sizeof(ib));

-     r = amdgpu_ib_get(adev, NULL, 20, AMDGPU_IB_POOL_DIRECT, &ib);
+     r = amdgpu_job_alloc_with_ib(ring->adev, NULL, NULL, 20,
+                                  AMDGPU_IB_POOL_DIRECT, &job,
+                                  AMDGPU_KERNEL_JOB_ID_GFX_RING_TEST);
       if (r)
               goto err1;

-     ib.ptr[0] = PACKET3(PACKET3_WRITE_DATA, 3);
-     ib.ptr[1] = WRITE_DATA_DST_SEL(5) | WR_CONFIRM;
-     ib.ptr[2] = lower_32_bits(gpu_addr);
-     ib.ptr[3] = upper_32_bits(gpu_addr);
-     ib.ptr[4] = 0xDEADBEEF;
-     ib.length_dw = 5;
+     ib = &job->ibs[0];
+     ib->ptr[0] = PACKET3(PACKET3_WRITE_DATA, 3);
+     ib->ptr[1] = WRITE_DATA_DST_SEL(5) | WR_CONFIRM;
+     ib->ptr[2] = lower_32_bits(gpu_addr);
+     ib->ptr[3] = upper_32_bits(gpu_addr);
+     ib->ptr[4] = 0xDEADBEEF;
+     ib->length_dw = 5;

-     r = amdgpu_ib_schedule(ring, 1, &ib, NULL, &f);
-     if (r)
+     r = amdgpu_job_submit_direct(job, ring, &f);
+     if (r) {
+             amdgpu_job_free(job);
               goto err2;
+     }

       r = dma_fence_wait_timeout(f, false, timeout);
       if (r == 0) {
@@ -1270,7 +1274,6 @@ static int gfx_v9_0_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
               r = -EINVAL;

   err2:
-     amdgpu_ib_free(&ib, NULL);
       dma_fence_put(f);
   err1:
       amdgpu_device_wb_free(adev, index);
@@ -4624,7 +4627,8 @@ static int gfx_v9_0_do_edc_gds_workarounds(struct 
amdgpu_device *adev)
   static int gfx_v9_0_do_edc_gpr_workarounds(struct amdgpu_device *adev)
   {
       struct amdgpu_ring *ring = &adev->gfx.compute_ring[0];
-     struct amdgpu_ib ib;
+     struct amdgpu_job *job;
+     struct amdgpu_ib *ib;
       struct dma_fence *f = NULL;
       int r, i;
       unsigned total_size, vgpr_offset, sgpr_offset;
@@ -4670,9 +4674,9 @@ static int gfx_v9_0_do_edc_gpr_workarounds(struct 
amdgpu_device *adev)
       total_size += sizeof(sgpr_init_compute_shader);

       /* allocate an indirect buffer to put the commands in */
-     memset(&ib, 0, sizeof(ib));
-     r = amdgpu_ib_get(adev, NULL, total_size,
-                                     AMDGPU_IB_POOL_DIRECT, &ib);
+     r = amdgpu_job_alloc_with_ib(ring->adev, NULL, NULL, total_size,
+                                  AMDGPU_IB_POOL_DIRECT, &job,
+                                  AMDGPU_KERNEL_JOB_ID_RUN_SHADER);
       if (r) {
               drm_err(adev_to_drm(adev), "failed to get ib (%d).\n", r);
               return r;
@@ -4680,102 +4684,103 @@ static int gfx_v9_0_do_edc_gpr_workarounds(struct 
amdgpu_device *adev)

       /* load the compute shaders */
       for (i = 0; i < vgpr_init_shader_size/sizeof(u32); i++)
-             ib.ptr[i + (vgpr_offset / 4)] = vgpr_init_shader_ptr[i];
+             ib->ptr[i + (vgpr_offset / 4)] = vgpr_init_shader_ptr[i];

       for (i = 0; i < ARRAY_SIZE(sgpr_init_compute_shader); i++)
-             ib.ptr[i + (sgpr_offset / 4)] = sgpr_init_compute_shader[i];
+             ib->ptr[i + (sgpr_offset / 4)] = sgpr_init_compute_shader[i];

       /* init the ib length to 0 */
-     ib.length_dw = 0;
+     ib->length_dw = 0;

       /* VGPR */
       /* write the register state for the compute dispatch */
       for (i = 0; i < gpr_reg_size; i++) {
-             ib.ptr[ib.length_dw++] = PACKET3(PACKET3_SET_SH_REG, 1);
-             ib.ptr[ib.length_dw++] = 
SOC15_REG_ENTRY_OFFSET(vgpr_init_regs_ptr[i])
+             ib->ptr[ib->length_dw++] = PACKET3(PACKET3_SET_SH_REG, 1);
+             ib->ptr[ib->length_dw++] = 
SOC15_REG_ENTRY_OFFSET(vgpr_init_regs_ptr[i])
                                                               - 
PACKET3_SET_SH_REG_START;
-             ib.ptr[ib.length_dw++] = vgpr_init_regs_ptr[i].reg_value;
+             ib->ptr[ib->length_dw++] = vgpr_init_regs_ptr[i].reg_value;
       }
       /* write the shader start address: mmCOMPUTE_PGM_LO, mmCOMPUTE_PGM_HI */
-     gpu_addr = (ib.gpu_addr + (u64)vgpr_offset) >> 8;
-     ib.ptr[ib.length_dw++] = PACKET3(PACKET3_SET_SH_REG, 2);
-     ib.ptr[ib.length_dw++] = SOC15_REG_OFFSET(GC, 0, mmCOMPUTE_PGM_LO)
+     gpu_addr = (ib->gpu_addr + (u64)vgpr_offset) >> 8;
+     ib->ptr[ib->length_dw++] = PACKET3(PACKET3_SET_SH_REG, 2);
+     ib->ptr[ib->length_dw++] = SOC15_REG_OFFSET(GC, 0, mmCOMPUTE_PGM_LO)
                                                       - 
PACKET3_SET_SH_REG_START;
-     ib.ptr[ib.length_dw++] = lower_32_bits(gpu_addr);
-     ib.ptr[ib.length_dw++] = upper_32_bits(gpu_addr);
+     ib->ptr[ib->length_dw++] = lower_32_bits(gpu_addr);
+     ib->ptr[ib->length_dw++] = upper_32_bits(gpu_addr);

       /* write dispatch packet */
-     ib.ptr[ib.length_dw++] = PACKET3(PACKET3_DISPATCH_DIRECT, 3);
-     ib.ptr[ib.length_dw++] = compute_dim_x * 2; /* x */
-     ib.ptr[ib.length_dw++] = 1; /* y */
-     ib.ptr[ib.length_dw++] = 1; /* z */
-     ib.ptr[ib.length_dw++] =
+     ib->ptr[ib->length_dw++] = PACKET3(PACKET3_DISPATCH_DIRECT, 3);
+     ib->ptr[ib->length_dw++] = compute_dim_x * 2; /* x */
+     ib->ptr[ib->length_dw++] = 1; /* y */
+     ib->ptr[ib->length_dw++] = 1; /* z */
+     ib->ptr[ib->length_dw++] =
               REG_SET_FIELD(0, COMPUTE_DISPATCH_INITIATOR, COMPUTE_SHADER_EN, 
1);

       /* write CS partial flush packet */
-     ib.ptr[ib.length_dw++] = PACKET3(PACKET3_EVENT_WRITE, 0);
-     ib.ptr[ib.length_dw++] = EVENT_TYPE(7) | EVENT_INDEX(4);
+     ib->ptr[ib->length_dw++] = PACKET3(PACKET3_EVENT_WRITE, 0);
+     ib->ptr[ib->length_dw++] = EVENT_TYPE(7) | EVENT_INDEX(4);

       /* SGPR1 */
       /* write the register state for the compute dispatch */
       for (i = 0; i < gpr_reg_size; i++) {
-             ib.ptr[ib.length_dw++] = PACKET3(PACKET3_SET_SH_REG, 1);
-             ib.ptr[ib.length_dw++] = 
SOC15_REG_ENTRY_OFFSET(sgpr1_init_regs[i])
+             ib->ptr[ib->length_dw++] = PACKET3(PACKET3_SET_SH_REG, 1);
+             ib->ptr[ib->length_dw++] = 
SOC15_REG_ENTRY_OFFSET(sgpr1_init_regs[i])
                                                               - 
PACKET3_SET_SH_REG_START;
-             ib.ptr[ib.length_dw++] = sgpr1_init_regs[i].reg_value;
+             ib->ptr[ib->length_dw++] = sgpr1_init_regs[i].reg_value;
       }
       /* write the shader start address: mmCOMPUTE_PGM_LO, mmCOMPUTE_PGM_HI */
-     gpu_addr = (ib.gpu_addr + (u64)sgpr_offset) >> 8;
-     ib.ptr[ib.length_dw++] = PACKET3(PACKET3_SET_SH_REG, 2);
-     ib.ptr[ib.length_dw++] = SOC15_REG_OFFSET(GC, 0, mmCOMPUTE_PGM_LO)
+     gpu_addr = (ib->gpu_addr + (u64)sgpr_offset) >> 8;
+     ib->ptr[ib->length_dw++] = PACKET3(PACKET3_SET_SH_REG, 2);
+     ib->ptr[ib->length_dw++] = SOC15_REG_OFFSET(GC, 0, mmCOMPUTE_PGM_LO)
                                                       - 
PACKET3_SET_SH_REG_START;
-     ib.ptr[ib.length_dw++] = lower_32_bits(gpu_addr);
-     ib.ptr[ib.length_dw++] = upper_32_bits(gpu_addr);
+     ib->ptr[ib->length_dw++] = lower_32_bits(gpu_addr);
+     ib->ptr[ib->length_dw++] = upper_32_bits(gpu_addr);

       /* write dispatch packet */
-     ib.ptr[ib.length_dw++] = PACKET3(PACKET3_DISPATCH_DIRECT, 3);
-     ib.ptr[ib.length_dw++] = compute_dim_x / 2 * sgpr_work_group_size; /* x */
-     ib.ptr[ib.length_dw++] = 1; /* y */
-     ib.ptr[ib.length_dw++] = 1; /* z */
-     ib.ptr[ib.length_dw++] =
+     ib->ptr[ib->length_dw++] = PACKET3(PACKET3_DISPATCH_DIRECT, 3);
+     ib->ptr[ib->length_dw++] = compute_dim_x / 2 * sgpr_work_group_size; /* x 
*/
+     ib->ptr[ib->length_dw++] = 1; /* y */
+     ib->ptr[ib->length_dw++] = 1; /* z */
+     ib->ptr[ib->length_dw++] =
               REG_SET_FIELD(0, COMPUTE_DISPATCH_INITIATOR, COMPUTE_SHADER_EN, 
1);

       /* write CS partial flush packet */
-     ib.ptr[ib.length_dw++] = PACKET3(PACKET3_EVENT_WRITE, 0);
-     ib.ptr[ib.length_dw++] = EVENT_TYPE(7) | EVENT_INDEX(4);
+     ib->ptr[ib->length_dw++] = PACKET3(PACKET3_EVENT_WRITE, 0);
+     ib->ptr[ib->length_dw++] = EVENT_TYPE(7) | EVENT_INDEX(4);

       /* SGPR2 */
       /* write the register state for the compute dispatch */
       for (i = 0; i < gpr_reg_size; i++) {
-             ib.ptr[ib.length_dw++] = PACKET3(PACKET3_SET_SH_REG, 1);
-             ib.ptr[ib.length_dw++] = 
SOC15_REG_ENTRY_OFFSET(sgpr2_init_regs[i])
+             ib->ptr[ib->length_dw++] = PACKET3(PACKET3_SET_SH_REG, 1);
+             ib->ptr[ib->length_dw++] = 
SOC15_REG_ENTRY_OFFSET(sgpr2_init_regs[i])
                                                               - 
PACKET3_SET_SH_REG_START;
-             ib.ptr[ib.length_dw++] = sgpr2_init_regs[i].reg_value;
+             ib->ptr[ib->length_dw++] = sgpr2_init_regs[i].reg_value;
       }
       /* write the shader start address: mmCOMPUTE_PGM_LO, mmCOMPUTE_PGM_HI */
-     gpu_addr = (ib.gpu_addr + (u64)sgpr_offset) >> 8;
-     ib.ptr[ib.length_dw++] = PACKET3(PACKET3_SET_SH_REG, 2);
-     ib.ptr[ib.length_dw++] = SOC15_REG_OFFSET(GC, 0, mmCOMPUTE_PGM_LO)
+     gpu_addr = (ib->gpu_addr + (u64)sgpr_offset) >> 8;
+     ib->ptr[ib->length_dw++] = PACKET3(PACKET3_SET_SH_REG, 2);
+     ib->ptr[ib->length_dw++] = SOC15_REG_OFFSET(GC, 0, mmCOMPUTE_PGM_LO)
                                                       - 
PACKET3_SET_SH_REG_START;
-     ib.ptr[ib.length_dw++] = lower_32_bits(gpu_addr);
-     ib.ptr[ib.length_dw++] = upper_32_bits(gpu_addr);
+     ib->ptr[ib->length_dw++] = lower_32_bits(gpu_addr);
+     ib->ptr[ib->length_dw++] = upper_32_bits(gpu_addr);

       /* write dispatch packet */
-     ib.ptr[ib.length_dw++] = PACKET3(PACKET3_DISPATCH_DIRECT, 3);
-     ib.ptr[ib.length_dw++] = compute_dim_x / 2 * sgpr_work_group_size; /* x */
-     ib.ptr[ib.length_dw++] = 1; /* y */
-     ib.ptr[ib.length_dw++] = 1; /* z */
-     ib.ptr[ib.length_dw++] =
+     ib->ptr[ib->length_dw++] = PACKET3(PACKET3_DISPATCH_DIRECT, 3);
+     ib->ptr[ib->length_dw++] = compute_dim_x / 2 * sgpr_work_group_size; /* x 
*/
+     ib->ptr[ib->length_dw++] = 1; /* y */
+     ib->ptr[ib->length_dw++] = 1; /* z */
+     ib->ptr[ib->length_dw++] =
               REG_SET_FIELD(0, COMPUTE_DISPATCH_INITIATOR, COMPUTE_SHADER_EN, 
1);

       /* write CS partial flush packet */
-     ib.ptr[ib.length_dw++] = PACKET3(PACKET3_EVENT_WRITE, 0);
-     ib.ptr[ib.length_dw++] = EVENT_TYPE(7) | EVENT_INDEX(4);
+     ib->ptr[ib->length_dw++] = PACKET3(PACKET3_EVENT_WRITE, 0);
+     ib->ptr[ib->length_dw++] = EVENT_TYPE(7) | EVENT_INDEX(4);

       /* shedule the ib on the ring */
-     r = amdgpu_ib_schedule(ring, 1, &ib, NULL, &f);
+     r = amdgpu_job_submit_direct(job, ring, &f);
       if (r) {
               drm_err(adev_to_drm(adev), "ib schedule failed (%d).\n", r);
+             amdgpu_job_free(job);
               goto fail;
       }

@@ -4787,7 +4792,6 @@ static int gfx_v9_0_do_edc_gpr_workarounds(struct 
amdgpu_device *adev)
       }

   fail:
-     amdgpu_ib_free(&ib, NULL);
       dma_fence_put(f);

       return r;
diff --git a/drivers/gpu/drm/amd/amdgpu/gfx_v9_4_2.c 
b/drivers/gpu/drm/amd/amdgpu/gfx_v9_4_2.c
index 8058ea91ecafd..424b05b84ea74 100644
--- a/drivers/gpu/drm/amd/amdgpu/gfx_v9_4_2.c
+++ b/drivers/gpu/drm/amd/amdgpu/gfx_v9_4_2.c
@@ -345,12 +345,13 @@ const struct soc15_reg_entry sgpr64_init_regs_aldebaran[] 
= {

   static int gfx_v9_4_2_run_shader(struct amdgpu_device *adev,
                                struct amdgpu_ring *ring,
-                              struct amdgpu_ib *ib,
                                const u32 *shader_ptr, u32 shader_size,
                                const struct soc15_reg_entry *init_regs, u32 
regs_size,
                                u32 compute_dim_x, u64 wb_gpu_addr, u32 pattern,
                                struct dma_fence **fence_ptr)
   {
+     struct amdgpu_job *job;
+     struct amdgpu_ib *ib;
       int r, i;
       uint32_t total_size, shader_offset;
       u64 gpu_addr;
@@ -360,10 +361,9 @@ static int gfx_v9_4_2_run_shader(struct amdgpu_device 
*adev,
       shader_offset = total_size;
       total_size += ALIGN(shader_size, 256);

-     /* allocate an indirect buffer to put the commands in */
-     memset(ib, 0, sizeof(*ib));
-     r = amdgpu_ib_get(adev, NULL, total_size,
-                                     AMDGPU_IB_POOL_DIRECT, ib);
+     r = amdgpu_job_alloc_with_ib(ring->adev, NULL, NULL, total_size,
+                                  AMDGPU_IB_POOL_DIRECT, &job,
+                                  AMDGPU_KERNEL_JOB_ID_RUN_SHADER);
       if (r) {
               dev_err(adev->dev, "failed to get ib (%d).\n", r);
               return r;
@@ -408,11 +408,11 @@ static int gfx_v9_4_2_run_shader(struct amdgpu_device 
*adev,
       ib->ptr[ib->length_dw++] =
               REG_SET_FIELD(0, COMPUTE_DISPATCH_INITIATOR, COMPUTE_SHADER_EN, 
1);

-     /* shedule the ib on the ring */
-     r = amdgpu_ib_schedule(ring, 1, ib, NULL, fence_ptr);
+     /* schedule the ib on the ring */
+     r = amdgpu_job_submit_direct(job, ring, fence_ptr);
       if (r) {
               dev_err(adev->dev, "ib submit failed (%d).\n", r);
-             amdgpu_ib_free(ib, NULL);
+             amdgpu_job_free(job);
       }
       return r;
   }
@@ -493,7 +493,6 @@ static int gfx_v9_4_2_do_sgprs_init(struct amdgpu_device 
*adev)
       int wb_size = adev->gfx.config.max_shader_engines *
                        CU_ID_MAX * SIMD_ID_MAX * WAVE_ID_MAX;
       struct amdgpu_ib wb_ib;
-     struct amdgpu_ib disp_ibs[3];
       struct dma_fence *fences[3];
       u32 pattern[3] = { 0x1, 0x5, 0xa };

@@ -514,7 +513,6 @@ static int gfx_v9_4_2_do_sgprs_init(struct amdgpu_device 
*adev)

       r = gfx_v9_4_2_run_shader(adev,
                       &adev->gfx.compute_ring[0],
-                     &disp_ibs[0],
                       sgpr112_init_compute_shader_aldebaran,
                       sizeof(sgpr112_init_compute_shader_aldebaran),
                       sgpr112_init_regs_aldebaran,
@@ -539,7 +537,6 @@ static int gfx_v9_4_2_do_sgprs_init(struct amdgpu_device 
*adev)

       r = gfx_v9_4_2_run_shader(adev,
                       &adev->gfx.compute_ring[1],
-                     &disp_ibs[1],
                       sgpr96_init_compute_shader_aldebaran,
                       sizeof(sgpr96_init_compute_shader_aldebaran),
                       sgpr96_init_regs_aldebaran,
@@ -579,7 +576,6 @@ static int gfx_v9_4_2_do_sgprs_init(struct amdgpu_device 
*adev)
       memset(wb_ib.ptr, 0, (1 + wb_size) * sizeof(uint32_t));
       r = gfx_v9_4_2_run_shader(adev,
                       &adev->gfx.compute_ring[0],
-                     &disp_ibs[2],
                       sgpr64_init_compute_shader_aldebaran,
                       sizeof(sgpr64_init_compute_shader_aldebaran),
                       sgpr64_init_regs_aldebaran,
@@ -611,13 +607,10 @@ static int gfx_v9_4_2_do_sgprs_init(struct amdgpu_device 
*adev)
       }

   disp2_failed:
-     amdgpu_ib_free(&disp_ibs[2], NULL);
       dma_fence_put(fences[2]);
   disp1_failed:
-     amdgpu_ib_free(&disp_ibs[1], NULL);
       dma_fence_put(fences[1]);
   disp0_failed:
-     amdgpu_ib_free(&disp_ibs[0], NULL);
       dma_fence_put(fences[0]);
   pro_end:
       amdgpu_ib_free(&wb_ib, NULL);
@@ -637,7 +630,6 @@ static int gfx_v9_4_2_do_vgprs_init(struct amdgpu_device 
*adev)
       int wb_size = adev->gfx.config.max_shader_engines *
                        CU_ID_MAX * SIMD_ID_MAX * WAVE_ID_MAX;
       struct amdgpu_ib wb_ib;
-     struct amdgpu_ib disp_ib;
       struct dma_fence *fence;
       u32 pattern = 0xa;

@@ -657,7 +649,6 @@ static int gfx_v9_4_2_do_vgprs_init(struct amdgpu_device 
*adev)

       r = gfx_v9_4_2_run_shader(adev,
                       &adev->gfx.compute_ring[0],
-                     &disp_ib,
                       vgpr_init_compute_shader_aldebaran,
                       sizeof(vgpr_init_compute_shader_aldebaran),
                       vgpr_init_regs_aldebaran,
@@ -687,7 +678,6 @@ static int gfx_v9_4_2_do_vgprs_init(struct amdgpu_device 
*adev)
       }

   disp_failed:
-     amdgpu_ib_free(&disp_ib, NULL);
       dma_fence_put(fence);
   pro_end:
       amdgpu_ib_free(&wb_ib, NULL);
diff --git a/drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c 
b/drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
index ad4d442e7345e..d78b2c2ae13a3 100644
--- a/drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
+++ b/drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
@@ -451,9 +451,9 @@ static int gfx_v9_4_3_ring_test_ring(struct amdgpu_ring 
*ring)
   static int gfx_v9_4_3_ring_test_ib(struct amdgpu_ring *ring, long timeout)
   {
       struct amdgpu_device *adev = ring->adev;
-     struct amdgpu_ib ib;
+     struct amdgpu_job *job;
+     struct amdgpu_ib *ib;
       struct dma_fence *f = NULL;
-
       unsigned index;
       uint64_t gpu_addr;
       uint32_t tmp;
@@ -465,22 +465,26 @@ static int gfx_v9_4_3_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)

       gpu_addr = adev->wb.gpu_addr + (index * 4);
       adev->wb.wb[index] = cpu_to_le32(0xCAFEDEAD);
-     memset(&ib, 0, sizeof(ib));

-     r = amdgpu_ib_get(adev, NULL, 20, AMDGPU_IB_POOL_DIRECT, &ib);
+     r = amdgpu_job_alloc_with_ib(ring->adev, NULL, NULL, 20,
+                                  AMDGPU_IB_POOL_DIRECT, &job,
+                                  AMDGPU_KERNEL_JOB_ID_GFX_RING_TEST);
       if (r)
               goto err1;

-     ib.ptr[0] = PACKET3(PACKET3_WRITE_DATA, 3);
-     ib.ptr[1] = WRITE_DATA_DST_SEL(5) | WR_CONFIRM;
-     ib.ptr[2] = lower_32_bits(gpu_addr);
-     ib.ptr[3] = upper_32_bits(gpu_addr);
-     ib.ptr[4] = 0xDEADBEEF;
-     ib.length_dw = 5;
+     ib = &job->ibs[0];
+     ib->ptr[0] = PACKET3(PACKET3_WRITE_DATA, 3);
+     ib->ptr[1] = WRITE_DATA_DST_SEL(5) | WR_CONFIRM;
+     ib->ptr[2] = lower_32_bits(gpu_addr);
+     ib->ptr[3] = upper_32_bits(gpu_addr);
+     ib->ptr[4] = 0xDEADBEEF;
+     ib->length_dw = 5;

-     r = amdgpu_ib_schedule(ring, 1, &ib, NULL, &f);
-     if (r)
+     r = amdgpu_job_submit_direct(job, ring, &f);
+     if (r) {
+             amdgpu_job_free(job);
               goto err2;
+     }

       r = dma_fence_wait_timeout(f, false, timeout);
       if (r == 0) {
@@ -497,7 +501,6 @@ static int gfx_v9_4_3_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
               r = -EINVAL;

   err2:
-     amdgpu_ib_free(&ib, NULL);
       dma_fence_put(f);
   err1:
       amdgpu_device_wb_free(adev, index);
diff --git a/drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 
b/drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c
index 92ce580647cdc..46263d50cc9ef 100644
--- a/drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c
+++ b/drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c
@@ -584,7 +584,8 @@ static int sdma_v2_4_ring_test_ring(struct amdgpu_ring 
*ring)
   static int sdma_v2_4_ring_test_ib(struct amdgpu_ring *ring, long timeout)
   {
       struct amdgpu_device *adev = ring->adev;
-     struct amdgpu_ib ib;
+     struct amdgpu_job *job;
+     struct amdgpu_ib *ib;
       struct dma_fence *f = NULL;
       unsigned index;
       u32 tmp = 0;
@@ -598,26 +599,30 @@ static int sdma_v2_4_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
       gpu_addr = adev->wb.gpu_addr + (index * 4);
       tmp = 0xCAFEDEAD;
       adev->wb.wb[index] = cpu_to_le32(tmp);
-     memset(&ib, 0, sizeof(ib));
-     r = amdgpu_ib_get(adev, NULL, 256,
-                                     AMDGPU_IB_POOL_DIRECT, &ib);
+
+     r = amdgpu_job_alloc_with_ib(ring->adev, NULL, NULL, 256,
+                                  AMDGPU_IB_POOL_DIRECT, &job,
+                                  AMDGPU_KERNEL_JOB_ID_SDMA_RING_TEST);
       if (r)
               goto err0;

-     ib.ptr[0] = SDMA_PKT_HEADER_OP(SDMA_OP_WRITE) |
+     ib = &job->ibs[0];
+     ib->ptr[0] = SDMA_PKT_HEADER_OP(SDMA_OP_WRITE) |
               SDMA_PKT_HEADER_SUB_OP(SDMA_SUBOP_WRITE_LINEAR);
-     ib.ptr[1] = lower_32_bits(gpu_addr);
-     ib.ptr[2] = upper_32_bits(gpu_addr);
-     ib.ptr[3] = SDMA_PKT_WRITE_UNTILED_DW_3_COUNT(1);
-     ib.ptr[4] = 0xDEADBEEF;
-     ib.ptr[5] = SDMA_PKT_HEADER_OP(SDMA_OP_NOP);
-     ib.ptr[6] = SDMA_PKT_HEADER_OP(SDMA_OP_NOP);
-     ib.ptr[7] = SDMA_PKT_HEADER_OP(SDMA_OP_NOP);
-     ib.length_dw = 8;
-
-     r = amdgpu_ib_schedule(ring, 1, &ib, NULL, &f);
-     if (r)
+     ib->ptr[1] = lower_32_bits(gpu_addr);
+     ib->ptr[2] = upper_32_bits(gpu_addr);
+     ib->ptr[3] = SDMA_PKT_WRITE_UNTILED_DW_3_COUNT(1);
+     ib->ptr[4] = 0xDEADBEEF;
+     ib->ptr[5] = SDMA_PKT_HEADER_OP(SDMA_OP_NOP);
+     ib->ptr[6] = SDMA_PKT_HEADER_OP(SDMA_OP_NOP);
+     ib->ptr[7] = SDMA_PKT_HEADER_OP(SDMA_OP_NOP);
+     ib->length_dw = 8;
+
+     r = amdgpu_job_submit_direct(job, ring, &f);
+     if (r) {
+             amdgpu_job_free(job);
               goto err1;
+     }

       r = dma_fence_wait_timeout(f, false, timeout);
       if (r == 0) {
@@ -633,7 +638,6 @@ static int sdma_v2_4_ring_test_ib(struct amdgpu_ring *ring, 
long timeout)
               r = -EINVAL;

   err1:
-     amdgpu_ib_free(&ib, NULL);
       dma_fence_put(f);
   err0:
       amdgpu_device_wb_free(adev, index);
diff --git a/drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 
b/drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c
index 1c076bd1cf73e..f9f05768072ad 100644
--- a/drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c
+++ b/drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c
@@ -858,7 +858,8 @@ static int sdma_v3_0_ring_test_ring(struct amdgpu_ring 
*ring)
   static int sdma_v3_0_ring_test_ib(struct amdgpu_ring *ring, long timeout)
   {
       struct amdgpu_device *adev = ring->adev;
-     struct amdgpu_ib ib;
+     struct amdgpu_job *job;
+     struct amdgpu_ib *ib;
       struct dma_fence *f = NULL;
       unsigned index;
       u32 tmp = 0;
@@ -872,26 +873,30 @@ static int sdma_v3_0_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
       gpu_addr = adev->wb.gpu_addr + (index * 4);
       tmp = 0xCAFEDEAD;
       adev->wb.wb[index] = cpu_to_le32(tmp);
-     memset(&ib, 0, sizeof(ib));
-     r = amdgpu_ib_get(adev, NULL, 256,
-                                     AMDGPU_IB_POOL_DIRECT, &ib);
+
+     r = amdgpu_job_alloc_with_ib(ring->adev, NULL, NULL, 256,
+                                  AMDGPU_IB_POOL_DIRECT, &job,
+                                  AMDGPU_KERNEL_JOB_ID_SDMA_RING_TEST);
       if (r)
               goto err0;

-     ib.ptr[0] = SDMA_PKT_HEADER_OP(SDMA_OP_WRITE) |
+     ib = &job->ibs[0];
+     ib->ptr[0] = SDMA_PKT_HEADER_OP(SDMA_OP_WRITE) |
               SDMA_PKT_HEADER_SUB_OP(SDMA_SUBOP_WRITE_LINEAR);
-     ib.ptr[1] = lower_32_bits(gpu_addr);
-     ib.ptr[2] = upper_32_bits(gpu_addr);
-     ib.ptr[3] = SDMA_PKT_WRITE_UNTILED_DW_3_COUNT(1);
-     ib.ptr[4] = 0xDEADBEEF;
-     ib.ptr[5] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
-     ib.ptr[6] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
-     ib.ptr[7] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
-     ib.length_dw = 8;
-
-     r = amdgpu_ib_schedule(ring, 1, &ib, NULL, &f);
-     if (r)
+     ib->ptr[1] = lower_32_bits(gpu_addr);
+     ib->ptr[2] = upper_32_bits(gpu_addr);
+     ib->ptr[3] = SDMA_PKT_WRITE_UNTILED_DW_3_COUNT(1);
+     ib->ptr[4] = 0xDEADBEEF;
+     ib->ptr[5] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
+     ib->ptr[6] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
+     ib->ptr[7] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
+     ib->length_dw = 8;
+
+     r = amdgpu_job_submit_direct(job, ring, &f);
+     if (r) {
+             amdgpu_job_free(job);
               goto err1;
+     }

       r = dma_fence_wait_timeout(f, false, timeout);
       if (r == 0) {
@@ -906,7 +911,6 @@ static int sdma_v3_0_ring_test_ib(struct amdgpu_ring *ring, 
long timeout)
       else
               r = -EINVAL;
   err1:
-     amdgpu_ib_free(&ib, NULL);
       dma_fence_put(f);
   err0:
       amdgpu_device_wb_free(adev, index);
diff --git a/drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 
b/drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c
index f38004e6064e5..56d2832ccba2d 100644
--- a/drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c
+++ b/drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c
@@ -1516,7 +1516,8 @@ static int sdma_v4_0_ring_test_ring(struct amdgpu_ring 
*ring)
   static int sdma_v4_0_ring_test_ib(struct amdgpu_ring *ring, long timeout)
   {
       struct amdgpu_device *adev = ring->adev;
-     struct amdgpu_ib ib;
+     struct amdgpu_job *job;
+     struct amdgpu_ib *ib;
       struct dma_fence *f = NULL;
       unsigned index;
       long r;
@@ -1530,26 +1531,30 @@ static int sdma_v4_0_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
       gpu_addr = adev->wb.gpu_addr + (index * 4);
       tmp = 0xCAFEDEAD;
       adev->wb.wb[index] = cpu_to_le32(tmp);
-     memset(&ib, 0, sizeof(ib));
-     r = amdgpu_ib_get(adev, NULL, 256,
-                                     AMDGPU_IB_POOL_DIRECT, &ib);
+
+     r = amdgpu_job_alloc_with_ib(ring->adev, NULL, NULL, 256,
+                                  AMDGPU_IB_POOL_DIRECT, &job,
+                                  AMDGPU_KERNEL_JOB_ID_SDMA_RING_TEST);
       if (r)
               goto err0;

-     ib.ptr[0] = SDMA_PKT_HEADER_OP(SDMA_OP_WRITE) |
+     ib = &job->ibs[0];
+     ib->ptr[0] = SDMA_PKT_HEADER_OP(SDMA_OP_WRITE) |
               SDMA_PKT_HEADER_SUB_OP(SDMA_SUBOP_WRITE_LINEAR);
-     ib.ptr[1] = lower_32_bits(gpu_addr);
-     ib.ptr[2] = upper_32_bits(gpu_addr);
-     ib.ptr[3] = SDMA_PKT_WRITE_UNTILED_DW_3_COUNT(0);
-     ib.ptr[4] = 0xDEADBEEF;
-     ib.ptr[5] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
-     ib.ptr[6] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
-     ib.ptr[7] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
-     ib.length_dw = 8;
-
-     r = amdgpu_ib_schedule(ring, 1, &ib, NULL, &f);
-     if (r)
+     ib->ptr[1] = lower_32_bits(gpu_addr);
+     ib->ptr[2] = upper_32_bits(gpu_addr);
+     ib->ptr[3] = SDMA_PKT_WRITE_UNTILED_DW_3_COUNT(0);
+     ib->ptr[4] = 0xDEADBEEF;
+     ib->ptr[5] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
+     ib->ptr[6] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
+     ib->ptr[7] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
+     ib->length_dw = 8;
+
+     r = amdgpu_job_submit_direct(job, ring, &f);
+     if (r) {
+             amdgpu_job_free(job);
               goto err1;
+     }

       r = dma_fence_wait_timeout(f, false, timeout);
       if (r == 0) {
@@ -1565,7 +1570,6 @@ static int sdma_v4_0_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
               r = -EINVAL;

   err1:
-     amdgpu_ib_free(&ib, NULL);
       dma_fence_put(f);
   err0:
       amdgpu_device_wb_free(adev, index);
diff --git a/drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c 
b/drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c
index a1443990d5c60..dd8d6a572710d 100644
--- a/drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c
+++ b/drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c
@@ -1112,7 +1112,8 @@ static int sdma_v4_4_2_ring_test_ring(struct amdgpu_ring 
*ring)
   static int sdma_v4_4_2_ring_test_ib(struct amdgpu_ring *ring, long timeout)
   {
       struct amdgpu_device *adev = ring->adev;
-     struct amdgpu_ib ib;
+     struct amdgpu_job *job;
+     struct amdgpu_ib *ib;
       struct dma_fence *f = NULL;
       unsigned index;
       long r;
@@ -1126,26 +1127,30 @@ static int sdma_v4_4_2_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
       gpu_addr = adev->wb.gpu_addr + (index * 4);
       tmp = 0xCAFEDEAD;
       adev->wb.wb[index] = cpu_to_le32(tmp);
-     memset(&ib, 0, sizeof(ib));
-     r = amdgpu_ib_get(adev, NULL, 256,
-                                     AMDGPU_IB_POOL_DIRECT, &ib);
+
+     r = amdgpu_job_alloc_with_ib(ring->adev, NULL, NULL, 256,
+                                  AMDGPU_IB_POOL_DIRECT, &job,
+                                  AMDGPU_KERNEL_JOB_ID_SDMA_RING_TEST);
       if (r)
               goto err0;

-     ib.ptr[0] = SDMA_PKT_HEADER_OP(SDMA_OP_WRITE) |
+     ib = &job->ibs[0];
+     ib->ptr[0] = SDMA_PKT_HEADER_OP(SDMA_OP_WRITE) |
               SDMA_PKT_HEADER_SUB_OP(SDMA_SUBOP_WRITE_LINEAR);
-     ib.ptr[1] = lower_32_bits(gpu_addr);
-     ib.ptr[2] = upper_32_bits(gpu_addr);
-     ib.ptr[3] = SDMA_PKT_WRITE_UNTILED_DW_3_COUNT(0);
-     ib.ptr[4] = 0xDEADBEEF;
-     ib.ptr[5] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
-     ib.ptr[6] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
-     ib.ptr[7] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
-     ib.length_dw = 8;
-
-     r = amdgpu_ib_schedule(ring, 1, &ib, NULL, &f);
-     if (r)
+     ib->ptr[1] = lower_32_bits(gpu_addr);
+     ib->ptr[2] = upper_32_bits(gpu_addr);
+     ib->ptr[3] = SDMA_PKT_WRITE_UNTILED_DW_3_COUNT(0);
+     ib->ptr[4] = 0xDEADBEEF;
+     ib->ptr[5] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
+     ib->ptr[6] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
+     ib->ptr[7] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
+     ib->length_dw = 8;
+
+     r = amdgpu_job_submit_direct(job, ring, &f);
+     if (r) {
+             amdgpu_job_free(job);
               goto err1;
+     }

       r = dma_fence_wait_timeout(f, false, timeout);
       if (r == 0) {
@@ -1161,7 +1166,6 @@ static int sdma_v4_4_2_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
               r = -EINVAL;

   err1:
-     amdgpu_ib_free(&ib, NULL);
       dma_fence_put(f);
   err0:
       amdgpu_device_wb_free(adev, index);
diff --git a/drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 
b/drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c
index e77e079fe8339..edb1e4d3f9292 100644
--- a/drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c
+++ b/drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c
@@ -1074,7 +1074,8 @@ static int sdma_v5_0_ring_test_ring(struct amdgpu_ring 
*ring)
   static int sdma_v5_0_ring_test_ib(struct amdgpu_ring *ring, long timeout)
   {
       struct amdgpu_device *adev = ring->adev;
-     struct amdgpu_ib ib;
+     struct amdgpu_job *job;
+     struct amdgpu_ib *ib;
       struct dma_fence *f = NULL;
       unsigned index;
       long r;
@@ -1082,7 +1083,6 @@ static int sdma_v5_0_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
       u64 gpu_addr;

       tmp = 0xCAFEDEAD;
-     memset(&ib, 0, sizeof(ib));

       r = amdgpu_device_wb_get(adev, &index);
       if (r) {
@@ -1093,27 +1093,31 @@ static int sdma_v5_0_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
       gpu_addr = adev->wb.gpu_addr + (index * 4);
       adev->wb.wb[index] = cpu_to_le32(tmp);

-     r = amdgpu_ib_get(adev, NULL, 256,
-                       AMDGPU_IB_POOL_DIRECT, &ib);
+     r = amdgpu_job_alloc_with_ib(ring->adev, NULL, NULL, 256,
+                                  AMDGPU_IB_POOL_DIRECT, &job,
+                                  AMDGPU_KERNEL_JOB_ID_SDMA_RING_TEST);
       if (r) {
               drm_err(adev_to_drm(adev), "failed to get ib (%ld).\n", r);
               goto err0;
       }

-     ib.ptr[0] = SDMA_PKT_HEADER_OP(SDMA_OP_WRITE) |
+     ib = &job->ibs[0];
+     ib->ptr[0] = SDMA_PKT_HEADER_OP(SDMA_OP_WRITE) |
               SDMA_PKT_HEADER_SUB_OP(SDMA_SUBOP_WRITE_LINEAR);
-     ib.ptr[1] = lower_32_bits(gpu_addr);
-     ib.ptr[2] = upper_32_bits(gpu_addr);
-     ib.ptr[3] = SDMA_PKT_WRITE_UNTILED_DW_3_COUNT(0);
-     ib.ptr[4] = 0xDEADBEEF;
-     ib.ptr[5] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
-     ib.ptr[6] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
-     ib.ptr[7] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
-     ib.length_dw = 8;
-
-     r = amdgpu_ib_schedule(ring, 1, &ib, NULL, &f);
-     if (r)
+     ib->ptr[1] = lower_32_bits(gpu_addr);
+     ib->ptr[2] = upper_32_bits(gpu_addr);
+     ib->ptr[3] = SDMA_PKT_WRITE_UNTILED_DW_3_COUNT(0);
+     ib->ptr[4] = 0xDEADBEEF;
+     ib->ptr[5] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
+     ib->ptr[6] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
+     ib->ptr[7] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
+     ib->length_dw = 8;
+
+     r = amdgpu_job_submit_direct(job, ring, &f);
+     if (r) {
+             amdgpu_job_free(job);
               goto err1;
+     }

       r = dma_fence_wait_timeout(f, false, timeout);
       if (r == 0) {
@@ -1133,7 +1137,6 @@ static int sdma_v5_0_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
               r = -EINVAL;

   err1:
-     amdgpu_ib_free(&ib, NULL);
       dma_fence_put(f);
   err0:
       amdgpu_device_wb_free(adev, index);
diff --git a/drivers/gpu/drm/amd/amdgpu/sdma_v5_2.c 
b/drivers/gpu/drm/amd/amdgpu/sdma_v5_2.c
index 50b51965c211e..a57fe976cccaf 100644
--- a/drivers/gpu/drm/amd/amdgpu/sdma_v5_2.c
+++ b/drivers/gpu/drm/amd/amdgpu/sdma_v5_2.c
@@ -974,7 +974,8 @@ static int sdma_v5_2_ring_test_ring(struct amdgpu_ring 
*ring)
   static int sdma_v5_2_ring_test_ib(struct amdgpu_ring *ring, long timeout)
   {
       struct amdgpu_device *adev = ring->adev;
-     struct amdgpu_ib ib;
+     struct amdgpu_job *job;
+     struct amdgpu_ib *ib;
       struct dma_fence *f = NULL;
       unsigned index;
       long r;
@@ -982,7 +983,6 @@ static int sdma_v5_2_ring_test_ib(struct amdgpu_ring *ring, 
long timeout)
       u64 gpu_addr;

       tmp = 0xCAFEDEAD;
-     memset(&ib, 0, sizeof(ib));

       r = amdgpu_device_wb_get(adev, &index);
       if (r) {
@@ -993,26 +993,31 @@ static int sdma_v5_2_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
       gpu_addr = adev->wb.gpu_addr + (index * 4);
       adev->wb.wb[index] = cpu_to_le32(tmp);

-     r = amdgpu_ib_get(adev, NULL, 256, AMDGPU_IB_POOL_DIRECT, &ib);
+     r = amdgpu_job_alloc_with_ib(ring->adev, NULL, NULL, 256,
+                                  AMDGPU_IB_POOL_DIRECT, &job,
+                                  AMDGPU_KERNEL_JOB_ID_SDMA_RING_TEST);
       if (r) {
               drm_err(adev_to_drm(adev), "failed to get ib (%ld).\n", r);
               goto err0;
       }

-     ib.ptr[0] = SDMA_PKT_HEADER_OP(SDMA_OP_WRITE) |
+     ib = &job->ibs[0];
+     ib->ptr[0] = SDMA_PKT_HEADER_OP(SDMA_OP_WRITE) |
               SDMA_PKT_HEADER_SUB_OP(SDMA_SUBOP_WRITE_LINEAR);
-     ib.ptr[1] = lower_32_bits(gpu_addr);
-     ib.ptr[2] = upper_32_bits(gpu_addr);
-     ib.ptr[3] = SDMA_PKT_WRITE_UNTILED_DW_3_COUNT(0);
-     ib.ptr[4] = 0xDEADBEEF;
-     ib.ptr[5] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
-     ib.ptr[6] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
-     ib.ptr[7] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
-     ib.length_dw = 8;
-
-     r = amdgpu_ib_schedule(ring, 1, &ib, NULL, &f);
-     if (r)
+     ib->ptr[1] = lower_32_bits(gpu_addr);
+     ib->ptr[2] = upper_32_bits(gpu_addr);
+     ib->ptr[3] = SDMA_PKT_WRITE_UNTILED_DW_3_COUNT(0);
+     ib->ptr[4] = 0xDEADBEEF;
+     ib->ptr[5] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
+     ib->ptr[6] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
+     ib->ptr[7] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
+     ib->length_dw = 8;
+
+     r = amdgpu_job_submit_direct(job, ring, &f);
+     if (r) {
+             amdgpu_job_free(job);
               goto err1;
+     }

       r = dma_fence_wait_timeout(f, false, timeout);
       if (r == 0) {
@@ -1032,7 +1037,6 @@ static int sdma_v5_2_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
               r = -EINVAL;

   err1:
-     amdgpu_ib_free(&ib, NULL);
       dma_fence_put(f);
   err0:
       amdgpu_device_wb_free(adev, index);
diff --git a/drivers/gpu/drm/amd/amdgpu/sdma_v6_0.c 
b/drivers/gpu/drm/amd/amdgpu/sdma_v6_0.c
index eec659194718d..210ea6ba6212f 100644
--- a/drivers/gpu/drm/amd/amdgpu/sdma_v6_0.c
+++ b/drivers/gpu/drm/amd/amdgpu/sdma_v6_0.c
@@ -981,7 +981,8 @@ static int sdma_v6_0_ring_test_ring(struct amdgpu_ring 
*ring)
   static int sdma_v6_0_ring_test_ib(struct amdgpu_ring *ring, long timeout)
   {
       struct amdgpu_device *adev = ring->adev;
-     struct amdgpu_ib ib;
+     struct amdgpu_job *job;
+     struct amdgpu_ib *ib;
       struct dma_fence *f = NULL;
       unsigned index;
       long r;
@@ -989,7 +990,6 @@ static int sdma_v6_0_ring_test_ib(struct amdgpu_ring *ring, 
long timeout)
       u64 gpu_addr;

       tmp = 0xCAFEDEAD;
-     memset(&ib, 0, sizeof(ib));

       r = amdgpu_device_wb_get(adev, &index);
       if (r) {
@@ -1000,26 +1000,31 @@ static int sdma_v6_0_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
       gpu_addr = adev->wb.gpu_addr + (index * 4);
       adev->wb.wb[index] = cpu_to_le32(tmp);

-     r = amdgpu_ib_get(adev, NULL, 256, AMDGPU_IB_POOL_DIRECT, &ib);
+     r = amdgpu_job_alloc_with_ib(ring->adev, NULL, NULL, 256,
+                                  AMDGPU_IB_POOL_DIRECT, &job,
+                                  AMDGPU_KERNEL_JOB_ID_SDMA_RING_TEST);
       if (r) {
               drm_err(adev_to_drm(adev), "failed to get ib (%ld).\n", r);
               goto err0;
       }

-     ib.ptr[0] = SDMA_PKT_COPY_LINEAR_HEADER_OP(SDMA_OP_WRITE) |
+     ib = &job->ibs[0];
+     ib->ptr[0] = SDMA_PKT_COPY_LINEAR_HEADER_OP(SDMA_OP_WRITE) |
               SDMA_PKT_COPY_LINEAR_HEADER_SUB_OP(SDMA_SUBOP_WRITE_LINEAR);
-     ib.ptr[1] = lower_32_bits(gpu_addr);
-     ib.ptr[2] = upper_32_bits(gpu_addr);
-     ib.ptr[3] = SDMA_PKT_WRITE_UNTILED_DW_3_COUNT(0);
-     ib.ptr[4] = 0xDEADBEEF;
-     ib.ptr[5] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
-     ib.ptr[6] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
-     ib.ptr[7] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
-     ib.length_dw = 8;
-
-     r = amdgpu_ib_schedule(ring, 1, &ib, NULL, &f);
-     if (r)
+     ib->ptr[1] = lower_32_bits(gpu_addr);
+     ib->ptr[2] = upper_32_bits(gpu_addr);
+     ib->ptr[3] = SDMA_PKT_WRITE_UNTILED_DW_3_COUNT(0);
+     ib->ptr[4] = 0xDEADBEEF;
+     ib->ptr[5] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
+     ib->ptr[6] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
+     ib->ptr[7] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
+     ib->length_dw = 8;
+
+     r = amdgpu_job_submit_direct(job, ring, &f);
+     if (r) {
+             amdgpu_job_free(job);
               goto err1;
+     }

       r = dma_fence_wait_timeout(f, false, timeout);
       if (r == 0) {
@@ -1039,7 +1044,6 @@ static int sdma_v6_0_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
               r = -EINVAL;

   err1:
-     amdgpu_ib_free(&ib, NULL);
       dma_fence_put(f);
   err0:
       amdgpu_device_wb_free(adev, index);
diff --git a/drivers/gpu/drm/amd/amdgpu/sdma_v7_0.c 
b/drivers/gpu/drm/amd/amdgpu/sdma_v7_0.c
index 8d16ef257bcb9..3b4417d19212e 100644
--- a/drivers/gpu/drm/amd/amdgpu/sdma_v7_0.c
+++ b/drivers/gpu/drm/amd/amdgpu/sdma_v7_0.c
@@ -997,7 +997,8 @@ static int sdma_v7_0_ring_test_ring(struct amdgpu_ring 
*ring)
   static int sdma_v7_0_ring_test_ib(struct amdgpu_ring *ring, long timeout)
   {
       struct amdgpu_device *adev = ring->adev;
-     struct amdgpu_ib ib;
+     struct amdgpu_job *job;
+     struct amdgpu_ib *ib;
       struct dma_fence *f = NULL;
       unsigned index;
       long r;
@@ -1005,7 +1006,6 @@ static int sdma_v7_0_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
       u64 gpu_addr;

       tmp = 0xCAFEDEAD;
-     memset(&ib, 0, sizeof(ib));

       r = amdgpu_device_wb_get(adev, &index);
       if (r) {
@@ -1016,26 +1016,31 @@ static int sdma_v7_0_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
       gpu_addr = adev->wb.gpu_addr + (index * 4);
       adev->wb.wb[index] = cpu_to_le32(tmp);

-     r = amdgpu_ib_get(adev, NULL, 256, AMDGPU_IB_POOL_DIRECT, &ib);
+     r = amdgpu_job_alloc_with_ib(ring->adev, NULL, NULL, 256,
+                                  AMDGPU_IB_POOL_DIRECT, &job,
+                                  AMDGPU_KERNEL_JOB_ID_SDMA_RING_TEST);
       if (r) {
               drm_err(adev_to_drm(adev), "failed to get ib (%ld).\n", r);
               goto err0;
       }

-     ib.ptr[0] = SDMA_PKT_COPY_LINEAR_HEADER_OP(SDMA_OP_WRITE) |
+     ib = &job->ibs[0];
+     ib->ptr[0] = SDMA_PKT_COPY_LINEAR_HEADER_OP(SDMA_OP_WRITE) |
               SDMA_PKT_COPY_LINEAR_HEADER_SUB_OP(SDMA_SUBOP_WRITE_LINEAR);
-     ib.ptr[1] = lower_32_bits(gpu_addr);
-     ib.ptr[2] = upper_32_bits(gpu_addr);
-     ib.ptr[3] = SDMA_PKT_WRITE_UNTILED_DW_3_COUNT(0);
-     ib.ptr[4] = 0xDEADBEEF;
-     ib.ptr[5] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
-     ib.ptr[6] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
-     ib.ptr[7] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
-     ib.length_dw = 8;
-
-     r = amdgpu_ib_schedule(ring, 1, &ib, NULL, &f);
-     if (r)
+     ib->ptr[1] = lower_32_bits(gpu_addr);
+     ib->ptr[2] = upper_32_bits(gpu_addr);
+     ib->ptr[3] = SDMA_PKT_WRITE_UNTILED_DW_3_COUNT(0);
+     ib->ptr[4] = 0xDEADBEEF;
+     ib->ptr[5] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
+     ib->ptr[6] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
+     ib->ptr[7] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
+     ib->length_dw = 8;
+
+     r = amdgpu_job_submit_direct(job, ring, &f);
+     if (r) {
+             amdgpu_job_free(job);
               goto err1;
+     }

       r = dma_fence_wait_timeout(f, false, timeout);
       if (r == 0) {
@@ -1055,7 +1060,6 @@ static int sdma_v7_0_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
               r = -EINVAL;

   err1:
-     amdgpu_ib_free(&ib, NULL);
       dma_fence_put(f);
   err0:
       amdgpu_device_wb_free(adev, index);
diff --git a/drivers/gpu/drm/amd/amdgpu/sdma_v7_1.c 
b/drivers/gpu/drm/amd/amdgpu/sdma_v7_1.c
index 0824cba48f2e7..d8167ce18dbd8 100644
--- a/drivers/gpu/drm/amd/amdgpu/sdma_v7_1.c
+++ b/drivers/gpu/drm/amd/amdgpu/sdma_v7_1.c
@@ -987,7 +987,8 @@ static int sdma_v7_1_ring_test_ring(struct amdgpu_ring 
*ring)
   static int sdma_v7_1_ring_test_ib(struct amdgpu_ring *ring, long timeout)
   {
       struct amdgpu_device *adev = ring->adev;
-     struct amdgpu_ib ib;
+     struct amdgpu_job *job;
+     struct amdgpu_ib *ib;
       struct dma_fence *f = NULL;
       unsigned index;
       long r;
@@ -995,7 +996,6 @@ static int sdma_v7_1_ring_test_ib(struct amdgpu_ring *ring, 
long timeout)
       u64 gpu_addr;

       tmp = 0xCAFEDEAD;
-     memset(&ib, 0, sizeof(ib));

       r = amdgpu_device_wb_get(adev, &index);
       if (r) {
@@ -1006,26 +1006,31 @@ static int sdma_v7_1_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
       gpu_addr = adev->wb.gpu_addr + (index * 4);
       adev->wb.wb[index] = cpu_to_le32(tmp);

-     r = amdgpu_ib_get(adev, NULL, 256, AMDGPU_IB_POOL_DIRECT, &ib);
+     r = amdgpu_job_alloc_with_ib(ring->adev, NULL, NULL, 256,
+                                  AMDGPU_IB_POOL_DIRECT, &job,
+                                  AMDGPU_KERNEL_JOB_ID_SDMA_RING_TEST);
       if (r) {
               DRM_ERROR("amdgpu: failed to get ib (%ld).\n", r);
               goto err0;
       }

-     ib.ptr[0] = SDMA_PKT_COPY_LINEAR_HEADER_OP(SDMA_OP_WRITE) |
+     ib = &job->ibs[0];
+     ib->ptr[0] = SDMA_PKT_COPY_LINEAR_HEADER_OP(SDMA_OP_WRITE) |
               SDMA_PKT_COPY_LINEAR_HEADER_SUB_OP(SDMA_SUBOP_WRITE_LINEAR);
-     ib.ptr[1] = lower_32_bits(gpu_addr);
-     ib.ptr[2] = upper_32_bits(gpu_addr);
-     ib.ptr[3] = SDMA_PKT_WRITE_UNTILED_DW_3_COUNT(0);
-     ib.ptr[4] = 0xDEADBEEF;
-     ib.ptr[5] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
-     ib.ptr[6] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
-     ib.ptr[7] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
-     ib.length_dw = 8;
-
-     r = amdgpu_ib_schedule(ring, 1, &ib, NULL, &f);
-     if (r)
+     ib->ptr[1] = lower_32_bits(gpu_addr);
+     ib->ptr[2] = upper_32_bits(gpu_addr);
+     ib->ptr[3] = SDMA_PKT_WRITE_UNTILED_DW_3_COUNT(0);
+     ib->ptr[4] = 0xDEADBEEF;
+     ib->ptr[5] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
+     ib->ptr[6] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
+     ib->ptr[7] = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP);
+     ib->length_dw = 8;
+
+     r = amdgpu_job_submit_direct(job, ring, &f);
+     if (r) {
+             amdgpu_job_free(job);
               goto err1;
+     }

       r = dma_fence_wait_timeout(f, false, timeout);
       if (r == 0) {
@@ -1045,7 +1050,6 @@ static int sdma_v7_1_ring_test_ib(struct amdgpu_ring 
*ring, long timeout)
               r = -EINVAL;

   err1:
-     amdgpu_ib_free(&ib, NULL);
       dma_fence_put(f);
   err0:
       amdgpu_device_wb_free(adev, index);
diff --git a/drivers/gpu/drm/amd/amdgpu/si_dma.c 
b/drivers/gpu/drm/amd/amdgpu/si_dma.c
index 74fcaa340d9b1..b67bd343f795f 100644
--- a/drivers/gpu/drm/amd/amdgpu/si_dma.c
+++ b/drivers/gpu/drm/amd/amdgpu/si_dma.c
@@ -259,7 +259,8 @@ static int si_dma_ring_test_ring(struct amdgpu_ring *ring)
   static int si_dma_ring_test_ib(struct amdgpu_ring *ring, long timeout)
   {
       struct amdgpu_device *adev = ring->adev;
-     struct amdgpu_ib ib;
+     struct amdgpu_job *job;
+     struct amdgpu_ib *ib;
       struct dma_fence *f = NULL;
       unsigned index;
       u32 tmp = 0;
@@ -273,20 +274,25 @@ static int si_dma_ring_test_ib(struct amdgpu_ring *ring, 
long timeout)
       gpu_addr = adev->wb.gpu_addr + (index * 4);
       tmp = 0xCAFEDEAD;
       adev->wb.wb[index] = cpu_to_le32(tmp);
-     memset(&ib, 0, sizeof(ib));
-     r = amdgpu_ib_get(adev, NULL, 256,
-                                     AMDGPU_IB_POOL_DIRECT, &ib);
+
+     r = amdgpu_job_alloc_with_ib(ring->adev, NULL, NULL, 256,
+                                  AMDGPU_IB_POOL_DIRECT, &job,
+                                  AMDGPU_KERNEL_JOB_ID_SDMA_RING_TEST);
       if (r)
               goto err0;

-     ib.ptr[0] = DMA_PACKET(DMA_PACKET_WRITE, 0, 0, 0, 1);
-     ib.ptr[1] = lower_32_bits(gpu_addr);
-     ib.ptr[2] = upper_32_bits(gpu_addr) & 0xff;
-     ib.ptr[3] = 0xDEADBEEF;
-     ib.length_dw = 4;
-     r = amdgpu_ib_schedule(ring, 1, &ib, NULL, &f);
-     if (r)
+     ib = &job->ibs[0];
+     ib->ptr[0] = DMA_PACKET(DMA_PACKET_WRITE, 0, 0, 0, 1);
+     ib->ptr[1] = lower_32_bits(gpu_addr);
+     ib->ptr[2] = upper_32_bits(gpu_addr) & 0xff;
+     ib->ptr[3] = 0xDEADBEEF;
+     ib->length_dw = 4;
+
+     r = amdgpu_job_submit_direct(job, ring, &f);
+     if (r) {
+             amdgpu_job_free(job);
               goto err1;
+     }

       r = dma_fence_wait_timeout(f, false, timeout);
       if (r == 0) {
@@ -302,7 +308,6 @@ static int si_dma_ring_test_ib(struct amdgpu_ring *ring, 
long timeout)
               r = -EINVAL;

   err1:
-     amdgpu_ib_free(&ib, NULL);
       dma_fence_put(f);
   err0:
       amdgpu_device_wb_free(adev, index);

Reply via email to