The benefits of using multiple entities is that multiple fill jobs
can run in parallel. Otherwise, even if the entity has access
to multiple engines, a burst of N independent jobs will all
run on the same engine because an entity guarantees the ordering
of execution matches the ordering of the submission.

Callers can opt-out of this behavior by passing the entity they
want to use (see amdgpu_move_blit).

Signed-off-by: Pierre-Eric Pelloux-Prayer <[email protected]>
---
 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c | 84 ++++++++++++++++++-------
 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.h |  1 +
 2 files changed, 64 insertions(+), 21 deletions(-)

diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 
b/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
index c357a6d9763a..839ea8c7f6be 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
@@ -2224,6 +2224,7 @@ u32 amdgpu_ttm_set_buffer_funcs_status(struct 
amdgpu_device *adev, bool enable)
                adev->mman.clear_entities = kcalloc(num_clear_entities,
                                                    sizeof(struct 
amdgpu_ttm_entity),
                                                    GFP_KERNEL);
+               atomic_set(&adev->mman.next_clear_entity, 0);
                if (!adev->mman.clear_entities)
                        goto error_free_entity;
 
@@ -2498,10 +2499,12 @@ int amdgpu_fill_buffer(struct amdgpu_ttm_entity *entity,
 {
        struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
        struct amdgpu_ring *ring = adev->mman.buffer_funcs_ring;
+       struct dma_fence *fences[TTM_FENCES_MAX_SLOT_COUNT] = {};
        struct dma_fence *fence = NULL;
        struct dma_resv *resv = NULL;
        struct amdgpu_res_cursor dst;
-       int r;
+       uint64_t cur_size, to;
+       int r, e, n_fences;
 
        /* The fences will be either added to the resv object or the last fence
         * will be returned to the caller. In the latter case, all fill jobs 
will
@@ -2515,53 +2518,92 @@ int amdgpu_fill_buffer(struct amdgpu_ttm_entity *entity,
        }
 
        if (!entity) {
-               entity = &adev->mman.clear_entities[0];
                resv = &bo->tbo.base._resv;
-               r = dma_resv_reserve_fences(resv, 1);
+
+               /* Determine how much fences we're going to add to the
+                * resv object.
+                */
+               n_fences = 0;
+               amdgpu_res_first(bo->tbo.resource, 0, amdgpu_bo_size(bo), &dst);
+               while (dst.remaining) {
+                       cur_size = min(dst.size, 256ULL << 20);
+
+                       n_fences += 1;
+                       amdgpu_res_next(&dst, cur_size);
+               }
+               if (n_fences == 0)
+                       return 0;
+
+               /* One slot per entity at most. */
+               n_fences = MIN(n_fences, adev->mman.num_clear_entities);
+
+               r = dma_resv_reserve_fences(resv, n_fences);
                if (r)
                        return r;
+       } else {
+               mutex_lock(&entity->gart_window_lock);
        }
 
        amdgpu_res_first(bo->tbo.resource, 0, amdgpu_bo_size(bo), &dst);
 
-       mutex_lock(&entity->gart_window_lock);
        while (dst.remaining) {
-               struct dma_fence *next;
-               uint64_t cur_size, to;
-
                /* Never fill more than 256MiB at once to avoid timeouts */
                cur_size = min(dst.size, 256ULL << 20);
 
+               if (resv) {
+                       /* Pick a new entity for each partial clear so they can
+                        * execute in parallel.
+                        */
+                       e = atomic_inc_return(&adev->mman.next_clear_entity) %
+                               adev->mman.num_clear_entities;
+                       entity = &adev->mman.clear_entities[e];
+                       mutex_lock(&entity->gart_window_lock);
+               }
+
                r = amdgpu_ttm_map_buffer(&entity->base,
                                          &bo->tbo, bo->tbo.resource, &dst,
                                          entity->gart_window_id1, ring, false,
                                          &cur_size, &to,
                                          dependency,
                                          resv);
-               if (r)
+               if (r) {
+                       mutex_unlock(&entity->gart_window_lock);
                        goto error;
+               }
 
                r = amdgpu_ttm_fill_mem(ring, &entity->base,
                                        src_data, to, cur_size, resv,
-                                       &next, true, k_job_id);
-               if (r)
+                                       &fence, true, k_job_id);
+               if (r) {
+                       mutex_unlock(&entity->gart_window_lock);
                        goto error;
-
-               if (resv) {
-                       dma_resv_add_fence(resv, next, DMA_RESV_USAGE_KERNEL);
-                       dma_fence_put(next);
-               } else {
-                       dma_fence_put(fence);
-                       fence = next;
                }
 
                amdgpu_res_next(&dst, cur_size);
+
+               if (resv) {
+                       /* Delay the addition of the fences to resv, otherwise 
the next partial
+                        * clears will depend on this one.
+                        */
+                       fences[e] = fence;
+                       mutex_unlock(&entity->gart_window_lock);
+               } else {
+                       dma_fence_put(*f);
+                       *f = fence;
+               }
        }
 error:
-       mutex_unlock(&entity->gart_window_lock);
-       if (f)
-               *f = dma_fence_get(fence);
-       dma_fence_put(fence);
+       if (resv) {
+               for (e = 0; e < adev->mman.num_clear_entities; e++) {
+                       if (fences[e]) {
+                               dma_resv_add_fence(resv, fences[e], 
DMA_RESV_USAGE_KERNEL);
+                               dma_fence_put(fences[e]);
+                       }
+               }
+       } else {
+               mutex_unlock(&entity->gart_window_lock);
+       }
+
        return r;
 }
 
diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.h 
b/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.h
index 38df2b5b4bc7..3fc31c7c6bfe 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.h
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.h
@@ -73,6 +73,7 @@ struct amdgpu_mman {
 
        struct amdgpu_ttm_entity default_entity; /* has no gart windows */
        struct amdgpu_ttm_entity *clear_entities;
+       atomic_t next_clear_entity;
        u32 num_clear_entities;
        struct amdgpu_ttm_entity move_entities[TTM_FENCES_MAX_SLOT_COUNT];
        u32 num_move_entities;
-- 
2.43.0

Reply via email to