Add the userq object virtual address get(),mapped() and put() helpers for tracking the userq obj va address usage.
Signed-off-by: Prike Liang <prike.li...@amd.com> --- drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c | 172 ++++++++++++++++++++- drivers/gpu/drm/amd/amdgpu/amdgpu_userq.h | 14 ++ drivers/gpu/drm/amd/amdgpu/mes_userqueue.c | 4 + 3 files changed, 189 insertions(+), 1 deletion(-) diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c index 0aeb7a96ccbf..562d12f9d0d2 100644 --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c @@ -76,6 +76,174 @@ int amdgpu_userq_input_va_validate(struct amdgpu_vm *vm, u64 addr, return r; } +int amdgpu_userq_buffer_va_get(struct amdgpu_vm *vm, u64 addr) +{ + struct amdgpu_bo_va_mapping *mapping; + u64 user_addr; + int r; + + user_addr = (addr & AMDGPU_GMC_HOLE_MASK) >> AMDGPU_GPU_PAGE_SHIFT; + r = amdgpu_bo_reserve(vm->root.bo, false); + if (r) + return r; + + mapping = amdgpu_vm_bo_lookup_mapping(vm, user_addr); + if (!mapping) + goto out_err; + + /* + * Need to unify the following userq va reference. + * mqd bo + * rptr bo + * wptr bo + * eop bo + * shadow bo + * csa bo + */ + /*amdgpu_bo_ref(mapping->bo_va->base.bo);*/ + mapping->bo_va->queue_refcount++; + + amdgpu_bo_unreserve(vm->root.bo); + return 0; + +out_err: + amdgpu_bo_unreserve(vm->root.bo); + return -EINVAL; +} + +bool amdgpu_userq_buffer_va_mapped(struct amdgpu_vm *vm, u64 addr) +{ + struct amdgpu_bo_va_mapping *mapping; + u64 user_addr; + bool r; + + user_addr = (addr & AMDGPU_GMC_HOLE_MASK) >> AMDGPU_GPU_PAGE_SHIFT; + + if (amdgpu_bo_reserve(vm->root.bo, false)) + return false; + + mapping = amdgpu_vm_bo_lookup_mapping(vm, user_addr); + if (!IS_ERR_OR_NULL(mapping) && mapping->bo_va->queue_refcount > 0) + r = true; + else + r = false; + amdgpu_bo_unreserve(vm->root.bo); + + return r; +} + +bool amdgpu_userq_buffer_vas_mapped(struct amdgpu_vm *vm, + struct amdgpu_usermode_queue *queue) +{ + + switch (queue->queue_type) { + case AMDGPU_HW_IP_GFX: + if (amdgpu_userq_buffer_va_mapped(vm, queue->queue_va) || + amdgpu_userq_buffer_va_mapped(vm, queue->rptr_va) || + amdgpu_userq_buffer_va_mapped(vm, queue->wptr_va) || + amdgpu_userq_buffer_va_mapped(vm, queue->shadow_va) || + amdgpu_userq_buffer_va_mapped(vm, queue->csa_va)) + return true; + break; + case AMDGPU_HW_IP_COMPUTE: + if (amdgpu_userq_buffer_va_mapped(vm, queue->queue_va) || + amdgpu_userq_buffer_va_mapped(vm, queue->rptr_va) || + amdgpu_userq_buffer_va_mapped(vm, queue->wptr_va) || + amdgpu_userq_buffer_va_mapped(vm, queue->eop_va)) + return true; + break; + case AMDGPU_HW_IP_DMA: + if (amdgpu_userq_buffer_va_mapped(vm, queue->queue_va) || + amdgpu_userq_buffer_va_mapped(vm, queue->rptr_va) || + amdgpu_userq_buffer_va_mapped(vm, queue->wptr_va) || + amdgpu_userq_buffer_va_mapped(vm, queue->csa_va)) + return true; + break; + default: + break; + } + + return false; +} + +int amdgpu_userq_buffer_va_put(struct amdgpu_vm *vm, u64 addr) +{ + struct amdgpu_bo_va_mapping *mapping; + u64 user_addr; + int r; + + user_addr = (addr & AMDGPU_GMC_HOLE_MASK) >> AMDGPU_GPU_PAGE_SHIFT; + r = amdgpu_bo_reserve(vm->root.bo, false); + if (r) + return r; + + mapping = amdgpu_vm_bo_lookup_mapping(vm, user_addr); + if (!mapping) + goto out_err; + /* + * TODO: It requires figuring out the root cause of userq va mapping + * reference imbalance issue. + */ + /*amdgpu_bo_unref(&mapping->bo_va->base.bo);*/ + mapping->bo_va->queue_refcount--; + + amdgpu_bo_unreserve(vm->root.bo); + return 0; + +out_err: + amdgpu_bo_unreserve(vm->root.bo); + return -EINVAL; +} + +static void amdgpu_userq_buffer_vas_get(struct amdgpu_vm *vm, + struct amdgpu_usermode_queue *queue) +{ + + + amdgpu_userq_buffer_va_get(vm, queue->queue_va); + amdgpu_userq_buffer_va_get(vm, queue->rptr_va); + amdgpu_userq_buffer_va_get(vm, queue->wptr_va); + + switch (queue->queue_type) { + case AMDGPU_HW_IP_GFX: + amdgpu_userq_buffer_va_get(vm, queue->shadow_va); + amdgpu_userq_buffer_va_get(vm, queue->csa_va); + break; + case AMDGPU_HW_IP_COMPUTE: + amdgpu_userq_buffer_va_get(vm, queue->eop_va); + break; + case AMDGPU_HW_IP_DMA: + amdgpu_userq_buffer_va_get(vm, queue->csa_va); + break; + default: + break; + } +} + +int amdgpu_userq_buffer_vas_put(struct amdgpu_vm *vm, + struct amdgpu_usermode_queue *queue) +{ + amdgpu_userq_buffer_va_put(vm, queue->queue_va); + amdgpu_userq_buffer_va_put(vm, queue->rptr_va); + amdgpu_userq_buffer_va_put(vm, queue->wptr_va); + + switch (queue->queue_type) { + case AMDGPU_HW_IP_GFX: + amdgpu_userq_buffer_va_put(vm, queue->shadow_va); + amdgpu_userq_buffer_va_put(vm, queue->csa_va); + break; + case AMDGPU_HW_IP_COMPUTE: + amdgpu_userq_buffer_va_put(vm, queue->eop_va); + break; + case AMDGPU_HW_IP_DMA: + amdgpu_userq_buffer_va_put(vm, queue->csa_va); + break; + default: + break; + } + return 0; +} + static int amdgpu_userq_unmap_helper(struct amdgpu_userq_mgr *uq_mgr, struct amdgpu_usermode_queue *queue) @@ -445,6 +613,9 @@ amdgpu_userq_create(struct drm_file *filp, union drm_amdgpu_userq *args) queue->vm = &fpriv->vm; queue->priority = priority; queue->generation = amdgpu_vm_generation(adev, &fpriv->vm); + queue->queue_va = args->in.queue_va; + queue->rptr_va = args->in.rptr_va; + queue->wptr_va = args->in.wptr_va; db_info.queue_type = queue->queue_type; db_info.doorbell_handle = queue->doorbell_handle; @@ -475,7 +646,6 @@ amdgpu_userq_create(struct drm_file *filp, union drm_amdgpu_userq *args) goto unlock; } - qid = idr_alloc(&uq_mgr->userq_idr, queue, 1, AMDGPU_MAX_USERQ_COUNT, GFP_KERNEL); if (qid < 0) { drm_file_err(uq_mgr->file, "Failed to allocate a queue id\n"); diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_userq.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_userq.h index 0eb2a9c2e340..30067f80eadf 100644 --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_userq.h +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_userq.h @@ -54,6 +54,13 @@ struct amdgpu_usermode_queue { enum amdgpu_userq_state state; uint64_t doorbell_handle; uint64_t doorbell_index; + uint64_t queue_va; + uint64_t rptr_va; + uint64_t wptr_va; + uint64_t eop_va; + uint64_t shadow_va; + uint64_t csa_va; + uint64_t flags; struct amdgpu_mqd_prop *userq_prop; struct amdgpu_userq_mgr *userq_mgr; @@ -137,4 +144,11 @@ int amdgpu_userq_start_sched_for_enforce_isolation(struct amdgpu_device *adev, int amdgpu_userq_input_va_validate(struct amdgpu_vm *vm, u64 addr, u64 expected_size); +int amdgpu_userq_buffer_va_get(struct amdgpu_vm *vm, u64 addr); +bool amdgpu_userq_buffer_va_mapped(struct amdgpu_vm *vm, u64 addr); +bool amdgpu_userq_buffer_vas_mapped(struct amdgpu_vm *vm, + struct amdgpu_usermode_queue *queue); +int amdgpu_userq_buffer_va_put(struct amdgpu_vm *vm, u64 addr); +int amdgpu_userq_buffer_vas_put(struct amdgpu_vm *vm, + struct amdgpu_usermode_queue *queue); #endif diff --git a/drivers/gpu/drm/amd/amdgpu/mes_userqueue.c b/drivers/gpu/drm/amd/amdgpu/mes_userqueue.c index 6e29e85bbf9f..42d6cd90be59 100644 --- a/drivers/gpu/drm/amd/amdgpu/mes_userqueue.c +++ b/drivers/gpu/drm/amd/amdgpu/mes_userqueue.c @@ -262,6 +262,7 @@ static int mes_userq_mqd_create(struct amdgpu_userq_mgr *uq_mgr, userq_props->hqd_active = false; userq_props->tmz_queue = mqd_user->flags & AMDGPU_USERQ_CREATE_FLAGS_QUEUE_SECURE; + queue->eop_va = compute_mqd->eop_va; kfree(compute_mqd); } else if (queue->queue_type == AMDGPU_HW_IP_GFX) { struct drm_amdgpu_userq_mqd_gfx11 *mqd_gfx_v11; @@ -283,6 +284,8 @@ static int mes_userq_mqd_create(struct amdgpu_userq_mgr *uq_mgr, userq_props->csa_addr = mqd_gfx_v11->csa_va; userq_props->tmz_queue = mqd_user->flags & AMDGPU_USERQ_CREATE_FLAGS_QUEUE_SECURE; + queue->shadow_va = mqd_gfx_v11->shadow_va; + queue->csa_va = mqd_gfx_v11->csa_va; if (amdgpu_userq_input_va_validate(queue->vm, mqd_gfx_v11->shadow_va, shadow_info.shadow_size)) { @@ -314,6 +317,7 @@ static int mes_userq_mqd_create(struct amdgpu_userq_mgr *uq_mgr, } userq_props->csa_addr = mqd_sdma_v11->csa_va; + queue->csa_va = mqd_sdma_v11->csa_va; kfree(mqd_sdma_v11); } -- 2.34.1