On Thu, Aug 07, 2025 at 10:13:13PM +0530, Himal Prasad Ghimiray wrote:
> From: Boris Brezillon <boris.brezil...@collabora.com>
> 
> We are about to pass more arguments to drm_gpuvm_sm_map[_ops_create](),
> so, before we do that, let's pass arguments through a struct instead
> of changing each call site every time a new optional argument is added.
> 
> v5
>  - Use drm_gpuva_op_map—same as drm_gpuvm_map_req
>  - Rebase changes for drm_gpuvm_sm_map_exec_lock()
>  - Fix kernel-docs
> 
> v6
>  - Use drm_gpuvm_map_req (Danilo/Matt)
> 
> Cc: Danilo Krummrich <d...@kernel.org>
> Cc: Brendan King <brendan.k...@imgtec.com>
> Cc: Boris Brezillon <bbrezil...@kernel.org>
> Cc: Caterina Shablia <caterina.shab...@collabora.com>
> Cc: Rob Clark <robin.cl...@oss.qualcomm.com>
> Cc: Matthew Brost <matthew.br...@intel.com>

Reviewed-by: Matthew Brost <matthew.br...@intel.com>

> Cc: <dri-devel@lists.freedesktop.org>
> Signed-off-by: Boris Brezillon <boris.brezil...@collabora.com>
> Signed-off-by: Caterina Shablia <caterina.shab...@collabora.com>
> Signed-off-by: Himal Prasad Ghimiray <himal.prasad.ghimi...@intel.com>
> ---
>  drivers/gpu/drm/drm_gpuvm.c            | 105 ++++++++++---------------
>  drivers/gpu/drm/imagination/pvr_vm.c   |  15 ++--
>  drivers/gpu/drm/msm/msm_gem_vma.c      |  25 ++++--
>  drivers/gpu/drm/nouveau/nouveau_uvmm.c |  11 ++-
>  drivers/gpu/drm/panthor/panthor_mmu.c  |  13 ++-
>  drivers/gpu/drm/xe/xe_vm.c             |  13 ++-
>  include/drm/drm_gpuvm.h                |  20 +++--
>  7 files changed, 114 insertions(+), 88 deletions(-)
> 
> diff --git a/drivers/gpu/drm/drm_gpuvm.c b/drivers/gpu/drm/drm_gpuvm.c
> index bbc7fecb6f4a..b3a01c40001b 100644
> --- a/drivers/gpu/drm/drm_gpuvm.c
> +++ b/drivers/gpu/drm/drm_gpuvm.c
> @@ -486,13 +486,18 @@
>   *                             u64 addr, u64 range,
>   *                             struct drm_gem_object *obj, u64 offset)
>   *   {
> + *           struct drm_gpuvm_map_req map_req = {
> + *                   .op_map.va.addr = addr,
> + *                   .op_map.va.range = range,
> + *                   .op_map.gem.obj = obj,
> + *                   .op_map.gem.offset = offset,
> + *              };
>   *           struct drm_gpuva_ops *ops;
>   *           struct drm_gpuva_op *op
>   *           struct drm_gpuvm_bo *vm_bo;
>   *
>   *           driver_lock_va_space();
> - *           ops = drm_gpuvm_sm_map_ops_create(gpuvm, addr, range,
> - *                                             obj, offset);
> + *           ops = drm_gpuvm_sm_map_ops_create(gpuvm, &map_req);
>   *           if (IS_ERR(ops))
>   *                   return PTR_ERR(ops);
>   *
> @@ -2054,16 +2059,15 @@ EXPORT_SYMBOL_GPL(drm_gpuva_unmap);
>  
>  static int
>  op_map_cb(const struct drm_gpuvm_ops *fn, void *priv,
> -       u64 addr, u64 range,
> -       struct drm_gem_object *obj, u64 offset)
> +       const struct drm_gpuvm_map_req *req)
>  {
>       struct drm_gpuva_op op = {};
>  
>       op.op = DRM_GPUVA_OP_MAP;
> -     op.map.va.addr = addr;
> -     op.map.va.range = range;
> -     op.map.gem.obj = obj;
> -     op.map.gem.offset = offset;
> +     op.map.va.addr = req->op_map.va.addr;
> +     op.map.va.range = req->op_map.va.range;
> +     op.map.gem.obj = req->op_map.gem.obj;
> +     op.map.gem.offset = req->op_map.gem.offset;
>  
>       return fn->sm_step_map(&op, priv);
>  }
> @@ -2102,17 +2106,16 @@ op_unmap_cb(const struct drm_gpuvm_ops *fn, void 
> *priv,
>  static int
>  __drm_gpuvm_sm_map(struct drm_gpuvm *gpuvm,
>                  const struct drm_gpuvm_ops *ops, void *priv,
> -                u64 req_addr, u64 req_range,
> -                struct drm_gem_object *req_obj, u64 req_offset)
> +                const struct drm_gpuvm_map_req *req)
>  {
>       struct drm_gpuva *va, *next;
> -     u64 req_end = req_addr + req_range;
> +     u64 req_end = req->op_map.va.addr + req->op_map.va.range;
>       int ret;
>  
> -     if (unlikely(!drm_gpuvm_range_valid(gpuvm, req_addr, req_range)))
> +     if (unlikely(!drm_gpuvm_range_valid(gpuvm, req->op_map.va.addr, 
> req->op_map.va.range)))
>               return -EINVAL;
>  
> -     drm_gpuvm_for_each_va_range_safe(va, next, gpuvm, req_addr, req_end) {
> +     drm_gpuvm_for_each_va_range_safe(va, next, gpuvm, req->op_map.va.addr, 
> req_end) {
>               struct drm_gem_object *obj = va->gem.obj;
>               u64 offset = va->gem.offset;
>               u64 addr = va->va.addr;
> @@ -2120,9 +2123,9 @@ __drm_gpuvm_sm_map(struct drm_gpuvm *gpuvm,
>               u64 end = addr + range;
>               bool merge = !!va->gem.obj;
>  
> -             if (addr == req_addr) {
> -                     merge &= obj == req_obj &&
> -                              offset == req_offset;
> +             if (addr == req->op_map.va.addr) {
> +                     merge &= obj == req->op_map.gem.obj &&
> +                              offset == req->op_map.gem.offset;
>  
>                       if (end == req_end) {
>                               ret = op_unmap_cb(ops, priv, va, merge);
> @@ -2141,9 +2144,9 @@ __drm_gpuvm_sm_map(struct drm_gpuvm *gpuvm,
>                       if (end > req_end) {
>                               struct drm_gpuva_op_map n = {
>                                       .va.addr = req_end,
> -                                     .va.range = range - req_range,
> +                                     .va.range = range - 
> req->op_map.va.range,
>                                       .gem.obj = obj,
> -                                     .gem.offset = offset + req_range,
> +                                     .gem.offset = offset + 
> req->op_map.va.range,
>                               };
>                               struct drm_gpuva_op_unmap u = {
>                                       .va = va,
> @@ -2155,8 +2158,8 @@ __drm_gpuvm_sm_map(struct drm_gpuvm *gpuvm,
>                                       return ret;
>                               break;
>                       }
> -             } else if (addr < req_addr) {
> -                     u64 ls_range = req_addr - addr;
> +             } else if (addr < req->op_map.va.addr) {
> +                     u64 ls_range = req->op_map.va.addr - addr;
>                       struct drm_gpuva_op_map p = {
>                               .va.addr = addr,
>                               .va.range = ls_range,
> @@ -2165,8 +2168,8 @@ __drm_gpuvm_sm_map(struct drm_gpuvm *gpuvm,
>                       };
>                       struct drm_gpuva_op_unmap u = { .va = va };
>  
> -                     merge &= obj == req_obj &&
> -                              offset + ls_range == req_offset;
> +                     merge &= obj == req->op_map.gem.obj &&
> +                              offset + ls_range == req->op_map.gem.offset;
>                       u.keep = merge;
>  
>                       if (end == req_end) {
> @@ -2189,7 +2192,7 @@ __drm_gpuvm_sm_map(struct drm_gpuvm *gpuvm,
>                                       .va.range = end - req_end,
>                                       .gem.obj = obj,
>                                       .gem.offset = offset + ls_range +
> -                                                   req_range,
> +                                                   req->op_map.va.range,
>                               };
>  
>                               ret = op_remap_cb(ops, priv, &p, &n, &u);
> @@ -2197,10 +2200,10 @@ __drm_gpuvm_sm_map(struct drm_gpuvm *gpuvm,
>                                       return ret;
>                               break;
>                       }
> -             } else if (addr > req_addr) {
> -                     merge &= obj == req_obj &&
> -                              offset == req_offset +
> -                                        (addr - req_addr);
> +             } else if (addr > req->op_map.va.addr) {
> +                     merge &= obj == req->op_map.gem.obj &&
> +                              offset == req->op_map.gem.offset +
> +                                        (addr - req->op_map.va.addr);
>  
>                       if (end == req_end) {
>                               ret = op_unmap_cb(ops, priv, va, merge);
> @@ -2236,9 +2239,7 @@ __drm_gpuvm_sm_map(struct drm_gpuvm *gpuvm,
>               }
>       }
>  
> -     return op_map_cb(ops, priv,
> -                      req_addr, req_range,
> -                      req_obj, req_offset);
> +     return op_map_cb(ops, priv, req);
>  }
>  
>  static int
> @@ -2303,10 +2304,7 @@ __drm_gpuvm_sm_unmap(struct drm_gpuvm *gpuvm,
>   * drm_gpuvm_sm_map() - calls the &drm_gpuva_op split/merge steps
>   * @gpuvm: the &drm_gpuvm representing the GPU VA space
>   * @priv: pointer to a driver private data structure
> - * @req_addr: the start address of the new mapping
> - * @req_range: the range of the new mapping
> - * @req_obj: the &drm_gem_object to map
> - * @req_offset: the offset within the &drm_gem_object
> + * @req: ptr to struct drm_gpuvm_map_req
>   *
>   * This function iterates the given range of the GPU VA space. It utilizes 
> the
>   * &drm_gpuvm_ops to call back into the driver providing the split and merge
> @@ -2333,8 +2331,7 @@ __drm_gpuvm_sm_unmap(struct drm_gpuvm *gpuvm,
>   */
>  int
>  drm_gpuvm_sm_map(struct drm_gpuvm *gpuvm, void *priv,
> -              u64 req_addr, u64 req_range,
> -              struct drm_gem_object *req_obj, u64 req_offset)
> +              const struct drm_gpuvm_map_req *req)
>  {
>       const struct drm_gpuvm_ops *ops = gpuvm->ops;
>  
> @@ -2343,9 +2340,7 @@ drm_gpuvm_sm_map(struct drm_gpuvm *gpuvm, void *priv,
>                      ops->sm_step_unmap)))
>               return -EINVAL;
>  
> -     return __drm_gpuvm_sm_map(gpuvm, ops, priv,
> -                               req_addr, req_range,
> -                               req_obj, req_offset);
> +     return __drm_gpuvm_sm_map(gpuvm, ops, priv, req);
>  }
>  EXPORT_SYMBOL_GPL(drm_gpuvm_sm_map);
>  
> @@ -2421,10 +2416,7 @@ static const struct drm_gpuvm_ops lock_ops = {
>   * @gpuvm: the &drm_gpuvm representing the GPU VA space
>   * @exec: the &drm_exec locking context
>   * @num_fences: for newly mapped objects, the # of fences to reserve
> - * @req_addr: the start address of the range to unmap
> - * @req_range: the range of the mappings to unmap
> - * @req_obj: the &drm_gem_object to map
> - * @req_offset: the offset within the &drm_gem_object
> + * @req: ptr to drm_gpuvm_map_req struct
>   *
>   * This function locks (drm_exec_lock_obj()) objects that will be unmapped/
>   * remapped, and locks+prepares (drm_exec_prepare_object()) objects that
> @@ -2445,9 +2437,7 @@ static const struct drm_gpuvm_ops lock_ops = {
>   *                    ret = drm_gpuvm_sm_unmap_exec_lock(gpuvm, &exec, 
> op->addr, op->range);
>   *                    break;
>   *                case DRIVER_OP_MAP:
> - *                    ret = drm_gpuvm_sm_map_exec_lock(gpuvm, &exec, 
> num_fences,
> - *                                                     op->addr, op->range,
> - *                                                     obj, op->obj_offset);
> + *                    ret = drm_gpuvm_sm_map_exec_lock(gpuvm, &exec, 
> num_fences, &req);
>   *                    break;
>   *                }
>   *
> @@ -2478,18 +2468,15 @@ static const struct drm_gpuvm_ops lock_ops = {
>  int
>  drm_gpuvm_sm_map_exec_lock(struct drm_gpuvm *gpuvm,
>                          struct drm_exec *exec, unsigned int num_fences,
> -                        u64 req_addr, u64 req_range,
> -                        struct drm_gem_object *req_obj, u64 req_offset)
> +                        struct drm_gpuvm_map_req *req)
>  {
> -     if (req_obj) {
> -             int ret = drm_exec_prepare_obj(exec, req_obj, num_fences);
> +     if (req->op_map.gem.obj) {
> +             int ret = drm_exec_prepare_obj(exec, req->op_map.gem.obj, 
> num_fences);
>               if (ret)
>                       return ret;
>       }
>  
> -     return __drm_gpuvm_sm_map(gpuvm, &lock_ops, exec,
> -                               req_addr, req_range,
> -                               req_obj, req_offset);
> +     return __drm_gpuvm_sm_map(gpuvm, &lock_ops, exec, req);
>  
>  }
>  EXPORT_SYMBOL_GPL(drm_gpuvm_sm_map_exec_lock);
> @@ -2611,10 +2598,7 @@ static const struct drm_gpuvm_ops gpuvm_list_ops = {
>  /**
>   * drm_gpuvm_sm_map_ops_create() - creates the &drm_gpuva_ops to split and 
> merge
>   * @gpuvm: the &drm_gpuvm representing the GPU VA space
> - * @req_addr: the start address of the new mapping
> - * @req_range: the range of the new mapping
> - * @req_obj: the &drm_gem_object to map
> - * @req_offset: the offset within the &drm_gem_object
> + * @req: map request arguments
>   *
>   * This function creates a list of operations to perform splitting and 
> merging
>   * of existent mapping(s) with the newly requested one.
> @@ -2642,8 +2626,7 @@ static const struct drm_gpuvm_ops gpuvm_list_ops = {
>   */
>  struct drm_gpuva_ops *
>  drm_gpuvm_sm_map_ops_create(struct drm_gpuvm *gpuvm,
> -                         u64 req_addr, u64 req_range,
> -                         struct drm_gem_object *req_obj, u64 req_offset)
> +                         const struct drm_gpuvm_map_req *req)
>  {
>       struct drm_gpuva_ops *ops;
>       struct {
> @@ -2661,9 +2644,7 @@ drm_gpuvm_sm_map_ops_create(struct drm_gpuvm *gpuvm,
>       args.vm = gpuvm;
>       args.ops = ops;
>  
> -     ret = __drm_gpuvm_sm_map(gpuvm, &gpuvm_list_ops, &args,
> -                              req_addr, req_range,
> -                              req_obj, req_offset);
> +     ret = __drm_gpuvm_sm_map(gpuvm, &gpuvm_list_ops, &args, req);
>       if (ret)
>               goto err_free_ops;
>  
> diff --git a/drivers/gpu/drm/imagination/pvr_vm.c 
> b/drivers/gpu/drm/imagination/pvr_vm.c
> index 2896fa7501b1..0f6b4cdb5fd8 100644
> --- a/drivers/gpu/drm/imagination/pvr_vm.c
> +++ b/drivers/gpu/drm/imagination/pvr_vm.c
> @@ -185,12 +185,17 @@ struct pvr_vm_bind_op {
>  static int pvr_vm_bind_op_exec(struct pvr_vm_bind_op *bind_op)
>  {
>       switch (bind_op->type) {
> -     case PVR_VM_BIND_TYPE_MAP:
> +     case PVR_VM_BIND_TYPE_MAP: {
> +             const struct drm_gpuvm_map_req map_req = {
> +                     .op_map.va.addr = bind_op->device_addr,
> +                     .op_map.va.range = bind_op->size,
> +                     .op_map.gem.obj = gem_from_pvr_gem(bind_op->pvr_obj),
> +                     .op_map.gem.offset = bind_op->offset,
> +             };
> +
>               return drm_gpuvm_sm_map(&bind_op->vm_ctx->gpuvm_mgr,
> -                                     bind_op, bind_op->device_addr,
> -                                     bind_op->size,
> -                                     gem_from_pvr_gem(bind_op->pvr_obj),
> -                                     bind_op->offset);
> +                                     bind_op, &map_req);
> +     }
>  
>       case PVR_VM_BIND_TYPE_UNMAP:
>               return drm_gpuvm_sm_unmap(&bind_op->vm_ctx->gpuvm_mgr,
> diff --git a/drivers/gpu/drm/msm/msm_gem_vma.c 
> b/drivers/gpu/drm/msm/msm_gem_vma.c
> index 3cd8562a5109..2ca408c40369 100644
> --- a/drivers/gpu/drm/msm/msm_gem_vma.c
> +++ b/drivers/gpu/drm/msm/msm_gem_vma.c
> @@ -1172,10 +1172,17 @@ vm_bind_job_lock_objects(struct msm_vm_bind_job *job, 
> struct drm_exec *exec)
>                               break;
>                       case MSM_VM_BIND_OP_MAP:
>                       case MSM_VM_BIND_OP_MAP_NULL:
> -                             ret = drm_gpuvm_sm_map_exec_lock(job->vm, exec, 
> 1,
> -                                                         op->iova, op->range,
> -                                                         op->obj, 
> op->obj_offset);
> +                     {
> +                             struct drm_gpuvm_map_req map_req = {
> +                                     .op_map.va.addr = op->iova,
> +                                     .op_map.va.range = op->range,
> +                                     .op_map.gem.obj = op->obj,
> +                                     .op_map.gem.offset = op->obj_offset,
> +                             };
> +
> +                             ret = drm_gpuvm_sm_map_exec_lock(job->vm, exec, 
> 1, &map_req);
>                               break;
> +                     }
>                       default:
>                               /*
>                                * lookup_op() should have already thrown an 
> error for
> @@ -1283,9 +1290,17 @@ vm_bind_job_prepare(struct msm_vm_bind_job *job)
>                               arg.flags |= MSM_VMA_DUMP;
>                       fallthrough;
>               case MSM_VM_BIND_OP_MAP_NULL:
> -                     ret = drm_gpuvm_sm_map(job->vm, &arg, op->iova,
> -                                            op->range, op->obj, 
> op->obj_offset);
> +             {
> +                     struct drm_gpuvm_map_req map_req = {
> +                             .op_map.va.addr = op->iova,
> +                             .op_map.va.range = op->range,
> +                             .op_map.gem.obj = op->obj,
> +                             .op_map.gem.offset = op->obj_offset,
> +                     };
> +
> +                     ret = drm_gpuvm_sm_map(job->vm, &arg, &map_req);
>                       break;
> +             }
>               default:
>                       /*
>                        * lookup_op() should have already thrown an error for
> diff --git a/drivers/gpu/drm/nouveau/nouveau_uvmm.c 
> b/drivers/gpu/drm/nouveau/nouveau_uvmm.c
> index ddfc46bc1b3e..92f87520eeb8 100644
> --- a/drivers/gpu/drm/nouveau/nouveau_uvmm.c
> +++ b/drivers/gpu/drm/nouveau/nouveau_uvmm.c
> @@ -1276,6 +1276,12 @@ nouveau_uvmm_bind_job_submit(struct nouveau_job *job,
>                       break;
>               case OP_MAP: {
>                       struct nouveau_uvma_region *reg;
> +                     struct drm_gpuvm_map_req map_req = {
> +                             .op_map.va.addr = op->va.addr,
> +                             .op_map.va.range = op->va.range,
> +                             .op_map.gem.obj = op->gem.obj,
> +                             .op_map.gem.offset = op->gem.offset,
> +                     };
>  
>                       reg = nouveau_uvma_region_find_first(uvmm,
>                                                            op->va.addr,
> @@ -1301,10 +1307,7 @@ nouveau_uvmm_bind_job_submit(struct nouveau_job *job,
>                       }
>  
>                       op->ops = drm_gpuvm_sm_map_ops_create(&uvmm->base,
> -                                                           op->va.addr,
> -                                                           op->va.range,
> -                                                           op->gem.obj,
> -                                                           op->gem.offset);
> +                                                           &map_req);
>                       if (IS_ERR(op->ops)) {
>                               ret = PTR_ERR(op->ops);
>                               goto unwind_continue;
> diff --git a/drivers/gpu/drm/panthor/panthor_mmu.c 
> b/drivers/gpu/drm/panthor/panthor_mmu.c
> index 4140f697ba5a..5ed4573b3a6b 100644
> --- a/drivers/gpu/drm/panthor/panthor_mmu.c
> +++ b/drivers/gpu/drm/panthor/panthor_mmu.c
> @@ -2169,15 +2169,22 @@ panthor_vm_exec_op(struct panthor_vm *vm, struct 
> panthor_vm_op_ctx *op,
>       mutex_lock(&vm->op_lock);
>       vm->op_ctx = op;
>       switch (op_type) {
> -     case DRM_PANTHOR_VM_BIND_OP_TYPE_MAP:
> +     case DRM_PANTHOR_VM_BIND_OP_TYPE_MAP: {
> +             const struct drm_gpuvm_map_req map_req = {
> +                     .op_map.va.addr = op->va.addr,
> +                     .op_map.va.range = op->va.range,
> +                     .op_map.gem.obj = op->map.vm_bo->obj,
> +                     .op_map.gem.offset = op->map.bo_offset,
> +             };
> +
>               if (vm->unusable) {
>                       ret = -EINVAL;
>                       break;
>               }
>  
> -             ret = drm_gpuvm_sm_map(&vm->base, vm, op->va.addr, op->va.range,
> -                                    op->map.vm_bo->obj, op->map.bo_offset);
> +             ret = drm_gpuvm_sm_map(&vm->base, vm, &map_req);
>               break;
> +     }
>  
>       case DRM_PANTHOR_VM_BIND_OP_TYPE_UNMAP:
>               ret = drm_gpuvm_sm_unmap(&vm->base, vm, op->va.addr, 
> op->va.range);
> diff --git a/drivers/gpu/drm/xe/xe_vm.c b/drivers/gpu/drm/xe/xe_vm.c
> index 432ea325677d..9fcc52032a1d 100644
> --- a/drivers/gpu/drm/xe/xe_vm.c
> +++ b/drivers/gpu/drm/xe/xe_vm.c
> @@ -2316,10 +2316,17 @@ vm_bind_ioctl_ops_create(struct xe_vm *vm, struct 
> xe_vma_ops *vops,
>  
>       switch (operation) {
>       case DRM_XE_VM_BIND_OP_MAP:
> -     case DRM_XE_VM_BIND_OP_MAP_USERPTR:
> -             ops = drm_gpuvm_sm_map_ops_create(&vm->gpuvm, addr, range,
> -                                               obj, bo_offset_or_userptr);
> +     case DRM_XE_VM_BIND_OP_MAP_USERPTR: {
> +             struct drm_gpuvm_map_req map_req = {
> +                     .op_map.va.addr = addr,
> +                     .op_map.va.range = range,
> +                     .op_map.gem.obj = obj,
> +                     .op_map.gem.offset = bo_offset_or_userptr,
> +             };
> +
> +             ops = drm_gpuvm_sm_map_ops_create(&vm->gpuvm, &map_req);
>               break;
> +     }
>       case DRM_XE_VM_BIND_OP_UNMAP:
>               ops = drm_gpuvm_sm_unmap_ops_create(&vm->gpuvm, addr, range);
>               break;
> diff --git a/include/drm/drm_gpuvm.h b/include/drm/drm_gpuvm.h
> index 274532facfd6..3cf0a84b8b08 100644
> --- a/include/drm/drm_gpuvm.h
> +++ b/include/drm/drm_gpuvm.h
> @@ -1058,10 +1058,20 @@ struct drm_gpuva_ops {
>   */
>  #define drm_gpuva_next_op(op) list_next_entry(op, entry)
>  
> +/**
> + * struct drm_gpuvm_map_req - arguments passed to 
> drm_gpuvm_sm_map[_ops_create]()
> + */
> +struct drm_gpuvm_map_req {
> +     /**
> +      * @op_map: struct drm_gpuva_op_map
> +      */
> +     struct drm_gpuva_op_map op_map;
> +};
> +
>  struct drm_gpuva_ops *
>  drm_gpuvm_sm_map_ops_create(struct drm_gpuvm *gpuvm,
> -                         u64 addr, u64 range,
> -                         struct drm_gem_object *obj, u64 offset);
> +                         const struct drm_gpuvm_map_req *req);
> +
>  struct drm_gpuva_ops *
>  drm_gpuvm_sm_unmap_ops_create(struct drm_gpuvm *gpuvm,
>                             u64 addr, u64 range);
> @@ -1205,16 +1215,14 @@ struct drm_gpuvm_ops {
>  };
>  
>  int drm_gpuvm_sm_map(struct drm_gpuvm *gpuvm, void *priv,
> -                  u64 addr, u64 range,
> -                  struct drm_gem_object *obj, u64 offset);
> +                  const struct drm_gpuvm_map_req *req);
>  
>  int drm_gpuvm_sm_unmap(struct drm_gpuvm *gpuvm, void *priv,
>                      u64 addr, u64 range);
>  
>  int drm_gpuvm_sm_map_exec_lock(struct drm_gpuvm *gpuvm,
>                         struct drm_exec *exec, unsigned int num_fences,
> -                       u64 req_addr, u64 req_range,
> -                       struct drm_gem_object *obj, u64 offset);
> +                       struct drm_gpuvm_map_req *req);
>  
>  int drm_gpuvm_sm_unmap_exec_lock(struct drm_gpuvm *gpuvm, struct drm_exec 
> *exec,
>                                u64 req_addr, u64 req_range);
> -- 
> 2.34.1
> 

Reply via email to