On Mon, May 24, 2021 at 03:59:50PM -0500, Jason Ekstrand wrote:
> None of these helpers actually leak any RCU details to the caller.  They
> all assume you have a genuine reference, take the RCU read lock, and
> retry if needed.  Naming them with an _rcu is likely to cause callers
> more panic than needed.
> 
> Signed-off-by: Jason Ekstrand <ja...@jlekstrand.net>
> Suggested-by: Daniel Vetter <daniel.vet...@ffwll.ch>

scripts/get_maintainer.pl gives you ideas whom to send this to. Somewhat
applies to most patches here since I don't think you've cc'ed Christian on
the entire pile.
-Daniel

> ---
>  drivers/dma-buf/dma-buf.c                     |  2 +-
>  drivers/dma-buf/dma-resv.c                    | 28 +++++++++----------
>  drivers/gpu/drm/amd/amdgpu/amdgpu_display.c   |  2 +-
>  drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c   |  2 +-
>  drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c       |  2 +-
>  drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c       |  4 +--
>  drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c        |  2 +-
>  drivers/gpu/drm/amd/amdgpu/amdgpu_object.c    |  2 +-
>  drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c       |  2 +-
>  drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c        |  8 +++---
>  .../gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c |  2 +-
>  drivers/gpu/drm/drm_gem.c                     |  6 ++--
>  drivers/gpu/drm/drm_gem_atomic_helper.c       |  2 +-
>  drivers/gpu/drm/etnaviv/etnaviv_gem.c         |  4 +--
>  drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c  |  4 +--
>  drivers/gpu/drm/i915/display/intel_display.c  |  2 +-
>  drivers/gpu/drm/i915/dma_resv_utils.c         |  2 +-
>  drivers/gpu/drm/i915/gem/i915_gem_busy.c      |  2 +-
>  .../gpu/drm/i915/gem/i915_gem_execbuffer.c    |  2 +-
>  drivers/gpu/drm/i915/gem/i915_gem_object.h    |  2 +-
>  drivers/gpu/drm/i915/gem/i915_gem_userptr.c   |  2 +-
>  drivers/gpu/drm/i915/gem/i915_gem_wait.c      | 10 +++----
>  drivers/gpu/drm/i915/i915_request.c           |  4 +--
>  drivers/gpu/drm/i915/i915_sw_fence.c          |  4 +--
>  drivers/gpu/drm/msm/msm_gem.c                 |  2 +-
>  drivers/gpu/drm/nouveau/dispnv50/wndw.c       |  2 +-
>  drivers/gpu/drm/nouveau/nouveau_gem.c         |  2 +-
>  drivers/gpu/drm/panfrost/panfrost_drv.c       |  2 +-
>  drivers/gpu/drm/panfrost/panfrost_job.c       |  2 +-
>  drivers/gpu/drm/radeon/radeon_gem.c           |  6 ++--
>  drivers/gpu/drm/radeon/radeon_mn.c            |  2 +-
>  drivers/gpu/drm/ttm/ttm_bo.c                  | 12 ++++----
>  drivers/gpu/drm/vgem/vgem_fence.c             |  2 +-
>  drivers/gpu/drm/virtio/virtgpu_ioctl.c        |  4 +--
>  drivers/gpu/drm/vmwgfx/vmwgfx_bo.c            |  2 +-
>  include/linux/dma-resv.h                      | 18 ++++++------
>  36 files changed, 79 insertions(+), 79 deletions(-)
> 
> diff --git a/drivers/dma-buf/dma-buf.c b/drivers/dma-buf/dma-buf.c
> index f264b70c383eb..d4529aa9d1a5a 100644
> --- a/drivers/dma-buf/dma-buf.c
> +++ b/drivers/dma-buf/dma-buf.c
> @@ -1147,7 +1147,7 @@ static int __dma_buf_begin_cpu_access(struct dma_buf 
> *dmabuf,
>       long ret;
>  
>       /* Wait on any implicit rendering fences */
> -     ret = dma_resv_wait_timeout_rcu(resv, write, true,
> +     ret = dma_resv_wait_timeout_unlocked(resv, write, true,
>                                                 MAX_SCHEDULE_TIMEOUT);
>       if (ret < 0)
>               return ret;
> diff --git a/drivers/dma-buf/dma-resv.c b/drivers/dma-buf/dma-resv.c
> index 6ddbeb5dfbf65..d6f1ed4cd4d55 100644
> --- a/drivers/dma-buf/dma-resv.c
> +++ b/drivers/dma-buf/dma-resv.c
> @@ -417,7 +417,7 @@ int dma_resv_copy_fences(struct dma_resv *dst, struct 
> dma_resv *src)
>  EXPORT_SYMBOL(dma_resv_copy_fences);
>  
>  /**
> - * dma_resv_get_fences_rcu - Get an object's shared and exclusive
> + * dma_resv_get_fences_unlocked - Get an object's shared and exclusive
>   * fences without update side lock held
>   * @obj: the reservation object
>   * @pfence_excl: the returned exclusive fence (or NULL)
> @@ -429,10 +429,10 @@ EXPORT_SYMBOL(dma_resv_copy_fences);
>   * exclusive fence is not specified the fence is put into the array of the
>   * shared fences as well. Returns either zero or -ENOMEM.
>   */
> -int dma_resv_get_fences_rcu(struct dma_resv *obj,
> -                         struct dma_fence **pfence_excl,
> -                         unsigned *pshared_count,
> -                         struct dma_fence ***pshared)
> +int dma_resv_get_fences_unlocked(struct dma_resv *obj,
> +                              struct dma_fence **pfence_excl,
> +                              unsigned *pshared_count,
> +                              struct dma_fence ***pshared)
>  {
>       struct dma_fence **shared = NULL;
>       struct dma_fence *fence_excl;
> @@ -515,10 +515,10 @@ int dma_resv_get_fences_rcu(struct dma_resv *obj,
>       *pshared = shared;
>       return ret;
>  }
> -EXPORT_SYMBOL_GPL(dma_resv_get_fences_rcu);
> +EXPORT_SYMBOL_GPL(dma_resv_get_fences_unlocked);
>  
>  /**
> - * dma_resv_wait_timeout_rcu - Wait on reservation's objects
> + * dma_resv_wait_timeout_unlocked - Wait on reservation's objects
>   * shared and/or exclusive fences.
>   * @obj: the reservation object
>   * @wait_all: if true, wait on all fences, else wait on just exclusive fence
> @@ -529,9 +529,9 @@ EXPORT_SYMBOL_GPL(dma_resv_get_fences_rcu);
>   * Returns -ERESTARTSYS if interrupted, 0 if the wait timed out, or
>   * greater than zer on success.
>   */
> -long dma_resv_wait_timeout_rcu(struct dma_resv *obj,
> -                            bool wait_all, bool intr,
> -                            unsigned long timeout)
> +long dma_resv_wait_timeout_unlocked(struct dma_resv *obj,
> +                                 bool wait_all, bool intr,
> +                                 unsigned long timeout)
>  {
>       struct dma_fence *fence;
>       unsigned seq, shared_count;
> @@ -602,7 +602,7 @@ long dma_resv_wait_timeout_rcu(struct dma_resv *obj,
>       rcu_read_unlock();
>       goto retry;
>  }
> -EXPORT_SYMBOL_GPL(dma_resv_wait_timeout_rcu);
> +EXPORT_SYMBOL_GPL(dma_resv_wait_timeout_unlocked);
>  
>  
>  static inline int dma_resv_test_signaled_single(struct dma_fence 
> *passed_fence)
> @@ -622,7 +622,7 @@ static inline int dma_resv_test_signaled_single(struct 
> dma_fence *passed_fence)
>  }
>  
>  /**
> - * dma_resv_test_signaled_rcu - Test if a reservation object's
> + * dma_resv_test_signaled_unlocked - Test if a reservation object's
>   * fences have been signaled.
>   * @obj: the reservation object
>   * @test_all: if true, test all fences, otherwise only test the exclusive
> @@ -631,7 +631,7 @@ static inline int dma_resv_test_signaled_single(struct 
> dma_fence *passed_fence)
>   * RETURNS
>   * true if all fences signaled, else false
>   */
> -bool dma_resv_test_signaled_rcu(struct dma_resv *obj, bool test_all)
> +bool dma_resv_test_signaled_unlocked(struct dma_resv *obj, bool test_all)
>  {
>       unsigned seq, shared_count;
>       int ret;
> @@ -680,4 +680,4 @@ bool dma_resv_test_signaled_rcu(struct dma_resv *obj, 
> bool test_all)
>       rcu_read_unlock();
>       return ret;
>  }
> -EXPORT_SYMBOL_GPL(dma_resv_test_signaled_rcu);
> +EXPORT_SYMBOL_GPL(dma_resv_test_signaled_unlocked);
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_display.c 
> b/drivers/gpu/drm/amd/amdgpu/amdgpu_display.c
> index 8a1fb8b6606e5..3b0df434e0ca3 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_display.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_display.c
> @@ -203,7 +203,7 @@ int amdgpu_display_crtc_page_flip_target(struct drm_crtc 
> *crtc,
>               goto unpin;
>       }
>  
> -     r = dma_resv_get_fences_rcu(new_abo->tbo.base.resv, &work->excl,
> +     r = dma_resv_get_fences_unlocked(new_abo->tbo.base.resv, &work->excl,
>                                             &work->shared_count,
>                                             &work->shared);
>       if (unlikely(r != 0)) {
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c 
> b/drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c
> index baa980a477d94..0d0319bc51577 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c
> @@ -98,7 +98,7 @@ __dma_resv_make_exclusive(struct dma_resv *obj)
>       if (!dma_resv_get_list(obj)) /* no shared fences to convert */
>               return 0;
>  
> -     r = dma_resv_get_fences_rcu(obj, NULL, &count, &fences);
> +     r = dma_resv_get_fences_unlocked(obj, NULL, &count, &fences);
>       if (r)
>               return r;
>  
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 
> b/drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
> index 18974bd081f00..a71f98ae1d72f 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
> @@ -471,7 +471,7 @@ int amdgpu_gem_wait_idle_ioctl(struct drm_device *dev, 
> void *data,
>               return -ENOENT;
>       }
>       robj = gem_to_amdgpu_bo(gobj);
> -     ret = dma_resv_wait_timeout_rcu(robj->tbo.base.resv, true, true,
> +     ret = dma_resv_wait_timeout_unlocked(robj->tbo.base.resv, true, true,
>                                                 timeout);
>  
>       /* ret == 0 means not signaled,
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 
> b/drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c
> index b4971e90b98cf..7045b104a33ae 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c
> @@ -112,7 +112,7 @@ void amdgpu_pasid_free_delayed(struct dma_resv *resv,
>       unsigned count;
>       int r;
>  
> -     r = dma_resv_get_fences_rcu(resv, NULL, &count, &fences);
> +     r = dma_resv_get_fences_unlocked(resv, NULL, &count, &fences);
>       if (r)
>               goto fallback;
>  
> @@ -156,7 +156,7 @@ void amdgpu_pasid_free_delayed(struct dma_resv *resv,
>       /* Not enough memory for the delayed delete, as last resort
>        * block for all the fences to complete.
>        */
> -     dma_resv_wait_timeout_rcu(resv, true, false,
> +     dma_resv_wait_timeout_unlocked(resv, true, false,
>                                           MAX_SCHEDULE_TIMEOUT);
>       amdgpu_pasid_free(pasid);
>  }
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 
> b/drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c
> index 828b5167ff128..58fb1de81c0c5 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c
> @@ -75,7 +75,7 @@ static bool amdgpu_mn_invalidate_gfx(struct 
> mmu_interval_notifier *mni,
>  
>       mmu_interval_set_seq(mni, cur_seq);
>  
> -     r = dma_resv_wait_timeout_rcu(bo->tbo.base.resv, true, false,
> +     r = dma_resv_wait_timeout_unlocked(bo->tbo.base.resv, true, false,
>                                     MAX_SCHEDULE_TIMEOUT);
>       mutex_unlock(&adev->notifier_lock);
>       if (r <= 0)
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 
> b/drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
> index 0adffcace3263..81db9ea391c1c 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
> @@ -791,7 +791,7 @@ int amdgpu_bo_kmap(struct amdgpu_bo *bo, void **ptr)
>               return 0;
>       }
>  
> -     r = dma_resv_wait_timeout_rcu(bo->tbo.base.resv, false, false,
> +     r = dma_resv_wait_timeout_unlocked(bo->tbo.base.resv, false, false,
>                                               MAX_SCHEDULE_TIMEOUT);
>       if (r < 0)
>               return r;
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c 
> b/drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c
> index c6dbc08016045..af7b667d3226d 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c
> @@ -1115,7 +1115,7 @@ static int amdgpu_uvd_send_msg(struct amdgpu_ring 
> *ring, struct amdgpu_bo *bo,
>       ib->length_dw = 16;
>  
>       if (direct) {
> -             r = dma_resv_wait_timeout_rcu(bo->tbo.base.resv,
> +             r = dma_resv_wait_timeout_unlocked(bo->tbo.base.resv,
>                                                       true, false,
>                                                       msecs_to_jiffies(10));
>               if (r == 0)
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 
> b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
> index 4a3e3f72e1277..33dbe3fcaf706 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
> @@ -2007,13 +2007,13 @@ static void amdgpu_vm_prt_fini(struct amdgpu_device 
> *adev, struct amdgpu_vm *vm)
>       unsigned i, shared_count;
>       int r;
>  
> -     r = dma_resv_get_fences_rcu(resv, &excl,
> +     r = dma_resv_get_fences_unlocked(resv, &excl,
>                                             &shared_count, &shared);
>       if (r) {
>               /* Not enough memory to grab the fence list, as last resort
>                * block for all the fences to complete.
>                */
> -             dma_resv_wait_timeout_rcu(resv, true, false,
> +             dma_resv_wait_timeout_unlocked(resv, true, false,
>                                                   MAX_SCHEDULE_TIMEOUT);
>               return;
>       }
> @@ -2625,7 +2625,7 @@ bool amdgpu_vm_evictable(struct amdgpu_bo *bo)
>               return true;
>  
>       /* Don't evict VM page tables while they are busy */
> -     if (!dma_resv_test_signaled_rcu(bo->tbo.base.resv, true))
> +     if (!dma_resv_test_signaled_unlocked(bo->tbo.base.resv, true))
>               return false;
>  
>       /* Try to block ongoing updates */
> @@ -2805,7 +2805,7 @@ void amdgpu_vm_adjust_size(struct amdgpu_device *adev, 
> uint32_t min_vm_size,
>   */
>  long amdgpu_vm_wait_idle(struct amdgpu_vm *vm, long timeout)
>  {
> -     timeout = dma_resv_wait_timeout_rcu(vm->root.base.bo->tbo.base.resv,
> +     timeout = 
> dma_resv_wait_timeout_unlocked(vm->root.base.bo->tbo.base.resv,
>                                           true, true, timeout);
>       if (timeout <= 0)
>               return timeout;
> diff --git a/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c 
> b/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
> index 9ca517b658546..e74fef044b301 100644
> --- a/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
> +++ b/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
> @@ -8276,7 +8276,7 @@ static void amdgpu_dm_commit_planes(struct 
> drm_atomic_state *state,
>                * deadlock during GPU reset when this fence will not signal
>                * but we hold reservation lock for the BO.
>                */
> -             r = dma_resv_wait_timeout_rcu(abo->tbo.base.resv, true,
> +             r = dma_resv_wait_timeout_unlocked(abo->tbo.base.resv, true,
>                                                       false,
>                                                       msecs_to_jiffies(5000));
>               if (unlikely(r <= 0))
> diff --git a/drivers/gpu/drm/drm_gem.c b/drivers/gpu/drm/drm_gem.c
> index 9989425e9875a..42a432708c2fe 100644
> --- a/drivers/gpu/drm/drm_gem.c
> +++ b/drivers/gpu/drm/drm_gem.c
> @@ -770,7 +770,7 @@ long drm_gem_dma_resv_wait(struct drm_file *filep, u32 
> handle,
>               return -EINVAL;
>       }
>  
> -     ret = dma_resv_wait_timeout_rcu(obj->resv, wait_all,
> +     ret = dma_resv_wait_timeout_unlocked(obj->resv, wait_all,
>                                                 true, timeout);
>       if (ret == 0)
>               ret = -ETIME;
> @@ -1375,12 +1375,12 @@ int drm_gem_fence_array_add_implicit(struct xarray 
> *fence_array,
>  
>       if (!write) {
>               struct dma_fence *fence =
> -                     dma_resv_get_excl_rcu(obj->resv);
> +                     dma_resv_get_excl_unlocked(obj->resv);
>  
>               return drm_gem_fence_array_add(fence_array, fence);
>       }
>  
> -     ret = dma_resv_get_fences_rcu(obj->resv, NULL,
> +     ret = dma_resv_get_fences_unlocked(obj->resv, NULL,
>                                               &fence_count, &fences);
>       if (ret || !fence_count)
>               return ret;
> diff --git a/drivers/gpu/drm/drm_gem_atomic_helper.c 
> b/drivers/gpu/drm/drm_gem_atomic_helper.c
> index a005c5a0ba46a..a27135084ae5c 100644
> --- a/drivers/gpu/drm/drm_gem_atomic_helper.c
> +++ b/drivers/gpu/drm/drm_gem_atomic_helper.c
> @@ -147,7 +147,7 @@ int drm_gem_plane_helper_prepare_fb(struct drm_plane 
> *plane, struct drm_plane_st
>               return 0;
>  
>       obj = drm_gem_fb_get_obj(state->fb, 0);
> -     fence = dma_resv_get_excl_rcu(obj->resv);
> +     fence = dma_resv_get_excl_unlocked(obj->resv);
>       drm_atomic_set_fence_for_plane(state, fence);
>  
>       return 0;
> diff --git a/drivers/gpu/drm/etnaviv/etnaviv_gem.c 
> b/drivers/gpu/drm/etnaviv/etnaviv_gem.c
> index db69f19ab5bca..b271e00480246 100644
> --- a/drivers/gpu/drm/etnaviv/etnaviv_gem.c
> +++ b/drivers/gpu/drm/etnaviv/etnaviv_gem.c
> @@ -390,13 +390,13 @@ int etnaviv_gem_cpu_prep(struct drm_gem_object *obj, 
> u32 op,
>       }
>  
>       if (op & ETNA_PREP_NOSYNC) {
> -             if (!dma_resv_test_signaled_rcu(obj->resv,
> +             if (!dma_resv_test_signaled_unlocked(obj->resv,
>                                                         write))
>                       return -EBUSY;
>       } else {
>               unsigned long remain = etnaviv_timeout_to_jiffies(timeout);
>  
> -             ret = dma_resv_wait_timeout_rcu(obj->resv,
> +             ret = dma_resv_wait_timeout_unlocked(obj->resv,
>                                                         write, true, remain);
>               if (ret <= 0)
>                       return ret == 0 ? -ETIMEDOUT : ret;
> diff --git a/drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c 
> b/drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c
> index d05c359945799..b4ac4c7ab144d 100644
> --- a/drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c
> +++ b/drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c
> @@ -189,13 +189,13 @@ static int submit_fence_sync(struct etnaviv_gem_submit 
> *submit)
>                       continue;
>  
>               if (bo->flags & ETNA_SUBMIT_BO_WRITE) {
> -                     ret = dma_resv_get_fences_rcu(robj, &bo->excl,
> +                     ret = dma_resv_get_fences_unlocked(robj, &bo->excl,
>                                                               &bo->nr_shared,
>                                                               &bo->shared);
>                       if (ret)
>                               return ret;
>               } else {
> -                     bo->excl = dma_resv_get_excl_rcu(robj);
> +                     bo->excl = dma_resv_get_excl_unlocked(robj);
>               }
>  
>       }
> diff --git a/drivers/gpu/drm/i915/display/intel_display.c 
> b/drivers/gpu/drm/i915/display/intel_display.c
> index 422b59ebf6dce..5f0b85a102159 100644
> --- a/drivers/gpu/drm/i915/display/intel_display.c
> +++ b/drivers/gpu/drm/i915/display/intel_display.c
> @@ -11040,7 +11040,7 @@ intel_prepare_plane_fb(struct drm_plane *_plane,
>               if (ret < 0)
>                       goto unpin_fb;
>  
> -             fence = dma_resv_get_excl_rcu(obj->base.resv);
> +             fence = dma_resv_get_excl_unlocked(obj->base.resv);
>               if (fence) {
>                       add_rps_boost_after_vblank(new_plane_state->hw.crtc,
>                                                  fence);
> diff --git a/drivers/gpu/drm/i915/dma_resv_utils.c 
> b/drivers/gpu/drm/i915/dma_resv_utils.c
> index 9e508e7d4629f..bdfc6bf16a4e9 100644
> --- a/drivers/gpu/drm/i915/dma_resv_utils.c
> +++ b/drivers/gpu/drm/i915/dma_resv_utils.c
> @@ -10,7 +10,7 @@
>  void dma_resv_prune(struct dma_resv *resv)
>  {
>       if (dma_resv_trylock(resv)) {
> -             if (dma_resv_test_signaled_rcu(resv, true))
> +             if (dma_resv_test_signaled_unlocked(resv, true))
>                       dma_resv_add_excl_fence(resv, NULL);
>               dma_resv_unlock(resv);
>       }
> diff --git a/drivers/gpu/drm/i915/gem/i915_gem_busy.c 
> b/drivers/gpu/drm/i915/gem/i915_gem_busy.c
> index 25235ef630c10..754ad6d1bace9 100644
> --- a/drivers/gpu/drm/i915/gem/i915_gem_busy.c
> +++ b/drivers/gpu/drm/i915/gem/i915_gem_busy.c
> @@ -105,7 +105,7 @@ i915_gem_busy_ioctl(struct drm_device *dev, void *data,
>        * Alternatively, we can trade that extra information on read/write
>        * activity with
>        *      args->busy =
> -      *              !dma_resv_test_signaled_rcu(obj->resv, true);
> +      *              !dma_resv_test_signaled_unlocked(obj->resv, true);
>        * to report the overall busyness. This is what the wait-ioctl does.
>        *
>        */
> diff --git a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 
> b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
> index 297143511f99b..e8f323564e57b 100644
> --- a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
> +++ b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
> @@ -1481,7 +1481,7 @@ static inline bool use_reloc_gpu(struct i915_vma *vma)
>       if (DBG_FORCE_RELOC)
>               return false;
>  
> -     return !dma_resv_test_signaled_rcu(vma->resv, true);
> +     return !dma_resv_test_signaled_unlocked(vma->resv, true);
>  }
>  
>  static unsigned long vma_phys_addr(struct i915_vma *vma, u32 offset)
> diff --git a/drivers/gpu/drm/i915/gem/i915_gem_object.h 
> b/drivers/gpu/drm/i915/gem/i915_gem_object.h
> index 2ebd79537aea9..7c0eb425cb3b3 100644
> --- a/drivers/gpu/drm/i915/gem/i915_gem_object.h
> +++ b/drivers/gpu/drm/i915/gem/i915_gem_object.h
> @@ -500,7 +500,7 @@ i915_gem_object_last_write_engine(struct 
> drm_i915_gem_object *obj)
>       struct dma_fence *fence;
>  
>       rcu_read_lock();
> -     fence = dma_resv_get_excl_rcu(obj->base.resv);
> +     fence = dma_resv_get_excl_unlocked(obj->base.resv);
>       rcu_read_unlock();
>  
>       if (fence && dma_fence_is_i915(fence) && !dma_fence_is_signaled(fence))
> diff --git a/drivers/gpu/drm/i915/gem/i915_gem_userptr.c 
> b/drivers/gpu/drm/i915/gem/i915_gem_userptr.c
> index a657b99ec7606..bb5f44ed932aa 100644
> --- a/drivers/gpu/drm/i915/gem/i915_gem_userptr.c
> +++ b/drivers/gpu/drm/i915/gem/i915_gem_userptr.c
> @@ -85,7 +85,7 @@ static bool i915_gem_userptr_invalidate(struct 
> mmu_interval_notifier *mni,
>               return true;
>  
>       /* we will unbind on next submission, still have userptr pins */
> -     r = dma_resv_wait_timeout_rcu(obj->base.resv, true, false,
> +     r = dma_resv_wait_timeout_unlocked(obj->base.resv, true, false,
>                                     MAX_SCHEDULE_TIMEOUT);
>       if (r <= 0)
>               drm_err(&i915->drm, "(%ld) failed to wait for idle\n", r);
> diff --git a/drivers/gpu/drm/i915/gem/i915_gem_wait.c 
> b/drivers/gpu/drm/i915/gem/i915_gem_wait.c
> index 4b9856d5ba14f..5b6c52659ad4d 100644
> --- a/drivers/gpu/drm/i915/gem/i915_gem_wait.c
> +++ b/drivers/gpu/drm/i915/gem/i915_gem_wait.c
> @@ -45,7 +45,7 @@ i915_gem_object_wait_reservation(struct dma_resv *resv,
>               unsigned int count, i;
>               int ret;
>  
> -             ret = dma_resv_get_fences_rcu(resv, &excl, &count, &shared);
> +             ret = dma_resv_get_fences_unlocked(resv, &excl, &count, 
> &shared);
>               if (ret)
>                       return ret;
>  
> @@ -73,7 +73,7 @@ i915_gem_object_wait_reservation(struct dma_resv *resv,
>                */
>               prune_fences = count && timeout >= 0;
>       } else {
> -             excl = dma_resv_get_excl_rcu(resv);
> +             excl = dma_resv_get_excl_unlocked(resv);
>       }
>  
>       if (excl && timeout >= 0)
> @@ -158,8 +158,8 @@ i915_gem_object_wait_priority(struct drm_i915_gem_object 
> *obj,
>               unsigned int count, i;
>               int ret;
>  
> -             ret = dma_resv_get_fences_rcu(obj->base.resv,
> -                                           &excl, &count, &shared);
> +             ret = dma_resv_get_fences_unlocked(obj->base.resv,
> +                                                &excl, &count, &shared);
>               if (ret)
>                       return ret;
>  
> @@ -170,7 +170,7 @@ i915_gem_object_wait_priority(struct drm_i915_gem_object 
> *obj,
>  
>               kfree(shared);
>       } else {
> -             excl = dma_resv_get_excl_rcu(obj->base.resv);
> +             excl = dma_resv_get_excl_unlocked(obj->base.resv);
>       }
>  
>       if (excl) {
> diff --git a/drivers/gpu/drm/i915/i915_request.c 
> b/drivers/gpu/drm/i915/i915_request.c
> index 970d8f4986bbe..d101d702fbadc 100644
> --- a/drivers/gpu/drm/i915/i915_request.c
> +++ b/drivers/gpu/drm/i915/i915_request.c
> @@ -1594,7 +1594,7 @@ i915_request_await_object(struct i915_request *to,
>               struct dma_fence **shared;
>               unsigned int count, i;
>  
> -             ret = dma_resv_get_fences_rcu(obj->base.resv,
> +             ret = dma_resv_get_fences_unlocked(obj->base.resv,
>                                                       &excl, &count, &shared);
>               if (ret)
>                       return ret;
> @@ -1611,7 +1611,7 @@ i915_request_await_object(struct i915_request *to,
>                       dma_fence_put(shared[i]);
>               kfree(shared);
>       } else {
> -             excl = dma_resv_get_excl_rcu(obj->base.resv);
> +             excl = dma_resv_get_excl_unlocked(obj->base.resv);
>       }
>  
>       if (excl) {
> diff --git a/drivers/gpu/drm/i915/i915_sw_fence.c 
> b/drivers/gpu/drm/i915/i915_sw_fence.c
> index 2744558f30507..0bcb7ea44201e 100644
> --- a/drivers/gpu/drm/i915/i915_sw_fence.c
> +++ b/drivers/gpu/drm/i915/i915_sw_fence.c
> @@ -582,7 +582,7 @@ int i915_sw_fence_await_reservation(struct i915_sw_fence 
> *fence,
>               struct dma_fence **shared;
>               unsigned int count, i;
>  
> -             ret = dma_resv_get_fences_rcu(resv, &excl, &count, &shared);
> +             ret = dma_resv_get_fences_unlocked(resv, &excl, &count, 
> &shared);
>               if (ret)
>                       return ret;
>  
> @@ -606,7 +606,7 @@ int i915_sw_fence_await_reservation(struct i915_sw_fence 
> *fence,
>                       dma_fence_put(shared[i]);
>               kfree(shared);
>       } else {
> -             excl = dma_resv_get_excl_rcu(resv);
> +             excl = dma_resv_get_excl_unlocked(resv);
>       }
>  
>       if (ret >= 0 && excl && excl->ops != exclude) {
> diff --git a/drivers/gpu/drm/msm/msm_gem.c b/drivers/gpu/drm/msm/msm_gem.c
> index 56df86e5f7400..f8afb081770e4 100644
> --- a/drivers/gpu/drm/msm/msm_gem.c
> +++ b/drivers/gpu/drm/msm/msm_gem.c
> @@ -915,7 +915,7 @@ int msm_gem_cpu_prep(struct drm_gem_object *obj, uint32_t 
> op, ktime_t *timeout)
>               op & MSM_PREP_NOSYNC ? 0 : timeout_to_jiffies(timeout);
>       long ret;
>  
> -     ret = dma_resv_wait_timeout_rcu(obj->resv, write,
> +     ret = dma_resv_wait_timeout_unlocked(obj->resv, write,
>                                                 true,  remain);
>       if (ret == 0)
>               return remain == 0 ? -EBUSY : -ETIMEDOUT;
> diff --git a/drivers/gpu/drm/nouveau/dispnv50/wndw.c 
> b/drivers/gpu/drm/nouveau/dispnv50/wndw.c
> index 0cb1f9d848d3e..8d048bacd6f02 100644
> --- a/drivers/gpu/drm/nouveau/dispnv50/wndw.c
> +++ b/drivers/gpu/drm/nouveau/dispnv50/wndw.c
> @@ -561,7 +561,7 @@ nv50_wndw_prepare_fb(struct drm_plane *plane, struct 
> drm_plane_state *state)
>                       asyw->image.handle[0] = ctxdma->object.handle;
>       }
>  
> -     asyw->state.fence = dma_resv_get_excl_rcu(nvbo->bo.base.resv);
> +     asyw->state.fence = dma_resv_get_excl_unlocked(nvbo->bo.base.resv);
>       asyw->image.offset[0] = nvbo->offset;
>  
>       if (wndw->func->prepare) {
> diff --git a/drivers/gpu/drm/nouveau/nouveau_gem.c 
> b/drivers/gpu/drm/nouveau/nouveau_gem.c
> index a70e82413fa75..06ea1fed02467 100644
> --- a/drivers/gpu/drm/nouveau/nouveau_gem.c
> +++ b/drivers/gpu/drm/nouveau/nouveau_gem.c
> @@ -928,7 +928,7 @@ nouveau_gem_ioctl_cpu_prep(struct drm_device *dev, void 
> *data,
>               return -ENOENT;
>       nvbo = nouveau_gem_object(gem);
>  
> -     lret = dma_resv_wait_timeout_rcu(nvbo->bo.base.resv, write, true,
> +     lret = dma_resv_wait_timeout_unlocked(nvbo->bo.base.resv, write, true,
>                                                  no_wait ? 0 : 30 * HZ);
>       if (!lret)
>               ret = -EBUSY;
> diff --git a/drivers/gpu/drm/panfrost/panfrost_drv.c 
> b/drivers/gpu/drm/panfrost/panfrost_drv.c
> index ca07098a61419..53e1842fe8bf8 100644
> --- a/drivers/gpu/drm/panfrost/panfrost_drv.c
> +++ b/drivers/gpu/drm/panfrost/panfrost_drv.c
> @@ -311,7 +311,7 @@ panfrost_ioctl_wait_bo(struct drm_device *dev, void *data,
>       if (!gem_obj)
>               return -ENOENT;
>  
> -     ret = dma_resv_wait_timeout_rcu(gem_obj->resv, true,
> +     ret = dma_resv_wait_timeout_unlocked(gem_obj->resv, true,
>                                                 true, timeout);
>       if (!ret)
>               ret = timeout ? -ETIMEDOUT : -EBUSY;
> diff --git a/drivers/gpu/drm/panfrost/panfrost_job.c 
> b/drivers/gpu/drm/panfrost/panfrost_job.c
> index 6003cfeb13221..2df3e999a38d0 100644
> --- a/drivers/gpu/drm/panfrost/panfrost_job.c
> +++ b/drivers/gpu/drm/panfrost/panfrost_job.c
> @@ -203,7 +203,7 @@ static void panfrost_acquire_object_fences(struct 
> drm_gem_object **bos,
>       int i;
>  
>       for (i = 0; i < bo_count; i++)
> -             implicit_fences[i] = dma_resv_get_excl_rcu(bos[i]->resv);
> +             implicit_fences[i] = dma_resv_get_excl_unlocked(bos[i]->resv);
>  }
>  
>  static void panfrost_attach_object_fences(struct drm_gem_object **bos,
> diff --git a/drivers/gpu/drm/radeon/radeon_gem.c 
> b/drivers/gpu/drm/radeon/radeon_gem.c
> index 05ea2f39f6261..1a38b0bf36d11 100644
> --- a/drivers/gpu/drm/radeon/radeon_gem.c
> +++ b/drivers/gpu/drm/radeon/radeon_gem.c
> @@ -125,7 +125,7 @@ static int radeon_gem_set_domain(struct drm_gem_object 
> *gobj,
>       }
>       if (domain == RADEON_GEM_DOMAIN_CPU) {
>               /* Asking for cpu access wait for object idle */
> -             r = dma_resv_wait_timeout_rcu(robj->tbo.base.resv, true, true, 
> 30 * HZ);
> +             r = dma_resv_wait_timeout_unlocked(robj->tbo.base.resv, true, 
> true, 30 * HZ);
>               if (!r)
>                       r = -EBUSY;
>  
> @@ -474,7 +474,7 @@ int radeon_gem_busy_ioctl(struct drm_device *dev, void 
> *data,
>       }
>       robj = gem_to_radeon_bo(gobj);
>  
> -     r = dma_resv_test_signaled_rcu(robj->tbo.base.resv, true);
> +     r = dma_resv_test_signaled_unlocked(robj->tbo.base.resv, true);
>       if (r == 0)
>               r = -EBUSY;
>       else
> @@ -503,7 +503,7 @@ int radeon_gem_wait_idle_ioctl(struct drm_device *dev, 
> void *data,
>       }
>       robj = gem_to_radeon_bo(gobj);
>  
> -     ret = dma_resv_wait_timeout_rcu(robj->tbo.base.resv, true, true, 30 * 
> HZ);
> +     ret = dma_resv_wait_timeout_unlocked(robj->tbo.base.resv, true, true, 
> 30 * HZ);
>       if (ret == 0)
>               r = -EBUSY;
>       else if (ret < 0)
> diff --git a/drivers/gpu/drm/radeon/radeon_mn.c 
> b/drivers/gpu/drm/radeon/radeon_mn.c
> index e37c9a57a7c36..8016c9e568684 100644
> --- a/drivers/gpu/drm/radeon/radeon_mn.c
> +++ b/drivers/gpu/drm/radeon/radeon_mn.c
> @@ -66,7 +66,7 @@ static bool radeon_mn_invalidate(struct 
> mmu_interval_notifier *mn,
>               return true;
>       }
>  
> -     r = dma_resv_wait_timeout_rcu(bo->tbo.base.resv, true, false,
> +     r = dma_resv_wait_timeout_unlocked(bo->tbo.base.resv, true, false,
>                                     MAX_SCHEDULE_TIMEOUT);
>       if (r <= 0)
>               DRM_ERROR("(%ld) failed to wait for user bo\n", r);
> diff --git a/drivers/gpu/drm/ttm/ttm_bo.c b/drivers/gpu/drm/ttm/ttm_bo.c
> index ca1b098b6a561..6925de3f179e8 100644
> --- a/drivers/gpu/drm/ttm/ttm_bo.c
> +++ b/drivers/gpu/drm/ttm/ttm_bo.c
> @@ -294,7 +294,7 @@ static int ttm_bo_cleanup_refs(struct ttm_buffer_object 
> *bo,
>       struct dma_resv *resv = &bo->base._resv;
>       int ret;
>  
> -     if (dma_resv_test_signaled_rcu(resv, true))
> +     if (dma_resv_test_signaled_unlocked(resv, true))
>               ret = 0;
>       else
>               ret = -EBUSY;
> @@ -306,7 +306,7 @@ static int ttm_bo_cleanup_refs(struct ttm_buffer_object 
> *bo,
>                       dma_resv_unlock(bo->base.resv);
>               spin_unlock(&bo->bdev->lru_lock);
>  
> -             lret = dma_resv_wait_timeout_rcu(resv, true, interruptible,
> +             lret = dma_resv_wait_timeout_unlocked(resv, true, interruptible,
>                                                30 * HZ);
>  
>               if (lret < 0)
> @@ -409,7 +409,7 @@ static void ttm_bo_release(struct kref *kref)
>                       /* Last resort, if we fail to allocate memory for the
>                        * fences block for the BO to become idle
>                        */
> -                     dma_resv_wait_timeout_rcu(bo->base.resv, true, false,
> +                     dma_resv_wait_timeout_unlocked(bo->base.resv, true, 
> false,
>                                                 30 * HZ);
>               }
>  
> @@ -420,7 +420,7 @@ static void ttm_bo_release(struct kref *kref)
>               ttm_mem_io_free(bdev, &bo->mem);
>       }
>  
> -     if (!dma_resv_test_signaled_rcu(bo->base.resv, true) ||
> +     if (!dma_resv_test_signaled_unlocked(bo->base.resv, true) ||
>           !dma_resv_trylock(bo->base.resv)) {
>               /* The BO is not idle, resurrect it for delayed destroy */
>               ttm_bo_flush_all_fences(bo);
> @@ -1116,13 +1116,13 @@ int ttm_bo_wait(struct ttm_buffer_object *bo,
>       long timeout = 15 * HZ;
>  
>       if (no_wait) {
> -             if (dma_resv_test_signaled_rcu(bo->base.resv, true))
> +             if (dma_resv_test_signaled_unlocked(bo->base.resv, true))
>                       return 0;
>               else
>                       return -EBUSY;
>       }
>  
> -     timeout = dma_resv_wait_timeout_rcu(bo->base.resv, true,
> +     timeout = dma_resv_wait_timeout_unlocked(bo->base.resv, true,
>                                                     interruptible, timeout);
>       if (timeout < 0)
>               return timeout;
> diff --git a/drivers/gpu/drm/vgem/vgem_fence.c 
> b/drivers/gpu/drm/vgem/vgem_fence.c
> index 2902dc6e64faf..110927edb9df7 100644
> --- a/drivers/gpu/drm/vgem/vgem_fence.c
> +++ b/drivers/gpu/drm/vgem/vgem_fence.c
> @@ -151,7 +151,7 @@ int vgem_fence_attach_ioctl(struct drm_device *dev,
>  
>       /* Check for a conflicting fence */
>       resv = obj->resv;
> -     if (!dma_resv_test_signaled_rcu(resv,
> +     if (!dma_resv_test_signaled_unlocked(resv,
>                                                 arg->flags & 
> VGEM_FENCE_WRITE)) {
>               ret = -EBUSY;
>               goto err_fence;
> diff --git a/drivers/gpu/drm/virtio/virtgpu_ioctl.c 
> b/drivers/gpu/drm/virtio/virtgpu_ioctl.c
> index 669f2ee395154..763a51686819c 100644
> --- a/drivers/gpu/drm/virtio/virtgpu_ioctl.c
> +++ b/drivers/gpu/drm/virtio/virtgpu_ioctl.c
> @@ -451,9 +451,9 @@ static int virtio_gpu_wait_ioctl(struct drm_device *dev, 
> void *data,
>               return -ENOENT;
>  
>       if (args->flags & VIRTGPU_WAIT_NOWAIT) {
> -             ret = dma_resv_test_signaled_rcu(obj->resv, true);
> +             ret = dma_resv_test_signaled_unlocked(obj->resv, true);
>       } else {
> -             ret = dma_resv_wait_timeout_rcu(obj->resv, true, true,
> +             ret = dma_resv_wait_timeout_unlocked(obj->resv, true, true,
>                                               timeout);
>       }
>       if (ret == 0)
> diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_bo.c 
> b/drivers/gpu/drm/vmwgfx/vmwgfx_bo.c
> index 04dd49c4c2572..19e1ce23842a9 100644
> --- a/drivers/gpu/drm/vmwgfx/vmwgfx_bo.c
> +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_bo.c
> @@ -743,7 +743,7 @@ static int vmw_user_bo_synccpu_grab(struct 
> vmw_user_buffer_object *user_bo,
>       if (flags & drm_vmw_synccpu_allow_cs) {
>               long lret;
>  
> -             lret = dma_resv_wait_timeout_rcu
> +             lret = dma_resv_wait_timeout_unlocked
>                       (bo->base.resv, true, true,
>                        nonblock ? 0 : MAX_SCHEDULE_TIMEOUT);
>               if (!lret)
> diff --git a/include/linux/dma-resv.h b/include/linux/dma-resv.h
> index d44a77e8a7e34..99926680c3964 100644
> --- a/include/linux/dma-resv.h
> +++ b/include/linux/dma-resv.h
> @@ -246,7 +246,7 @@ dma_resv_get_excl(struct dma_resv *obj)
>  }
>  
>  /**
> - * dma_resv_get_excl_rcu - get the reservation object's
> + * dma_resv_get_excl - get the reservation object's
>   * exclusive fence, without lock held.
>   * @obj: the reservation object
>   *
> @@ -257,7 +257,7 @@ dma_resv_get_excl(struct dma_resv *obj)
>   * The exclusive fence or NULL if none
>   */
>  static inline struct dma_fence *
> -dma_resv_get_excl_rcu(struct dma_resv *obj)
> +dma_resv_get_excl_unlocked(struct dma_resv *obj)
>  {
>       struct dma_fence *fence;
>  
> @@ -278,16 +278,16 @@ void dma_resv_add_shared_fence(struct dma_resv *obj, 
> struct dma_fence *fence);
>  
>  void dma_resv_add_excl_fence(struct dma_resv *obj, struct dma_fence *fence);
>  
> -int dma_resv_get_fences_rcu(struct dma_resv *obj,
> -                         struct dma_fence **pfence_excl,
> -                         unsigned *pshared_count,
> -                         struct dma_fence ***pshared);
> +int dma_resv_get_fences_unlocked(struct dma_resv *obj,
> +                              struct dma_fence **pfence_excl,
> +                              unsigned *pshared_count,
> +                              struct dma_fence ***pshared);
>  
>  int dma_resv_copy_fences(struct dma_resv *dst, struct dma_resv *src);
>  
> -long dma_resv_wait_timeout_rcu(struct dma_resv *obj, bool wait_all, bool 
> intr,
> -                            unsigned long timeout);
> +long dma_resv_wait_timeout_unlocked(struct dma_resv *obj, bool wait_all, 
> bool intr,
> +                                 unsigned long timeout);
>  
> -bool dma_resv_test_signaled_rcu(struct dma_resv *obj, bool test_all);
> +bool dma_resv_test_signaled_unlocked(struct dma_resv *obj, bool test_all);
>  
>  #endif /* _LINUX_RESERVATION_H */
> -- 
> 2.31.1
> 

-- 
Daniel Vetter
Software Engineer, Intel Corporation
http://blog.ffwll.ch
_______________________________________________
Intel-gfx mailing list
Intel-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/intel-gfx

Reply via email to