From: Honglei Huang <[email protected]>

- kmem_cache management for amdgpu_svm_range
- Reference counting: kref-based release for async safety
- XNACK helper.
- TLB flush helper for compute mode
- amdgpu_svm_init_with_ops: allocate SVM context, initialize
  attr tree, work queues, and drm_gpusvm with configurable
  chunk sizes and notifier size
- amdgpu_svm_init/close/fini: public lifecycle API

Signed-off-by: Honglei Huang <[email protected]>
---
 drivers/gpu/drm/amd/amdgpu/amdgpu_svm.c | 270 ++++++++++++++++++++++++
 1 file changed, 270 insertions(+)
 create mode 100644 drivers/gpu/drm/amd/amdgpu/amdgpu_svm.c

diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_svm.c 
b/drivers/gpu/drm/amd/amdgpu/amdgpu_svm.c
new file mode 100644
index 000000000..aa40e1126
--- /dev/null
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_svm.c
@@ -0,0 +1,270 @@
+/* SPDX-License-Identifier: GPL-2.0 OR MIT */
+/*
+ * Copyright 2026 Advanced Micro Devices, Inc.
+ *
+ * Permission is hereby granted, free of charge, to any person obtaining a
+ * copy of this software and associated documentation files (the "Software"),
+ * to deal in the Software without restriction, including without limitation
+ * the rights to use, copy, modify, merge, publish, distribute, sublicense,
+ * and/or sell copies of the Software, and to permit persons to whom the
+ * Software is furnished to do so, subject to the following conditions:
+ *
+ * The above copyright notice and this permission notice shall be included in
+ * all copies or substantial portions of the Software.
+ *
+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
+ * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
+ * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
+ * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
+ * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
+ * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
+ * OTHER DEALINGS IN THE SOFTWARE.
+ *
+ */
+
+#include <linux/sched/mm.h>
+#include <linux/uaccess.h>
+#include <linux/xarray.h>
+
+#include <drm/drm_file.h>
+
+#include "amdgpu.h"
+#include "amdgpu_svm.h"
+#include "amdgpu_svm_attr.h"
+#include "amdgpu_svm_range.h"
+#include "amdgpu_vm.h"
+
+#if IS_ENABLED(CONFIG_DRM_AMDGPU_SVM)
+
+#define AMDGPU_SVM_MAX_ATTRS 64
+#define AMDGPU_SVM_DEFAULT_SVM_NOTIFIER_SIZE 512
+
+static const unsigned long amdgpu_svm_chunk_sizes[] = {
+       SZ_2M,
+       SZ_64K,
+       SZ_4K,
+};
+
+static struct kmem_cache *amdgpu_svm_range_cache;
+
+static void amdgpu_svm_invalidate(struct drm_gpusvm *gpusvm,
+                                 struct drm_gpusvm_notifier *notifier,
+                                 const struct mmu_notifier_range *mmu_range)
+{
+       amdgpu_svm_range_invalidate(to_amdgpu_svm(gpusvm), notifier, mmu_range);
+}
+
+static struct drm_gpusvm_range *amdgpu_svm_range_alloc(struct drm_gpusvm 
*gpusvm)
+{
+       struct amdgpu_svm_range *range;
+
+       range = kmem_cache_zalloc(amdgpu_svm_range_cache, GFP_KERNEL);
+       if (!range)
+               return NULL;
+
+       INIT_LIST_HEAD(&range->gc_node);
+       range->pending_start = ULONG_MAX;
+       return &range->base;
+}
+
+static void amdgpu_svm_range_free(struct drm_gpusvm_range *range)
+{
+       kmem_cache_free(amdgpu_svm_range_cache, to_amdgpu_svm_range(range));
+}
+
+static const struct drm_gpusvm_ops amdgpu_gpusvm_ops = {
+       .range_alloc = amdgpu_svm_range_alloc,
+       .range_free = amdgpu_svm_range_free,
+       .invalidate = amdgpu_svm_invalidate,
+};
+
+static void amdgpu_svm_release(struct kref *ref)
+{
+       kfree(container_of(ref, struct amdgpu_svm, refcount));
+}
+
+static void amdgpu_svm_put(struct amdgpu_svm *svm)
+{
+       if (svm)
+               kref_put(&svm->refcount, amdgpu_svm_release);
+}
+
+int amdgpu_svm_cache_init(void)
+{
+       int ret = 0;
+
+       if (amdgpu_svm_range_cache)
+               return 0;
+
+       amdgpu_svm_range_cache = 
AMDGPU_SVM_KMEM_CACHE_CREATE("amdgpu_svm_range_cache",
+                                                                struct 
amdgpu_svm_range);
+       if (!amdgpu_svm_range_cache)
+               return -ENOMEM;
+
+       ret = amdgpu_svm_attr_cache_init();
+       if (ret)
+               goto free_out;
+
+       return 0;
+free_out:
+       amdgpu_svm_attr_cache_fini();
+       AMDGPU_SVM_KMEM_CACHE_DESTROY(amdgpu_svm_range_cache);
+       return ret;
+}
+
+void amdgpu_svm_cache_fini(void)
+{
+       if (!amdgpu_svm_range_cache)
+               return;
+
+       amdgpu_svm_attr_cache_fini();
+       AMDGPU_SVM_KMEM_CACHE_DESTROY(amdgpu_svm_range_cache);
+}
+
+static bool amdgpu_svm_default_xnack_enabled(struct amdgpu_device *adev)
+{
+       uint32_t gc_ver = amdgpu_ip_version(adev, GC_HWIP, 0);
+
+       if (gc_ver < IP_VERSION(9, 0, 1))
+               return false;
+       if (!amdgpu_sriov_xnack_support(adev))
+               return false;
+
+       switch (gc_ver) {
+       case IP_VERSION(9, 4, 2):
+       case IP_VERSION(9, 4, 3):
+       case IP_VERSION(9, 4, 4):
+       case IP_VERSION(9, 5, 0):
+               return true;
+       default:
+               break;
+       }
+       if (gc_ver >= IP_VERSION(10, 1, 1))
+               return false;
+       return !adev->gmc.noretry;
+}
+
+static void amdgpu_svm_flush_tlb_compute(struct amdgpu_svm *svm)
+{
+       amdgpu_vm_flush_compute_tlb(svm->adev, svm->vm, TLB_FLUSH_HEAVYWEIGHT,
+                                   svm->adev->gfx.xcc_mask);
+}
+
+static int amdgpu_svm_init_with_ops(struct amdgpu_device *adev,
+                                   struct amdgpu_vm *vm,
+                                   void (*begin_restore)(struct amdgpu_svm *),
+                                   void (*end_restore)(struct amdgpu_svm *),
+                                   void (*flush_tlb)(struct amdgpu_svm *))
+{
+       struct amdgpu_svm *svm;
+       int ret;
+
+       if (vm->svm)
+               return 0;
+
+       ret = amdgpu_svm_cache_init();
+       if (ret)
+               return ret;
+
+       svm = kzalloc(sizeof(*svm), GFP_KERNEL);
+       if (!svm)
+               return -ENOMEM;
+
+       kref_init(&svm->refcount);
+       svm->adev = adev;
+       svm->vm = vm;
+
+       svm->default_granularity = min_t(u8, amdgpu_svm_default_granularity, 
0x3f);
+       svm->xnack_enabled = amdgpu_svm_default_xnack_enabled(adev);
+       svm->xnack_enabled = false; // WA/POC: force to disable xnack
+       svm->begin_restore = begin_restore;
+       svm->end_restore = end_restore;
+       svm->flush_tlb = flush_tlb;
+       atomic_set(&svm->kfd_queues_quiesced, 0);
+       atomic_set(&svm->evicted_ranges, 0);
+       atomic_set(&svm->exiting, 0);
+
+       ret = amdgpu_svm_range_work_init(svm);
+       if (ret)
+               goto err_free;
+
+       svm->attr_tree = amdgpu_svm_attr_tree_create(svm);
+       if (!svm->attr_tree) {
+               ret = -ENOMEM;
+               goto err_range_work_fini;
+       }
+
+       ret = drm_gpusvm_init(&svm->gpusvm, "AMDGPU SVM",
+                                               adev_to_drm(adev), current->mm, 
0,
+                                               adev->vm_manager.max_pfn << 
AMDGPU_GPU_PAGE_SHIFT,
+                                               
AMDGPU_SVM_DEFAULT_SVM_NOTIFIER_SIZE * SZ_1M,
+                                               &amdgpu_gpusvm_ops,
+                                               amdgpu_svm_chunk_sizes,
+                                               
ARRAY_SIZE(amdgpu_svm_chunk_sizes));
+
+       if (ret)
+               goto err_attr_tree_destroy;
+
+       drm_gpusvm_driver_set_lock(&svm->gpusvm, &svm->svm_lock);
+       vm->svm = svm;
+       return 0;
+
+err_attr_tree_destroy:
+       amdgpu_svm_attr_tree_destroy(svm->attr_tree);
+err_range_work_fini:
+       amdgpu_svm_range_work_fini(svm);
+err_free:
+       kfree(svm);
+       return ret;
+}
+
+static int amdgpu_svm_init_compute(struct amdgpu_device *adev, struct 
amdgpu_vm *vm)
+{
+       return amdgpu_svm_init_with_ops(adev, vm,
+                                       amdgpu_svm_range_restore_begin_compute,
+                                       amdgpu_svm_range_restore_end_compute,
+                                       amdgpu_svm_flush_tlb_compute);
+}
+
+int amdgpu_svm_init(struct amdgpu_device *adev, struct amdgpu_vm *vm)
+{
+       /* graphics svm init maybe different */
+
+       return amdgpu_svm_init_compute(adev, vm);
+}
+
+void amdgpu_svm_close(struct amdgpu_vm *vm)
+{
+       if (!vm->svm)
+               return;
+
+       if (atomic_xchg(&vm->svm->exiting, 1))
+               return;
+
+       amdgpu_svm_range_sync_work(vm->svm);
+}
+
+void amdgpu_svm_fini(struct amdgpu_vm *vm)
+{
+       struct amdgpu_svm *svm = vm->svm;
+
+       if (!svm)
+               return;
+
+       amdgpu_svm_close(vm);
+       down_write(&svm->svm_lock);
+       drm_gpusvm_fini(&svm->gpusvm);
+       up_write(&svm->svm_lock);
+
+       amdgpu_svm_range_work_fini(svm);
+       amdgpu_svm_attr_tree_destroy(svm->attr_tree);
+       vm->svm = NULL;
+       amdgpu_svm_put(svm);
+}
+
+bool amdgpu_svm_is_enabled(struct amdgpu_vm *vm)
+{
+       return vm->svm != NULL;
+}
+
+#endif /* CONFIG_DRM_AMDGPU_SVM */
-- 
2.34.1

Reply via email to