Add an initial test suit covering query device properties, allocating memory, binding and unbinding VA ranges through VM_BIND and submitting a simple piece of work through GROUP_SUBMIT.
Signed-off-by: Daniel Almeida <daniel.alme...@collabora.com> --- lib/igt_panthor.c | 188 +++++++++++++++++++++++ lib/igt_panthor.h | 18 +++ tests/panthor/meson.build | 5 +- tests/panthor/panthor_gem.c | 66 ++++++++ tests/panthor/panthor_group.c | 276 ++++++++++++++++++++++++++++++++++ tests/panthor/panthor_vm.c | 80 ++++++++++ 6 files changed, 632 insertions(+), 1 deletion(-) create mode 100644 tests/panthor/panthor_gem.c create mode 100644 tests/panthor/panthor_group.c create mode 100644 tests/panthor/panthor_vm.c diff --git a/lib/igt_panthor.c b/lib/igt_panthor.c index 0b690f796..5d3b572b4 100644 --- a/lib/igt_panthor.c +++ b/lib/igt_panthor.c @@ -39,3 +39,191 @@ void igt_panthor_query(int fd, int32_t type, void *data, size_t size, int err) else do_ioctl(fd, DRM_IOCTL_PANTHOR_DEV_QUERY, &query); } + +/** + * igt_panthor_vm_create: + * @fd: device file descriptor + * @vm_id: pointer to store the created VM ID + * @err: expected error code, or 0 for success + * + * Creates a VM. + */ +void igt_panthor_vm_create(int fd, uint32_t *vm_id, int err) +{ + struct drm_panthor_vm_create vm_create = {}; + + if (err) { + do_ioctl_err(fd, DRM_IOCTL_PANTHOR_VM_CREATE, &vm_create, err); + } else { + do_ioctl(fd, DRM_IOCTL_PANTHOR_VM_CREATE, &vm_create); + *vm_id = vm_create.id; + } +} + +/** + * igt_panthor_vm_destroy: + * @fd: device file descriptor + * @vm_id: VM ID to destroy + * @err: expected error code, or 0 for success + * + * Destroys a VM. + */ +void igt_panthor_vm_destroy(int fd, uint32_t vm_id, int err) +{ + struct drm_panthor_vm_destroy vm_destroy = { + .id = vm_id, + }; + + if (err) + do_ioctl_err(fd, DRM_IOCTL_PANTHOR_VM_DESTROY, &vm_destroy, err); + else + do_ioctl(fd, DRM_IOCTL_PANTHOR_VM_DESTROY, &vm_destroy); +} + +/** + * igt_panthor_vm_bind: + * @fd: device file descriptor + * @vm_id: VM ID to bind the buffer to + * @bo_handle: buffer object handle to bind + * @va: virtual address to bind at + * @size: size of the binding + * @flags: binding flags + * @err: expected error code, or 0 for success + * + * Bind a buffer object to a virtual address in the specified VM. + */ +void igt_panthor_vm_bind(int fd, uint32_t vm_id, uint32_t bo_handle, + uint64_t va, uint64_t size, uint32_t flags, int err) +{ + struct drm_panthor_vm_bind_op bind_op = { + .flags = flags, + .bo_handle = bo_handle, + .va = va, + .size = size, + }; + + struct drm_panthor_vm_bind vm_bind = { + .vm_id = vm_id, + .flags = 0, + .ops = DRM_PANTHOR_OBJ_ARRAY(1, &bind_op), + }; + + if (err) + do_ioctl_err(fd, DRM_IOCTL_PANTHOR_VM_BIND, &vm_bind, err); + else + do_ioctl(fd, DRM_IOCTL_PANTHOR_VM_BIND, &vm_bind); +} + +/** + * igt_panthor_bo_create: + * @fd: device file descriptor + * @bo: pointer to panthor_bo structure to initialize + * @size: requested buffer size in bytes + * @flags: buffer object creation flags + * @err: expected error code, or 0 for success + * + * Creates a new buffer object + */ +void igt_panthor_bo_create(int fd, struct panthor_bo *bo, + uint64_t size, uint32_t flags, int err) +{ + struct drm_panthor_bo_create bo_create = { + .size = size, + .flags = flags, + }; + + if (err) + do_ioctl_err(fd, DRM_IOCTL_PANTHOR_BO_CREATE, &bo_create, err); + else + do_ioctl(fd, DRM_IOCTL_PANTHOR_BO_CREATE, &bo_create); + + bo->handle = bo_create.handle; + bo->size = bo_create.size; + bo->offset = 0; + bo->map = NULL; +} + +/** + * igt_panthor_bo_mmap_offset: + * @fd: device file descriptor + * @handle: buffer object handle + * @err: expected error code, or 0 for success + * + * Get the mmap offset for a buffer object. + * + * Returns: the mmap offset for the buffer object + */ +uint64_t igt_panthor_bo_mmap_offset(int fd, uint32_t handle, int err) +{ + struct drm_panthor_bo_mmap_offset bo_mmap_offset = { + .handle = handle, + }; + + if (err) + do_ioctl_err(fd, DRM_IOCTL_PANTHOR_BO_MMAP_OFFSET, &bo_mmap_offset, err); + else + do_ioctl(fd, DRM_IOCTL_PANTHOR_BO_MMAP_OFFSET, &bo_mmap_offset); + + return bo_mmap_offset.offset; +} + +/** + * igt_panthor_mmap_bo: + * @fd: device file descriptor + * @handle: buffer object handle + * @size: size of the buffer to map + * @prot: memory protection flags (e.g., PROT_READ | PROT_WRITE) + * @offset: mmap offset for the buffer object + * + * Map a buffer object into the process address space. + * + * Returns: pointer to the mapped memory, or NULL on failure + */ +void *igt_panthor_mmap_bo(int fd, uint32_t handle, uint64_t size, + unsigned int prot, uint64_t offset) +{ + void *ptr; + + ptr = mmap(0, size, prot, MAP_SHARED, fd, offset); + if (ptr == MAP_FAILED) + return NULL; + return ptr; +} + +/** + * igt_panthor_bo_create_mapped: + * @fd: device file descriptor + * @bo: pointer to panthor_bo structure to initialize + * @size: requested buffer size in bytes + * @flags: buffer object creation flags + * @err: expected error code, or 0 for success + * + * Create a new buffer object on the panthor device and map it into + * the process address space. + */ +void igt_panthor_bo_create_mapped(int fd, struct panthor_bo *bo, uint64_t size, + uint32_t flags, int err) +{ + igt_panthor_bo_create(fd, bo, size, flags, err); + bo->offset = igt_panthor_bo_mmap_offset(fd, bo->handle, err); + bo->map = igt_panthor_mmap_bo(fd, bo->handle, bo->size, + PROT_READ | PROT_WRITE, bo->offset); +} + +/** + * igt_panthor_free_bo: + * @fd: panthor device file descriptor + * @bo: pointer to panthor_bo structure to free + * + * Free a buffer object and unmap it if it was mapped. + */ +void igt_panthor_free_bo(int fd, struct panthor_bo *bo) +{ + if (!bo) + return; + + if (bo->map) + munmap(bo->map, bo->size); + + gem_close(fd, bo->handle); +} diff --git a/lib/igt_panthor.h b/lib/igt_panthor.h index a99b7102d..275ad54d6 100644 --- a/lib/igt_panthor.h +++ b/lib/igt_panthor.h @@ -7,6 +7,24 @@ #include <stddef.h> #include <stdint.h> +struct panthor_bo { + int handle; + uint64_t offset; + uint64_t size; + void *map; +}; + void igt_panthor_query(int fd, int32_t type, void *data, size_t size, int err); +void igt_panthor_vm_create(int fd, uint32_t *vm_id, int err); +void igt_panthor_vm_destroy(int fd, uint32_t vm_id, int err); +void igt_panthor_vm_bind(int fd, uint32_t vm_id, uint32_t bo_handle, + uint64_t va, uint64_t size, uint32_t flags, int err); +void igt_panthor_bo_create(int fd, struct panthor_bo *bo, uint64_t size, uint32_t flags, int err); +uint64_t igt_panthor_bo_mmap_offset(int fd, uint32_t handle, int err); +void igt_panthor_free_bo(int fd, struct panthor_bo *bo); +void igt_panthor_bo_create_mapped(int fd, struct panthor_bo *bo, uint64_t size, + uint32_t flags, int err); +void *igt_panthor_mmap_bo(int fd, uint32_t handle, uint64_t size, + unsigned int prot, uint64_t offset); #endif /* IGT_PANTHOR_H */ diff --git a/tests/panthor/meson.build b/tests/panthor/meson.build index ce13aebaa..42a46e993 100644 --- a/tests/panthor/meson.build +++ b/tests/panthor/meson.build @@ -1,5 +1,8 @@ panthor_progs = [ - 'panthor_query' + 'panthor_gem', + 'panthor_group', + 'panthor_query', + 'panthor_vm', ] foreach prog : panthor_progs diff --git a/tests/panthor/panthor_gem.c b/tests/panthor/panthor_gem.c new file mode 100644 index 000000000..7e8e2084d --- /dev/null +++ b/tests/panthor/panthor_gem.c @@ -0,0 +1,66 @@ +// SPDX-License-Identifier: MIT +// SPDX-FileCopyrightText: Copyright (C) 2025 Collabora Ltd. + +#include "igt.h" +#include "igt_core.h" +#include "igt_panthor.h" + +igt_main { + int fd; + + igt_fixture { fd = drm_open_driver(DRIVER_PANTHOR); } + + igt_describe("Create a buffer object"); + igt_subtest("bo_create") { + struct panthor_bo bo; + + igt_panthor_bo_create(fd, &bo, 4096, 0, 0); + igt_assert(bo.handle != 0); + + igt_panthor_free_bo(fd, &bo); + } + + igt_describe("Create a fake mmap offset for a buffer object"); + igt_subtest("bo_mmap_offset") { + struct panthor_bo bo; + uint64_t mmap_offset; + + igt_panthor_bo_create(fd, &bo, 4096, 0, 0); + igt_assert(bo.handle != 0); + + mmap_offset = igt_panthor_bo_mmap_offset(fd, bo.handle, 0); + igt_assert(mmap_offset != 0); + + igt_panthor_free_bo(fd, &bo); + } + + igt_describe("Same as bo_mmap_offset but with an invalid handle"); + igt_subtest("bo_mmap_offset_invalid_handle") { + struct panthor_bo bo; + uint64_t mmap_offset; + + igt_panthor_bo_create(fd, &bo, 4096, 0, 0); + igt_assert(bo.handle != 0); + + mmap_offset = igt_panthor_bo_mmap_offset(fd, 0xdeadbeef, ENOENT); + igt_assert(mmap_offset == 0); + + igt_panthor_free_bo(fd, &bo); + } + + igt_describe_f("Create a buffer object whose size is not page-aligned, and check " + "that the allocated size is rounded up to the next page size %lu.", + 8192UL); + igt_subtest("bo_create_round_size") { + struct panthor_bo bo; + uint64_t expected_size = 8192; + + igt_panthor_bo_create(fd, &bo, 5000, 0, 0); + igt_assert(bo.handle != 0); + igt_assert(bo.size == expected_size); + + igt_panthor_free_bo(fd, &bo); + } + + igt_fixture { drm_close_driver(fd); } +} diff --git a/tests/panthor/panthor_group.c b/tests/panthor/panthor_group.c new file mode 100644 index 000000000..0a3b746d8 --- /dev/null +++ b/tests/panthor/panthor_group.c @@ -0,0 +1,276 @@ +// SPDX-License-Identifier: MIT +// SPDX-FileCopyrightText: Copyright (C) 2025 Collabora Ltd. + +#include <stdint.h> +#include <sys/mman.h> +#include <endian.h> // For htole64 +#include <unistd.h> + +#include "drm.h" +#include "igt.h" +#include "igt_core.h" +#include "igt_panthor.h" +#include "panthor_drm.h" + +static void +issue_store_multiple(u8 *command_stream, uint64_t kernel_va, uint32_t constant) +{ + uint64_t opcode, reg_num, mov48, store_multiple, flush; + uint64_t sr, src0, register_bitmap, offset; + + // MOV48: Load the source register ([r68; r69]) with the kernel address + opcode = 0x1; + reg_num = 68; + mov48 = (opcode << 56) | (reg_num << 48) | kernel_va; + mov48 = htole64(mov48); + memcpy(&command_stream[0], &mov48, sizeof(mov48)); + + // MOV48: Load a known constant into r70 + opcode = 0x1; + reg_num = 70; + mov48 = (opcode << 56) | (reg_num << 48) | constant; + mov48 = htole64(mov48); + memcpy(&command_stream[8], &mov48, sizeof(mov48)); + + // STORE_MULTIPLE: Store the first register to the address pointed to by [r68; r69] + opcode = 0x15; // STORE_MULTIPLE + sr = 70; // Starting from register r70 + src0 = 68; // Address pointed to by [r68; r69] + register_bitmap = 1; // Store the first register + offset = 0; // Offset + store_multiple = (opcode << 56) | (sr << 48) | (src0 << 40) | + (register_bitmap << 16) | offset; + store_multiple = htole64(store_multiple); + memcpy(&command_stream[16], &store_multiple, sizeof(store_multiple)); + + opcode = 0x1; + reg_num = 68; + mov48 = (opcode << 56) | (reg_num << 48) | 0; + mov48 = htole64(mov48); + memcpy(&command_stream[24], &mov48, sizeof(mov48)); + + opcode = 36; + flush = opcode << 56 | 0ull << 48 | reg_num << 40 | 0ull << 16 | 0x233; + flush = htole64(flush); + memcpy(&command_stream[32], &flush, sizeof(flush)); +} + +igt_main { + int fd; + + igt_fixture { fd = drm_open_driver(DRIVER_PANTHOR); } + + igt_describe("Create and destroy a CSF group."); + igt_subtest("group_create") { + struct drm_panthor_gpu_info gpu_info = {}; + struct drm_panthor_vm_create vm_create = {}; + struct drm_panthor_group_create group_create = {}; + struct drm_panthor_queue_create queue = {}; + struct drm_panthor_obj_array queues = {}; + struct drm_panthor_group_destroy group_destroy = {}; + struct drm_panthor_vm_destroy vm_destroy = {}; + + igt_panthor_query(fd, DRM_PANTHOR_DEV_QUERY_GPU_INFO, + &gpu_info, sizeof(gpu_info), 0); + igt_assert(gpu_info.gpu_id != 0); + + vm_create.flags = 0; + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_VM_CREATE, &vm_create), 0); + igt_assert(vm_create.id != 0); + + queue.priority = 0; // Low priority + queue.ringbuf_size = 4096; // Example size + queues = (struct drm_panthor_obj_array)DRM_PANTHOR_OBJ_ARRAY(1, &queue); + + group_create.queues = queues; + group_create.max_compute_cores = 1; + group_create.max_fragment_cores = 1; + group_create.max_tiler_cores = 1; + group_create.priority = PANTHOR_GROUP_PRIORITY_MEDIUM; + group_create.compute_core_mask = gpu_info.shader_present & 0x1; // Use first core + group_create.fragment_core_mask = gpu_info.shader_present & 0x1; // Use first core + group_create.tiler_core_mask = gpu_info.tiler_present & 0x1; // Use first tiler + group_create.vm_id = vm_create.id; + + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_GROUP_CREATE, &group_create), 0); + igt_assert(group_create.group_handle != 0); + + // Cleanup: Destroy the group and VM + group_destroy = (struct drm_panthor_group_destroy){ + .group_handle = group_create.group_handle + }; + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_GROUP_DESTROY, &group_destroy), 0); + + vm_destroy = (struct drm_panthor_vm_destroy) { .id = vm_create.id }; + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_VM_DESTROY, &vm_destroy), 0); + } + + igt_describe("Submit a job to a group and wait for completion. " + "The job writes a known value to a buffer object that is then " + "mmaped and checked."); + igt_subtest("group_submit") { + struct drm_panthor_gpu_info gpu_info = {}; + struct drm_panthor_vm_create vm_create = {}; + struct drm_panthor_group_create group_create = {}; + struct drm_panthor_queue_create queue = {}; + struct drm_panthor_obj_array queues = {}; + struct drm_panthor_group_submit group_submit = {}; + struct drm_panthor_queue_submit queue_submit = {}; + struct drm_panthor_group_destroy group_destroy = {}; + struct drm_panthor_obj_array queue_submits = {}; + struct drm_panthor_vm_destroy vm_destroy = {}; + struct drm_panthor_bo_create bo_create = {}; + struct drm_panthor_vm_bind vm_bind = {}; + struct drm_panthor_vm_bind_op vm_bind_op = {}; + struct drm_syncobj_wait wait = {}; + struct drm_syncobj_create syncobj_create = {}; + struct drm_panthor_sync_op sync_op = {}; + struct drm_gem_close gem_close = {}; + struct drm_syncobj_destroy syncobj_destroy = {}; + uint64_t command_stream_gpu_addr; + uint32_t command_stream_size; + uint64_t result_gpu_addr; + uint32_t cmd_buf_bo_handle; + uint32_t result_bo_handle; + uint32_t syncobj_handle; + uint8_t command_stream[64] = {0}; + uint8_t *bo_cpu_addr; + uint8_t *result_cpu_addr; + const int INITIAL_VA = 0x1000000; + uint64_t bo_mmap_offset; + + igt_panthor_query(fd, DRM_PANTHOR_DEV_QUERY_GPU_INFO, + &gpu_info, sizeof(gpu_info), 0); + igt_assert(gpu_info.gpu_id != 0); + + vm_create.flags = 0; + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_VM_CREATE, &vm_create), 0); + igt_assert(vm_create.id != 0); + + bo_create.size = 4096; + bo_create.flags = 0; + bo_create.exclusive_vm_id = vm_create.id; + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_BO_CREATE, &bo_create), 0); + igt_assert(bo_create.handle != 0); + cmd_buf_bo_handle = bo_create.handle; + + vm_bind_op.flags = DRM_PANTHOR_VM_BIND_OP_TYPE_MAP; + vm_bind_op.bo_handle = cmd_buf_bo_handle; + vm_bind_op.bo_offset = 0; + vm_bind_op.va = INITIAL_VA; + vm_bind_op.size = bo_create.size; + vm_bind.ops = (struct drm_panthor_obj_array)DRM_PANTHOR_OBJ_ARRAY(1, &vm_bind_op); + vm_bind.vm_id = vm_create.id; + vm_bind.flags = 0; + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_VM_BIND, &vm_bind), 0); + + command_stream_gpu_addr = vm_bind_op.va; + command_stream_size = sizeof(command_stream); + + bo_mmap_offset = igt_panthor_bo_mmap_offset(fd, cmd_buf_bo_handle, 0); + bo_cpu_addr = igt_panthor_mmap_bo(fd, cmd_buf_bo_handle, + bo_create.size, PROT_READ | PROT_WRITE, + bo_mmap_offset); + igt_assert(bo_cpu_addr); + + // Create the BO to receive the result of the store. + memset(&bo_create, 0, sizeof(bo_create)); + bo_create.size = 4096; + bo_create.flags = 0; + bo_create.exclusive_vm_id = vm_create.id; + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_BO_CREATE, &bo_create), 0); + igt_assert(bo_create.handle != 0); + result_bo_handle = bo_create.handle; + + // Also bind the result BO. + vm_bind_op.flags = DRM_PANTHOR_VM_BIND_OP_TYPE_MAP; + vm_bind_op.bo_handle = result_bo_handle; + vm_bind_op.bo_offset = 0; + vm_bind_op.va = INITIAL_VA + 4096; + vm_bind_op.size = bo_create.size; + vm_bind.ops = (struct drm_panthor_obj_array)DRM_PANTHOR_OBJ_ARRAY(1, &vm_bind_op); + vm_bind.vm_id = vm_create.id; + vm_bind.flags = 0; + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_VM_BIND, &vm_bind), 0); + result_gpu_addr = vm_bind_op.va; + + issue_store_multiple(command_stream, result_gpu_addr, 0xdeadbeef); + memcpy(bo_cpu_addr, command_stream, command_stream_size); + munmap(bo_cpu_addr, bo_create.size); + + queue.priority = 0; + queue.ringbuf_size = 4096; + queues = (struct drm_panthor_obj_array)DRM_PANTHOR_OBJ_ARRAY(1, &queue); + + group_create.queues = queues; + group_create.max_compute_cores = 1; + group_create.max_fragment_cores = 1; + group_create.max_tiler_cores = 1; + group_create.priority = PANTHOR_GROUP_PRIORITY_MEDIUM; + group_create.compute_core_mask = gpu_info.shader_present & 0x1; + group_create.fragment_core_mask = gpu_info.shader_present & 0x1; + group_create.tiler_core_mask = gpu_info.tiler_present & 0x1; + group_create.vm_id = vm_create.id; + + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_GROUP_CREATE, &group_create), 0); + igt_assert(group_create.group_handle != 0); + + syncobj_create = (struct drm_syncobj_create){ + .flags = 0, + }; + + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_SYNCOBJ_CREATE, &syncobj_create), 0); + syncobj_handle = syncobj_create.handle; + + sync_op = (struct drm_panthor_sync_op) { + .handle = syncobj_handle, + .flags = DRM_PANTHOR_SYNC_OP_SIGNAL, + }; + + queue_submit.syncs = (struct drm_panthor_obj_array)DRM_PANTHOR_OBJ_ARRAY(1, &sync_op); + + queue_submit.queue_index = 0; + queue_submit.stream_size = command_stream_size; + queue_submit.stream_addr = command_stream_gpu_addr; + queue_submit.latest_flush = 0; + queue_submits = (struct drm_panthor_obj_array)DRM_PANTHOR_OBJ_ARRAY(1, &queue_submit); + + group_submit.group_handle = group_create.group_handle; + group_submit.queue_submits = queue_submits; + + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_GROUP_SUBMIT, &group_submit), 0); + + wait = (struct drm_syncobj_wait) { + .handles = (uint64_t)&syncobj_handle, + .count_handles = 1, + .timeout_nsec = INT64_MAX, + .flags = 0, + }; + + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_SYNCOBJ_WAIT, &wait), 0); + + bo_mmap_offset = igt_panthor_bo_mmap_offset(fd, result_bo_handle, 0); + result_cpu_addr = igt_panthor_mmap_bo(fd, result_bo_handle, + bo_create.size, PROT_READ | PROT_WRITE, bo_mmap_offset); + + igt_assert(*(uint32_t *)result_cpu_addr == 0xdeadbeef); + munmap(result_cpu_addr, bo_create.size); + + syncobj_destroy.handle = syncobj_handle; + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_SYNCOBJ_DESTROY, &syncobj_destroy), 0); + + group_destroy.group_handle = group_create.group_handle; + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_GROUP_DESTROY, &group_destroy), 0); + + vm_destroy.id = vm_create.id; + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_VM_DESTROY, &vm_destroy), 0); + + gem_close.handle = cmd_buf_bo_handle; + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_GEM_CLOSE, &gem_close), 0); + + gem_close.handle = result_bo_handle; + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_GEM_CLOSE, &gem_close), 0); +} + + igt_fixture { drm_close_driver(fd); } +} diff --git a/tests/panthor/panthor_vm.c b/tests/panthor/panthor_vm.c new file mode 100644 index 000000000..1b641b601 --- /dev/null +++ b/tests/panthor/panthor_vm.c @@ -0,0 +1,80 @@ +// SPDX-License-Identifier: MIT +// SPDX-FileCopyrightText: Copyright (C) 2025 Collabora Ltd. + +#include "igt.h" +#include "igt_core.h" +#include "igt_panthor.h" +#include "panthor_drm.h" + +igt_main { + int fd; + + igt_fixture { fd = drm_open_driver(DRIVER_PANTHOR); } + + igt_describe("Create and destroy a VM"); + igt_subtest("vm_create_destroy") { + uint32_t vm_id; + + igt_panthor_vm_create(fd, &vm_id, 0); + igt_assert(vm_id != 0); + + igt_panthor_vm_destroy(fd, vm_id, 0); + } + + igt_subtest("vm_destroy_invalid") { + igt_panthor_vm_destroy(fd, 0xdeadbeef, EINVAL); + } + + igt_describe("Test the VM_BIND API synchronously"); + igt_subtest("vm_bind") { + uint32_t vm_id; + struct panthor_bo bo; + uint64_t bo_size = 0x1000; + + igt_panthor_vm_create(fd, &vm_id, 0); + igt_assert(vm_id != 0); + + igt_panthor_bo_create(fd, &bo, bo_size, 0, 0); + igt_panthor_vm_bind(fd, vm_id, bo.handle, + 0x1000, 0x1000, DRM_PANTHOR_VM_BIND_OP_TYPE_MAP, 0); + + igt_panthor_vm_destroy(fd, vm_id, 0); + } + + igt_describe("Test unbinding a previously bound range"); + igt_subtest("vm_unbind") { + uint32_t vm_id; + struct panthor_bo bo; + uint64_t bo_size = 0x1000; + + igt_panthor_vm_create(fd, &vm_id, 0); + igt_assert(vm_id != 0); + + igt_panthor_bo_create(fd, &bo, bo_size, 0, 0); + igt_panthor_vm_bind(fd, vm_id, bo.handle, + 0x1000, 0x1000, DRM_PANTHOR_VM_BIND_OP_TYPE_MAP, 0); + igt_panthor_vm_bind(fd, vm_id, 0, + 0x1000, 0x1000, DRM_PANTHOR_VM_BIND_OP_TYPE_UNMAP, 0); + + igt_panthor_vm_destroy(fd, vm_id, 0); + } + + igt_describe("Test unbinding an address range that was not previously bound"); + igt_subtest("vm_unbind_invalid_address") { + uint32_t vm_id; + struct panthor_bo bo; + uint64_t bo_size = 0x1000; + + igt_panthor_vm_create(fd, &vm_id, 0); + igt_assert(vm_id != 0); + + igt_panthor_bo_create(fd, &bo, bo_size, 0, 0); + + /* This was not bound previously*/ + igt_panthor_vm_bind(fd, vm_id, bo.handle, + 0x1000, 0x1000, DRM_PANTHOR_VM_BIND_OP_TYPE_UNMAP, EINVAL); + igt_panthor_vm_destroy(fd, vm_id, 0); + } + + igt_fixture { drm_close_driver(fd); } +} -- 2.51.0