Add an initial test suit covering query device properties, allocating memory, binding and unbinding VA ranges through VM_BIND and submitting a simple piece of work through GROUP_SUBMIT. --- lib/igt_panthor.c | 136 ++++++++++++++++++ lib/igt_panthor.h | 20 +++ tests/panthor/meson.build | 4 + tests/panthor/panthor_gem.c | 59 ++++++++ tests/panthor/panthor_group.c | 264 ++++++++++++++++++++++++++++++++++ tests/panthor/panthor_query.c | 25 ++++ tests/panthor/panthor_vm.c | 73 ++++++++++ 7 files changed, 581 insertions(+) create mode 100644 tests/panthor/panthor_gem.c create mode 100644 tests/panthor/panthor_group.c create mode 100644 tests/panthor/panthor_query.c create mode 100644 tests/panthor/panthor_vm.c
diff --git a/lib/igt_panthor.c b/lib/igt_panthor.c index 3e2c29b17..c422320c5 100644 --- a/lib/igt_panthor.c +++ b/lib/igt_panthor.c @@ -2,6 +2,9 @@ // SPDX-FileCopyrightText: Copyright (C) 2025 Collabora Ltd. #include "igt_panthor.h" +#include "drmtest.h" +#include "ioctl_wrappers.h" +#include "panthor_drm.h" /** * SECTION:igt_panthor @@ -12,3 +15,136 @@ * This library provides various auxiliary helper functions for writing Panthor * tests. */ + +void igt_panthor_query(int fd, int32_t type, void* data, size_t size, int err) +{ + struct drm_panthor_dev_query query = { + .type = type, + .pointer = (uintptr_t)data, + .size = size, + }; + + if (err) { + do_ioctl_err(fd, DRM_IOCTL_PANTHOR_DEV_QUERY, &query, err); + } else { + do_ioctl(fd, DRM_IOCTL_PANTHOR_DEV_QUERY, &query); + } +} + +void igt_panthor_vm_create(int fd, uint32_t *vm_id, int err) +{ + struct drm_panthor_vm_create vm_create = {}; + + if (err) { + do_ioctl_err(fd, DRM_IOCTL_PANTHOR_VM_CREATE, &vm_create, err); + } else { + do_ioctl(fd, DRM_IOCTL_PANTHOR_VM_CREATE, &vm_create); + *vm_id = vm_create.id; + } +} + +void igt_panthor_vm_destroy(int fd, uint32_t vm_id, int err) +{ + struct drm_panthor_vm_destroy vm_destroy = { + .id = vm_id, + }; + + if (err) { + do_ioctl_err(fd, DRM_IOCTL_PANTHOR_VM_DESTROY, &vm_destroy, err); + } else { + do_ioctl(fd, DRM_IOCTL_PANTHOR_VM_DESTROY, &vm_destroy); + } +} + +void igt_panthor_vm_bind(int fd, uint32_t vm_id, uint32_t bo_handle, + uint64_t va, uint64_t size, uint32_t flags, int err) +{ + struct drm_panthor_vm_bind_op bind_op = { + .flags = flags, + .bo_handle = bo_handle, + .va = va, + .size = size, + }; + + struct drm_panthor_vm_bind vm_bind = { + .vm_id = vm_id, + .flags = 0, + .ops = DRM_PANTHOR_OBJ_ARRAY(1, &bind_op), + }; + + if (err) { + do_ioctl_err(fd, DRM_IOCTL_PANTHOR_VM_BIND, &vm_bind, err); + } else { + do_ioctl(fd, DRM_IOCTL_PANTHOR_VM_BIND, &vm_bind); + } +} + +void igt_panthor_bo_create(int fd, struct panthor_bo *bo, + uint64_t size, uint32_t flags, int err) +{ + struct drm_panthor_bo_create bo_create = { + .size = size, + .flags = flags, + }; + + if (err) { + do_ioctl_err(fd, DRM_IOCTL_PANTHOR_BO_CREATE, &bo_create, err); + } else { + do_ioctl(fd, DRM_IOCTL_PANTHOR_BO_CREATE, &bo_create); + } + + bo->handle = bo_create.handle; + bo->size = bo_create.size; + bo->offset = 0; + bo->map = NULL; +} + +uint64_t igt_panthor_bo_mmap_offset(int fd, uint32_t handle, int err) +{ + struct drm_panthor_bo_mmap_offset bo_mmap_offset = { + .handle = handle, + }; + + if (err) { + do_ioctl_err(fd, DRM_IOCTL_PANTHOR_BO_MMAP_OFFSET, &bo_mmap_offset, err); + } else { + do_ioctl(fd, DRM_IOCTL_PANTHOR_BO_MMAP_OFFSET, &bo_mmap_offset); + } + return bo_mmap_offset.offset; +} + +void *igt_panthor_mmap_bo(int fd, uint32_t handle, uint64_t size, unsigned prot) +{ + struct drm_panthor_bo_mmap_offset mmap_bo = { + .handle = handle, + }; + void *ptr; + + do_ioctl(fd, DRM_IOCTL_PANTHOR_BO_MMAP_OFFSET, &mmap_bo); + + ptr = mmap(0, size, prot, MAP_SHARED, fd, mmap_bo.offset); + if (ptr == MAP_FAILED) + return NULL; + else + return ptr; +} + +void igt_panthor_bo_create_mapped(int fd, struct panthor_bo *bo, uint64_t size, + uint32_t flags, int err) { + igt_panthor_bo_create(fd, bo, size, flags, err); + bo->offset = igt_panthor_bo_mmap_offset(fd, bo->handle, err); + bo->map = igt_panthor_mmap_bo(fd, bo->handle, bo->size, + PROT_READ | PROT_WRITE); +} + +void igt_panthor_free_bo(int fd, struct panthor_bo *bo) +{ + if (!bo) + return; + + if (bo->map) { + munmap(bo->map, bo->size); + } + + gem_close(fd, bo->handle); +} \ No newline at end of file diff --git a/lib/igt_panthor.h b/lib/igt_panthor.h index c4bee1838..421f44a33 100644 --- a/lib/igt_panthor.h +++ b/lib/igt_panthor.h @@ -4,5 +4,25 @@ #ifndef IGT_PANTHOR_H #define IGT_PANTHOR_H +#include <stdint.h> +#include <stddef.h> + +struct panthor_bo { + int handle; + uint64_t offset; + uint64_t size; + void *map; +}; + +void igt_panthor_query(int fd, int32_t type, void* data, size_t size, int err); +void igt_panthor_vm_create(int fd, uint32_t *vm_id, int err); +void igt_panthor_vm_destroy(int fd, uint32_t vm_id, int err); +void igt_panthor_vm_bind(int fd, uint32_t vm_id, uint32_t bo_handle, uint64_t va, uint64_t size, uint32_t flags, int err); +void igt_panthor_bo_create(int fd, struct panthor_bo *bo, uint64_t size, uint32_t flags, int err); +uint64_t igt_panthor_bo_mmap_offset(int fd, uint32_t handle, int err); +void igt_panthor_free_bo(int fd, struct panthor_bo *bo); +void igt_panthor_bo_create_mapped(int fd, struct panthor_bo *bo, uint64_t size, + uint32_t flags, int err); +void *igt_panthor_mmap_bo(int fd, uint32_t handle, uint64_t size, unsigned prot); #endif /* IGT_PANTHOR_H */ diff --git a/tests/panthor/meson.build b/tests/panthor/meson.build index 979ae91e0..89edcc844 100644 --- a/tests/panthor/meson.build +++ b/tests/panthor/meson.build @@ -1,4 +1,8 @@ panthor_progs = [ + 'panthor_gem', + 'panthor_query', + 'panthor_vm', + 'panthor_group', ] foreach prog : panthor_progs diff --git a/tests/panthor/panthor_gem.c b/tests/panthor/panthor_gem.c new file mode 100644 index 000000000..0bdaa3495 --- /dev/null +++ b/tests/panthor/panthor_gem.c @@ -0,0 +1,59 @@ +// SPDX-License-Identifier: MIT +// SPDX-FileCopyrightText: Copyright (C) 2025 Collabora Ltd. + +#include "igt.h" +#include "igt_core.h" +#include "igt_panthor.h" + +igt_main { + int fd; + + igt_fixture { fd = drm_open_driver(DRIVER_PANTHOR); } + + igt_subtest("bo_create") { + struct panthor_bo bo; + igt_panthor_bo_create(fd, &bo, 4096, 0, 0); + igt_assert(bo.handle != 0); + + igt_panthor_free_bo(fd, &bo); + } + + igt_subtest("bo_mmap_offset") { + struct panthor_bo bo; + uint64_t mmap_offset; + + igt_panthor_bo_create(fd, &bo, 4096, 0, 0); + igt_assert(bo.handle != 0); + + mmap_offset = igt_panthor_bo_mmap_offset(fd, bo.handle, 0); + igt_assert(mmap_offset != 0); + + igt_panthor_free_bo(fd, &bo); + } + + igt_subtest("bo_mmap_offset_invalid_handle") { + struct panthor_bo bo; + uint64_t mmap_offset; + + igt_panthor_bo_create(fd, &bo, 4096, 0, 0); + igt_assert(bo.handle != 0); + + mmap_offset = igt_panthor_bo_mmap_offset(fd, 0xdeadbeef, ENOENT); + igt_assert(mmap_offset == 0); + + igt_panthor_free_bo(fd, &bo); + } + + igt_subtest("bo_create_round_size") { + struct panthor_bo bo; + uint64_t expected_size = 8192; + + igt_panthor_bo_create(fd, &bo, 5000, 0, 0); + igt_assert(bo.handle != 0); + igt_assert(bo.size == expected_size); + + igt_panthor_free_bo(fd, &bo); + } + + igt_fixture { drm_close_driver(fd); } +} diff --git a/tests/panthor/panthor_group.c b/tests/panthor/panthor_group.c new file mode 100644 index 000000000..b7e3cf9c3 --- /dev/null +++ b/tests/panthor/panthor_group.c @@ -0,0 +1,264 @@ +// SPDX-License-Identifier: MIT +// SPDX-FileCopyrightText: Copyright (C) 2025 Collabora Ltd. + +#include <stdint.h> +#include <sys/mman.h> +#include <endian.h> // For htole64 +#include <unistd.h> + +#include "drm.h" +#include "igt.h" +#include "igt_core.h" +#include "igt_panthor.h" +#include "panthor_drm.h" + +static void +issue_store_multiple(u8 *command_stream, uint64_t kernel_va, uint32_t constant) +{ + uint64_t opcode, reg_num, mov48, store_multiple, flush; + uint64_t sr, src0, register_bitmap, offset; + + // MOV48: Load the source register ([r68; r69]) with the kernel address + opcode = 0x1; + reg_num = 68; + mov48 = (opcode << 56) | (reg_num << 48) | kernel_va; + mov48 = htole64(mov48); + memcpy(&command_stream[0], &mov48, sizeof(mov48)); + + // MOV48: Load a known constant into r70 + opcode = 0x1; + reg_num = 70; + mov48 = (opcode << 56) | (reg_num << 48) | constant; + mov48 = htole64(mov48); + memcpy(&command_stream[8], &mov48, sizeof(mov48)); + + // STORE_MULTIPLE: Store the first register to the address pointed to by [r68; r69] + opcode = 0x15; // STORE_MULTIPLE + sr = 70; // Starting from register r70 + src0 = 68; // Address pointed to by [r68; r69] + register_bitmap = 1; // Store the first register + offset = 0; // Offset + store_multiple = (opcode << 56) | (sr << 48) | (src0 << 40) | (register_bitmap << 16) | offset; + store_multiple = htole64(store_multiple); + memcpy(&command_stream[16], &store_multiple, sizeof(store_multiple)); + + opcode = 0x1; + reg_num = 68; + mov48 = (opcode << 56) | (reg_num << 48) | 0; + mov48 = htole64(mov48); + memcpy(&command_stream[24], &mov48, sizeof(mov48)); + + opcode = 36; + flush = opcode << 56 | 0ull << 48 | reg_num << 40 | 0ull << 16 | 0x233; + flush = htole64(flush); + memcpy(&command_stream[32], &flush, sizeof(flush)); +} + +igt_main { + int fd; + + igt_fixture { fd = drm_open_driver(DRIVER_PANTHOR); } + + igt_subtest("group_create") { + struct drm_panthor_gpu_info gpu_info = {}; + struct drm_panthor_vm_create vm_create = {}; + struct drm_panthor_group_create group_create = {}; + struct drm_panthor_queue_create queue = {}; + struct drm_panthor_obj_array queues; + struct drm_panthor_group_destroy group_destroy; + struct drm_panthor_vm_destroy vm_destroy; + + igt_panthor_query(fd, DRM_PANTHOR_DEV_QUERY_GPU_INFO, &gpu_info, sizeof(gpu_info), 0); + igt_assert(gpu_info.gpu_id != 0); + + vm_create.flags = 0; + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_VM_CREATE, &vm_create), 0); + igt_assert(vm_create.id != 0); + + queue.priority = 0; // Low priority + queue.ringbuf_size = 4096; // Example size + queues = (struct drm_panthor_obj_array)DRM_PANTHOR_OBJ_ARRAY(1, &queue); + + group_create.queues = queues; + group_create.max_compute_cores = 1; + group_create.max_fragment_cores = 1; + group_create.max_tiler_cores = 1; + group_create.priority = PANTHOR_GROUP_PRIORITY_MEDIUM; + group_create.compute_core_mask = gpu_info.shader_present & 0x1; // Use first core + group_create.fragment_core_mask = gpu_info.shader_present & 0x1; // Use first core + group_create.tiler_core_mask = gpu_info.tiler_present & 0x1; // Use first tiler + group_create.vm_id = vm_create.id; + + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_GROUP_CREATE, &group_create), 0); + igt_assert(group_create.group_handle != 0); + + // Cleanup: Destroy the group and VM + group_destroy = (struct drm_panthor_group_destroy){ .group_handle = group_create.group_handle }; + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_GROUP_DESTROY, &group_destroy), 0); + + vm_destroy = (struct drm_panthor_vm_destroy) { .id = vm_create.id }; + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_VM_DESTROY, &vm_destroy), 0); + } + + igt_subtest("group_submit") { + struct drm_panthor_gpu_info gpu_info = {}; + struct drm_panthor_vm_create vm_create = {}; + struct drm_panthor_group_create group_create = {}; + struct drm_panthor_queue_create queue = {}; + struct drm_panthor_obj_array queues; + struct drm_panthor_group_submit group_submit = {}; + struct drm_panthor_queue_submit queue_submit = {}; + struct drm_panthor_group_destroy group_destroy; + struct drm_panthor_obj_array queue_submits; + struct drm_panthor_vm_destroy vm_destroy; + struct drm_panthor_bo_create bo_create = {}; + struct drm_panthor_vm_bind vm_bind = {}; + struct drm_panthor_vm_bind_op vm_bind_op = {}; + struct drm_syncobj_wait wait = {}; + struct drm_syncobj_create syncobj_create = {}; + struct drm_panthor_sync_op sync_op = {}; + struct drm_gem_close gem_close = {}; + struct drm_syncobj_destroy syncobj_destroy = {}; + uint64_t command_stream_gpu_addr; + uint32_t command_stream_size; + uint64_t result_gpu_addr; + uint32_t cmd_buf_bo_handle; + uint32_t result_bo_handle; + uint32_t syncobj_handle; + uint8_t command_stream[64] = {0}; + uint8_t *bo_cpu_addr; + uint8_t *result_cpu_addr; + const int INITIAL_VA = 0x1000000; + + + igt_panthor_query(fd, DRM_PANTHOR_DEV_QUERY_GPU_INFO, &gpu_info, sizeof(gpu_info), 0); + igt_assert(gpu_info.gpu_id != 0); + + vm_create.flags = 0; + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_VM_CREATE, &vm_create), 0); + igt_assert(vm_create.id != 0); + + bo_create.size = 4096; + bo_create.flags = 0; + bo_create.exclusive_vm_id = vm_create.id; + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_BO_CREATE, &bo_create), 0); + igt_assert(bo_create.handle != 0); + cmd_buf_bo_handle = bo_create.handle; + + vm_bind_op.flags = DRM_PANTHOR_VM_BIND_OP_TYPE_MAP; + vm_bind_op.bo_handle = cmd_buf_bo_handle; + vm_bind_op.bo_offset = 0; + vm_bind_op.va = INITIAL_VA; + vm_bind_op.size = bo_create.size; + vm_bind.ops = (struct drm_panthor_obj_array)DRM_PANTHOR_OBJ_ARRAY(1, &vm_bind_op); + vm_bind.vm_id = vm_create.id; + vm_bind.flags = 0; + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_VM_BIND, &vm_bind), 0); + + command_stream_gpu_addr = vm_bind_op.va; + command_stream_size = sizeof(command_stream); + + bo_cpu_addr = igt_panthor_mmap_bo(fd, bo_create.handle, bo_create.size, PROT_READ | PROT_WRITE); + igt_assert(bo_cpu_addr != MAP_FAILED); + + // Create the BO to receive the result of the store. + memset(&bo_create, 0, sizeof(bo_create)); + bo_create.size = 4096; + bo_create.flags = 0; + bo_create.exclusive_vm_id = vm_create.id; + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_BO_CREATE, &bo_create), 0); + igt_assert(bo_create.handle != 0); + result_bo_handle = bo_create.handle; + + // Also bind the result BO. + vm_bind_op.flags = DRM_PANTHOR_VM_BIND_OP_TYPE_MAP; + vm_bind_op.bo_handle = result_bo_handle; + vm_bind_op.bo_offset = 0; + vm_bind_op.va = INITIAL_VA + 4096; + vm_bind_op.size = bo_create.size; + vm_bind.ops = (struct drm_panthor_obj_array)DRM_PANTHOR_OBJ_ARRAY(1, &vm_bind_op); + vm_bind.vm_id = vm_create.id; + vm_bind.flags = 0; + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_VM_BIND, &vm_bind), 0); + result_gpu_addr = vm_bind_op.va; + + issue_store_multiple(command_stream, result_gpu_addr, 0xdeadbeef); + memcpy(bo_cpu_addr, command_stream, command_stream_size); + munmap(bo_cpu_addr, bo_create.size); + + queue.priority = 0; + queue.ringbuf_size = 4096; + queues = (struct drm_panthor_obj_array)DRM_PANTHOR_OBJ_ARRAY(1, &queue); + + group_create.queues = queues; + group_create.max_compute_cores = 1; + group_create.max_fragment_cores = 1; + group_create.max_tiler_cores = 1; + group_create.priority = PANTHOR_GROUP_PRIORITY_MEDIUM; + group_create.compute_core_mask = gpu_info.shader_present & 0x1; + group_create.fragment_core_mask = gpu_info.shader_present & 0x1; + group_create.tiler_core_mask = gpu_info.tiler_present & 0x1; + group_create.vm_id = vm_create.id; + + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_GROUP_CREATE, &group_create), 0); + igt_assert(group_create.group_handle != 0); + + syncobj_create = (struct drm_syncobj_create){ + .flags = 0, + }; + + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_SYNCOBJ_CREATE, &syncobj_create), 0); + syncobj_handle = syncobj_create.handle; + + sync_op = (struct drm_panthor_sync_op) { + .handle = syncobj_handle, + .flags = DRM_PANTHOR_SYNC_OP_SIGNAL, + }; + + queue_submit.syncs = (struct drm_panthor_obj_array)DRM_PANTHOR_OBJ_ARRAY(1, &sync_op); + + queue_submit.queue_index = 0; + queue_submit.stream_size = command_stream_size; + queue_submit.stream_addr = command_stream_gpu_addr; + queue_submit.latest_flush = 0; + queue_submits = (struct drm_panthor_obj_array)DRM_PANTHOR_OBJ_ARRAY(1, &queue_submit); + + group_submit.group_handle = group_create.group_handle; + group_submit.queue_submits = queue_submits; + + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_GROUP_SUBMIT, &group_submit), 0); + + wait = (struct drm_syncobj_wait) { + .handles = (uint64_t)&syncobj_handle, + .count_handles = 1, + .timeout_nsec = INT64_MAX, + .flags = 0, + }; + + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_SYNCOBJ_WAIT, &wait), 0); + + result_cpu_addr = igt_panthor_mmap_bo(fd, bo_create.handle, bo_create.size, PROT_READ | PROT_WRITE); + + igt_assert(*(uint32_t *)result_cpu_addr == 0xdeadbeef); + munmap(result_cpu_addr, bo_create.size); + + syncobj_destroy.handle = syncobj_handle; + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_SYNCOBJ_DESTROY, &syncobj_destroy), 0); + + + group_destroy.group_handle = group_create.group_handle; + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_GROUP_DESTROY, &group_destroy), 0); + + vm_destroy.id = vm_create.id; + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_VM_DESTROY, &vm_destroy), 0); + + gem_close.handle = cmd_buf_bo_handle; + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_GEM_CLOSE, &gem_close), 0); + + gem_close.handle = result_bo_handle; + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_GEM_CLOSE, &gem_close), 0); + +} + + igt_fixture { drm_close_driver(fd); } +} diff --git a/tests/panthor/panthor_query.c b/tests/panthor/panthor_query.c new file mode 100644 index 000000000..3bbecf3a6 --- /dev/null +++ b/tests/panthor/panthor_query.c @@ -0,0 +1,25 @@ +// SPDX-License-Identifier: MIT +// SPDX-FileCopyrightText: Copyright (C) 2025 Collabora Ltd. + +#include "igt.h" +#include "igt_core.h" +#include "igt_panthor.h" +#include "panthor_drm.h" +#include <stdint.h> + +igt_main { + int fd; + + igt_fixture { fd = drm_open_driver(DRIVER_PANTHOR); } + + igt_subtest("query") { + struct drm_panthor_gpu_info gpu = {}; + + igt_panthor_query(fd, DRM_PANTHOR_DEV_QUERY_GPU_INFO, &gpu, sizeof(gpu), 0); + + igt_assert(gpu.gpu_id != 0); + igt_assert(gpu.gpu_rev != 0); + } + + igt_fixture { drm_close_driver(fd); } +} diff --git a/tests/panthor/panthor_vm.c b/tests/panthor/panthor_vm.c new file mode 100644 index 000000000..484602de3 --- /dev/null +++ b/tests/panthor/panthor_vm.c @@ -0,0 +1,73 @@ +// SPDX-License-Identifier: MIT +// SPDX-FileCopyrightText: Copyright (C) 2025 Collabora Ltd. + +#include "igt.h" +#include "igt_core.h" +#include "igt_panthor.h" +#include "panthor_drm.h" + +igt_main { + int fd; + + igt_fixture { fd = drm_open_driver(DRIVER_PANTHOR); } + + igt_subtest("vm_create_destroy") { + uint32_t vm_id; + + igt_panthor_vm_create(fd, &vm_id, 0); + igt_assert(vm_id != 0); + + igt_panthor_vm_destroy(fd, vm_id, 0); + } + + igt_subtest("vm_destroy_invalid") { + igt_panthor_vm_destroy(fd, 0xdeadbeef, EINVAL); + } + + igt_subtest("vm_bind") { + uint32_t vm_id; + struct panthor_bo bo; + uint64_t bo_size = 0x1000; + + igt_panthor_vm_create(fd, &vm_id, 0); + igt_assert(vm_id != 0); + + igt_panthor_bo_create(fd, &bo, bo_size, 0, 0); + igt_panthor_vm_bind(fd, vm_id, bo.handle, 0x1000, 0x1000, DRM_PANTHOR_VM_BIND_OP_TYPE_MAP, 0); + + igt_panthor_vm_destroy(fd, vm_id, 0); + } + + igt_subtest("vm_unbind") { + uint32_t vm_id; + struct panthor_bo bo; + uint64_t bo_size = 0x1000; + + igt_panthor_vm_create(fd, &vm_id, 0); + igt_assert(vm_id != 0); + + igt_panthor_bo_create(fd, &bo, bo_size, 0, 0); + igt_panthor_vm_bind(fd, vm_id, bo.handle, 0x1000, 0x1000, DRM_PANTHOR_VM_BIND_OP_TYPE_MAP, 0); + igt_panthor_vm_bind(fd, vm_id, 0, 0x1000, 0x1000, DRM_PANTHOR_VM_BIND_OP_TYPE_UNMAP, 0); + + igt_panthor_vm_destroy(fd, vm_id, 0); + } + + igt_subtest("vm_unbind_invalid_address") { + uint32_t vm_id; + struct panthor_bo bo; + uint64_t bo_size = 0x1000; + + igt_panthor_vm_create(fd, &vm_id, 0); + igt_assert(vm_id != 0); + + igt_panthor_bo_create(fd, &bo, bo_size, 0, 0); + + /* This was not bound previously*/ + igt_panthor_vm_bind(fd, vm_id, bo.handle, 0x1000, 0x1000, DRM_PANTHOR_VM_BIND_OP_TYPE_UNMAP, EINVAL); + igt_panthor_vm_destroy(fd, vm_id, 0); + } + + + igt_fixture { drm_close_driver(fd); } +} -- 2.50.1