[PATCH i-g-t 4/4] tests/panthor: add panthor tests

Daniel Almeida daniel.almeida at collabora.com
Thu Aug 28 13:04:00 UTC 2025


Add an initial test suit covering query device properties, allocating
memory, binding and unbinding VA ranges through VM_BIND and submitting a
simple piece of work through GROUP_SUBMIT.
---
 lib/igt_panthor.c             | 136 ++++++++++++++++++
 lib/igt_panthor.h             |  20 +++
 tests/panthor/meson.build     |   4 +
 tests/panthor/panthor_gem.c   |  59 ++++++++
 tests/panthor/panthor_group.c | 264 ++++++++++++++++++++++++++++++++++
 tests/panthor/panthor_query.c |  25 ++++
 tests/panthor/panthor_vm.c    |  73 ++++++++++
 7 files changed, 581 insertions(+)
 create mode 100644 tests/panthor/panthor_gem.c
 create mode 100644 tests/panthor/panthor_group.c
 create mode 100644 tests/panthor/panthor_query.c
 create mode 100644 tests/panthor/panthor_vm.c

diff --git a/lib/igt_panthor.c b/lib/igt_panthor.c
index 3e2c29b17..c422320c5 100644
--- a/lib/igt_panthor.c
+++ b/lib/igt_panthor.c
@@ -2,6 +2,9 @@
 // SPDX-FileCopyrightText: Copyright (C) 2025 Collabora Ltd.
 
 #include "igt_panthor.h"
+#include "drmtest.h"
+#include "ioctl_wrappers.h"
+#include "panthor_drm.h"
 
 /**
  * SECTION:igt_panthor
@@ -12,3 +15,136 @@
  * This library provides various auxiliary helper functions for writing Panthor
  * tests.
  */
+
+void igt_panthor_query(int fd, int32_t type, void* data, size_t size, int err)
+{
+    struct drm_panthor_dev_query query = {
+        .type = type,
+        .pointer = (uintptr_t)data,
+        .size = size,
+    };
+
+    if (err) {
+        do_ioctl_err(fd, DRM_IOCTL_PANTHOR_DEV_QUERY, &query, err);
+    } else {
+        do_ioctl(fd, DRM_IOCTL_PANTHOR_DEV_QUERY, &query);
+    }
+}
+
+void igt_panthor_vm_create(int fd, uint32_t *vm_id, int err)
+{
+    struct drm_panthor_vm_create vm_create = {};
+
+    if (err) {
+        do_ioctl_err(fd, DRM_IOCTL_PANTHOR_VM_CREATE, &vm_create, err);
+    } else {
+        do_ioctl(fd, DRM_IOCTL_PANTHOR_VM_CREATE, &vm_create);
+        *vm_id = vm_create.id;
+    }
+}
+
+void igt_panthor_vm_destroy(int fd, uint32_t vm_id, int err)
+{
+    struct drm_panthor_vm_destroy vm_destroy = {
+        .id = vm_id,
+    };
+
+    if (err) {
+        do_ioctl_err(fd, DRM_IOCTL_PANTHOR_VM_DESTROY, &vm_destroy, err);
+    } else {
+        do_ioctl(fd, DRM_IOCTL_PANTHOR_VM_DESTROY, &vm_destroy);
+    }
+}
+
+void igt_panthor_vm_bind(int fd, uint32_t vm_id, uint32_t bo_handle,
+                         uint64_t va, uint64_t size, uint32_t flags, int err)
+{
+    struct drm_panthor_vm_bind_op bind_op = {
+        .flags = flags,
+        .bo_handle = bo_handle,
+        .va = va,
+        .size = size,
+    };
+
+    struct drm_panthor_vm_bind vm_bind = {
+        .vm_id = vm_id,
+        .flags = 0,
+        .ops = DRM_PANTHOR_OBJ_ARRAY(1, &bind_op),
+    };
+
+    if (err) {
+        do_ioctl_err(fd, DRM_IOCTL_PANTHOR_VM_BIND, &vm_bind, err);
+    } else {
+        do_ioctl(fd, DRM_IOCTL_PANTHOR_VM_BIND, &vm_bind);
+    }
+}
+
+void igt_panthor_bo_create(int fd, struct panthor_bo *bo,
+                           uint64_t size, uint32_t flags, int err)
+{
+    struct drm_panthor_bo_create bo_create = {
+        .size = size,
+        .flags = flags,
+    };
+
+    if (err) {
+        do_ioctl_err(fd, DRM_IOCTL_PANTHOR_BO_CREATE, &bo_create, err);
+    } else {
+        do_ioctl(fd, DRM_IOCTL_PANTHOR_BO_CREATE, &bo_create);
+    }
+
+    bo->handle = bo_create.handle;
+    bo->size = bo_create.size;
+    bo->offset = 0;
+    bo->map = NULL;
+}
+
+uint64_t igt_panthor_bo_mmap_offset(int fd, uint32_t handle, int err)
+{
+    struct drm_panthor_bo_mmap_offset bo_mmap_offset = {
+        .handle = handle,
+    };
+
+    if (err) {
+        do_ioctl_err(fd, DRM_IOCTL_PANTHOR_BO_MMAP_OFFSET, &bo_mmap_offset, err);
+    } else {
+        do_ioctl(fd, DRM_IOCTL_PANTHOR_BO_MMAP_OFFSET, &bo_mmap_offset);
+    }
+    return bo_mmap_offset.offset;
+}
+
+void *igt_panthor_mmap_bo(int fd, uint32_t handle, uint64_t size, unsigned prot)
+{
+    struct drm_panthor_bo_mmap_offset mmap_bo = {
+        .handle = handle,
+    };
+    void *ptr;
+
+    do_ioctl(fd, DRM_IOCTL_PANTHOR_BO_MMAP_OFFSET, &mmap_bo);
+
+    ptr = mmap(0, size, prot, MAP_SHARED, fd, mmap_bo.offset);
+    if (ptr == MAP_FAILED)
+        return NULL;
+    else
+        return ptr;
+}
+
+void igt_panthor_bo_create_mapped(int fd, struct panthor_bo *bo, uint64_t size,
+                                  uint32_t flags, int err) {
+  igt_panthor_bo_create(fd, bo, size, flags, err);
+  bo->offset = igt_panthor_bo_mmap_offset(fd, bo->handle, err);
+  bo->map = igt_panthor_mmap_bo(fd, bo->handle, bo->size,
+                           PROT_READ | PROT_WRITE);
+}
+
+void igt_panthor_free_bo(int fd, struct panthor_bo *bo)
+{
+    if (!bo)
+        return;
+
+    if (bo->map) {
+        munmap(bo->map, bo->size);
+    }
+
+    gem_close(fd, bo->handle);
+}
\ No newline at end of file
diff --git a/lib/igt_panthor.h b/lib/igt_panthor.h
index c4bee1838..421f44a33 100644
--- a/lib/igt_panthor.h
+++ b/lib/igt_panthor.h
@@ -4,5 +4,25 @@
 #ifndef IGT_PANTHOR_H
 #define IGT_PANTHOR_H
 
+#include <stdint.h>
+#include <stddef.h>
+
+struct panthor_bo {
+    int handle;
+    uint64_t offset;
+    uint64_t size;
+    void *map;
+};
+
+void igt_panthor_query(int fd, int32_t type, void* data, size_t size, int err);
+void igt_panthor_vm_create(int fd, uint32_t *vm_id, int err);
+void igt_panthor_vm_destroy(int fd, uint32_t vm_id, int err);
+void igt_panthor_vm_bind(int fd, uint32_t vm_id, uint32_t bo_handle, uint64_t va, uint64_t size, uint32_t flags, int err);
+void igt_panthor_bo_create(int fd, struct panthor_bo *bo, uint64_t size, uint32_t flags, int err);
+uint64_t igt_panthor_bo_mmap_offset(int fd, uint32_t handle, int err);
+void igt_panthor_free_bo(int fd, struct panthor_bo *bo);
+void igt_panthor_bo_create_mapped(int fd, struct panthor_bo *bo, uint64_t size,
+                                  uint32_t flags, int err);
+void *igt_panthor_mmap_bo(int fd, uint32_t handle, uint64_t size, unsigned prot);
 
 #endif /* IGT_PANTHOR_H */
diff --git a/tests/panthor/meson.build b/tests/panthor/meson.build
index 979ae91e0..89edcc844 100644
--- a/tests/panthor/meson.build
+++ b/tests/panthor/meson.build
@@ -1,4 +1,8 @@
 panthor_progs = [
+	'panthor_gem',
+	'panthor_query',
+	'panthor_vm',
+	'panthor_group',
 ]
 
 foreach prog : panthor_progs
diff --git a/tests/panthor/panthor_gem.c b/tests/panthor/panthor_gem.c
new file mode 100644
index 000000000..0bdaa3495
--- /dev/null
+++ b/tests/panthor/panthor_gem.c
@@ -0,0 +1,59 @@
+// SPDX-License-Identifier: MIT
+// SPDX-FileCopyrightText: Copyright (C) 2025 Collabora Ltd.
+
+#include "igt.h"
+#include "igt_core.h"
+#include "igt_panthor.h"
+
+igt_main {
+  int fd;
+
+  igt_fixture { fd = drm_open_driver(DRIVER_PANTHOR); }
+
+  igt_subtest("bo_create") {
+    struct panthor_bo bo;
+    igt_panthor_bo_create(fd, &bo, 4096, 0, 0);
+    igt_assert(bo.handle != 0);
+
+    igt_panthor_free_bo(fd, &bo);
+  }
+
+  igt_subtest("bo_mmap_offset") {
+    struct panthor_bo bo;
+    uint64_t mmap_offset;
+
+    igt_panthor_bo_create(fd, &bo, 4096, 0, 0);
+    igt_assert(bo.handle != 0);
+
+    mmap_offset = igt_panthor_bo_mmap_offset(fd, bo.handle, 0);
+    igt_assert(mmap_offset != 0);
+
+    igt_panthor_free_bo(fd, &bo);
+  }
+
+  igt_subtest("bo_mmap_offset_invalid_handle") {
+    struct panthor_bo bo;
+    uint64_t mmap_offset;
+
+    igt_panthor_bo_create(fd, &bo, 4096, 0, 0);
+    igt_assert(bo.handle != 0);
+
+    mmap_offset = igt_panthor_bo_mmap_offset(fd, 0xdeadbeef, ENOENT);
+    igt_assert(mmap_offset == 0);
+
+    igt_panthor_free_bo(fd, &bo);
+  }
+
+  igt_subtest("bo_create_round_size") {
+    struct panthor_bo bo;
+    uint64_t expected_size = 8192;
+
+    igt_panthor_bo_create(fd, &bo, 5000, 0, 0);
+    igt_assert(bo.handle != 0);
+    igt_assert(bo.size == expected_size);
+
+    igt_panthor_free_bo(fd, &bo);
+  }
+
+  igt_fixture { drm_close_driver(fd); }
+}
diff --git a/tests/panthor/panthor_group.c b/tests/panthor/panthor_group.c
new file mode 100644
index 000000000..b7e3cf9c3
--- /dev/null
+++ b/tests/panthor/panthor_group.c
@@ -0,0 +1,264 @@
+// SPDX-License-Identifier: MIT
+// SPDX-FileCopyrightText: Copyright (C) 2025 Collabora Ltd.
+
+#include <stdint.h>
+#include <sys/mman.h>
+#include <endian.h> // For htole64
+#include <unistd.h>
+
+#include "drm.h"
+#include "igt.h"
+#include "igt_core.h"
+#include "igt_panthor.h"
+#include "panthor_drm.h"
+
+static void
+issue_store_multiple(u8 *command_stream, uint64_t kernel_va, uint32_t constant)
+{
+    uint64_t opcode, reg_num, mov48, store_multiple, flush;
+    uint64_t sr, src0, register_bitmap, offset;
+
+    // MOV48: Load the source register ([r68; r69]) with the kernel address
+    opcode = 0x1;
+    reg_num = 68;
+    mov48 = (opcode << 56) | (reg_num << 48) | kernel_va;
+    mov48 = htole64(mov48);
+    memcpy(&command_stream[0], &mov48, sizeof(mov48));
+
+    // MOV48: Load a known constant into r70
+    opcode = 0x1;
+    reg_num = 70;
+    mov48 = (opcode << 56) | (reg_num << 48) | constant;
+    mov48 = htole64(mov48);
+    memcpy(&command_stream[8], &mov48, sizeof(mov48));
+
+    // STORE_MULTIPLE: Store the first register to the address pointed to by [r68; r69]
+    opcode = 0x15; // STORE_MULTIPLE
+    sr = 70; // Starting from register r70
+    src0 = 68; // Address pointed to by [r68; r69]
+    register_bitmap = 1; // Store the first register
+    offset = 0; // Offset
+    store_multiple = (opcode << 56) | (sr << 48) | (src0 << 40) | (register_bitmap << 16) | offset;
+    store_multiple = htole64(store_multiple);
+    memcpy(&command_stream[16], &store_multiple, sizeof(store_multiple));
+
+    opcode = 0x1;
+    reg_num = 68;
+    mov48 = (opcode << 56) | (reg_num << 48) | 0;
+    mov48 = htole64(mov48);
+    memcpy(&command_stream[24], &mov48, sizeof(mov48));
+
+    opcode = 36;
+    flush = opcode << 56 | 0ull << 48 | reg_num << 40 | 0ull << 16 | 0x233;
+    flush = htole64(flush);
+    memcpy(&command_stream[32], &flush, sizeof(flush));
+}
+
+igt_main {
+  int fd;
+
+  igt_fixture { fd = drm_open_driver(DRIVER_PANTHOR); }
+
+  igt_subtest("group_create") {
+    struct drm_panthor_gpu_info gpu_info = {};
+    struct drm_panthor_vm_create vm_create = {};
+    struct drm_panthor_group_create group_create = {};
+    struct drm_panthor_queue_create queue = {};
+    struct drm_panthor_obj_array queues;
+    struct drm_panthor_group_destroy group_destroy;
+    struct drm_panthor_vm_destroy vm_destroy;
+
+    igt_panthor_query(fd, DRM_PANTHOR_DEV_QUERY_GPU_INFO, &gpu_info, sizeof(gpu_info), 0);
+    igt_assert(gpu_info.gpu_id != 0);
+
+    vm_create.flags = 0;
+    igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_VM_CREATE, &vm_create), 0);
+    igt_assert(vm_create.id != 0);
+
+    queue.priority = 0; // Low priority
+    queue.ringbuf_size = 4096; // Example size
+    queues = (struct drm_panthor_obj_array)DRM_PANTHOR_OBJ_ARRAY(1, &queue);
+
+    group_create.queues = queues;
+    group_create.max_compute_cores = 1;
+    group_create.max_fragment_cores = 1;
+    group_create.max_tiler_cores = 1;
+    group_create.priority = PANTHOR_GROUP_PRIORITY_MEDIUM;
+    group_create.compute_core_mask = gpu_info.shader_present & 0x1; // Use first core
+    group_create.fragment_core_mask = gpu_info.shader_present & 0x1; // Use first core
+    group_create.tiler_core_mask = gpu_info.tiler_present & 0x1; // Use first tiler
+    group_create.vm_id = vm_create.id;
+
+    igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_GROUP_CREATE, &group_create), 0);
+    igt_assert(group_create.group_handle != 0);
+
+    // Cleanup: Destroy the group and VM
+    group_destroy = (struct drm_panthor_group_destroy){ .group_handle = group_create.group_handle };
+    igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_GROUP_DESTROY, &group_destroy), 0);
+
+    vm_destroy = (struct drm_panthor_vm_destroy) { .id = vm_create.id };
+    igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_VM_DESTROY, &vm_destroy), 0);
+  }
+
+  igt_subtest("group_submit") {
+    struct drm_panthor_gpu_info gpu_info = {};
+    struct drm_panthor_vm_create vm_create = {};
+    struct drm_panthor_group_create group_create = {};
+    struct drm_panthor_queue_create queue = {};
+    struct drm_panthor_obj_array queues;
+    struct drm_panthor_group_submit group_submit = {};
+    struct drm_panthor_queue_submit queue_submit = {};
+    struct drm_panthor_group_destroy group_destroy;
+    struct drm_panthor_obj_array queue_submits;
+    struct drm_panthor_vm_destroy vm_destroy;
+    struct drm_panthor_bo_create bo_create = {};
+    struct drm_panthor_vm_bind vm_bind = {};
+    struct drm_panthor_vm_bind_op vm_bind_op = {};
+    struct drm_syncobj_wait wait = {};
+    struct drm_syncobj_create syncobj_create = {};
+    struct drm_panthor_sync_op sync_op = {};
+    struct drm_gem_close gem_close = {};
+    struct drm_syncobj_destroy syncobj_destroy = {};
+    uint64_t command_stream_gpu_addr;
+    uint32_t command_stream_size;
+    uint64_t result_gpu_addr;
+    uint32_t cmd_buf_bo_handle;
+    uint32_t result_bo_handle;
+    uint32_t syncobj_handle;
+    uint8_t command_stream[64] = {0};
+    uint8_t *bo_cpu_addr;
+    uint8_t *result_cpu_addr;
+    const int INITIAL_VA = 0x1000000;
+
+
+    igt_panthor_query(fd, DRM_PANTHOR_DEV_QUERY_GPU_INFO, &gpu_info, sizeof(gpu_info), 0);
+    igt_assert(gpu_info.gpu_id != 0);
+
+    vm_create.flags = 0;
+    igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_VM_CREATE, &vm_create), 0);
+    igt_assert(vm_create.id != 0);
+
+    bo_create.size = 4096;
+    bo_create.flags = 0;
+    bo_create.exclusive_vm_id = vm_create.id;
+    igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_BO_CREATE, &bo_create), 0);
+    igt_assert(bo_create.handle != 0);
+    cmd_buf_bo_handle = bo_create.handle;
+
+    vm_bind_op.flags = DRM_PANTHOR_VM_BIND_OP_TYPE_MAP;
+    vm_bind_op.bo_handle = cmd_buf_bo_handle;
+    vm_bind_op.bo_offset = 0;
+    vm_bind_op.va = INITIAL_VA;
+    vm_bind_op.size = bo_create.size;
+    vm_bind.ops = (struct drm_panthor_obj_array)DRM_PANTHOR_OBJ_ARRAY(1, &vm_bind_op);
+    vm_bind.vm_id = vm_create.id;
+    vm_bind.flags = 0;
+    igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_VM_BIND, &vm_bind), 0);
+
+    command_stream_gpu_addr = vm_bind_op.va;
+    command_stream_size = sizeof(command_stream);
+
+    bo_cpu_addr = igt_panthor_mmap_bo(fd, bo_create.handle, bo_create.size, PROT_READ | PROT_WRITE);
+    igt_assert(bo_cpu_addr != MAP_FAILED);
+
+    // Create the BO to receive the result of the store.
+    memset(&bo_create, 0, sizeof(bo_create));
+    bo_create.size = 4096;
+    bo_create.flags = 0;
+    bo_create.exclusive_vm_id = vm_create.id;
+    igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_BO_CREATE, &bo_create), 0);
+    igt_assert(bo_create.handle != 0);
+    result_bo_handle = bo_create.handle;
+
+    // Also bind the result BO.
+    vm_bind_op.flags = DRM_PANTHOR_VM_BIND_OP_TYPE_MAP;
+    vm_bind_op.bo_handle = result_bo_handle;
+    vm_bind_op.bo_offset = 0;
+    vm_bind_op.va = INITIAL_VA + 4096;
+    vm_bind_op.size = bo_create.size;
+    vm_bind.ops = (struct drm_panthor_obj_array)DRM_PANTHOR_OBJ_ARRAY(1, &vm_bind_op);
+    vm_bind.vm_id = vm_create.id;
+    vm_bind.flags = 0;
+    igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_VM_BIND, &vm_bind), 0);
+    result_gpu_addr = vm_bind_op.va;
+
+    issue_store_multiple(command_stream, result_gpu_addr, 0xdeadbeef);
+    memcpy(bo_cpu_addr, command_stream, command_stream_size);
+    munmap(bo_cpu_addr, bo_create.size);
+
+    queue.priority = 0;
+    queue.ringbuf_size = 4096;
+    queues = (struct drm_panthor_obj_array)DRM_PANTHOR_OBJ_ARRAY(1, &queue);
+
+    group_create.queues = queues;
+    group_create.max_compute_cores = 1;
+    group_create.max_fragment_cores = 1;
+    group_create.max_tiler_cores = 1;
+    group_create.priority = PANTHOR_GROUP_PRIORITY_MEDIUM;
+    group_create.compute_core_mask = gpu_info.shader_present & 0x1;
+    group_create.fragment_core_mask = gpu_info.shader_present & 0x1;
+    group_create.tiler_core_mask = gpu_info.tiler_present & 0x1;
+    group_create.vm_id = vm_create.id;
+
+    igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_GROUP_CREATE, &group_create), 0);
+    igt_assert(group_create.group_handle != 0);
+
+    syncobj_create = (struct drm_syncobj_create){
+        .flags = 0,
+    };
+
+    igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_SYNCOBJ_CREATE, &syncobj_create), 0);
+    syncobj_handle = syncobj_create.handle;
+
+    sync_op = (struct drm_panthor_sync_op) {
+        .handle = syncobj_handle,
+        .flags = DRM_PANTHOR_SYNC_OP_SIGNAL,
+    };
+
+    queue_submit.syncs = (struct drm_panthor_obj_array)DRM_PANTHOR_OBJ_ARRAY(1, &sync_op);
+
+    queue_submit.queue_index = 0;
+    queue_submit.stream_size = command_stream_size;
+    queue_submit.stream_addr = command_stream_gpu_addr;
+    queue_submit.latest_flush = 0;
+    queue_submits = (struct drm_panthor_obj_array)DRM_PANTHOR_OBJ_ARRAY(1, &queue_submit);
+
+    group_submit.group_handle = group_create.group_handle;
+    group_submit.queue_submits = queue_submits;
+
+    igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_GROUP_SUBMIT, &group_submit), 0);
+
+    wait = (struct drm_syncobj_wait) {
+      .handles = (uint64_t)&syncobj_handle,
+      .count_handles = 1,
+      .timeout_nsec = INT64_MAX,
+      .flags = 0,
+    };
+
+    igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_SYNCOBJ_WAIT, &wait), 0);
+
+    result_cpu_addr = igt_panthor_mmap_bo(fd, bo_create.handle, bo_create.size, PROT_READ | PROT_WRITE);
+
+    igt_assert(*(uint32_t *)result_cpu_addr == 0xdeadbeef);
+    munmap(result_cpu_addr, bo_create.size);
+
+    syncobj_destroy.handle = syncobj_handle;
+    igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_SYNCOBJ_DESTROY, &syncobj_destroy), 0);
+
+
+    group_destroy.group_handle = group_create.group_handle;
+    igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_GROUP_DESTROY, &group_destroy), 0);
+
+    vm_destroy.id = vm_create.id;
+    igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_PANTHOR_VM_DESTROY, &vm_destroy), 0);
+
+    gem_close.handle = cmd_buf_bo_handle;
+    igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_GEM_CLOSE, &gem_close), 0);
+
+    gem_close.handle = result_bo_handle;
+    igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_GEM_CLOSE, &gem_close), 0);
+
+}
+
+  igt_fixture { drm_close_driver(fd); }
+}
diff --git a/tests/panthor/panthor_query.c b/tests/panthor/panthor_query.c
new file mode 100644
index 000000000..3bbecf3a6
--- /dev/null
+++ b/tests/panthor/panthor_query.c
@@ -0,0 +1,25 @@
+// SPDX-License-Identifier: MIT
+// SPDX-FileCopyrightText: Copyright (C) 2025 Collabora Ltd.
+
+#include "igt.h"
+#include "igt_core.h"
+#include "igt_panthor.h"
+#include "panthor_drm.h"
+#include <stdint.h>
+
+igt_main {
+  int fd;
+
+  igt_fixture { fd = drm_open_driver(DRIVER_PANTHOR); }
+
+  igt_subtest("query") {
+    struct drm_panthor_gpu_info gpu = {};
+
+    igt_panthor_query(fd, DRM_PANTHOR_DEV_QUERY_GPU_INFO, &gpu, sizeof(gpu), 0);
+
+    igt_assert(gpu.gpu_id != 0);
+    igt_assert(gpu.gpu_rev != 0);
+  }
+
+  igt_fixture { drm_close_driver(fd); }
+}
diff --git a/tests/panthor/panthor_vm.c b/tests/panthor/panthor_vm.c
new file mode 100644
index 000000000..484602de3
--- /dev/null
+++ b/tests/panthor/panthor_vm.c
@@ -0,0 +1,73 @@
+// SPDX-License-Identifier: MIT
+// SPDX-FileCopyrightText: Copyright (C) 2025 Collabora Ltd.
+
+#include "igt.h"
+#include "igt_core.h"
+#include "igt_panthor.h"
+#include "panthor_drm.h"
+
+igt_main {
+  int fd;
+
+  igt_fixture { fd = drm_open_driver(DRIVER_PANTHOR); }
+
+  igt_subtest("vm_create_destroy") {
+    uint32_t vm_id;
+
+    igt_panthor_vm_create(fd, &vm_id, 0);
+    igt_assert(vm_id != 0);
+
+    igt_panthor_vm_destroy(fd, vm_id, 0);
+  }
+
+  igt_subtest("vm_destroy_invalid") {
+    igt_panthor_vm_destroy(fd, 0xdeadbeef, EINVAL);
+  }
+
+  igt_subtest("vm_bind") {
+    uint32_t vm_id;
+    struct panthor_bo bo;
+    uint64_t bo_size = 0x1000;
+
+    igt_panthor_vm_create(fd, &vm_id, 0);
+    igt_assert(vm_id != 0);
+
+    igt_panthor_bo_create(fd, &bo, bo_size, 0, 0);
+    igt_panthor_vm_bind(fd, vm_id, bo.handle, 0x1000, 0x1000, DRM_PANTHOR_VM_BIND_OP_TYPE_MAP, 0);
+
+    igt_panthor_vm_destroy(fd, vm_id, 0);
+  }
+
+  igt_subtest("vm_unbind") {
+    uint32_t vm_id;
+    struct panthor_bo bo;
+    uint64_t bo_size = 0x1000;
+
+    igt_panthor_vm_create(fd, &vm_id, 0);
+    igt_assert(vm_id != 0);
+
+    igt_panthor_bo_create(fd, &bo, bo_size, 0, 0);
+    igt_panthor_vm_bind(fd, vm_id, bo.handle, 0x1000, 0x1000, DRM_PANTHOR_VM_BIND_OP_TYPE_MAP, 0);
+    igt_panthor_vm_bind(fd, vm_id, 0, 0x1000, 0x1000, DRM_PANTHOR_VM_BIND_OP_TYPE_UNMAP, 0);
+
+    igt_panthor_vm_destroy(fd, vm_id, 0);
+  }
+
+  igt_subtest("vm_unbind_invalid_address") {
+    uint32_t vm_id;
+    struct panthor_bo bo;
+    uint64_t bo_size = 0x1000;
+
+    igt_panthor_vm_create(fd, &vm_id, 0);
+    igt_assert(vm_id != 0);
+
+    igt_panthor_bo_create(fd, &bo, bo_size, 0, 0);
+
+    /* This was not bound previously*/
+    igt_panthor_vm_bind(fd, vm_id, bo.handle, 0x1000, 0x1000, DRM_PANTHOR_VM_BIND_OP_TYPE_UNMAP, EINVAL);
+    igt_panthor_vm_destroy(fd, vm_id, 0);
+  }
+
+
+  igt_fixture { drm_close_driver(fd); }
+}
-- 
2.50.1



More information about the igt-dev mailing list