[PATCH i-g-t] tests/intel/xe_exec_compute_mode: Add non-blocking subtest
sai.gowtham.ch at intel.com
sai.gowtham.ch at intel.com
Fri Jan 5 20:42:06 UTC 2024
From: Sai Gowtham Ch <sai.gowtham.ch at intel.com>
Fill the ring with maximum workload and expecte kernel to return
-EWOULDBLOCK error.
v2: Fixing batch buffer submission, updating sync mapping addresses
of VM, SPIN & EXEC for better readability. (Matthew Brost)
Signed-off-by: Matthew Brost <matthew.brost at intel.com>
Signed-off-by: Sai Gowtham Ch <sai.gowtham.ch at intel.com>
---
tests/intel/xe_exec_compute_mode.c | 111 +++++++++++++++++++++++++++++
1 file changed, 111 insertions(+)
diff --git a/tests/intel/xe_exec_compute_mode.c b/tests/intel/xe_exec_compute_mode.c
index b935ee0bc..6ebfa1a86 100644
--- a/tests/intel/xe_exec_compute_mode.c
+++ b/tests/intel/xe_exec_compute_mode.c
@@ -19,6 +19,7 @@
#include "xe/xe_ioctl.h"
#include "xe/xe_query.h"
+#include "xe/xe_spin.h"
#include <string.h>
#define MAX_N_EXECQUEUES 16
@@ -289,6 +290,112 @@ test_exec(int fd, struct drm_xe_engine_class_instance *eci,
close(map_fd);
}
+/**
+ * SUBTEST: non-blocking
+ * Description: Fill the ring and check we get expected errors.
+ * Test category: functionality test
+ */
+#define ONE_SEC MS_TO_NS(1000)
+#define VM_DATA 0
+#define SPIN_DATA 1
+#define EXEC_DATA 2
+#define DATA_COUNT 3
+static void non_block(int fd, int expect)
+{
+ uint64_t addr = 0x1a0000;
+ struct drm_xe_sync sync[1] = {
+ { .type = DRM_XE_SYNC_TYPE_USER_FENCE,
+ .flags = DRM_XE_SYNC_FLAG_SIGNAL,
+ .timeline_value = USER_FENCE_VALUE},
+ };
+ struct drm_xe_exec exec = {
+ .num_batch_buffer = 1,
+ .num_syncs = 1,
+ .syncs = to_user_pointer(&sync),
+ };
+ struct {
+ struct xe_spin spin;
+ uint32_t batch[16];
+ uint64_t vm_sync;
+ uint32_t data;
+ uint64_t exec_sync;
+ } *data = NULL;
+ struct xe_spin_opts spin_opts = { .addr = addr, .preempt = false };
+ struct drm_xe_engine *engine;
+ uint64_t batch_offset = (char *)&data[EXEC_DATA].batch - (char *)data;
+ uint64_t batch_addr = addr + batch_offset;
+ uint64_t sdi_offset = (char *)&data[EXEC_DATA].data - (char *)data;
+ uint64_t sdi_addr = addr + sdi_offset;
+ size_t bo_size;
+ uint32_t vm;
+ uint32_t exec_queue;
+ int value = 0x123456;
+ uint32_t bo = 0;
+ int b, err;
+
+ vm = xe_vm_create(fd, DRM_XE_VM_CREATE_FLAG_LR_MODE, 0);
+ bo_size = sizeof(*data) * DATA_COUNT;
+ bo_size = ALIGN(bo_size + xe_cs_prefetch_size(fd), xe_get_default_alignment(fd));
+
+ engine = xe_engine(fd, 1);
+ bo = xe_bo_create(fd, vm, bo_size, vram_if_possible(fd, engine->instance.gt_id), 0);
+
+ data = xe_bo_map(fd, bo, bo_size);
+ memset(data, 0, bo_size);
+ exec_queue = xe_exec_queue_create(fd, vm, &engine->instance, 0);
+
+ sync[0].addr = to_user_pointer(&data[VM_DATA].vm_sync);
+ xe_vm_bind_async(fd, vm, engine->instance.gt_id, bo, 0,
+ addr, bo_size, sync, 1);
+ xe_wait_ufence(fd, &data[VM_DATA].vm_sync, USER_FENCE_VALUE, 0, ONE_SEC);
+ data[VM_DATA].vm_sync = 0;
+
+ xe_spin_init(&data[SPIN_DATA].spin, &spin_opts);
+ sync[0].addr = addr + (char *)&data[SPIN_DATA].exec_sync - (char *)data;
+ exec.exec_queue_id = exec_queue;
+ exec.address = spin_opts.addr;
+ xe_exec(fd, &exec);
+ xe_spin_wait_started(&data[SPIN_DATA].spin);
+
+ b = 0;
+ data[0].batch[b++] = MI_STORE_DWORD_IMM_GEN4;
+ data[0].batch[b++] = sdi_addr;
+ data[0].batch[b++] = sdi_addr >> 32;
+ data[0].batch[b++] = value;
+ data[0].batch[b++] = MI_BATCH_BUFFER_END;
+ igt_assert(b <= ARRAY_SIZE(data->batch));
+
+ sync[0].addr = addr + (char *)&data[EXEC_DATA].exec_sync - (char *)data;
+ exec.num_syncs = 0;
+ exec.exec_queue_id = exec_queue;
+ exec.address = batch_addr;
+
+ while (1) {
+ err = __xe_exec(fd, &exec);
+
+ if (err == -EWOULDBLOCK)
+ break;
+ }
+ igt_assert_eq(err, expect);
+
+ xe_spin_end(&data[SPIN_DATA].spin);
+ xe_wait_ufence(fd, &data[SPIN_DATA].exec_sync, USER_FENCE_VALUE, 0, ONE_SEC);
+ usleep(50000);
+
+ exec.num_syncs = 1;
+ xe_exec(fd, &exec);
+ xe_wait_ufence(fd, &data[EXEC_DATA].exec_sync, USER_FENCE_VALUE, 0, ONE_SEC);
+
+ sync[0].addr = to_user_pointer(&data[VM_DATA].vm_sync);
+ xe_vm_unbind_async(fd, vm, 0, 0, addr, bo_size, sync, 1);
+ xe_wait_ufence(fd, &data[VM_DATA].vm_sync, USER_FENCE_VALUE, 0, ONE_SEC);
+ munmap(data, bo_size);
+ gem_close(fd, bo);
+
+ xe_exec_queue_destroy(fd, exec_queue);
+ xe_vm_destroy(fd, vm);
+}
+
igt_main
{
struct drm_xe_engine_class_instance *hwe;
@@ -346,6 +453,10 @@ igt_main
s->flags);
}
+ igt_subtest("non-blocking")
+ non_block(fd, -EWOULDBLOCK);
+
+
igt_fixture
drm_close_driver(fd);
}
--
2.39.1
More information about the igt-dev
mailing list