[igt-dev] [PATCH i-g-t] tests/i915/gem_sync: Add support for local memory

sai.gowtham.ch at intel.com sai.gowtham.ch at intel.com
Tue Jan 11 16:56:00 UTC 2022


From: Ch Sai Gowtham <sai.gowtham.ch at intel.com>

Add support for local memory region (Device memory)

Signed-off-by: Ch Sai Gowtham <sai.gowtham.ch at intel.com>
Cc: Zbigniew Kempczyński <zbigniew.kempczynski at intel.com>
---
 lib/igt_dummyload.h   |   1 +
 tests/i915/gem_sync.c | 206 ++++++++++++++++++++++++++----------------
 2 files changed, 129 insertions(+), 78 deletions(-)

diff --git a/lib/igt_dummyload.h b/lib/igt_dummyload.h
index f0205261..649b6ca4 100644
--- a/lib/igt_dummyload.h
+++ b/lib/igt_dummyload.h
@@ -81,6 +81,7 @@ typedef struct igt_spin_factory {
 	unsigned int flags;
 	int fence;
 	uint64_t ahnd;
+	uint32_t region;
 } igt_spin_factory_t;
 
 #define IGT_SPIN_FENCE_IN      (1 << 0)
diff --git a/tests/i915/gem_sync.c b/tests/i915/gem_sync.c
index 8c435845..43626a99 100644
--- a/tests/i915/gem_sync.c
+++ b/tests/i915/gem_sync.c
@@ -143,9 +143,20 @@ static void xchg_engine(void *array, unsigned i, unsigned j)
 	igt_swap(E[i], E[j]);
 }
 
+static uint32_t batch_create(int fd, uint32_t batch_size, uint32_t region)
+{
+	const uint32_t bbe = MI_BATCH_BUFFER_END;
+	uint32_t handle;
+
+	handle = gem_create_in_memory_regions(fd, batch_size, region);
+	gem_write(fd, handle, 0, &bbe, sizeof(bbe));
+
+	return handle;
+}
+
 static void
 sync_ring(int fd, const intel_ctx_t *ctx,
-	  unsigned ring, int num_children, int timeout)
+	  unsigned ring, int num_children, int timeout, uint32_t region)
 {
 	struct intel_engine_data ied;
 
@@ -155,15 +166,13 @@ sync_ring(int fd, const intel_ctx_t *ctx,
 
 	intel_detect_and_clear_missed_interrupts(fd);
 	igt_fork(child, num_children) {
-		const uint32_t bbe = MI_BATCH_BUFFER_END;
 		struct drm_i915_gem_exec_object2 object;
 		struct drm_i915_gem_execbuffer2 execbuf;
 		double start, elapsed;
 		unsigned long cycles;
 
 		memset(&object, 0, sizeof(object));
-		object.handle = gem_create(fd, 4096);
-		gem_write(fd, object.handle, 0, &bbe, sizeof(bbe));
+		object.handle = batch_create(fd, 4096, region);
 
 		memset(&execbuf, 0, sizeof(execbuf));
 		execbuf.buffers_ptr = to_user_pointer(&object);
@@ -193,9 +202,8 @@ sync_ring(int fd, const intel_ctx_t *ctx,
 
 static void
 idle_ring(int fd, const intel_ctx_t *ctx, unsigned int ring,
-	  int num_children, int timeout)
+	  int num_children, int timeout, uint32_t region)
 {
-	const uint32_t bbe = MI_BATCH_BUFFER_END;
 	struct drm_i915_gem_exec_object2 object;
 	struct drm_i915_gem_execbuffer2 execbuf;
 	double start, elapsed;
@@ -204,8 +212,7 @@ idle_ring(int fd, const intel_ctx_t *ctx, unsigned int ring,
 	gem_require_ring(fd, ring);
 
 	memset(&object, 0, sizeof(object));
-	object.handle = gem_create(fd, 4096);
-	gem_write(fd, object.handle, 0, &bbe, sizeof(bbe));
+	object.handle = batch_create(fd, 4096, region);
 
 	memset(&execbuf, 0, sizeof(execbuf));
 	execbuf.buffers_ptr = to_user_pointer(&object);
@@ -234,7 +241,7 @@ idle_ring(int fd, const intel_ctx_t *ctx, unsigned int ring,
 
 static void
 wakeup_ring(int fd, const intel_ctx_t *ctx, unsigned ring,
-	    int timeout, int wlen)
+	    int timeout, int wlen, uint32_t region)
 {
 	struct intel_engine_data ied;
 	uint64_t ahnd = get_reloc_ahnd(fd, ctx->id);
@@ -244,7 +251,6 @@ wakeup_ring(int fd, const intel_ctx_t *ctx, unsigned ring,
 
 	intel_detect_and_clear_missed_interrupts(fd);
 	igt_fork(child, ied.nengines) {
-		const uint32_t bbe = MI_BATCH_BUFFER_END;
 		struct drm_i915_gem_exec_object2 object;
 		struct drm_i915_gem_execbuffer2 execbuf;
 		double end, this, elapsed, now, baseline;
@@ -254,11 +260,10 @@ wakeup_ring(int fd, const intel_ctx_t *ctx, unsigned ring,
 		ahnd = get_reloc_ahnd(fd, ctx->id);
 
 		memset(&object, 0, sizeof(object));
-		object.handle = gem_create(fd, 4096);
+		object.handle = batch_create(fd, 4096, region);
 		object.offset = get_offset(ahnd, object.handle, 4096, 0);
 		if (ahnd)
 			object.flags = EXEC_OBJECT_PINNED;
-		gem_write(fd, object.handle, 0, &bbe, sizeof(bbe));
 
 		memset(&execbuf, 0, sizeof(execbuf));
 		execbuf.buffers_ptr = to_user_pointer(&object);
@@ -339,7 +344,7 @@ wakeup_ring(int fd, const intel_ctx_t *ctx, unsigned ring,
 }
 
 static void active_ring(int fd, const intel_ctx_t *ctx, unsigned int ring,
-			int num_children, int timeout)
+			int num_children, int timeout, uint32_t region)
 {
 	struct intel_engine_data ied;
 	uint64_t ahnd = get_reloc_ahnd(fd, ctx->id);
@@ -359,13 +364,15 @@ static void active_ring(int fd, const intel_ctx_t *ctx, unsigned int ring,
 					 .ahnd = ahnd,
 					 .ctx = ctx,
 					 .engine = ied_flags(&ied, child),
-					 .flags = IGT_SPIN_FAST);
+					 .flags = IGT_SPIN_FAST,
+					 .region = region);
 
 		spin[1] = __igt_spin_new(fd,
 					 .ahnd = ahnd,
 					 .ctx = ctx,
 					 .engine = ied_flags(&ied, child),
-					 .flags = IGT_SPIN_FAST);
+					 .flags = IGT_SPIN_FAST,
+					 .region = region);
 
 		start = gettime();
 		end = start + timeout;
@@ -398,7 +405,7 @@ static void active_ring(int fd, const intel_ctx_t *ctx, unsigned int ring,
 
 static void
 active_wakeup_ring(int fd, const intel_ctx_t *ctx, unsigned ring,
-		   int timeout, int wlen)
+		   int timeout, int wlen, uint32_t region)
 {
 	struct intel_engine_data ied;
 	uint64_t ahnd0 = get_reloc_ahnd(fd, 0);
@@ -409,7 +416,6 @@ active_wakeup_ring(int fd, const intel_ctx_t *ctx, unsigned ring,
 
 	intel_detect_and_clear_missed_interrupts(fd);
 	igt_fork(child, ied.nengines) {
-		const uint32_t bbe = MI_BATCH_BUFFER_END;
 		struct drm_i915_gem_exec_object2 object;
 		struct drm_i915_gem_execbuffer2 execbuf;
 		double end, this, elapsed, now, baseline;
@@ -420,11 +426,10 @@ active_wakeup_ring(int fd, const intel_ctx_t *ctx, unsigned ring,
 		ahnd = get_reloc_ahnd(fd, ctx->id);
 
 		memset(&object, 0, sizeof(object));
-		object.handle = gem_create(fd, 4096);
+		object.handle = batch_create(fd, 4096, region);
 		object.offset = get_offset(ahnd, object.handle, 4096, 0);
 		if (ahnd)
 			object.offset = EXEC_OBJECT_PINNED;
-		gem_write(fd, object.handle, 0, &bbe, sizeof(bbe));
 
 		memset(&execbuf, 0, sizeof(execbuf));
 		execbuf.buffers_ptr = to_user_pointer(&object);
@@ -529,7 +534,7 @@ active_wakeup_ring(int fd, const intel_ctx_t *ctx, unsigned ring,
 
 static void
 store_ring(int fd, const intel_ctx_t *ctx, unsigned ring,
-	   int num_children, int timeout)
+	   int num_children, int timeout, uint32_t region)
 {
 	const unsigned int gen = intel_gen(intel_get_drm_devid(fd));
 	struct intel_engine_data ied;
@@ -541,7 +546,6 @@ store_ring(int fd, const intel_ctx_t *ctx, unsigned ring,
 
 	intel_detect_and_clear_missed_interrupts(fd);
 	igt_fork(child, num_children) {
-		const uint32_t bbe = MI_BATCH_BUFFER_END;
 		struct drm_i915_gem_exec_object2 object[2];
 		struct drm_i915_gem_relocation_entry reloc[1024];
 		struct drm_i915_gem_execbuffer2 execbuf;
@@ -559,14 +563,13 @@ store_ring(int fd, const intel_ctx_t *ctx, unsigned ring,
 		execbuf.rsvd1 = ctx->id;
 
 		memset(object, 0, sizeof(object));
-		object[0].handle = gem_create(fd, 4096);
-		gem_write(fd, object[0].handle, 0, &bbe, sizeof(bbe));
+		object[0].handle = batch_create(fd, 4096, region);
 		execbuf.buffer_count = 1;
 		gem_execbuf(fd, &execbuf);
 
 		object[0].flags |= EXEC_OBJECT_WRITE;
 		object[0].flags |= has_relocs ? 0 : EXEC_OBJECT_PINNED;
-		object[1].handle = gem_create(fd, 20*1024);
+		object[1].handle = gem_create_in_memory_regions(fd, 20*1024, region);
 
 		object[1].relocs_ptr = to_user_pointer(reloc);
 		object[1].relocation_count = has_relocs ? 1024 : 0;
@@ -629,7 +632,7 @@ store_ring(int fd, const intel_ctx_t *ctx, unsigned ring,
 
 static void
 switch_ring(int fd, const intel_ctx_t *ctx, unsigned ring,
-	    int num_children, int timeout)
+	    int num_children, int timeout, uint32_t region)
 {
 	const unsigned int gen = intel_gen(intel_get_drm_devid(fd));
 	struct intel_engine_data ied;
@@ -653,7 +656,6 @@ switch_ring(int fd, const intel_ctx_t *ctx, unsigned ring,
 		unsigned long cycles;
 
 		for (int i = 0; i < ARRAY_SIZE(contexts); i++) {
-			const uint32_t bbe = MI_BATCH_BUFFER_END;
 			const uint32_t sz = 32 << 10;
 			struct context *c = &contexts[i];
 			uint32_t *batch, *b;
@@ -670,13 +672,12 @@ switch_ring(int fd, const intel_ctx_t *ctx, unsigned ring,
 			c->execbuf.rsvd1 = c->ctx->id;
 
 			memset(c->object, 0, sizeof(c->object));
-			c->object[0].handle = gem_create(fd, 4096);
-			gem_write(fd, c->object[0].handle, 0, &bbe, sizeof(bbe));
+			c->object[0].handle = batch_create(fd, 4096, region);
 			c->execbuf.buffer_count = 1;
 			gem_execbuf(fd, &c->execbuf);
 
 			c->object[0].flags |= EXEC_OBJECT_WRITE;
-			c->object[1].handle = gem_create(fd, sz);
+			c->object[1].handle = gem_create_in_memory_regions(fd, sz, region);
 
 			c->object[1].relocs_ptr = to_user_pointer(c->reloc);
 			c->object[1].relocation_count = has_relocs ? 1024 * i : 0;
@@ -813,10 +814,9 @@ static void *waiter(void *arg)
 
 static void
 __store_many(int fd, const intel_ctx_t *ctx, unsigned ring,
-	     int timeout, unsigned long *cycles)
+	     int timeout, unsigned long *cycles, uint32_t region)
 {
 	const unsigned int gen = intel_gen(intel_get_drm_devid(fd));
-	const uint32_t bbe = MI_BATCH_BUFFER_END;
 	struct drm_i915_gem_exec_object2 object[2];
 	struct drm_i915_gem_execbuffer2 execbuf;
 	struct drm_i915_gem_relocation_entry reloc[1024];
@@ -836,8 +836,7 @@ __store_many(int fd, const intel_ctx_t *ctx, unsigned ring,
 	execbuf.rsvd1 = ctx->id;
 
 	memset(object, 0, sizeof(object));
-	object[0].handle = gem_create(fd, 4096);
-	gem_write(fd, object[0].handle, 0, &bbe, sizeof(bbe));
+	object[0].handle = batch_create(fd, 4096, region);
 	execbuf.buffer_count = 1;
 	gem_execbuf(fd, &execbuf);
 	object[0].flags |= EXEC_OBJECT_WRITE;
@@ -945,7 +944,7 @@ __store_many(int fd, const intel_ctx_t *ctx, unsigned ring,
 
 static void
 store_many(int fd, const intel_ctx_t *ctx, unsigned int ring,
-	   int num_children, int timeout)
+	   int num_children, int timeout, uint32_t region)
 {
 	struct intel_engine_data ied;
 	unsigned long *shared;
@@ -963,7 +962,8 @@ store_many(int fd, const intel_ctx_t *ctx, unsigned int ring,
 			__store_many(fd, ctx,
 				     ied_flags(&ied, n),
 				     timeout,
-				     &shared[n]);
+				     &shared[n],
+				     region);
 	}
 	igt_waitchildren();
 
@@ -976,7 +976,7 @@ store_many(int fd, const intel_ctx_t *ctx, unsigned int ring,
 }
 
 static void
-sync_all(int fd, const intel_ctx_t *ctx, int num_children, int timeout)
+sync_all(int fd, const intel_ctx_t *ctx, int num_children, int timeout, uint32_t region)
 {
 	struct intel_engine_data ied;
 
@@ -985,15 +985,13 @@ sync_all(int fd, const intel_ctx_t *ctx, int num_children, int timeout)
 
 	intel_detect_and_clear_missed_interrupts(fd);
 	igt_fork(child, num_children) {
-		const uint32_t bbe = MI_BATCH_BUFFER_END;
 		struct drm_i915_gem_exec_object2 object;
 		struct drm_i915_gem_execbuffer2 execbuf;
 		double start, elapsed;
 		unsigned long cycles;
 
 		memset(&object, 0, sizeof(object));
-		object.handle = gem_create(fd, 4096);
-		gem_write(fd, object.handle, 0, &bbe, sizeof(bbe));
+		object.handle = batch_create(fd, 4096, region);
 
 		memset(&execbuf, 0, sizeof(execbuf));
 		execbuf.buffers_ptr = to_user_pointer(&object);
@@ -1023,7 +1021,7 @@ sync_all(int fd, const intel_ctx_t *ctx, int num_children, int timeout)
 }
 
 static void
-store_all(int fd, const intel_ctx_t *ctx, int num_children, int timeout)
+store_all(int fd, const intel_ctx_t *ctx, int num_children, int timeout, uint32_t region)
 {
 	const unsigned int gen = intel_gen(intel_get_drm_devid(fd));
 	struct intel_engine_data ied;
@@ -1034,7 +1032,6 @@ store_all(int fd, const intel_ctx_t *ctx, int num_children, int timeout)
 
 	intel_detect_and_clear_missed_interrupts(fd);
 	igt_fork(child, num_children) {
-		const uint32_t bbe = MI_BATCH_BUFFER_END;
 		struct drm_i915_gem_exec_object2 object[2];
 		struct drm_i915_gem_relocation_entry reloc[1024];
 		struct drm_i915_gem_execbuffer2 execbuf;
@@ -1051,14 +1048,13 @@ store_all(int fd, const intel_ctx_t *ctx, int num_children, int timeout)
 		execbuf.rsvd1 = ctx->id;
 
 		memset(object, 0, sizeof(object));
-		object[0].handle = gem_create(fd, 4096);
-		gem_write(fd, object[0].handle, 0, &bbe, sizeof(bbe));
+		object[0].handle = batch_create(fd, 4096, region);
 		execbuf.buffer_count = 1;
 		gem_execbuf(fd, &execbuf);
 
 		object[0].flags |= EXEC_OBJECT_WRITE;
 		object[0].flags |= has_relocs ? 0 : EXEC_OBJECT_PINNED;
-		object[1].handle = gem_create(fd, 1024*16 + 4096);
+		object[1].handle = gem_create_in_memory_regions(fd, 1024*16 + 4096, region);
 
 		object[1].relocs_ptr = to_user_pointer(reloc);
 		object[1].relocation_count = has_relocs ? 1024 : 0;
@@ -1126,7 +1122,7 @@ store_all(int fd, const intel_ctx_t *ctx, int num_children, int timeout)
 
 static void
 preempt(int fd, const intel_ctx_t *ctx, unsigned ring,
-	int num_children, int timeout)
+	int num_children, int timeout, uint32_t region)
 {
 	struct intel_engine_data ied;
 	const intel_ctx_t *tmp_ctx[2];
@@ -1144,7 +1140,6 @@ preempt(int fd, const intel_ctx_t *ctx, unsigned ring,
 
 	intel_detect_and_clear_missed_interrupts(fd);
 	igt_fork(child, num_children) {
-		const uint32_t bbe = MI_BATCH_BUFFER_END;
 		struct drm_i915_gem_exec_object2 object;
 		struct drm_i915_gem_execbuffer2 execbuf;
 		double start, elapsed;
@@ -1153,11 +1148,10 @@ preempt(int fd, const intel_ctx_t *ctx, unsigned ring,
 		ahnd = get_reloc_ahnd(fd, 0);
 
 		memset(&object, 0, sizeof(object));
-		object.handle = gem_create(fd, 4096);
+		object.handle = batch_create(fd, 4096, region);
 		object.offset = get_offset(ahnd, object.handle, 4096, 0);
 		if (ahnd)
 			object.flags = EXEC_OBJECT_PINNED;
-		gem_write(fd, object.handle, 0, &bbe, sizeof(bbe));
 
 		memset(&execbuf, 0, sizeof(execbuf));
 		execbuf.buffers_ptr = to_user_pointer(&object);
@@ -1205,7 +1199,7 @@ igt_main
 	const struct {
 		const char *name;
 		void (*func)(int fd, const intel_ctx_t *ctx, unsigned int engine,
-			     int num_children, int timeout);
+			     int num_children, int timeout, uint32_t memregion);
 		int num_children;
 		int timeout;
 	} all[] = {
@@ -1236,11 +1230,40 @@ igt_main
 		{ "forked-store", store_ring, ncpus, 20 },
 		{}
 	};
+	const struct test {
+		const char *name;
+		void (*func) (int, const intel_ctx_t *, int, int, uint32_t);
+		int num_children;
+		int timeout;
+	} *test, tests[] = {
+		{ "basic-all", sync_all, 1, 2},
+		{ "basic-store-all", store_all, 1, 2},
+		{ "all", sync_all, 1, 20},
+		{ "store-all", store_all, 1, 20},
+		{ "forked-all", sync_all, ncpus, 20},
+		{ "forked-store-all", store_all, ncpus, 20},
+		{}
+	};
+#define subtest_for_each_combination(__name, __ctx, __num_children, __timeout, __func) \
+	igt_subtest_with_dynamic(__name) { \
+		for_each_combination(regions, 1, set) { \
+			sub_name = memregion_dynamic_subtest_name(regions); \
+			region = igt_collection_get_value(regions, 0); \
+			igt_dynamic_f("%s", sub_name) \
+				(__func)(fd, (__ctx), (__num_children), (__timeout), region); \
+			free(sub_name); \
+		} \
+	}
+
 #define for_each_test(t, T) for(typeof(*T) *t = T; t->name; t++)
 
 	const struct intel_execution_engine2 *e;
 	const intel_ctx_t *ctx;
 	int fd = -1;
+	struct drm_i915_query_memory_regions *query_info;
+	struct igt_collection *regions, *set;
+	char *sub_name;
+	uint32_t region;
 
 	igt_fixture {
 		fd = drm_open_driver(DRIVER_INTEL);
@@ -1250,6 +1273,11 @@ igt_main
 		ctx = intel_ctx_create_all_physical(fd);
 
 		igt_fork_hang_detector(fd);
+		query_info = gem_get_query_memory_regions(fd);
+		igt_assert(query_info);
+		set = get_memory_region_set(query_info,
+					I915_SYSTEM_MEMORY,
+					I915_DEVICE_MEMORY);
 		intel_allocator_multiprocess_start();
 	}
 
@@ -1257,41 +1285,48 @@ igt_main
 	for_each_test(t, individual) {
 		igt_subtest_with_dynamic_f("legacy-%s", t->name) {
 			for (const struct intel_execution_ring *l = intel_execution_rings; l->name; l++) {
-				igt_dynamic_f("%s", l->name) {
-					t->func(fd, intel_ctx_0(fd), eb_ring(l),
-						t->num_children, t->timeout);
+				for_each_combination(regions, 1, set) {
+					sub_name = memregion_dynamic_subtest_name(regions);
+					region = igt_collection_get_value(regions, 0);
+					igt_dynamic_f("%s-%s", l->name, sub_name) {
+						t->func(fd, intel_ctx_0(fd), eb_ring(l),
+								t->num_children,
+								t->timeout,
+								region);
+					}
+					free(sub_name);
 				}
 			}
 		}
 	}
-
-	igt_subtest("basic-all")
-		sync_all(fd, ctx, 1, 2);
-	igt_subtest("basic-store-all")
-		store_all(fd, ctx, 1, 2);
-
-	igt_subtest("all")
-		sync_all(fd, ctx, 1, 20);
-	igt_subtest("store-all")
-		store_all(fd, ctx, 1, 20);
-	igt_subtest("forked-all")
-		sync_all(fd, ctx, ncpus, 20);
-	igt_subtest("forked-store-all")
-		store_all(fd, ctx, ncpus, 20);
+	for (test= tests; test->name; test++)
+		subtest_for_each_combination(test->name, ctx, test->num_children, test->timeout, test->func);
 
 	for_each_test(t, all) {
-		igt_subtest_f("%s", t->name)
-			t->func(fd, ctx, ALL_ENGINES, t->num_children, t->timeout);
+		igt_subtest_with_dynamic_f("%s", t->name) {
+			for_each_combination(regions, 1, set) {
+				sub_name = memregion_dynamic_subtest_name(regions);
+				region = igt_collection_get_value(regions, 0);
+				igt_dynamic_f("%s", sub_name)
+					t->func(fd, ctx, ALL_ENGINES, t->num_children,
+							t->timeout, region);
+				free(sub_name);
+			}
+		}
 	}
 
 	/* New way of selecting engines. */
 	for_each_test(t, individual) {
 		igt_subtest_with_dynamic_f("%s", t->name) {
-			for_each_ctx_engine(fd, ctx, e) {
-				igt_dynamic_f("%s", e->name) {
-					t->func(fd, ctx, e->flags,
-						t->num_children, t->timeout);
-				}
+			for_each_combination(regions, 1, set) {
+				sub_name = memregion_dynamic_subtest_name(regions);
+				region = igt_collection_get_value(regions, 0);
+				for_each_ctx_engine(fd, ctx, e)
+					igt_dynamic_f("%s-%s", e->name, sub_name)
+						t->func(fd, ctx, e->flags,
+								t->num_children,
+								t->timeout, region);
+				free(sub_name);
 			}
 		}
 	}
@@ -1303,17 +1338,32 @@ igt_main
 			igt_require(gem_scheduler_has_preemption(fd));
 		}
 
-		igt_subtest("preempt-all")
-			preempt(fd, ctx, ALL_ENGINES, 1, 20);
+		igt_subtest_with_dynamic("preempt-all") {
+			for_each_combination(regions, 1, set) {
+				sub_name = memregion_dynamic_subtest_name(regions);
+				region = igt_collection_get_value(regions, 0);
+				igt_dynamic_f("%s", sub_name)
+					preempt(fd, ctx, ALL_ENGINES, 1, 20, region);
+				free(sub_name);
+			}
+		}
+
 		igt_subtest_with_dynamic("preempt") {
-			for_each_ctx_engine(fd, ctx, e) {
-				igt_dynamic_f("%s", e->name)
-					preempt(fd, ctx, e->flags, ncpus, 20);
+			for_each_combination(regions, 1, set) {
+				sub_name = memregion_dynamic_subtest_name(regions);
+				region = igt_collection_get_value(regions, 0);
+				for_each_ctx_engine(fd, ctx, e) {
+					igt_dynamic_f("%s-%s", e->name, sub_name)
+						preempt(fd, ctx, e->flags, ncpus, 20, region);
+					free(sub_name);
+				}
 			}
 		}
 	}
 
 	igt_fixture {
+		free(query_info);
+		igt_collection_destroy(set);
 		intel_allocator_multiprocess_stop();
 		igt_stop_hang_detector();
 		intel_ctx_destroy(fd, ctx);
-- 
2.32.0



More information about the igt-dev mailing list