[PATCH i-g-t 4/4] tests/xe_sriov_scheduling: nonpreempt-engine-resets subtest
Marcin Bernatowicz
marcin.bernatowicz at linux.intel.com
Mon Jan 20 20:34:45 UTC 2025
Verify the occurrence of engine resets
when non-preemptible workloads surpass the combined
duration of execution quantum and preemption timeout.
Signed-off-by: Marcin Bernatowicz <marcin.bernatowicz at linux.intel.com>
Cc: Adam Miszczak <adam.miszczak at linux.intel.com>
Cc: Jakub Kolakowski <jakub1.kolakowski at intel.com>
Cc: Lukasz Laguna <lukasz.laguna at intel.com>
Cc: Michał Wajdeczko <michal.wajdeczko at intel.com>
Cc: Michał Winiarski <michal.winiarski at intel.com>
Cc: Narasimha C V <narasimha.c.v at intel.com>
Cc: Piotr Piórkowski <piotr.piorkowski at intel.com>
Cc: Satyanarayana K V P <satyanarayana.k.v.p at intel.com>
Cc: Tomasz Lis <tomasz.lis at intel.com>
---
tests/intel/xe_sriov_scheduling.c | 126 ++++++++++++++++++++++++++++++
1 file changed, 126 insertions(+)
diff --git a/tests/intel/xe_sriov_scheduling.c b/tests/intel/xe_sriov_scheduling.c
index 20ec15b22..5999c3f98 100644
--- a/tests/intel/xe_sriov_scheduling.c
+++ b/tests/intel/xe_sriov_scheduling.c
@@ -605,6 +605,119 @@ static void throughput_ratio(int pf_fd, int num_vfs, const struct subm_opts *opt
igt_sriov_disable_vfs(pf_fd);
}
+static unsigned int select_random_exec_quantum_value(unsigned int min,
+ unsigned int num_vfs,
+ unsigned int job_timeout)
+{
+ int max = min(64u, job_timeout / (3 * (num_vfs + 1)));
+
+ igt_skip_on(max <= min);
+ /* random between min (inclusive) and max (exclusive) */
+ return rand() % (max - min) + min;
+}
+
+static struct vf_sched_params prepare_vf_sched_params(int num_vfs,
+ const struct subm_opts *opts)
+{
+ struct vf_sched_params params = {};
+
+ if (opts->exec_quantum_ms || opts->preempt_timeout_us) {
+ if (opts->exec_quantum_ms)
+ params.exec_quantum_ms = opts->exec_quantum_ms;
+ if (opts->preempt_timeout_us)
+ params.preempt_timeout_us = opts->preempt_timeout_us;
+ } else {
+ params.exec_quantum_ms =
+ select_random_exec_quantum_value(8, num_vfs, 5000);
+ params.preempt_timeout_us = 2 * params.exec_quantum_ms * 1000;
+ }
+
+ return params;
+}
+
+/**
+ * SUBTEST: nonpreempt-engine-resets
+ * Description:
+ * Check all VFs running a non-preemptible workload with a duration
+ * exceeding the sum of its execution quantum and preemption timeout,
+ * will experience engine reset due to preemption timeout.
+ */
+static void nonpreempt_engine_resets(int pf_fd, int num_vfs,
+ const struct subm_opts *opts)
+{
+ struct subm_set set_ = {}, *set = &set_;
+ struct vf_sched_params vf_sched_params =
+ prepare_vf_sched_params(num_vfs, opts);
+ uint64_t duration_ms = 2 * vf_sched_params.exec_quantum_ms +
+ vf_sched_params.preempt_timeout_us / 1000;
+ int preemptible_end = 1;
+ uint8_t vf_ids[num_vfs + 1 /*PF*/];
+
+ igt_info("eq=%ums pt=%uus duration=%lums num_vfs=%d\n",
+ vf_sched_params.exec_quantum_ms,
+ vf_sched_params.preempt_timeout_us, duration_ms, num_vfs);
+ igt_assert(duration_ms);
+ igt_assert_lt(duration_ms, 2000);
+
+ init_vf_ids(vf_ids, ARRAY_SIZE(vf_ids),
+ &(struct init_vf_ids_opts){ .shuffle = true,
+ .shuffle_pf = true });
+ xe_sriov_require_default_scheduling_attributes(pf_fd);
+ /* enable VFs */
+ igt_sriov_disable_driver_autoprobe(pf_fd);
+ igt_sriov_enable_vfs(pf_fd, num_vfs);
+ /* set scheduling params (PF and VFs) */
+ set_vfs_scheduling_params(pf_fd, num_vfs, &vf_sched_params);
+ /* probe VFs */
+ igt_sriov_enable_driver_autoprobe(pf_fd);
+ for (int vf = 1; vf <= num_vfs; ++vf)
+ igt_sriov_bind_vf_drm_driver(pf_fd, vf);
+
+ /* init subm_set */
+ subm_set_alloc_data(set, num_vfs + 1 /*PF*/);
+ subm_set_init_sync_method(set, opts->sync_method);
+
+ for (int n = 0; n < set->ndata; ++n) {
+ int vf_fd =
+ vf_ids[n] ?
+ igt_sriov_open_vf_drm_device(pf_fd, vf_ids[n]) :
+ drm_reopen_driver(pf_fd);
+
+ igt_assert_fd(vf_fd);
+ set->data[n].opts = opts;
+ subm_init(&set->data[n].subm, vf_fd, vf_ids[n], 0,
+ xe_engine(vf_fd, 0)->instance);
+ subm_workload_init(&set->data[n].subm,
+ &(struct subm_work_desc){
+ .duration_ms = duration_ms,
+ .preempt = (n < preemptible_end),
+ .repeats = 2000 / duration_ms });
+ igt_stats_init_with_size(&set->data[n].stats.samples,
+ set->data[n].subm.work.repeats);
+ if (set->sync_method == SYNC_BARRIER)
+ set->data[n].barrier = &set->barrier;
+ }
+
+ /* dispatch spinners, wait for results */
+ subm_set_dispatch_and_wait_threads(set);
+
+ /* verify results */
+ for (int n = 0; n < set->ndata; ++n) {
+ if (n < preemptible_end) {
+ igt_assert_eq(0, set->data[n].stats.num_early_finish);
+ igt_assert_eq(set->data[n].subm.work.repeats,
+ set->data[n].stats.samples.n_values);
+ } else {
+ igt_assert_eq(1, set->data[n].stats.num_early_finish);
+ }
+ }
+
+ /* cleanup */
+ subm_set_fini(set);
+ set_vfs_scheduling_params(pf_fd, num_vfs, &(struct vf_sched_params){});
+ igt_sriov_disable_vfs(pf_fd);
+}
+
static struct subm_opts subm_opts = {
.sync_method = SYNC_BARRIER,
.outlier_treshold = 0.1,
@@ -682,6 +795,19 @@ igt_main_args("s:e:p:", long_opts, help_str, subm_opts_handler, NULL)
throughput_ratio(pf_fd, vf, &subm_opts);
}
+ igt_describe("Check VFs experience engine reset due to preemption timeout");
+ igt_subtest_with_dynamic("nonpreempt-engine-resets") {
+ if (extended_scope)
+ for_each_sriov_num_vfs(pf_fd, vf)
+ igt_dynamic_f("numvfs-%d", vf)
+ nonpreempt_engine_resets(pf_fd, vf,
+ &subm_opts);
+
+ for_random_sriov_vf(pf_fd, vf)
+ igt_dynamic("numvfs-random")
+ nonpreempt_engine_resets(pf_fd, vf, &subm_opts);
+ }
+
igt_fixture {
set_vfs_scheduling_params(pf_fd, igt_sriov_get_total_vfs(pf_fd),
&(struct vf_sched_params){});
--
2.31.1
More information about the igt-dev
mailing list