[Intel-gfx] [PATCH i-g-t] i915/gem_exec_schedule: Trick semaphores into a GPU hang
Tvrtko Ursulin
tvrtko.ursulin at linux.intel.com
Tue Apr 9 15:39:43 UTC 2019
On 09/04/2019 14:56, Chris Wilson wrote:
> If we have two tasks running on xcs0 and xcs1 independently, but who
> queue subsequent work onto rcs, we may insert semaphores before the rcs
> work and pick unwisely which task to run first. To maximise throughput,
> we want to run on rcs whichever task is ready first. Conversely, if we
> pick wrongly that can be used to trigger a GPU hang with unaware
> userspace.
>
> Signed-off-by: Chris Wilson <chris at chris-wilson.co.uk>
> Cc: Tvrtko Ursulin <tvrtko.ursulin at intel.com>
> ---
> tests/i915/gem_exec_schedule.c | 61 ++++++++++++++++++++++++++++++++++
> 1 file changed, 61 insertions(+)
>
> diff --git a/tests/i915/gem_exec_schedule.c b/tests/i915/gem_exec_schedule.c
> index 3df319bcc..d6f109540 100644
> --- a/tests/i915/gem_exec_schedule.c
> +++ b/tests/i915/gem_exec_schedule.c
> @@ -404,6 +404,65 @@ static void semaphore_userlock(int i915)
> igt_spin_batch_free(i915, spin);
> }
>
> +static void semaphore_codependency(int i915)
> +{
> + struct {
> + igt_spin_t *xcs, *rcs;
> + } task[2];
> + unsigned int engine;
> + int i;
> +
> + /*
> + * Consider two tasks, task A runs on (xcs0, rcs0) and task B
> + * on (xcs1, rcs0). That is they must both run a dependent
> + * batch on rcs0, after first running in parallel on separate
> + * engines. To maximise throughput, we want the shorter xcs task
> + * to start on rcs first. However, if we insert semaphores we may
> + * pick wrongly and end up running the requests in the least
> + * optimal order.
> + */
> +
> + i = 0;
> + for_each_physical_engine(i915, engine) {
> + uint32_t ctx;
> +
> + if (engine == I915_EXEC_RENDER)
> + continue;
> +
> + ctx = gem_context_create(i915);
> +
> + task[i].xcs =
> + __igt_spin_batch_new(i915,
> + .ctx = ctx,
> + .engine = engine,
> + .flags = IGT_SPIN_POLL_RUN);
> + igt_spin_busywait_until_running(task[i].xcs);
> +
> + /* Common rcs tasks will be queued in FIFO */
> + task[i].rcs =
> + __igt_spin_batch_new(i915,
> + .ctx = ctx,
> + .engine = I915_EXEC_RENDER,
> + .dependency = task[i].xcs->handle);
> +
> + gem_context_destroy(i915, ctx);
> +
> + if (++i == ARRAY_SIZE(task))
> + break;
> + }
> + igt_require(i == ARRAY_SIZE(task));
> +
> + /* Since task[0] was queued first, it will be first in queue for rcs */
> + igt_spin_batch_end(task[1].xcs);
> + igt_spin_batch_end(task[1].rcs);
> + gem_sync(i915, task[1].rcs->handle); /* to hang if task[0] hogs rcs */
> +
> + for (i = 0; i < ARRAY_SIZE(task); i++) {
> + igt_spin_batch_free(i915, task[i].xcs);
> + igt_spin_batch_free(i915, task[i].rcs);
> + }
> +}
> +
> static void reorder(int fd, unsigned ring, unsigned flags)
> #define EQUAL 1
> {
> @@ -1393,6 +1452,8 @@ igt_main
>
> igt_subtest("semaphore-user")
> semaphore_userlock(fd);
> + igt_subtest("semaphore-codependency")
> + semaphore_codependency(fd);
>
> igt_subtest("smoketest-all")
> smoketest(fd, ALL_ENGINES, 30);
>
Just need can_store_dword for code correctness before IGT_SPIN_POLL_RUN
and with that:
Reviewed-by: Tvrtko Ursulin <tvrtko.ursulin at intel.com>
Regards,
Tvrtko
More information about the Intel-gfx
mailing list