[Intel-gfx] [PATCH 1/6] drm/i915: Individual request cancellation

Matthew Auld matthew.william.auld at gmail.com
Mon Mar 22 15:38:31 UTC 2021


On Thu, 18 Mar 2021 at 17:04, Tvrtko Ursulin
<tvrtko.ursulin at linux.intel.com> wrote:
>
> From: Chris Wilson <chris at chris-wilson.co.uk>
>
> Currently, we cancel outstanding requests within a context when the
> context is closed. We may also want to cancel individual requests using
> the same graceful preemption mechanism.
>
> v2 (Tvrtko):
>  * Cancel waiters carefully considering no timeline lock and RCU.
>  * Fixed selftests.
>
> v3 (Tvrtko):
>  * Remove error propagation to waiters for now.
>
> Signed-off-by: Chris Wilson <chris at chris-wilson.co.uk>
> Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin at intel.com>
> ---
>  .../gpu/drm/i915/gt/intel_engine_heartbeat.c  |   1 +
>  .../drm/i915/gt/intel_execlists_submission.c  |   9 +-
>  drivers/gpu/drm/i915/i915_request.c           |  52 ++++-
>  drivers/gpu/drm/i915/i915_request.h           |   4 +-
>  drivers/gpu/drm/i915/selftests/i915_request.c | 201 ++++++++++++++++++
>  5 files changed, 261 insertions(+), 6 deletions(-)
>
> diff --git a/drivers/gpu/drm/i915/gt/intel_engine_heartbeat.c b/drivers/gpu/drm/i915/gt/intel_engine_heartbeat.c
> index 0b062fad1837..e2fb3ae2aaf3 100644
> --- a/drivers/gpu/drm/i915/gt/intel_engine_heartbeat.c
> +++ b/drivers/gpu/drm/i915/gt/intel_engine_heartbeat.c
> @@ -314,6 +314,7 @@ int intel_engine_pulse(struct intel_engine_cs *engine)
>                 mutex_unlock(&ce->timeline->mutex);
>         }
>
> +       intel_engine_flush_scheduler(engine);
>         intel_engine_pm_put(engine);
>         return err;
>  }
> diff --git a/drivers/gpu/drm/i915/gt/intel_execlists_submission.c b/drivers/gpu/drm/i915/gt/intel_execlists_submission.c
> index 85ff5fe861b4..4c2acb5a6c0a 100644
> --- a/drivers/gpu/drm/i915/gt/intel_execlists_submission.c
> +++ b/drivers/gpu/drm/i915/gt/intel_execlists_submission.c
> @@ -421,6 +421,11 @@ static void reset_active(struct i915_request *rq,
>         ce->lrc.lrca = lrc_update_regs(ce, engine, head);
>  }
>
> +static bool bad_request(const struct i915_request *rq)
> +{
> +       return rq->fence.error && i915_request_started(rq);
> +}
> +
>  static struct intel_engine_cs *
>  __execlists_schedule_in(struct i915_request *rq)
>  {
> @@ -433,7 +438,7 @@ __execlists_schedule_in(struct i915_request *rq)
>                      !intel_engine_has_heartbeat(engine)))
>                 intel_context_set_banned(ce);
>
> -       if (unlikely(intel_context_is_banned(ce)))
> +       if (unlikely(intel_context_is_banned(ce) || bad_request(rq)))
>                 reset_active(rq, engine);
>
>         if (IS_ENABLED(CONFIG_DRM_I915_DEBUG_GEM))
> @@ -1112,7 +1117,7 @@ static unsigned long active_preempt_timeout(struct intel_engine_cs *engine,
>                 return 0;
>
>         /* Force a fast reset for terminated contexts (ignoring sysfs!) */
> -       if (unlikely(intel_context_is_banned(rq->context)))
> +       if (unlikely(intel_context_is_banned(rq->context) || bad_request(rq)))
>                 return 1;
>
>         return READ_ONCE(engine->props.preempt_timeout_ms);
> diff --git a/drivers/gpu/drm/i915/i915_request.c b/drivers/gpu/drm/i915/i915_request.c
> index e7b4c4bc41a6..b4511ac05e9a 100644
> --- a/drivers/gpu/drm/i915/i915_request.c
> +++ b/drivers/gpu/drm/i915/i915_request.c
> @@ -33,7 +33,10 @@
>  #include "gem/i915_gem_context.h"
>  #include "gt/intel_breadcrumbs.h"
>  #include "gt/intel_context.h"
> +#include "gt/intel_engine.h"
> +#include "gt/intel_engine_heartbeat.h"
>  #include "gt/intel_gpu_commands.h"
> +#include "gt/intel_reset.h"
>  #include "gt/intel_ring.h"
>  #include "gt/intel_rps.h"
>
> @@ -429,20 +432,22 @@ void __i915_request_skip(struct i915_request *rq)
>         rq->infix = rq->postfix;
>  }
>
> -void i915_request_set_error_once(struct i915_request *rq, int error)
> +bool i915_request_set_error_once(struct i915_request *rq, int error)
>  {
>         int old;
>
>         GEM_BUG_ON(!IS_ERR_VALUE((long)error));
>
>         if (i915_request_signaled(rq))
> -               return;
> +               return false;
>
>         old = READ_ONCE(rq->fence.error);
>         do {
>                 if (fatal_error(old))
> -                       return;
> +                       return false;
>         } while (!try_cmpxchg(&rq->fence.error, &old, error));
> +
> +       return true;
>  }
>
>  struct i915_request *i915_request_mark_eio(struct i915_request *rq)
> @@ -609,6 +614,47 @@ void i915_request_unsubmit(struct i915_request *request)
>         spin_unlock_irqrestore(&se->lock, flags);
>  }
>
> +static struct intel_engine_cs *active_engine(struct i915_request *rq)
> +{
> +       struct intel_engine_cs *engine, *locked;
> +
> +       locked = READ_ONCE(rq->engine);
> +       spin_lock_irq(&locked->sched.lock);
> +       while (unlikely(locked != (engine = READ_ONCE(rq->engine)))) {
> +               spin_unlock(&locked->sched.lock);
> +               locked = engine;
> +               spin_lock(&locked->sched.lock);
> +       }
> +
> +       engine = NULL;
> +       if (i915_request_is_active(rq) && !__i915_request_is_complete(rq))
> +               engine = locked;
> +
> +       spin_unlock_irq(&locked->sched.lock);
> +
> +       return engine;

Bad idea to reuse __active_engine() somehow?

Reviewed-by: Matthew Auld <matthew.auld at intel.com>


More information about the dri-devel mailing list