[Intel-gfx] [PATCH 3/3] drm/i915: Move find_active_request() to the engine
Tvrtko Ursulin
tvrtko.ursulin at linux.intel.com
Tue Mar 5 18:13:34 UTC 2019
On 05/03/2019 18:03, Chris Wilson wrote:
> To find the active request, we need only search along the individual
> engine for the right request. This does not require touching any global
> GEM state, so move it into the engine compartment.
>
> Signed-off-by: Chris Wilson <chris at chris-wilson.co.uk>
> ---
> drivers/gpu/drm/i915/i915_drv.h | 3 --
> drivers/gpu/drm/i915/i915_gem.c | 45 -----------------------
> drivers/gpu/drm/i915/i915_gpu_error.c | 2 +-
> drivers/gpu/drm/i915/intel_engine_cs.c | 47 ++++++++++++++++++++++++-
> drivers/gpu/drm/i915/intel_ringbuffer.h | 3 ++
> 5 files changed, 50 insertions(+), 50 deletions(-)
>
> diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h
> index 08ead854ac2d..ff039750069d 100644
> --- a/drivers/gpu/drm/i915/i915_drv.h
> +++ b/drivers/gpu/drm/i915/i915_drv.h
> @@ -2996,9 +2996,6 @@ void i915_gem_track_fb(struct drm_i915_gem_object *old,
>
> int __must_check i915_gem_set_global_seqno(struct drm_device *dev, u32 seqno);
>
> -struct i915_request *
> -i915_gem_find_active_request(struct intel_engine_cs *engine);
> -
> static inline bool __i915_wedged(struct i915_gpu_error *error)
> {
> return unlikely(test_bit(I915_WEDGED, &error->flags));
> diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c
> index 69413f99ed04..c67369bd145b 100644
> --- a/drivers/gpu/drm/i915/i915_gem.c
> +++ b/drivers/gpu/drm/i915/i915_gem.c
> @@ -2803,51 +2803,6 @@ i915_gem_object_pwrite_gtt(struct drm_i915_gem_object *obj,
> return 0;
> }
>
> -static bool match_ring(struct i915_request *rq)
> -{
> - struct drm_i915_private *dev_priv = rq->i915;
> - u32 ring = I915_READ(RING_START(rq->engine->mmio_base));
> -
> - return ring == i915_ggtt_offset(rq->ring->vma);
> -}
> -
> -struct i915_request *
> -i915_gem_find_active_request(struct intel_engine_cs *engine)
> -{
> - struct i915_request *request, *active = NULL;
> - unsigned long flags;
> -
> - /*
> - * We are called by the error capture, reset and to dump engine
> - * state at random points in time. In particular, note that neither is
> - * crucially ordered with an interrupt. After a hang, the GPU is dead
> - * and we assume that no more writes can happen (we waited long enough
> - * for all writes that were in transaction to be flushed) - adding an
> - * extra delay for a recent interrupt is pointless. Hence, we do
> - * not need an engine->irq_seqno_barrier() before the seqno reads.
> - * At all other times, we must assume the GPU is still running, but
> - * we only care about the snapshot of this moment.
> - */
> - spin_lock_irqsave(&engine->timeline.lock, flags);
> - list_for_each_entry(request, &engine->timeline.requests, link) {
> - if (i915_request_completed(request))
> - continue;
> -
> - if (!i915_request_started(request))
> - break;
> -
> - /* More than one preemptible request may match! */
> - if (!match_ring(request))
> - break;
> -
> - active = request;
> - break;
> - }
> - spin_unlock_irqrestore(&engine->timeline.lock, flags);
> -
> - return active;
> -}
> -
> static void
> i915_gem_retire_work_handler(struct work_struct *work)
> {
> diff --git a/drivers/gpu/drm/i915/i915_gpu_error.c b/drivers/gpu/drm/i915/i915_gpu_error.c
> index 5f1cdbc9eb5d..3d8020888604 100644
> --- a/drivers/gpu/drm/i915/i915_gpu_error.c
> +++ b/drivers/gpu/drm/i915/i915_gpu_error.c
> @@ -1411,7 +1411,7 @@ static void gem_record_rings(struct i915_gpu_state *error)
> error_record_engine_registers(error, engine, ee);
> error_record_engine_execlists(engine, ee);
>
> - request = i915_gem_find_active_request(engine);
> + request = intel_engine_find_active_request(engine);
> if (request) {
> struct i915_gem_context *ctx = request->gem_context;
> struct intel_ring *ring;
> diff --git a/drivers/gpu/drm/i915/intel_engine_cs.c b/drivers/gpu/drm/i915/intel_engine_cs.c
> index 62a2bbbbcc64..555a4590fa23 100644
> --- a/drivers/gpu/drm/i915/intel_engine_cs.c
> +++ b/drivers/gpu/drm/i915/intel_engine_cs.c
> @@ -1545,7 +1545,7 @@ void intel_engine_dump(struct intel_engine_cs *engine,
> if (&rq->link != &engine->timeline.requests)
> print_request(m, rq, "\t\tlast ");
>
> - rq = i915_gem_find_active_request(engine);
> + rq = intel_engine_find_active_request(engine);
> if (rq) {
> print_request(m, rq, "\t\tactive ");
>
> @@ -1712,6 +1712,51 @@ void intel_disable_engine_stats(struct intel_engine_cs *engine)
> write_sequnlock_irqrestore(&engine->stats.lock, flags);
> }
>
> +static bool match_ring(struct i915_request *rq)
> +{
> + struct drm_i915_private *dev_priv = rq->i915;
> + u32 ring = I915_READ(RING_START(rq->engine->mmio_base));
> +
> + return ring == i915_ggtt_offset(rq->ring->vma);
> +}
> +
> +struct i915_request *
> +intel_engine_find_active_request(struct intel_engine_cs *engine)
> +{
> + struct i915_request *request, *active = NULL;
> + unsigned long flags;
> +
> + /*
> + * We are called by the error capture, reset and to dump engine
> + * state at random points in time. In particular, note that neither is
> + * crucially ordered with an interrupt. After a hang, the GPU is dead
> + * and we assume that no more writes can happen (we waited long enough
> + * for all writes that were in transaction to be flushed) - adding an
> + * extra delay for a recent interrupt is pointless. Hence, we do
> + * not need an engine->irq_seqno_barrier() before the seqno reads.
> + * At all other times, we must assume the GPU is still running, but
> + * we only care about the snapshot of this moment.
> + */
> + spin_lock_irqsave(&engine->timeline.lock, flags);
> + list_for_each_entry(request, &engine->timeline.requests, link) {
> + if (i915_request_completed(request))
> + continue;
> +
> + if (!i915_request_started(request))
> + break;
> +
> + /* More than one preemptible request may match! */
> + if (!match_ring(request))
> + break;
> +
> + active = request;
> + break;
> + }
> + spin_unlock_irqrestore(&engine->timeline.lock, flags);
> +
> + return active;
> +}
> +
> #if IS_ENABLED(CONFIG_DRM_I915_SELFTEST)
> #include "selftests/mock_engine.c"
> #include "selftests/intel_engine_cs.c"
> diff --git a/drivers/gpu/drm/i915/intel_ringbuffer.h b/drivers/gpu/drm/i915/intel_ringbuffer.h
> index 18e865ff6637..84b7047e2df5 100644
> --- a/drivers/gpu/drm/i915/intel_ringbuffer.h
> +++ b/drivers/gpu/drm/i915/intel_ringbuffer.h
> @@ -1014,6 +1014,9 @@ void intel_disable_engine_stats(struct intel_engine_cs *engine);
>
> ktime_t intel_engine_get_busy_time(struct intel_engine_cs *engine);
>
> +struct i915_request *
> +intel_engine_find_active_request(struct intel_engine_cs *engine);
> +
> #if IS_ENABLED(CONFIG_DRM_I915_SELFTEST)
>
> static inline bool inject_preempt_hang(struct intel_engine_execlists *execlists)
>
intel_engine.h in 3...2...1.. ;)
Reviewed-by: Tvrtko Ursulin <tvrtko.ursulin at intel.com>
Regards,
Tvrtko
More information about the Intel-gfx
mailing list