[Intel-gfx] [PATCH] drm/i915/gt: Detect if we miss WaIdleLiteRestore
Chris Wilson
chris at chris-wilson.co.uk
Tue Dec 10 10:06:00 UTC 2019
Quoting Tvrtko Ursulin (2019-12-10 10:00:34)
>
> On 09/12/2019 02:32, Chris Wilson wrote:
> > In order to avoid confusing the HW, we must never submit an empty ring
> > during lite-restore, that is we should always advance the RING_TAIL
> > before submitting to stay ahead of the RING_HEAD.
> >
> > Normally this is prevented by keeping a couple of spare NOPs in the
> > request->wa_tail so that on resubmission we can advance the tail. This
> > relies on the request only being resubmitted once, which is the normal
> > condition as it is seen once for ELSP[1] and then later in ELSP[0]. On
> > preemption, the requests are unwound and the tail reset back to the
> > normal end point (as we know the request is incomplete and therefore its
> > RING_HEAD is even earlier).
> >
> > However, if this w/a should fail we would try and resubmit the request
> > with the RING_TAIL already set to the location of this request's wa_tail
> > potentially causing a GPU hang. We can spot when we do try and
> > incorrectly resubmit without advancing the RING_TAIL and spare any
> > embarrassment by forcing the context restore.
> >
> > In the case of preempt-to-busy, we leave the requests running on the HW
> > while we unwind. As the ring is still live, we cannot rewind our
> > rq->tail without forcing a reload so leave it set to rq->wa_tail and
> > only force a reload if we resubmit after a lite-restore. (Normally, the
> > forced reload will be a part of the preemption event.)
> >
> > Fixes: 22b7a426bbe1 ("drm/i915/execlists: Preempt-to-busy")
> > Closes: https://gitlab.freedesktop.org/drm/intel/issues/673
> > Signed-off-by: Chris Wilson <chris at chris-wilson.co.uk>
> > Cc: Mika Kuoppala <mika.kuoppala at linux.intel.com>
> > Cc: Tvrtko Ursulin <tvrtko.ursulin at intel.com>
> > ---
> > drivers/gpu/drm/i915/gt/intel_lrc.c | 46 +++++++++++++----------------
> > 1 file changed, 21 insertions(+), 25 deletions(-)
> >
> > diff --git a/drivers/gpu/drm/i915/gt/intel_lrc.c b/drivers/gpu/drm/i915/gt/intel_lrc.c
> > index c7ea8a055005..929f6bae4eba 100644
> > --- a/drivers/gpu/drm/i915/gt/intel_lrc.c
> > +++ b/drivers/gpu/drm/i915/gt/intel_lrc.c
> > @@ -845,12 +845,6 @@ static const u8 *reg_offsets(const struct intel_engine_cs *engine)
> > }
> > }
> >
> > -static void unwind_wa_tail(struct i915_request *rq)
> > -{
> > - rq->tail = intel_ring_wrap(rq->ring, rq->wa_tail - WA_TAIL_BYTES);
> > - assert_ring_tail_valid(rq->ring, rq->tail);
> > -}
> > -
> > static struct i915_request *
> > __unwind_incomplete_requests(struct intel_engine_cs *engine)
> > {
> > @@ -863,12 +857,10 @@ __unwind_incomplete_requests(struct intel_engine_cs *engine)
> > list_for_each_entry_safe_reverse(rq, rn,
> > &engine->active.requests,
> > sched.link) {
> > -
> > if (i915_request_completed(rq))
> > continue; /* XXX */
> >
> > __i915_request_unsubmit(rq);
> > - unwind_wa_tail(rq);
> >
> > /*
> > * Push the request back into the queue for later resubmission.
> > @@ -1216,13 +1208,29 @@ execlists_schedule_out(struct i915_request *rq)
> > i915_request_put(rq);
> > }
> >
> > -static u64 execlists_update_context(const struct i915_request *rq)
> > +static u64 execlists_update_context(struct i915_request *rq)
> > {
> > struct intel_context *ce = rq->hw_context;
> > - u64 desc;
> > + u64 desc = ce->lrc_desc;
> > + u32 tail;
> >
> > - ce->lrc_reg_state[CTX_RING_TAIL] =
> > - intel_ring_set_tail(rq->ring, rq->tail);
> > + /*
> > + * WaIdleLiteRestore:bdw,skl
> > + *
> > + * We should never submit the context with the same RING_TAIL twice
> > + * just in case we submit an empty ring, which confuses the HW.
> > + *
> > + * We append a couple of NOOPs (gen8_emit_wa_tail) after the end of
> > + * the normal request to be able to always advance the RING_TAIL on
> > + * subsequent resubmissions (for lite restore). Should that fail us,
> > + * and we try and submit the same tail again, force the context
> > + * reload.
> > + */
> > + tail = intel_ring_set_tail(rq->ring, rq->tail);
> > + if (unlikely(ce->lrc_reg_state[CTX_RING_TAIL] == tail))
> > + desc |= CTX_DESC_FORCE_RESTORE;
> > + ce->lrc_reg_state[CTX_RING_TAIL] = tail;
> > + rq->tail = rq->wa_tail;
> >
> > /*
> > * Make sure the context image is complete before we submit it to HW.
> > @@ -1236,13 +1244,11 @@ static u64 execlists_update_context(const struct i915_request *rq)
> > */
> > wmb();
> >
> > - desc = ce->lrc_desc;
> > - ce->lrc_desc &= ~CTX_DESC_FORCE_RESTORE;
> > -
> > /* Wa_1607138340:tgl */
> > if (IS_TGL_REVID(rq->i915, TGL_REVID_A0, TGL_REVID_A0))
> > desc |= CTX_DESC_FORCE_RESTORE;
> >
> > + ce->lrc_desc &= ~CTX_DESC_FORCE_RESTORE;
> > return desc;
> > }
> >
> > @@ -1776,16 +1782,6 @@ static void execlists_dequeue(struct intel_engine_cs *engine)
> >
> > return;
> > }
> > -
> > - /*
> > - * WaIdleLiteRestore:bdw,skl
> > - * Apply the wa NOOPs to prevent
> > - * ring:HEAD == rq:TAIL as we resubmit the
> > - * request. See gen8_emit_fini_breadcrumb() for
> > - * where we prepare the padding after the
> > - * end of the request.
> > - */
> > - last->tail = last->wa_tail;
> > }
> > }
> >
> >
>
> Reviewed-by: Tvrtko Ursulin <tvrtko.ursulin at intel.com>
>
> Is the shard failure in gem_ctx_persistence the unrelated fence status race?
It's a timing failure; the test is not quite coordinating its timer with
the reset-timer very well and decides the reset was not fast enough. Not
really sure why, and there's a very small element of doubt that I have a
genuine race in the kill_contexts().
-Chris
More information about the Intel-gfx
mailing list