[Intel-gfx] [PATCH v3 12/13] drm/i915/execlists: Preemption!
Mika Kuoppala
mika.kuoppala at linux.intel.com
Fri Sep 29 06:55:45 UTC 2017
Chris Wilson <chris at chris-wilson.co.uk> writes:
> When we write to ELSP, it triggers a context preemption at the earliest
> arbitration point (3DPRIMITIVE, some PIPECONTROLs, a few other
> operations and the explicit MI_ARB_CHECK). If this is to the same
> context, it triggers a LITE_RESTORE where the RING_TAIL is merely
> updated (used currently to chain requests from the same context
> together, avoiding bubbles). However, if it is to a different context, a
> full context-switch is performed and it will start to execute the new
> context saving the image of the old for later execution.
>
> Previously we avoided preemption by only submitting a new context when
> the old was idle. But now we wish embrace it, and if the new request has
> a higher priority than the currently executing request, we write to the
> ELSP regardless, thus triggering preemption, but we tell the GPU to
> switch to our special preemption context (not the target). In the
> context-switch interrupt handler, we know that the previous contexts
> have finished execution and so can unwind all the incomplete requests
> and compute the new highest priority request to execute.
>
> It would be feasible to avoid the switch-to-idle intermediate by
> programming the ELSP with the target context. The difficulty is in
> tracking which request that should be whilst maintaining the dependency
> change, the error comes in with coalesced requests. As we only track the
> most recent request and its priority, we may run into the issue of being
> tricked in preempting a high priority request that was followed by a
> low priority request from the same context (e.g. for PI); worse still
> that earlier request may be our own dependency and the order then broken
> by preemption. By injecting the switch-to-idle and then recomputing the
> priority queue, we avoid the issue with tracking in-flight coalesced
> requests. Having tried the preempt-to-busy approach, and failed to find
> a way around the coalesced priority issue, Michal's original proposal to
> inject an idle context (based on handling GuC preemption) succeeds.
>
> The current heuristic for deciding when to preempt are only if the new
> request is of higher priority, and has the privileged priority of
> greater than 0. Note that the scheduler remains unfair!
>
> v2: Disable for gen8 (bdw/bsw) as we need additional w/a for GPGPU.
> Since, the feature is now conditional and not always available when we
> have a scheduler, make it known via the HAS_SCHEDULER GETPARAM (now a
> capability mask).
> v3: Stylistic tweaks.
>
> Suggested-by: Michal Winiarski <michal.winiarski at intel.com>
> Signed-off-by: Chris Wilson <chris at chris-wilson.co.uk>
> Cc: Michal Winiarski <michal.winiarski at intel.com>
> Cc: Tvrtko Ursulin <tvrtko.ursulin at intel.com>
> Cc: Arkadiusz Hiler <arkadiusz.hiler at intel.com>
> Cc: Mika Kuoppala <mika.kuoppala at intel.com>
> Cc: Ben Widawsky <benjamin.widawsky at intel.com>
> Cc: Zhenyu Wang <zhenyuw at linux.intel.com>
> Cc: Zhi Wang <zhi.a.wang at intel.com>
> ---
> drivers/gpu/drm/i915/i915_drv.c | 9 ++-
> drivers/gpu/drm/i915/i915_irq.c | 6 +-
> drivers/gpu/drm/i915/i915_pci.c | 1 +
> drivers/gpu/drm/i915/intel_lrc.c | 137 ++++++++++++++++++++++++--------
> drivers/gpu/drm/i915/intel_ringbuffer.h | 2 +
> 5 files changed, 119 insertions(+), 36 deletions(-)
>
> diff --git a/drivers/gpu/drm/i915/i915_drv.c b/drivers/gpu/drm/i915/i915_drv.c
> index c5070f859c66..a3bb352a581b 100644
> --- a/drivers/gpu/drm/i915/i915_drv.c
> +++ b/drivers/gpu/drm/i915/i915_drv.c
> @@ -368,9 +368,16 @@ static int i915_getparam(struct drm_device *dev, void *data,
> break;
> case I915_PARAM_HAS_SCHEDULER:
> value = 0;
> - if (dev_priv->engine[RCS] && dev_priv->engine[RCS]->schedule)
> + if (dev_priv->engine[RCS] && dev_priv->engine[RCS]->schedule) {
> value |= I915_SCHEDULER_CAP_ENABLED;
> +
> + if (INTEL_INFO(dev_priv)->has_logical_ring_preemption &&
> + i915_modparams.enable_execlists &&
> + !i915_modparams.enable_guc_submission)
> + value |= I915_SCHEDULER_CAP_PREEMPTION;
> + }
> break;
> +
> case I915_PARAM_MMAP_VERSION:
> /* Remember to bump this if the version changes! */
> case I915_PARAM_HAS_GEM:
> diff --git a/drivers/gpu/drm/i915/i915_irq.c b/drivers/gpu/drm/i915/i915_irq.c
> index efd7827ff181..8620fbd22c55 100644
> --- a/drivers/gpu/drm/i915/i915_irq.c
> +++ b/drivers/gpu/drm/i915/i915_irq.c
> @@ -1382,10 +1382,8 @@ gen8_cs_irq_handler(struct intel_engine_cs *engine, u32 iir, int test_shift)
> bool tasklet = false;
>
> if (iir & (GT_CONTEXT_SWITCH_INTERRUPT << test_shift)) {
> - if (port_count(&execlists->port[0])) {
> - __set_bit(ENGINE_IRQ_EXECLIST, &engine->irq_posted);
> - tasklet = true;
> - }
> + __set_bit(ENGINE_IRQ_EXECLIST, &engine->irq_posted);
> + tasklet = true;
> }
>
> if (iir & (GT_RENDER_USER_INTERRUPT << test_shift)) {
> diff --git a/drivers/gpu/drm/i915/i915_pci.c b/drivers/gpu/drm/i915/i915_pci.c
> index 01d4b569b2cc..d11c1f5f5a97 100644
> --- a/drivers/gpu/drm/i915/i915_pci.c
> +++ b/drivers/gpu/drm/i915/i915_pci.c
> @@ -470,6 +470,7 @@ static const struct intel_device_info intel_skylake_gt4_info __initconst = {
> .has_rc6 = 1, \
> .has_dp_mst = 1, \
> .has_logical_ring_contexts = 1, \
> + .has_logical_ring_preemption = 1, \
> .has_guc = 1, \
> .has_aliasing_ppgtt = 1, \
> .has_full_ppgtt = 1, \
> diff --git a/drivers/gpu/drm/i915/intel_lrc.c b/drivers/gpu/drm/i915/intel_lrc.c
> index 02ea1e4e098b..bc3fc4cd039e 100644
> --- a/drivers/gpu/drm/i915/intel_lrc.c
> +++ b/drivers/gpu/drm/i915/intel_lrc.c
> @@ -208,9 +208,9 @@
>
> /* Typical size of the average request (2 pipecontrols and a MI_BB) */
> #define EXECLISTS_REQUEST_SIZE 64 /* bytes */
> -
> #define WA_TAIL_DWORDS 2
> #define WA_TAIL_BYTES (sizeof(u32) * WA_TAIL_DWORDS)
> +#define PREEMPT_ID 0x1
>
> static int execlists_context_deferred_alloc(struct i915_gem_context *ctx,
> struct intel_engine_cs *engine);
> @@ -430,6 +430,12 @@ static u64 execlists_update_context(struct drm_i915_gem_request *rq)
> return ce->lrc_desc;
> }
>
> +static inline void elsp_write(u64 desc, u32 __iomem *elsp)
> +{
> + writel(upper_32_bits(desc), elsp);
> + writel(lower_32_bits(desc), elsp);
> +}
> +
> static void execlists_submit_ports(struct intel_engine_cs *engine)
> {
> struct execlist_port *port = engine->execlists.port;
> @@ -455,8 +461,7 @@ static void execlists_submit_ports(struct intel_engine_cs *engine)
> desc = 0;
> }
>
> - writel(upper_32_bits(desc), elsp);
> - writel(lower_32_bits(desc), elsp);
> + elsp_write(desc, elsp);
> }
> }
>
> @@ -489,26 +494,43 @@ static void port_assign(struct execlist_port *port,
> port_set(port, port_pack(i915_gem_request_get(rq), port_count(port)));
> }
>
> +static void inject_preempt_context(struct intel_engine_cs *engine)
> +{
> + struct intel_context *ce =
> + &engine->i915->preempt_context->engine[engine->id];
> + u32 __iomem *elsp =
> + engine->i915->regs + i915_mmio_reg_offset(RING_ELSP(engine));
> + unsigned int n;
> +
> + GEM_BUG_ON(engine->i915->preempt_context->hw_id != PREEMPT_ID);
> + GEM_BUG_ON(!IS_ALIGNED(ce->ring->size, WA_TAIL_BYTES));
> +
> + memset(ce->ring->vaddr + ce->ring->tail, 0, WA_TAIL_BYTES);
> + ce->ring->tail += WA_TAIL_BYTES;
> + ce->ring->tail &= (ce->ring->size - 1);
> + ce->lrc_reg_state[CTX_RING_TAIL+1] = ce->ring->tail;
> +
> + for (n = execlists_num_ports(&engine->execlists); --n; )
> + elsp_write(0, elsp);
> +
> + elsp_write(ce->lrc_desc, elsp);
> +}
> +
> +static bool can_preempt(struct intel_engine_cs *engine)
> +{
> + return INTEL_INFO(engine->i915)->has_logical_ring_preemption;
> +}
> +
> static void execlists_dequeue(struct intel_engine_cs *engine)
> {
> - struct drm_i915_gem_request *last;
> struct intel_engine_execlists * const execlists = &engine->execlists;
> struct execlist_port *port = execlists->port;
> const struct execlist_port * const last_port =
> &execlists->port[execlists->port_mask];
> + struct drm_i915_gem_request *last = port_request(port);
> struct rb_node *rb;
> bool submit = false;
>
> - last = port_request(port);
> - if (last)
> - /* WaIdleLiteRestore:bdw,skl
> - * Apply the wa NOOPs to prevent ring:HEAD == req:TAIL
> - * as we resubmit the request. See gen8_emit_breadcrumb()
> - * for where we prepare the padding after the end of the
> - * request.
> - */
> - last->tail = last->wa_tail;
> -
> /* Hardware submission is through 2 ports. Conceptually each port
> * has a (RING_START, RING_HEAD, RING_TAIL) tuple. RING_START is
> * static for a context, and unique to each, so we only execute
> @@ -533,7 +555,45 @@ static void execlists_dequeue(struct intel_engine_cs *engine)
> spin_lock_irq(&engine->timeline->lock);
> rb = execlists->first;
> GEM_BUG_ON(rb_first(&execlists->queue) != rb);
> - while (rb) {
> + if (!rb)
> + goto unlock;
> +
> + if (last) {
> + /*
> + * Don't resubmit or switch until all outstanding
> + * preemptions (lite-restore) are seen. Then we
> + * know the next preemption status we see corresponds
> + * to this ELSP update.
> + */
> + if (port_count(&port[0]) > 1)
> + goto unlock;
> +
> + if (can_preempt(engine) &&
> + rb_entry(rb, struct i915_priolist, node)->priority >
> + max(last->priotree.priority, 0)) {
> + /*
> + * Switch to our empty preempt context so
> + * the state of the GPU is known (idle).
> + */
> + inject_preempt_context(engine);
> + execlists->preempt = true;
> + goto unlock;
> + } else {
> + if (port_count(&port[1]))
> + goto unlock;
I am assuming that this is check for hw
availability and nothing else?
- Mika (being paranoid)
> +
> + /* WaIdleLiteRestore:bdw,skl
> + * Apply the wa NOOPs to prevent
> + * ring:HEAD == req:TAIL as we resubmit the
> + * request. See gen8_emit_breadcrumb() for
> + * where we prepare the padding after the
> + * end of the request.
> + */
> + last->tail = last->wa_tail;
> + }
> + }
> +
> + do {
> struct i915_priolist *p = rb_entry(rb, typeof(*p), node);
> struct drm_i915_gem_request *rq, *rn;
>
> @@ -596,11 +656,12 @@ static void execlists_dequeue(struct intel_engine_cs *engine)
> INIT_LIST_HEAD(&p->requests);
> if (p->priority != I915_PRIORITY_NORMAL)
> kmem_cache_free(engine->i915->priorities, p);
> - }
> + } while (rb);
> done:
> execlists->first = rb;
> if (submit)
> port_assign(port, last);
> +unlock:
> spin_unlock_irq(&engine->timeline->lock);
>
> if (submit)
> @@ -681,13 +742,6 @@ static void execlists_cancel_requests(struct intel_engine_cs *engine)
> spin_unlock_irqrestore(&engine->timeline->lock, flags);
> }
>
> -static bool execlists_elsp_ready(const struct intel_engine_cs *engine)
> -{
> - const struct execlist_port *port = engine->execlists.port;
> -
> - return port_count(&port[0]) + port_count(&port[1]) < 2;
> -}
> -
> /*
> * Check the unread Context Status Buffers and manage the submission of new
> * contexts to the ELSP accordingly.
> @@ -696,7 +750,7 @@ static void intel_lrc_irq_handler(unsigned long data)
> {
> struct intel_engine_cs * const engine = (struct intel_engine_cs *)data;
> struct intel_engine_execlists * const execlists = &engine->execlists;
> - struct execlist_port *port = execlists->port;
> + struct execlist_port * const port = execlists->port;
> struct drm_i915_private *dev_priv = engine->i915;
>
> /* We can skip acquiring intel_runtime_pm_get() here as it was taken
> @@ -781,6 +835,23 @@ static void intel_lrc_irq_handler(unsigned long data)
> if (!(status & GEN8_CTX_STATUS_COMPLETED_MASK))
> continue;
>
> + if (status & GEN8_CTX_STATUS_ACTIVE_IDLE &&
> + buf[2*head + 1] == PREEMPT_ID) {
> + execlist_cancel_port_requests(execlists);
> +
> + spin_lock_irq(&engine->timeline->lock);
> + unwind_incomplete_requests(engine);
> + spin_unlock_irq(&engine->timeline->lock);
> +
> + GEM_BUG_ON(!execlists->preempt);
> + execlists->preempt = false;
> + continue;
> + }
> +
> + if (status & GEN8_CTX_STATUS_PREEMPTED &&
> + execlists->preempt)
> + continue;
> +
> /* Check the context/desc id for this event matches */
> GEM_DEBUG_BUG_ON(buf[2 * head + 1] != port->context_id);
>
> @@ -812,7 +883,7 @@ static void intel_lrc_irq_handler(unsigned long data)
> }
> }
>
> - if (execlists_elsp_ready(engine))
> + if (!execlists->preempt)
> execlists_dequeue(engine);
>
> intel_uncore_forcewake_put(dev_priv, execlists->fw_domains);
> @@ -825,7 +896,7 @@ static void insert_request(struct intel_engine_cs *engine,
> struct i915_priolist *p = lookup_priolist(engine, pt, prio);
>
> list_add_tail(&pt->link, &ptr_mask_bits(p, 1)->requests);
> - if (ptr_unmask_bits(p, 1) && execlists_elsp_ready(engine))
> + if (ptr_unmask_bits(p, 1))
> tasklet_hi_schedule(&engine->execlists.irq_tasklet);
> }
>
> @@ -955,8 +1026,6 @@ static void execlists_schedule(struct drm_i915_gem_request *request, int prio)
> }
>
> spin_unlock_irq(&engine->timeline->lock);
> -
> - /* XXX Do we need to preempt to make room for us and our deps? */
> }
>
> static struct intel_ring *
> @@ -1152,6 +1221,8 @@ static u32 *gen8_init_indirectctx_bb(struct intel_engine_cs *engine, u32 *batch)
> i915_ggtt_offset(engine->scratch) +
> 2 * CACHELINE_BYTES);
>
> + *batch++ = MI_ARB_ON_OFF | MI_ARB_ENABLE;
> +
> /* Pad to end of cacheline */
> while ((unsigned long)batch % CACHELINE_BYTES)
> *batch++ = MI_NOOP;
> @@ -1167,6 +1238,8 @@ static u32 *gen8_init_indirectctx_bb(struct intel_engine_cs *engine, u32 *batch)
>
> static u32 *gen9_init_indirectctx_bb(struct intel_engine_cs *engine, u32 *batch)
> {
> + *batch++ = MI_ARB_ON_OFF | MI_ARB_DISABLE;
> +
> /* WaFlushCoherentL3CacheLinesAtContextSwitch:skl,bxt,glk */
> batch = gen8_emit_flush_coherentl3_wa(engine, batch);
>
> @@ -1212,6 +1285,8 @@ static u32 *gen9_init_indirectctx_bb(struct intel_engine_cs *engine, u32 *batch)
> *batch++ = 0;
> }
>
> + *batch++ = MI_ARB_ON_OFF | MI_ARB_ENABLE;
> +
> /* Pad to end of cacheline */
> while ((unsigned long)batch % CACHELINE_BYTES)
> *batch++ = MI_NOOP;
> @@ -1365,6 +1440,7 @@ static int gen8_init_common_ring(struct intel_engine_cs *engine)
> GT_CONTEXT_SWITCH_INTERRUPT << engine->irq_shift);
> clear_bit(ENGINE_IRQ_EXECLIST, &engine->irq_posted);
> execlists->csb_head = -1;
> + execlists->preempt = false;
>
> /* After a GPU reset, we may have requests to replay */
> if (!i915_modparams.enable_guc_submission && execlists->first)
> @@ -1660,7 +1736,8 @@ static int gen8_emit_flush_render(struct drm_i915_gem_request *request,
> */
> static void gen8_emit_wa_tail(struct drm_i915_gem_request *request, u32 *cs)
> {
> - *cs++ = MI_NOOP;
> + /* Ensure there's always at least one preemption point per-request. */
> + *cs++ = MI_ARB_CHECK;
> *cs++ = MI_NOOP;
> request->wa_tail = intel_ring_offset(request, cs);
> }
> @@ -1681,7 +1758,6 @@ static void gen8_emit_breadcrumb(struct drm_i915_gem_request *request, u32 *cs)
>
> gen8_emit_wa_tail(request, cs);
> }
> -
> static const int gen8_emit_breadcrumb_sz = 6 + WA_TAIL_DWORDS;
>
> static void gen8_emit_breadcrumb_render(struct drm_i915_gem_request *request,
> @@ -1709,7 +1785,6 @@ static void gen8_emit_breadcrumb_render(struct drm_i915_gem_request *request,
>
> gen8_emit_wa_tail(request, cs);
> }
> -
> static const int gen8_emit_breadcrumb_render_sz = 8 + WA_TAIL_DWORDS;
>
> static int gen8_init_rcs_context(struct drm_i915_gem_request *req)
> diff --git a/drivers/gpu/drm/i915/intel_ringbuffer.h b/drivers/gpu/drm/i915/intel_ringbuffer.h
> index 56d7ae9f298b..891d9555dce6 100644
> --- a/drivers/gpu/drm/i915/intel_ringbuffer.h
> +++ b/drivers/gpu/drm/i915/intel_ringbuffer.h
> @@ -238,6 +238,8 @@ struct intel_engine_execlists {
> #define EXECLIST_MAX_PORTS 2
> } port[EXECLIST_MAX_PORTS];
>
> + bool preempt;
> +
> /**
> * @port_mask: number of execlist ports - 1
> */
> --
> 2.14.2
>
> _______________________________________________
> Intel-gfx mailing list
> Intel-gfx at lists.freedesktop.org
> https://lists.freedesktop.org/mailman/listinfo/intel-gfx
More information about the Intel-gfx
mailing list