[Intel-gfx] [RFC 09/11] HACK drm/i915/preempt: Actually send the preemption request
Michał Winiarski
michal.winiarski at intel.com
Thu Feb 23 19:14:19 UTC 2017
Now that we're able to post-process the preemption event, let's actually
send it to GuC. To identify that preemption has finished, we're using a
dummy request sent through kernel_context.
Signed-off-by: Michał Winiarski <michal.winiarski at intel.com>
---
drivers/gpu/drm/i915/i915_guc_submission.c | 90 ++++++++++++++++++++++++++++--
1 file changed, 86 insertions(+), 4 deletions(-)
diff --git a/drivers/gpu/drm/i915/i915_guc_submission.c b/drivers/gpu/drm/i915/i915_guc_submission.c
index 2a658b9..7ccc5b4 100644
--- a/drivers/gpu/drm/i915/i915_guc_submission.c
+++ b/drivers/gpu/drm/i915/i915_guc_submission.c
@@ -491,14 +491,96 @@ static int guc_ring_doorbell(struct i915_guc_client *client)
return ret;
}
-static int i915_guc_preempt_noop(struct intel_engine_cs *engine)
+static int add_preemption_workitem(struct intel_engine_cs *engine)
{
- engine->preempt_requested = false;
- intel_write_status_page(engine, I915_GEM_HWS_PREEMPT_INDEX, 0);
+ struct drm_i915_private *dev_priv = engine->i915;
+ struct intel_guc *guc = &engine->i915->guc;
+ struct i915_gem_context *ctx = engine->i915->kernel_context;
+ struct i915_guc_client *client = guc->client[PREEMPT];
+ struct drm_i915_gem_request rq;
+ u32 *cs;
+
+ rq.i915 = engine->i915;
+ rq.engine = engine;
+ rq.ctx = ctx;
+ rq.global_seqno = 0;
+ rq.file_priv = NULL;
+ rq.batch = NULL;
+ rq.reserved_space = 32;
+ rq.ring = ctx->engine[engine->id].ring;
+
+ if (engine->id == RCS) {
+ cs = intel_ring_begin(&rq, 8);
+ if (IS_ERR(cs))
+ return PTR_ERR(cs);
+
+ *cs++ = GFX_OP_PIPE_CONTROL(6);
+ *cs++ = PIPE_CONTROL_GLOBAL_GTT_IVB |
+ PIPE_CONTROL_CS_STALL |
+ PIPE_CONTROL_QW_WRITE;
+ *cs++ = engine->status_page.ggtt_offset +
+ I915_GEM_HWS_PREEMPT_ADDR;
+ *cs++ = 0;
+ *cs++ = 1;
+ *cs++ = 0;
+ *cs++ = MI_USER_INTERRUPT;
+ *cs++ = MI_NOOP;
+ intel_ring_advance(&rq, cs);
+ } else {
+ cs = intel_ring_begin(&rq, 6);
+ if (IS_ERR(cs))
+ return PTR_ERR(cs);
+
+ *cs++ = (MI_FLUSH_DW + 1) | MI_FLUSH_DW_OP_STOREDW;
+ *cs++ = (engine->status_page.ggtt_offset + I915_GEM_HWS_PREEMPT_ADDR) |
+ MI_FLUSH_DW_USE_GTT;
+ *cs++ = 0;
+ *cs++ = 1;
+ *cs++ = MI_USER_INTERRUPT;
+ *cs++ = MI_NOOP;
+ intel_ring_advance(&rq, cs);
+ }
+ rq.tail = intel_ring_offset(&rq, cs);
+
+ spin_lock_irq(&client->wq_lock);
+ client->wq_rsvd += sizeof(struct guc_wq_item);
+ guc_wq_item_append(client, &rq);
+ if (i915_vma_is_map_and_fenceable(rq.ring->vma))
+ POSTING_READ_FW(GUC_STATUS);
+ spin_unlock_irq(&client->wq_lock);
return 0;
}
+static int i915_guc_preempt(struct intel_engine_cs *engine)
+{
+ struct intel_guc *guc = &engine->i915->guc;
+ struct i915_guc_client *client = guc->client[PREEMPT];
+ struct guc_process_desc *desc;
+ u32 data[7];
+ int ret;
+
+ ret = add_preemption_workitem(engine);
+ if (ret)
+ return ret;
+
+ desc = client->vaddr + client->proc_desc_offset;
+ desc->tail = client->wq_tail;
+
+ data[0] = INTEL_GUC_ACTION_REQUEST_PREEMPTION;
+ data[1] = client->ctx_index;
+ data[2] = INTEL_GUC_PREEMPT_OPTION_IMMEDIATE |
+ INTEL_GUC_PREEMPT_OPTION_DROP_WORK_Q |
+ INTEL_GUC_PREEMPT_OPTION_DROP_SUBMIT_Q;
+ data[3] = engine->guc_id;
+ data[4] = guc->client[NORMAL]->priority;
+ data[5] = guc->client[NORMAL]->ctx_index;
+ data[6] = engine->status_page.ggtt_offset +
+ (LRC_GUCSHR_PN * PAGE_SIZE);
+
+ return intel_guc_send(guc, data, ARRAY_SIZE(data));
+}
+
/**
* __i915_guc_submit() - Submit commands through GuC
* @rq: request associated with the commands
@@ -1143,7 +1225,7 @@ int i915_guc_submission_enable(struct drm_i915_private *dev_priv)
engine->irq_tasklet.func = i915_guc_irq_handler;
if (i915.enable_preemption)
- engine->preempt = i915_guc_preempt_noop;
+ engine->preempt = i915_guc_preempt;
/* Replay the current set of previously submitted requests */
spin_lock_irqsave(&engine->timeline->lock, flags);
--
2.9.3
More information about the Intel-gfx
mailing list