[Intel-gfx] [PATCH 2/2] drm/i915/selftests: Include arbitration points in preemption smoketest
Chris Wilson
chris at chris-wilson.co.uk
Thu Sep 27 08:49:40 UTC 2018
Include a batch full of a page of arbitration points in order to provide
a window for inject_preempt_context() in the preemption smoketests.
Signed-off-by: Chris Wilson <chris at chris-wilson.co.uk>
---
drivers/gpu/drm/i915/selftests/intel_lrc.c | 104 +++++++++++++++++----
1 file changed, 85 insertions(+), 19 deletions(-)
diff --git a/drivers/gpu/drm/i915/selftests/intel_lrc.c b/drivers/gpu/drm/i915/selftests/intel_lrc.c
index 2ab142c9d9c2..9dc7346a312c 100644
--- a/drivers/gpu/drm/i915/selftests/intel_lrc.c
+++ b/drivers/gpu/drm/i915/selftests/intel_lrc.c
@@ -588,6 +588,7 @@ struct preempt_smoke {
struct drm_i915_private *i915;
struct i915_gem_context **contexts;
struct intel_engine_cs *engine;
+ struct drm_i915_gem_object *batch;
unsigned int ncontext;
struct rnd_state prng;
unsigned long count;
@@ -599,6 +600,47 @@ static struct i915_gem_context *smoke_context(struct preempt_smoke *smoke)
&smoke->prng)];
}
+static int smoke_submit(struct preempt_smoke *smoke,
+ struct i915_gem_context *ctx,
+ struct drm_i915_gem_object *batch)
+{
+ struct i915_request *rq;
+ struct i915_vma *vma = NULL;
+ int err = 0;
+
+ if (batch) {
+ vma = i915_vma_instance(batch, &ctx->ppgtt->vm, NULL);
+ if (IS_ERR(vma))
+ return PTR_ERR(vma);
+
+ err = i915_vma_pin(vma, 0, 0, PIN_USER);
+ if (err)
+ return err;
+ }
+
+ rq = i915_request_alloc(smoke->engine, ctx);
+ if (IS_ERR(rq)) {
+ err = PTR_ERR(rq);
+ goto unpin;
+ }
+
+ if (vma) {
+ err = rq->engine->emit_bb_start(rq,
+ vma->node.start,
+ PAGE_SIZE, 0);
+ if (!err)
+ err = i915_vma_move_to_active(vma, rq, 0);
+ }
+
+ i915_request_add(rq);
+
+unpin:
+ if (vma)
+ i915_vma_unpin(vma);
+
+ return err;
+}
+
static int smoke_crescendo_thread(void *arg)
{
struct preempt_smoke *smoke = arg;
@@ -608,19 +650,15 @@ static int smoke_crescendo_thread(void *arg)
count = 0;
do {
struct i915_gem_context *ctx = smoke_context(smoke);
- struct i915_request *rq;
+ int err;
ctx->sched.priority = count % I915_PRIORITY_MAX;
mutex_lock(&smoke->i915->drm.struct_mutex);
- rq = i915_request_alloc(smoke->engine, ctx);
- if (IS_ERR(rq)) {
- mutex_unlock(&smoke->i915->drm.struct_mutex);
- return PTR_ERR(rq);
- }
-
- i915_request_add(rq);
+ err = smoke_submit(smoke, ctx, smoke->batch);
mutex_unlock(&smoke->i915->drm.struct_mutex);
+ if (err)
+ return err;
count++;
} while (!__igt_timeout(end_time, NULL));
@@ -629,7 +667,8 @@ static int smoke_crescendo_thread(void *arg)
return 0;
}
-static int smoke_crescendo(struct preempt_smoke *smoke)
+static int smoke_crescendo(struct preempt_smoke *smoke, unsigned int flags)
+#define BATCH BIT(0)
{
struct task_struct *tsk[I915_NUM_ENGINES] = {};
struct preempt_smoke arg[I915_NUM_ENGINES];
@@ -643,6 +682,8 @@ static int smoke_crescendo(struct preempt_smoke *smoke)
for_each_engine(engine, smoke->i915, id) {
arg[id] = *smoke;
arg[id].engine = engine;
+ if (!(flags & BATCH))
+ arg[id].batch = NULL;
arg[id].count = 0;
tsk[id] = kthread_run(smoke_crescendo_thread, &arg,
@@ -669,31 +710,32 @@ static int smoke_crescendo(struct preempt_smoke *smoke)
mutex_lock(&smoke->i915->drm.struct_mutex);
- pr_info("Submitted %lu crescendo requests across %d engines and %d contexts\n",
- count, INTEL_INFO(smoke->i915)->num_rings, smoke->ncontext);
+ pr_info("Submitted %lu crescendo:%x requests across %d engines and %d contexts\n",
+ count, flags, INTEL_INFO(smoke->i915)->num_rings, smoke->ncontext);
return 0;
}
static int smoke_random(struct preempt_smoke *smoke)
{
- struct intel_engine_cs *engine;
enum intel_engine_id id;
IGT_TIMEOUT(end_time);
unsigned long count;
count = 0;
do {
- for_each_engine(engine, smoke->i915, id) {
+ for_each_engine(smoke->engine, smoke->i915, id) {
struct i915_gem_context *ctx = smoke_context(smoke);
- struct i915_request *rq;
+ struct drm_i915_gem_object *batch = smoke->batch;
+ int err;
ctx->sched.priority = random_priority(&smoke->prng);
+ if (prandom_u32_state(&smoke->prng) & 1)
+ batch = NULL;
- rq = i915_request_alloc(engine, ctx);
- if (IS_ERR(rq))
- return PTR_ERR(rq);
+ err = smoke_submit(smoke, ctx, batch);
+ if (err)
+ return err;
- i915_request_add(rq);
count++;
}
} while (!__igt_timeout(end_time, NULL));
@@ -711,6 +753,7 @@ static int live_preempt_smoke(void *arg)
.ncontext = 1024,
};
int err = -ENOMEM;
+ u32 *cs;
int n;
if (!HAS_LOGICAL_RING_PREEMPTION(smoke.i915))
@@ -725,13 +768,33 @@ static int live_preempt_smoke(void *arg)
mutex_lock(&smoke.i915->drm.struct_mutex);
intel_runtime_pm_get(smoke.i915);
+ smoke.batch = i915_gem_object_create_internal(smoke.i915, PAGE_SIZE);
+ if (IS_ERR(smoke.batch)) {
+ err = PTR_ERR(smoke.batch);
+ goto err_unlock;
+ }
+
+ cs = i915_gem_object_pin_map(smoke.batch, I915_MAP_WB);
+ if (IS_ERR(cs)) {
+ err = PTR_ERR(cs);
+ goto err_batch;
+ }
+ for (n = 0; n < PAGE_SIZE / sizeof(*cs) - 1; n++)
+ cs[n] = MI_ARB_CHECK;
+ cs[n] = MI_BATCH_BUFFER_END;
+ i915_gem_object_unpin_map(smoke.batch);
+
for (n = 0; n < smoke.ncontext; n++) {
smoke.contexts[n] = kernel_context(smoke.i915);
if (!smoke.contexts[n])
goto err_ctx;
}
- err = smoke_crescendo(&smoke);
+ err = smoke_crescendo(&smoke, 0);
+ if (err)
+ goto err_ctx;
+
+ err = smoke_crescendo(&smoke, BATCH);
if (err)
goto err_ctx;
@@ -749,6 +812,9 @@ static int live_preempt_smoke(void *arg)
kernel_context_close(smoke.contexts[n]);
}
+err_batch:
+ i915_gem_object_put(smoke.batch);
+err_unlock:
intel_runtime_pm_put(smoke.i915);
mutex_unlock(&smoke.i915->drm.struct_mutex);
kfree(smoke.contexts);
--
2.19.0
More information about the Intel-gfx
mailing list