[PATCH 74/75] drm/i915: Remove i915_request.ring/timeline
Chris Wilson
chris at chris-wilson.co.uk
Tue Feb 2 11:38:31 UTC 2021
We only use the ring/timeline during request construction, so let's
remove the direct access from i915_request and retrieve them from the
intel_context when needed.
Signed-off-by: Chris Wilson <chris at chris-wilson.co.uk>
---
.../gpu/drm/i915/gem/i915_gem_execbuffer.c | 4 +-
drivers/gpu/drm/i915/gt/gen2_engine_cs.c | 12 ++---
drivers/gpu/drm/i915/gt/gen6_engine_cs.c | 16 +++----
drivers/gpu/drm/i915/gt/gen8_engine_cs.c | 19 ++++----
drivers/gpu/drm/i915/gt/intel_context.c | 2 +-
drivers/gpu/drm/i915/gt/intel_engine_cs.c | 40 ++++-------------
.../drm/i915/gt/intel_execlists_submission.c | 16 +++----
drivers/gpu/drm/i915/gt/intel_ring.c | 11 ++---
drivers/gpu/drm/i915/gt/intel_ring.h | 10 +++--
.../gpu/drm/i915/gt/intel_ring_scheduler.c | 2 +-
.../gpu/drm/i915/gt/intel_ring_submission.c | 6 +--
drivers/gpu/drm/i915/gt/selftest_context.c | 2 +-
drivers/gpu/drm/i915/gt/selftest_engine_cs.c | 2 +-
drivers/gpu/drm/i915/gt/selftest_execlists.c | 4 +-
drivers/gpu/drm/i915/gt/selftest_timeline.c | 10 ++---
drivers/gpu/drm/i915/i915_active.c | 2 +-
drivers/gpu/drm/i915/i915_active.h | 4 +-
drivers/gpu/drm/i915/i915_gpu_error.c | 2 +-
drivers/gpu/drm/i915/i915_request.c | 44 +++++++++----------
drivers/gpu/drm/i915/i915_request.h | 29 ++++--------
drivers/gpu/drm/i915/i915_scheduler.c | 34 +++++++-------
drivers/gpu/drm/i915/selftests/i915_request.c | 7 +--
22 files changed, 122 insertions(+), 156 deletions(-)
diff --git a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
index 052ae889bfde..e20d2bfdf41f 100644
--- a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
+++ b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
@@ -2628,7 +2628,7 @@ static struct i915_request *eb_throttle(struct i915_execbuffer *eb, struct intel
* drains before we can submit our next request.
*/
list_for_each_entry(rq, &tl->requests, link) {
- if (rq->ring != ring)
+ if (i915_request_get_ring(rq) != ring)
continue;
if (__intel_ring_space(rq->postfix,
@@ -3097,7 +3097,7 @@ static void retire_requests(struct intel_timeline *tl, struct i915_request *end)
static int eb_request_add(struct i915_execbuffer *eb, int err)
{
struct i915_request *rq = eb->request;
- struct intel_timeline * const tl = i915_request_timeline(rq);
+ struct intel_timeline * const tl = i915_request_get_timeline(rq);
struct i915_sched_attr attr = {};
struct i915_request *prev;
diff --git a/drivers/gpu/drm/i915/gt/gen2_engine_cs.c b/drivers/gpu/drm/i915/gt/gen2_engine_cs.c
index 4da0a63c231d..5e499e47079d 100644
--- a/drivers/gpu/drm/i915/gt/gen2_engine_cs.c
+++ b/drivers/gpu/drm/i915/gt/gen2_engine_cs.c
@@ -149,8 +149,8 @@ static u32 *__gen2_emit_breadcrumb(const struct intel_engine_cs *engine,
struct i915_request *rq, u32 *cs,
int flush, int post)
{
- GEM_BUG_ON(i915_request_active_timeline(rq)->hwsp_ggtt != engine->status_page.vma);
- GEM_BUG_ON(offset_in_page(i915_request_active_timeline(rq)->hwsp_offset) != I915_GEM_HWS_SEQNO_ADDR);
+ GEM_BUG_ON(i915_request_get_timeline(rq)->hwsp_ggtt != engine->status_page.vma);
+ GEM_BUG_ON(offset_in_page(i915_request_get_timeline(rq)->hwsp_offset) != I915_GEM_HWS_SEQNO_ADDR);
*cs++ = MI_FLUSH;
@@ -169,7 +169,7 @@ static u32 *__gen2_emit_breadcrumb(const struct intel_engine_cs *engine,
*cs++ = MI_USER_INTERRUPT;
rq->tail = intel_ring_offset(rq, cs);
- assert_ring_tail_valid(rq->ring, rq->tail);
+ assert_ring_tail_valid(i915_request_get_ring(rq), rq->tail);
return cs;
}
@@ -184,7 +184,7 @@ u32 *gen3_emit_breadcrumb(const struct intel_engine_cs *engine,
static u32 *__gen4_emit_breadcrumb(struct i915_request *rq, u32 *cs,
int flush, int post)
{
- struct intel_timeline *tl = rcu_dereference_protected(rq->timeline, 1);
+ struct intel_timeline *tl = i915_request_get_timeline(rq);
u32 offset = __i915_request_hwsp_offset(rq);
GEM_BUG_ON(tl->mode == INTEL_TIMELINE_RELATIVE_CONTEXT);
@@ -217,7 +217,7 @@ static u32 *__gen4_emit_breadcrumb(struct i915_request *rq, u32 *cs,
*cs++ = MI_USER_INTERRUPT;
rq->tail = intel_ring_offset(rq, cs);
- assert_ring_tail_valid(rq->ring, rq->tail);
+ assert_ring_tail_valid(i915_request_get_ring(rq), rq->tail);
return cs;
}
@@ -232,7 +232,7 @@ u32 *gen4_emit_breadcrumb_xcs(const struct intel_engine_cs *engine,
int gen4_emit_init_breadcrumb_xcs(const struct intel_engine_cs *engine,
struct i915_request *rq)
{
- struct intel_timeline *tl = i915_request_timeline(rq);
+ struct intel_timeline *tl = i915_request_get_timeline(rq);
u32 *cs;
GEM_BUG_ON(i915_request_has_initial_breadcrumb(rq));
diff --git a/drivers/gpu/drm/i915/gt/gen6_engine_cs.c b/drivers/gpu/drm/i915/gt/gen6_engine_cs.c
index fe06554729f9..465978ee3a42 100644
--- a/drivers/gpu/drm/i915/gt/gen6_engine_cs.c
+++ b/drivers/gpu/drm/i915/gt/gen6_engine_cs.c
@@ -146,7 +146,7 @@ u32 *gen6_emit_breadcrumb_rcs(const struct intel_engine_cs *engine,
struct i915_request *rq,
u32 *cs)
{
- struct intel_timeline *tl = rcu_dereference_protected(rq->timeline, 1);
+ struct intel_timeline *tl = i915_request_get_timeline(rq);
u32 offset = __i915_request_hwsp_offset(rq);
unsigned int flags;
@@ -187,7 +187,7 @@ u32 *gen6_emit_breadcrumb_rcs(const struct intel_engine_cs *engine,
*cs++ = MI_NOOP;
rq->tail = intel_ring_offset(rq, cs);
- assert_ring_tail_valid(rq->ring, rq->tail);
+ assert_ring_tail_valid(i915_request_get_ring(rq), rq->tail);
return cs;
}
@@ -379,7 +379,7 @@ u32 *gen7_emit_breadcrumb_rcs(const struct intel_engine_cs *engine,
struct i915_request *rq,
u32 *cs)
{
- struct intel_timeline *tl = rcu_dereference_protected(rq->timeline, 1);
+ struct intel_timeline *tl = i915_request_get_timeline(rq);
u32 offset = __i915_request_hwsp_offset(rq);
unsigned int flags;
@@ -407,7 +407,7 @@ u32 *gen7_emit_breadcrumb_rcs(const struct intel_engine_cs *engine,
*cs++ = MI_NOOP;
rq->tail = intel_ring_offset(rq, cs);
- assert_ring_tail_valid(rq->ring, rq->tail);
+ assert_ring_tail_valid(i915_request_get_ring(rq), rq->tail);
return cs;
}
@@ -416,7 +416,7 @@ u32 *gen6_emit_breadcrumb_xcs(const struct intel_engine_cs *engine,
struct i915_request *rq,
u32 *cs)
{
- struct intel_timeline *tl = rcu_dereference_protected(rq->timeline, 1);
+ struct intel_timeline *tl = i915_request_get_timeline(rq);
u32 offset = __i915_request_hwsp_offset(rq);
unsigned int flags = 0;
@@ -436,7 +436,7 @@ u32 *gen6_emit_breadcrumb_xcs(const struct intel_engine_cs *engine,
*cs++ = MI_USER_INTERRUPT;
rq->tail = intel_ring_offset(rq, cs);
- assert_ring_tail_valid(rq->ring, rq->tail);
+ assert_ring_tail_valid(i915_request_get_ring(rq), rq->tail);
return cs;
}
@@ -446,7 +446,7 @@ u32 *gen7_emit_breadcrumb_xcs(const struct intel_engine_cs *engien,
struct i915_request *rq,
u32 *cs)
{
- struct intel_timeline *tl = rcu_dereference_protected(rq->timeline, 1);
+ struct intel_timeline *tl = i915_request_get_timeline(rq);
u32 offset = __i915_request_hwsp_offset(rq);
u32 cmd = MI_FLUSH_DW | MI_FLUSH_DW_OP_STOREDW;
int i;
@@ -473,7 +473,7 @@ u32 *gen7_emit_breadcrumb_xcs(const struct intel_engine_cs *engien,
*cs++ = MI_USER_INTERRUPT;
rq->tail = intel_ring_offset(rq, cs);
- assert_ring_tail_valid(rq->ring, rq->tail);
+ assert_ring_tail_valid(i915_request_get_ring(rq), rq->tail);
return cs;
}
diff --git a/drivers/gpu/drm/i915/gt/gen8_engine_cs.c b/drivers/gpu/drm/i915/gt/gen8_engine_cs.c
index bdc6e5e40b14..2086f10b607b 100644
--- a/drivers/gpu/drm/i915/gt/gen8_engine_cs.c
+++ b/drivers/gpu/drm/i915/gt/gen8_engine_cs.c
@@ -351,7 +351,7 @@ int gen8_emit_init_breadcrumb(const struct intel_engine_cs *engine,
u32 *cs;
GEM_BUG_ON(i915_request_has_initial_breadcrumb(rq));
- if (!intel_timeline_has_initial_breadcrumb(i915_request_timeline(rq)))
+ if (!intel_timeline_has_initial_breadcrumb(i915_request_get_timeline(rq)))
return 0;
cs = intel_ring_begin(rq, 6);
@@ -461,10 +461,9 @@ int gen8_emit_bb_start(const struct intel_engine_cs *engine,
static void assert_request_valid(struct i915_request *rq)
{
- struct intel_ring *ring __maybe_unused = rq->ring;
-
/* Can we unwind this request without appearing to go forwards? */
- GEM_BUG_ON(intel_ring_direction(ring, rq->wa_tail, rq->head) <= 0);
+ GEM_BUG_ON(intel_ring_direction(i915_request_get_ring(rq),
+ rq->wa_tail, rq->head) <= 0);
}
/*
@@ -514,14 +513,14 @@ gen8_emit_fini_breadcrumb_tail(const struct intel_engine_cs *engine,
cs = emit_preempt_busywait(engine, rq, cs);
rq->tail = intel_ring_offset(rq, cs);
- assert_ring_tail_valid(rq->ring, rq->tail);
+ assert_ring_tail_valid(i915_request_get_ring(rq), rq->tail);
return gen8_emit_wa_tail(rq, cs);
}
static u32 *emit_xcs_breadcrumb(struct i915_request *rq, u32 *cs)
{
- struct intel_timeline *tl = rcu_dereference_protected(rq->timeline, 1);
+ struct intel_timeline *tl = i915_request_get_timeline(rq);
unsigned int flags = MI_FLUSH_DW_OP_STOREDW;
u32 offset = __i915_request_hwsp_offset(rq);
@@ -548,7 +547,7 @@ u32 *gen8_emit_fini_breadcrumb_rcs(const struct intel_engine_cs *engine,
struct i915_request *rq,
u32 *cs)
{
- struct intel_timeline *tl = rcu_dereference_protected(rq->timeline, 1);
+ struct intel_timeline *tl = i915_request_get_timeline(rq);
unsigned int flags = PIPE_CONTROL_FLUSH_ENABLE | PIPE_CONTROL_CS_STALL;
u32 offset = __i915_request_hwsp_offset(rq);
@@ -576,7 +575,7 @@ u32 *gen11_emit_fini_breadcrumb_rcs(const struct intel_engine_cs *engine,
struct i915_request *rq,
u32 *cs)
{
- struct intel_timeline *tl = rcu_dereference_protected(rq->timeline, 1);
+ struct intel_timeline *tl = i915_request_get_timeline(rq);
u32 offset = __i915_request_hwsp_offset(rq);
unsigned int flags;
@@ -648,7 +647,7 @@ gen12_emit_fini_breadcrumb_tail(const struct intel_engine_cs *engine,
cs = gen12_emit_preempt_busywait(engine, rq, cs);
rq->tail = intel_ring_offset(rq, cs);
- assert_ring_tail_valid(rq->ring, rq->tail);
+ assert_ring_tail_valid(i915_request_get_ring(rq), rq->tail);
return gen8_emit_wa_tail(rq, cs);
}
@@ -666,7 +665,7 @@ u32 *gen12_emit_fini_breadcrumb_rcs(const struct intel_engine_cs *engine,
struct i915_request *rq,
u32 *cs)
{
- struct intel_timeline *tl = rcu_dereference_protected(rq->timeline, 1);
+ struct intel_timeline *tl = i915_request_get_timeline(rq);
u32 offset = __i915_request_hwsp_offset(rq);
unsigned int flags;
diff --git a/drivers/gpu/drm/i915/gt/intel_context.c b/drivers/gpu/drm/i915/gt/intel_context.c
index c7ab4ed92da4..014b59f18def 100644
--- a/drivers/gpu/drm/i915/gt/intel_context.c
+++ b/drivers/gpu/drm/i915/gt/intel_context.c
@@ -447,7 +447,7 @@ int intel_context_prepare_remote_request(struct intel_context *ce,
/* Only suitable for use in remotely modifying this context */
GEM_BUG_ON(rq->context == ce);
- if (rcu_access_pointer(rq->timeline) != tl) { /* timeline sharing! */
+ if (i915_request_get_timeline(rq) != tl) { /* timeline sharing! */
/* Queue this switch after current activity by this context. */
err = i915_active_fence_set(&tl->last_request, rq);
if (err)
diff --git a/drivers/gpu/drm/i915/gt/intel_engine_cs.c b/drivers/gpu/drm/i915/gt/intel_engine_cs.c
index 87b48e712f0e..d3352ebe977f 100644
--- a/drivers/gpu/drm/i915/gt/intel_engine_cs.c
+++ b/drivers/gpu/drm/i915/gt/intel_engine_cs.c
@@ -696,49 +696,27 @@ static int engine_setup_common(struct intel_engine_cs *engine)
return err;
}
-struct measure_breadcrumb {
- struct i915_request rq;
- struct intel_ring ring;
- u32 cs[2048];
-};
-
static int measure_breadcrumb_dw(struct intel_context *ce)
{
struct intel_engine_cs *engine = ce->engine;
- struct i915_sched *se = intel_engine_get_scheduler(engine);
- struct measure_breadcrumb *frame;
+ struct i915_request *rq;
int dw;
GEM_BUG_ON(!engine->gt->scratch);
+ GEM_BUG_ON(ce->ring->emit != 0);
- frame = kzalloc(sizeof(*frame), GFP_KERNEL);
- if (!frame)
+ rq = kzalloc(sizeof(*rq), GFP_KERNEL);
+ if (!rq)
return -ENOMEM;
- frame->rq.context = ce;
- rcu_assign_pointer(frame->rq.timeline, ce->timeline);
- frame->rq.sched.engine = se;
-
- frame->ring.vaddr = frame->cs;
- frame->ring.size = sizeof(frame->cs);
- frame->ring.wrap =
- BITS_PER_TYPE(frame->ring.size) - ilog2(frame->ring.size);
- frame->ring.effective_size = frame->ring.size;
- intel_ring_update_space(&frame->ring);
- frame->rq.ring = &frame->ring;
-
- mutex_lock(&ce->timeline->mutex);
- spin_lock_irq(&se->lock);
-
- dw = engine->emit_fini_breadcrumb(engine, &frame->rq, frame->cs) -
- frame->cs;
-
- spin_unlock_irq(&se->lock);
- mutex_unlock(&ce->timeline->mutex);
+ rq->context = ce;
+ rq->sched.engine = intel_engine_get_scheduler(engine);
+ dw = engine->emit_fini_breadcrumb(engine, rq, ce->ring->vaddr) -
+ (u32 *)ce->ring->vaddr;
GEM_BUG_ON(dw & 1); /* RING_TAIL must be qword aligned */
- kfree(frame);
+ kfree(rq);
return dw;
}
diff --git a/drivers/gpu/drm/i915/gt/intel_execlists_submission.c b/drivers/gpu/drm/i915/gt/intel_execlists_submission.c
index c77c7ae1dc5d..f85e175691d3 100644
--- a/drivers/gpu/drm/i915/gt/intel_execlists_submission.c
+++ b/drivers/gpu/drm/i915/gt/intel_execlists_submission.c
@@ -668,7 +668,7 @@ resubmit_virtual_request(struct intel_execlists *el,
if (__i915_request_is_complete(rq))
goto unlock;
- tl = i915_request_active_timeline(rq);
+ tl = i915_request_get_timeline(rq);
/* Rewind back to the start of this virtual engine queue */
list_for_each_entry_continue_reverse(rq, &tl->requests, link) {
@@ -822,10 +822,10 @@ static u64 execlists_update_context(struct i915_request *rq)
* HW has a tendency to ignore us rewinding the TAIL to the end of
* an earlier request.
*/
- GEM_BUG_ON(ce->lrc_reg_state[CTX_RING_TAIL] != rq->ring->tail);
- prev = rq->ring->tail;
- tail = intel_ring_set_tail(rq->ring, rq->tail);
- if (unlikely(intel_ring_direction(rq->ring, tail, prev) <= 0))
+ GEM_BUG_ON(ce->lrc_reg_state[CTX_RING_TAIL] != ce->ring->tail);
+ prev = ce->ring->tail;
+ tail = intel_ring_set_tail(ce->ring, rq->tail);
+ if (unlikely(intel_ring_direction(ce->ring, tail, prev) <= 0))
desc |= CTX_DESC_FORCE_RESTORE;
ce->lrc_reg_state[CTX_RING_TAIL] = tail;
rq->tail = rq->wa_tail;
@@ -2007,7 +2007,7 @@ process_csb(struct intel_execlists *el, struct i915_request **inactive)
ENGINE_READ(engine, RING_MI_MODE));
ENGINE_TRACE(engine,
"rq:{start:%08x, head:%04x, tail:%04x, seqno:%llx:%d, hwsp:%d}, ",
- i915_ggtt_offset(rq->ring->vma),
+ i915_ggtt_offset(i915_request_get_ring(rq)->vma),
rq->head, rq->tail,
rq->fence.context,
lower_32_bits(rq->fence.seqno),
@@ -3835,11 +3835,11 @@ static int print_ring(char *buf, int sz, struct i915_request *rq)
rcu_read_lock();
if (!i915_request_signaled(rq)) {
- struct intel_timeline *tl = rcu_dereference(rq->timeline);
+ struct intel_timeline *tl = i915_request_get_timeline(rq);
len = scnprintf(buf, sz,
"ring:{start:%08x, hwsp:%08x, seqno:%08x, runtime:%llums}, ",
- intel_ring_address(rq->ring),
+ intel_ring_address(i915_request_get_ring(rq)),
tl ? tl->ggtt_offset : 0,
hwsp_seqno(rq),
DIV_ROUND_CLOSEST_ULL(intel_context_get_total_runtime_ns(rq->context),
diff --git a/drivers/gpu/drm/i915/gt/intel_ring.c b/drivers/gpu/drm/i915/gt/intel_ring.c
index 521972c297a9..9310cbe23ea7 100644
--- a/drivers/gpu/drm/i915/gt/intel_ring.c
+++ b/drivers/gpu/drm/i915/gt/intel_ring.c
@@ -214,7 +214,7 @@ wait_for_space(struct intel_ring *ring,
GEM_BUG_ON(list_empty(&tl->requests));
list_for_each_entry(target, &tl->requests, link) {
- if (target->ring != ring)
+ if (i915_request_get_ring(target) != ring)
continue;
/* Would completion of this request free enough space? */
@@ -241,7 +241,7 @@ wait_for_space(struct intel_ring *ring,
u32 *intel_ring_begin(struct i915_request *rq, unsigned int num_dwords)
{
- struct intel_ring *ring = rq->ring;
+ struct intel_ring *ring = i915_request_get_ring(rq);
const unsigned int remain_usable = ring->effective_size - ring->emit;
const unsigned int bytes = num_dwords * sizeof(u32);
unsigned int need_wrap = 0;
@@ -291,7 +291,7 @@ u32 *intel_ring_begin(struct i915_request *rq, unsigned int num_dwords)
GEM_BUG_ON(!rq->reserved_space);
ret = wait_for_space(ring,
- i915_request_timeline(rq),
+ i915_request_get_timeline(rq),
total_bytes);
if (unlikely(ret))
return ERR_PTR(ret);
@@ -322,10 +322,11 @@ u32 *intel_ring_begin(struct i915_request *rq, unsigned int num_dwords)
/* Align the ring tail to a cacheline boundary */
int intel_ring_cacheline_align(struct i915_request *rq)
{
+ struct intel_ring *ring = i915_request_get_ring(rq);
int num_dwords;
void *cs;
- num_dwords = (rq->ring->emit & (CACHELINE_BYTES - 1)) / sizeof(u32);
+ num_dwords = (ring->emit & (CACHELINE_BYTES - 1)) / sizeof(u32);
if (num_dwords == 0)
return 0;
@@ -339,7 +340,7 @@ int intel_ring_cacheline_align(struct i915_request *rq)
memset64(cs, (u64)MI_NOOP << 32 | MI_NOOP, num_dwords / 2);
intel_ring_advance(rq, cs + num_dwords);
- GEM_BUG_ON(rq->ring->emit & (CACHELINE_BYTES - 1));
+ GEM_BUG_ON(ring->emit & (CACHELINE_BYTES - 1));
return 0;
}
diff --git a/drivers/gpu/drm/i915/gt/intel_ring.h b/drivers/gpu/drm/i915/gt/intel_ring.h
index 89d79c22fe9e..aac5ac3a8bc1 100644
--- a/drivers/gpu/drm/i915/gt/intel_ring.h
+++ b/drivers/gpu/drm/i915/gt/intel_ring.h
@@ -50,7 +50,8 @@ static inline void intel_ring_advance(struct i915_request *rq, u32 *cs)
* reserved for the command packet (i.e. the value passed to
* intel_ring_begin()).
*/
- GEM_BUG_ON((rq->ring->vaddr + rq->ring->emit) != cs);
+ GEM_BUG_ON((i915_request_get_ring(rq)->vaddr +
+ i915_request_get_ring(rq)->emit) != cs);
}
static inline u32 intel_ring_wrap(const struct intel_ring *ring, u32 pos)
@@ -82,10 +83,11 @@ intel_ring_offset_valid(const struct intel_ring *ring,
static inline u32 intel_ring_offset(const struct i915_request *rq, void *addr)
{
/* Don't write ring->size (equivalent to 0) as that hangs some GPUs. */
- u32 offset = addr - rq->ring->vaddr;
+ struct intel_ring *ring = i915_request_get_ring(rq);
+ u32 offset = addr - ring->vaddr;
- GEM_BUG_ON(offset > rq->ring->size);
- return intel_ring_wrap(rq->ring, offset);
+ GEM_BUG_ON(offset > ring->size);
+ return intel_ring_wrap(ring, offset);
}
static inline void
diff --git a/drivers/gpu/drm/i915/gt/intel_ring_scheduler.c b/drivers/gpu/drm/i915/gt/intel_ring_scheduler.c
index 9af5190928bd..26ecb44e2852 100644
--- a/drivers/gpu/drm/i915/gt/intel_ring_scheduler.c
+++ b/drivers/gpu/drm/i915/gt/intel_ring_scheduler.c
@@ -498,7 +498,7 @@ ring_submit(struct intel_engine_cs *engine, struct i915_request *rq)
set_current_context(&engine->legacy.context, rq->context);
}
- ring_copy(ring, rq->ring, rq->head, rq->tail);
+ ring_copy(ring, i915_request_get_ring(rq), rq->head, rq->tail);
return rq;
}
diff --git a/drivers/gpu/drm/i915/gt/intel_ring_submission.c b/drivers/gpu/drm/i915/gt/intel_ring_submission.c
index 9c7810e1db80..5ff2b8d315af 100644
--- a/drivers/gpu/drm/i915/gt/intel_ring_submission.c
+++ b/drivers/gpu/drm/i915/gt/intel_ring_submission.c
@@ -379,7 +379,7 @@ static void reset_rewind(struct intel_engine_cs *engine, bool stalled)
*/
__i915_request_reset(rq, stalled);
- GEM_BUG_ON(rq->ring != engine->legacy.ring);
+ GEM_BUG_ON(i915_request_get_ring(rq) != engine->legacy.ring);
head = rq->head;
} else {
head = engine->legacy.ring->tail;
@@ -891,7 +891,7 @@ ring_context_request(struct intel_context *ce, struct i915_request *rq)
int ret;
GEM_BUG_ON(!intel_context_is_pinned(ce));
- GEM_BUG_ON(intel_timeline_has_initial_breadcrumb(i915_request_timeline(rq)));
+ GEM_BUG_ON(intel_timeline_has_initial_breadcrumb(i915_request_get_timeline(rq)));
/*
* Flush enough space to reduce the likelihood of waiting after
@@ -1151,7 +1151,7 @@ static void __write_tail(struct intel_engine_cs *engine,
struct i915_request *rq)
{
ENGINE_WRITE(engine, RING_TAIL,
- intel_ring_set_tail(rq->ring, rq->tail));
+ intel_ring_set_tail(i915_request_get_ring(rq), rq->tail));
}
static void wa_write_tail(struct intel_engine_cs *engine,
diff --git a/drivers/gpu/drm/i915/gt/selftest_context.c b/drivers/gpu/drm/i915/gt/selftest_context.c
index a02fd70644e2..8a0388309218 100644
--- a/drivers/gpu/drm/i915/gt/selftest_context.c
+++ b/drivers/gpu/drm/i915/gt/selftest_context.c
@@ -14,7 +14,7 @@
static int request_sync(struct i915_request *rq)
{
- struct intel_timeline *tl = i915_request_timeline(rq);
+ struct intel_timeline *tl = i915_request_get_timeline(rq);
long timeout;
int err = 0;
diff --git a/drivers/gpu/drm/i915/gt/selftest_engine_cs.c b/drivers/gpu/drm/i915/gt/selftest_engine_cs.c
index c4da47f91d72..a86aab1b3354 100644
--- a/drivers/gpu/drm/i915/gt/selftest_engine_cs.c
+++ b/drivers/gpu/drm/i915/gt/selftest_engine_cs.c
@@ -54,7 +54,7 @@ static int write_timestamp(struct i915_request *rq, int slot)
cmd++;
*cs++ = cmd;
*cs++ = i915_mmio_reg_offset(RING_TIMESTAMP(engine->mmio_base));
- *cs++ = i915_request_timeline(rq)->ggtt_offset + slot * sizeof(u32);
+ *cs++ = i915_request_get_timeline(rq)->ggtt_offset + slot * sizeof(u32);
*cs++ = 0;
intel_ring_advance(rq, cs);
diff --git a/drivers/gpu/drm/i915/gt/selftest_execlists.c b/drivers/gpu/drm/i915/gt/selftest_execlists.c
index e46a16647d21..4fcbf3bf3394 100644
--- a/drivers/gpu/drm/i915/gt/selftest_execlists.c
+++ b/drivers/gpu/drm/i915/gt/selftest_execlists.c
@@ -2578,8 +2578,8 @@ static int live_chain_preempt(void *arg)
ring_size = rq->wa_tail - rq->head;
if (ring_size < 0)
- ring_size += rq->ring->size;
- ring_size = rq->ring->size / ring_size;
+ ring_size += i915_request_get_ring(rq)->size;
+ ring_size = i915_request_get_ring(rq)->size / ring_size;
pr_debug("%s(%s): Using maximum of %d requests\n",
__func__, engine->name, ring_size);
diff --git a/drivers/gpu/drm/i915/gt/selftest_timeline.c b/drivers/gpu/drm/i915/gt/selftest_timeline.c
index 1089171f5b58..f71034d91d08 100644
--- a/drivers/gpu/drm/i915/gt/selftest_timeline.c
+++ b/drivers/gpu/drm/i915/gt/selftest_timeline.c
@@ -1104,8 +1104,8 @@ static int live_hwsp_read(void *arg)
i915_request_put(rq);
/* Single requests are limited to half a ring at most */
- if (8 * watcher[1].rq->ring->emit >
- 3 * watcher[1].rq->ring->size)
+ if (8 * i915_request_get_ring(watcher[1].rq)->emit >
+ 3 * i915_request_get_ring(watcher[1].rq)->size)
break;
} while (!__igt_timeout(end_time, NULL));
@@ -1172,7 +1172,7 @@ static int live_hwsp_rollover_kernel(void *arg)
engine->name,
lower_32_bits(this->fence.seqno));
- GEM_BUG_ON(rcu_access_pointer(this->timeline) != tl);
+ GEM_BUG_ON(i915_request_get_timeline(this) != tl);
rq[i] = i915_request_get(this);
i915_request_add(this);
@@ -1257,7 +1257,7 @@ static int live_hwsp_rollover_user(void *arg)
engine->name,
lower_32_bits(this->fence.seqno));
- GEM_BUG_ON(rcu_access_pointer(this->timeline) != tl);
+ GEM_BUG_ON(i915_request_get_timeline(this) != tl);
rq[i] = i915_request_get(this);
i915_request_add(this);
@@ -1453,7 +1453,7 @@ static int live_hwsp_relative(void *arg)
if (IS_ERR(rq))
return PTR_ERR(rq);
- GEM_BUG_ON(rcu_access_pointer(rq->timeline) != tl);
+ GEM_BUG_ON(i915_request_get_timeline(rq) != tl);
i915_request_get(rq);
i915_request_add(rq);
diff --git a/drivers/gpu/drm/i915/i915_active.c b/drivers/gpu/drm/i915/i915_active.c
index bf467a9ca0fc..57cb0025a3cc 100644
--- a/drivers/gpu/drm/i915/i915_active.c
+++ b/drivers/gpu/drm/i915/i915_active.c
@@ -1020,7 +1020,7 @@ void i915_request_add_active_barriers(struct i915_request *rq)
GEM_BUG_ON(!intel_context_is_barrier(rq->context));
GEM_BUG_ON(intel_engine_is_virtual(engine));
- GEM_BUG_ON(i915_request_timeline(rq) != engine->kernel_context->timeline);
+ GEM_BUG_ON(i915_request_get_timeline(rq) != engine->kernel_context->timeline);
node = llist_del_all(&engine->barrier_tasks);
if (!node)
diff --git a/drivers/gpu/drm/i915/i915_active.h b/drivers/gpu/drm/i915/i915_active.h
index fb165d3f01cf..2ab503e1ebcc 100644
--- a/drivers/gpu/drm/i915/i915_active.h
+++ b/drivers/gpu/drm/i915/i915_active.h
@@ -170,9 +170,7 @@ int i915_active_ref(struct i915_active *ref, u64 idx, struct dma_fence *fence);
static inline int
i915_active_add_request(struct i915_active *ref, struct i915_request *rq)
{
- return i915_active_ref(ref,
- i915_request_timeline(rq)->fence_context,
- &rq->fence);
+ return i915_active_ref(ref, rq->fence.context, &rq->fence);
}
struct dma_fence *
diff --git a/drivers/gpu/drm/i915/i915_gpu_error.c b/drivers/gpu/drm/i915/i915_gpu_error.c
index ed20fb2181c2..7430f68d2ced 100644
--- a/drivers/gpu/drm/i915/i915_gpu_error.c
+++ b/drivers/gpu/drm/i915/i915_gpu_error.c
@@ -1320,7 +1320,7 @@ intel_engine_coredump_add_request(struct intel_engine_coredump *ee,
*/
vma = capture_vma(vma, rq->batch, "batch", gfp);
vma = capture_user(vma, rq, gfp);
- vma = capture_vma(vma, rq->ring->vma, "ring", gfp);
+ vma = capture_vma(vma, i915_request_get_ring(rq)->vma, "ring", gfp);
vma = capture_vma(vma, rq->context->state, "HW context", gfp);
ee->rq_head = rq->head;
diff --git a/drivers/gpu/drm/i915/i915_request.c b/drivers/gpu/drm/i915/i915_request.c
index e478bd68c073..b48107788b0d 100644
--- a/drivers/gpu/drm/i915/i915_request.c
+++ b/drivers/gpu/drm/i915/i915_request.c
@@ -225,12 +225,12 @@ static void free_capture_list(struct i915_request *request)
static void __i915_request_fill(struct i915_request *rq, u8 val)
{
- void *vaddr = rq->ring->vaddr;
+ void *vaddr = i915_request_get_ring(rq)->vaddr;
u32 head;
head = rq->infix;
if (rq->postfix < head) {
- memset(vaddr + head, val, rq->ring->size - head);
+ memset(vaddr + head, val, i915_request_get_ring(rq)->size - head);
head = 0;
}
memset(vaddr + head, val, rq->postfix - head);
@@ -287,11 +287,11 @@ bool i915_request_retire(struct i915_request *rq)
* completion order.
*/
GEM_BUG_ON(!list_is_first(&rq->link,
- &i915_request_timeline(rq)->requests));
+ &i915_request_get_timeline(rq)->requests));
if (IS_ENABLED(CONFIG_DRM_I915_DEBUG_GEM))
/* Poison before we release our space in the ring */
__i915_request_fill(rq, POISON_FREE);
- rq->ring->head = rq->postfix;
+ i915_request_get_ring(rq)->head = rq->postfix;
if (!i915_request_signaled(rq)) {
spin_lock_irq(&rq->lock);
@@ -330,7 +330,7 @@ bool i915_request_retire(struct i915_request *rq)
void i915_request_retire_upto(struct i915_request *rq)
{
- struct intel_timeline * const tl = i915_request_timeline(rq);
+ struct intel_timeline * const tl = i915_request_get_timeline(rq);
struct i915_request *tmp;
RQ_TRACE(rq, "\n");
@@ -511,7 +511,8 @@ bool __i915_request_submit(struct i915_request *request,
request->context->saturated |= request->sched.semaphores;
engine->emit_fini_breadcrumb(engine, request,
- request->ring->vaddr + request->postfix);
+ i915_request_get_ring(request)->vaddr +
+ request->postfix);
trace_i915_request_execute(request);
engine->serial++;
@@ -673,6 +674,7 @@ struct i915_request *
__i915_request_create(struct intel_context *ce, gfp_t gfp)
{
struct intel_timeline *tl = ce->timeline;
+ struct intel_ring *ring = ce->ring;
struct i915_request *rq;
u32 seqno;
int ret;
@@ -722,7 +724,6 @@ __i915_request_create(struct intel_context *ce, gfp_t gfp)
}
rq->context = ce;
- rq->ring = ce->ring;
kref_init(&rq->fence.refcount);
rq->fence.flags = 0;
@@ -736,7 +737,6 @@ __i915_request_create(struct intel_context *ce, gfp_t gfp)
rq->fence.context = tl->fence_context;
rq->fence.seqno = seqno;
- RCU_INIT_POINTER(rq->timeline, tl);
RCU_INIT_POINTER(rq->hwsp_cacheline, tl->hwsp_cacheline);
rq->hwsp_seqno = tl->hwsp_seqno;
GEM_BUG_ON(__i915_request_is_complete(rq));
@@ -771,13 +771,13 @@ __i915_request_create(struct intel_context *ce, gfp_t gfp)
* GPU processing the request, we never over-estimate the
* position of the head.
*/
- rq->head = rq->ring->emit;
+ rq->head = ring->emit;
ret = ce->ops->init_request(ce, rq);
if (ret)
goto err_unwind;
- rq->infix = rq->ring->emit; /* end of header; start of user payload */
+ rq->infix = ring->emit; /* end of header; start of user payload */
intel_context_mark_active(ce);
list_add_tail_rcu(&rq->link, &tl->requests);
@@ -785,7 +785,7 @@ __i915_request_create(struct intel_context *ce, gfp_t gfp)
return rq;
err_unwind:
- ce->ring->emit = rq->head;
+ ring->emit = rq->head;
/* Make sure we didn't add ourselves to external state before freeing */
GEM_BUG_ON(!list_empty(&rq->sched.signalers_list));
@@ -835,7 +835,7 @@ i915_request_await_start(struct i915_request *rq, struct i915_request *signal)
struct dma_fence *fence;
int err;
- if (i915_request_timeline(rq) == rcu_access_pointer(signal->timeline))
+ if (i915_request_get_timeline(rq) == i915_request_get_timeline(signal))
return 0;
if (i915_request_started(signal))
@@ -860,7 +860,7 @@ i915_request_await_start(struct i915_request *rq, struct i915_request *signal)
break;
/* Is signal the earliest request on its timeline? */
- if (pos == &rcu_dereference(signal->timeline)->requests)
+ if (pos == &i915_request_get_timeline(signal)->requests)
break;
/*
@@ -886,7 +886,7 @@ i915_request_await_start(struct i915_request *rq, struct i915_request *signal)
return 0;
err = 0;
- if (!intel_timeline_sync_is_later(i915_request_timeline(rq), fence))
+ if (!intel_timeline_sync_is_later(i915_request_get_timeline(rq), fence))
err = i915_sw_fence_await_dma_fence(&rq->sched.submit,
fence, 0,
I915_FENCE_GFP);
@@ -1043,7 +1043,7 @@ __i915_request_await_execution(struct i915_request *to,
return err;
/* Squash repeated depenendices to the same timelines */
- if (intel_timeline_sync_has_start(i915_request_timeline(to),
+ if (intel_timeline_sync_has_start(i915_request_get_timeline(to),
&from->fence))
return 0;
@@ -1096,7 +1096,7 @@ __i915_request_await_execution(struct i915_request *to,
return err;
}
- return intel_timeline_sync_set_start(i915_request_timeline(to),
+ return intel_timeline_sync_set_start(i915_request_get_timeline(to),
&from->fence);
}
@@ -1223,8 +1223,8 @@ i915_request_await_request(struct i915_request *to, struct i915_request *from)
int ret;
GEM_BUG_ON(to == from);
- GEM_BUG_ON(to->timeline == from->timeline);
GEM_BUG_ON(to->context == from->context);
+ GEM_BUG_ON(to->context->timeline == from->context->timeline);
if (i915_request_completed(from)) {
i915_sw_fence_set_error_once(&to->sched.submit,
@@ -1291,7 +1291,7 @@ i915_request_await_dma_fence(struct i915_request *rq, struct dma_fence *fence)
/* Squash repeated waits to the same timelines */
if (fence->context &&
- intel_timeline_sync_is_later(i915_request_timeline(rq),
+ intel_timeline_sync_is_later(i915_request_get_timeline(rq),
fence))
continue;
@@ -1304,7 +1304,7 @@ i915_request_await_dma_fence(struct i915_request *rq, struct dma_fence *fence)
/* Record the latest fence used against each timeline */
if (fence->context)
- intel_timeline_sync_set(i915_request_timeline(rq),
+ intel_timeline_sync_set(i915_request_get_timeline(rq),
fence);
} while (--nchild);
@@ -1385,7 +1385,7 @@ static bool in_order_submission(const struct i915_request *prev,
static struct i915_request *
__i915_request_add_to_timeline(struct i915_request *rq)
{
- struct intel_timeline *timeline = i915_request_timeline(rq);
+ struct intel_timeline *timeline = i915_request_get_timeline(rq);
struct i915_request *prev;
/*
@@ -1454,7 +1454,7 @@ __i915_request_add_to_timeline(struct i915_request *rq)
struct i915_request *__i915_request_commit(struct i915_request *rq)
{
struct intel_engine_cs *engine = i915_request_get_engine(rq);
- struct intel_ring *ring = rq->ring;
+ struct intel_ring *ring = i915_request_get_ring(rq);
u32 *cs;
RQ_TRACE(rq, "\n");
@@ -1511,7 +1511,7 @@ void __i915_request_queue(struct i915_request *rq,
void i915_request_add(struct i915_request *rq)
{
- struct intel_timeline * const tl = i915_request_timeline(rq);
+ struct intel_timeline * const tl = i915_request_get_timeline(rq);
struct i915_sched_attr attr = {};
struct i915_gem_context *ctx;
diff --git a/drivers/gpu/drm/i915/i915_request.h b/drivers/gpu/drm/i915/i915_request.h
index dfd74b32f10d..babe7ff8a181 100644
--- a/drivers/gpu/drm/i915/i915_request.h
+++ b/drivers/gpu/drm/i915/i915_request.h
@@ -177,9 +177,6 @@ struct i915_request {
* context.
*/
struct intel_context *context;
- struct intel_ring *ring;
- struct intel_timeline __rcu *timeline;
-
struct list_head signal_link;
struct llist_node signal_node;
@@ -603,11 +600,15 @@ i915_request_get_engine(const struct i915_request *rq)
}
static inline struct intel_timeline *
-i915_request_timeline(const struct i915_request *rq)
+i915_request_get_timeline(const struct i915_request *rq)
{
- /* Valid only while the request is being constructed (or retired). */
- return rcu_dereference_protected(rq->timeline,
- lockdep_is_held(&rcu_access_pointer(rq->timeline)->mutex));
+ return rq->context->timeline;
+}
+
+static inline struct intel_ring *
+i915_request_get_ring(const struct i915_request *rq)
+{
+ return rq->context->ring;
}
static inline struct i915_gem_context *
@@ -617,18 +618,6 @@ i915_request_gem_context(const struct i915_request *rq)
return rcu_dereference_protected(rq->context->gem_context, true);
}
-static inline struct intel_timeline *
-i915_request_active_timeline(const struct i915_request *rq)
-{
- /*
- * When in use during submission, we are protected by a guarantee that
- * the context/timeline is pinned and must remain pinned until after
- * this submission.
- */
- return rcu_dereference_protected(rq->timeline,
- lockdep_is_held(&i915_request_get_scheduler(rq)->lock));
-}
-
static inline u32 __i915_request_hwsp_offset(const struct i915_request *rq)
{
const struct intel_timeline_cacheline *cl;
@@ -639,7 +628,7 @@ static inline u32 __i915_request_hwsp_offset(const struct i915_request *rq)
if (cl)
return cl->ggtt_offset;
- return rcu_dereference_protected(rq->timeline, 1)->ggtt_offset;
+ return i915_request_get_timeline(rq)->ggtt_offset;
}
static inline bool i915_request_use_scheduler(const struct i915_request *rq)
diff --git a/drivers/gpu/drm/i915/i915_scheduler.c b/drivers/gpu/drm/i915/i915_scheduler.c
index 8de1fff57efc..3f1a81809561 100644
--- a/drivers/gpu/drm/i915/i915_scheduler.c
+++ b/drivers/gpu/drm/i915/i915_scheduler.c
@@ -134,7 +134,7 @@ static void i915_sched_init_ipi(struct i915_sched_ipi *ipi)
static bool match_ring(const struct i915_request *rq)
{
const struct intel_engine_cs *engine = i915_request_get_engine(rq);
- const struct intel_ring *ring = rq->ring;
+ const struct intel_ring *ring = i915_request_get_ring(rq);
if (intel_ring_is_internal(ring))
return true;
@@ -791,7 +791,7 @@ static u64 signal_deadline(const struct i915_request *rq)
if (__i915_request_is_complete(s))
continue;
- if (s->timeline == rq->timeline &&
+ if (i915_request_get_timeline(s) == i915_request_get_timeline(rq) &&
__i915_request_has_started(s))
continue;
@@ -1264,9 +1264,9 @@ struct i915_request *__i915_sched_rewind_requests(struct i915_sched *se)
set_bit(I915_FENCE_FLAG_PQUEUE, &rq->fence.flags);
/* Check in case we rollback so far we wrap [size/2] */
- if (intel_ring_direction(rq->ring,
+ if (intel_ring_direction(i915_request_get_ring(rq),
rq->tail,
- rq->ring->tail + 8) > 0)
+ i915_request_get_ring(rq)->tail + 8) > 0)
rq->context->lrc.desc |= CTX_DESC_FORCE_RESTORE;
active = rq;
@@ -1711,7 +1711,7 @@ void i915_request_show_with_schedule(struct drm_printer *m,
node_to_request(dep->signaler);
/* Dependencies along the same timeline are expected. */
- if (signaler->timeline == rq->timeline)
+ if (i915_request_get_timeline(signaler) == i915_request_get_timeline(rq))
continue;
if (__i915_request_is_complete(signaler))
@@ -1776,16 +1776,16 @@ print_request_ring(struct drm_printer *m, const struct i915_request *rq)
size = rq->tail - rq->head;
if (rq->tail < rq->head)
- size += rq->ring->size;
+ size += i915_request_get_ring(rq)->size;
ring = kmalloc(size, GFP_ATOMIC);
if (ring) {
- const void *vaddr = rq->ring->vaddr;
+ const void *vaddr = i915_request_get_ring(rq)->vaddr;
unsigned int head = rq->head;
unsigned int len = 0;
if (rq->tail < head) {
- len = rq->ring->size - head;
+ len = i915_request_get_ring(rq)->size - head;
memcpy(ring, vaddr + head, len);
head = 0;
}
@@ -1828,20 +1828,18 @@ void i915_sched_show(struct drm_printer *m,
rq = se->active_request(se);
if (rq) {
+ struct intel_ring *ring = i915_request_get_ring(rq);
+
i915_request_show(m, rq, "\t\tactive ", 0);
drm_printf(m, "\t\tring->start: 0x%08x\n",
- intel_ring_address(rq->ring));
- drm_printf(m, "\t\tring->head: 0x%08x\n",
- rq->ring->head);
- drm_printf(m, "\t\tring->tail: 0x%08x\n",
- rq->ring->tail);
- drm_printf(m, "\t\tring->emit: 0x%08x\n",
- rq->ring->emit);
- drm_printf(m, "\t\tring->space: 0x%08x\n",
- rq->ring->space);
+ intel_ring_address(ring));
+ drm_printf(m, "\t\tring->head: 0x%08x\n", ring->head);
+ drm_printf(m, "\t\tring->tail: 0x%08x\n", ring->tail);
+ drm_printf(m, "\t\tring->emit: 0x%08x\n", ring->emit);
+ drm_printf(m, "\t\tring->space: 0x%08x\n", ring->space);
drm_printf(m, "\t\tring->hwsp: 0x%08x\n",
- i915_request_active_timeline(rq)->ggtt_offset);
+ i915_request_get_timeline(rq)->ggtt_offset);
print_request_ring(m, rq);
diff --git a/drivers/gpu/drm/i915/selftests/i915_request.c b/drivers/gpu/drm/i915/selftests/i915_request.c
index 8059888f8eef..f7f7808eedea 100644
--- a/drivers/gpu/drm/i915/selftests/i915_request.c
+++ b/drivers/gpu/drm/i915/selftests/i915_request.c
@@ -1255,14 +1255,15 @@ max_batches(struct i915_gem_context *ctx, struct intel_engine_cs *engine)
if (IS_ERR(rq)) {
ret = PTR_ERR(rq);
} else {
+ struct intel_ring *ring = i915_request_get_ring(rq);
int sz;
- ret = rq->ring->size - rq->reserved_space;
+ ret = ring->size - rq->reserved_space;
i915_request_add(rq);
- sz = rq->ring->emit - rq->head;
+ sz = ring->emit - rq->head;
if (sz < 0)
- sz += rq->ring->size;
+ sz += ring->size;
ret /= sz;
ret /= 2; /* leave half spare, in case of emergency! */
}
--
2.20.1
More information about the Intel-gfx-trybot
mailing list