[Intel-gfx] [PATCH 4/5] drm/i915/gt: Adopt engine_park synchronisation rules for engine_retire
Tvrtko Ursulin
tvrtko.ursulin at linux.intel.com
Thu Nov 21 16:17:59 UTC 2019
On 21/11/2019 14:53, Chris Wilson wrote:
> Quoting Tvrtko Ursulin (2019-11-21 14:42:56)
>>
>> On 21/11/2019 13:51, Chris Wilson wrote:
>>> In the next patch, we will introduce a new asynchronous retirement
>>> worker, fed by execlists CS events. Here we may queue a retirement as
>>> soon as a request is submitted to HW (and completes instantly), and we
>>> also want to process that retirement as early as possible and cannot
>>> afford to postpone (as there may not be another opportunity to retire it
>>> for a few seconds). To allow the new async retirer to run in parallel
>>> with our submission, pull the __i915_request_queue (that passes the
>>> request to HW) inside the timelines spinlock so that the retirement
>>> cannot release the timeline before we have completed the submission.
>>>
>>> Signed-off-by: Chris Wilson <chris at chris-wilson.co.uk>
>>> Cc: Tvrtko Ursulin <tvrtko.ursulin at intel.com>
>>> ---
>>> drivers/gpu/drm/i915/gt/intel_engine_pm.c | 29 ++++++++++++++++-------
>>> 1 file changed, 21 insertions(+), 8 deletions(-)
>>>
>>> diff --git a/drivers/gpu/drm/i915/gt/intel_engine_pm.c b/drivers/gpu/drm/i915/gt/intel_engine_pm.c
>>> index 373a4b9f159c..bd0af02bea16 100644
>>> --- a/drivers/gpu/drm/i915/gt/intel_engine_pm.c
>>> +++ b/drivers/gpu/drm/i915/gt/intel_engine_pm.c
>>> @@ -74,18 +74,33 @@ static inline void __timeline_mark_unlock(struct intel_context *ce,
>>> #endif /* !IS_ENABLED(CONFIG_LOCKDEP) */
>>>
>>> static void
>>> -__intel_timeline_enter_and_release_pm(struct intel_timeline *tl,
>>> - struct intel_engine_cs *engine)
>>> +__queue_and_release_pm(struct i915_request *rq,
>>> + struct intel_timeline *tl,
>>> + struct intel_engine_cs *engine)
>>> {
>>> struct intel_gt_timelines *timelines = &engine->gt->timelines;
>>>
>>> + /*
>>> + * We have to serialise all potential retirement paths with our
>>> + * submission, as we don't want to underflow either the
>>> + * engine->wakeref.counter or our timeline->active_count.
>>> + *
>>> + * Equally, we cannot allow a new submission to start until
>>> + * after we finish queueing, nor could we allow that submitter
>>> + * to retire us before we are ready!
>>> + */
>>> spin_lock(&timelines->lock);
>>>
>>> - if (!atomic_fetch_inc(&tl->active_count))
>>> - list_add_tail(&tl->link, &timelines->active_list);
>>> + /* Hand the request over to HW and so engine_retire() */
>>> + __i915_request_queue(rq, NULL);
>>>
>>> + /* Let new submissions commence (and maybe retire this timeline) */
>>> __intel_wakeref_defer_park(&engine->wakeref);
>>>
>>> + /* Let intel_gt_retire_requests() retire us */
>>> + if (!atomic_fetch_inc(&tl->active_count))
>>> + list_add_tail(&tl->link, &timelines->active_list);
>>> +
>>> spin_unlock(&timelines->lock);
>>
>> Now that everything is under the lock the order of operation is not
>> important, or it still is?
>
> queue before unpark that is required.
>
> unpark and add_to_timeline, the order is flexible as the lock governors
> the interactions between those and retirers. So I chose to allow the
> next newcomer start a few instructions earlier.
Yes, because of different locks. So the comment above
__intel_wakeref_defer_park is not correct since timeline cannot be
retired until the lock is dropped.
It's only preservation of timeline ordering which mandates defer_park
after request_queue. As far as I am able to summon my own understanding
from yesterday.
Regards,
Tvrtko
More information about the Intel-gfx
mailing list