[Intel-gfx] [PATCH 3/5] drm/i915: Fixup preempt-to-busy vs resubmission of a virtual request

Tvrtko Ursulin tvrtko.ursulin at linux.intel.com
Mon Sep 23 13:58:32 UTC 2019


On 23/09/2019 14:39, Chris Wilson wrote:
> Quoting Tvrtko Ursulin (2019-09-23 14:32:23)
>>
>> On 21/09/2019 10:55, Chris Wilson wrote:
>>> As preempt-to-busy leaves the request on the HW as the resubmission is
>>> processed, that request may complete in the background and even cause a
>>> second virtual request to enter queue. This second virtual request
>>> breaks our "single request in the virtual pipeline" assumptions.
>>> Furthermore, as the virtual request may be completed and retired, we
>>> lose the reference the virtual engine assumes is held. Normally, just
>>> removing the request from the scheduler queue removes it from the
>>> engine, but the virtual engine keeps track of its singleton request via
>>> its ve->request. This pointer needs protecting with a reference.
>>>
>>> Fixes: 22b7a426bbe1 ("drm/i915/execlists: Preempt-to-busy")
>>> Signed-off-by: Chris Wilson <chris at chris-wilson.co.uk>
>>> Cc: Mika Kuoppala <mika.kuoppala at linux.intel.com>
>>> Cc: Tvrtko Ursulin <tvrtko.ursulin at intel.com>
>>> ---
>>>    drivers/gpu/drm/i915/gt/intel_lrc.c | 34 ++++++++++++++++++++++-------
>>>    1 file changed, 26 insertions(+), 8 deletions(-)
>>>
>>> diff --git a/drivers/gpu/drm/i915/gt/intel_lrc.c b/drivers/gpu/drm/i915/gt/intel_lrc.c
>>> index 53bc4308793c..1b2bacc60300 100644
>>> --- a/drivers/gpu/drm/i915/gt/intel_lrc.c
>>> +++ b/drivers/gpu/drm/i915/gt/intel_lrc.c
>>> @@ -529,7 +529,6 @@ __unwind_incomplete_requests(struct intel_engine_cs *engine)
>>>                                i915_request_cancel_breadcrumb(rq);
>>>                                spin_unlock(&rq->lock);
>>>                        }
>>> -                     rq->engine = owner;
>>>                        owner->submit_request(rq);
>>>                        active = NULL;
>>>                }
>>> @@ -1248,6 +1247,7 @@ static void execlists_dequeue(struct intel_engine_cs *engine)
>>>                                submit = true;
>>>                                last = rq;
>>>                        }
>>> +                     i915_request_put(rq);
>>>    
>>>                        if (!submit) {
>>>                                spin_unlock(&ve->base.active.lock);
>>> @@ -2535,6 +2535,7 @@ static void execlists_cancel_requests(struct intel_engine_cs *engine)
>>>    
>>>                        rq->engine = engine;
>>>                        __i915_request_submit(rq);
>>> +                     i915_request_put(rq);
>>>    
>>>                        ve->base.execlists.queue_priority_hint = INT_MIN;
>>>                }
>>> @@ -3787,7 +3788,9 @@ static void virtual_submission_tasklet(unsigned long data)
>>>    
>>>    static void virtual_submit_request(struct i915_request *rq)
>>>    {
>>> -     struct virtual_engine *ve = to_virtual_engine(rq->engine);
>>> +     struct virtual_engine *ve = to_virtual_engine(rq->hw_context->engine);
>>> +     struct i915_request *stale;
>>> +     unsigned long flags;
>>>    
>>>        GEM_TRACE("%s: rq=%llx:%lld\n",
>>>                  ve->base.name,
>>> @@ -3796,15 +3799,30 @@ static void virtual_submit_request(struct i915_request *rq)
>>>    
>>>        GEM_BUG_ON(ve->base.submit_request != virtual_submit_request);
>>>    
>>> -     GEM_BUG_ON(ve->request);
>>> -     GEM_BUG_ON(!list_empty(virtual_queue(ve)));
>>> +     spin_lock_irqsave(&ve->base.active.lock, flags);
>>> +
>>> +     stale = ve->request;
>>
>> fetch_and_zero so you don't have to set it to NULL a bit lower?
> 
> I iterated through xchg then fetch_and_zero, before settling on this
> variant. My feeling was setting ve->request in both sides of the if()
> was more balanced.

Ok.

> 
>> s/stale/completed/, plus a comment describing preempt-to-busy is to blame?
> 
> completed is a little long, old? Already added a comment to remind about
> the preempt-to-busy link :)

I disliked stale since it has negative connotations and this is actually 
normal/expected situation (albeit rare), but maybe it's just me. Old is 
perhaps better. But you can also keep stale I guess.

>>> +     if (stale) {
>>> +             GEM_BUG_ON(!i915_request_completed(stale));
>>> +             __i915_request_submit(stale);
>>> +             i915_request_put(stale);
>>> +     }
>>> +
>>> +     if (i915_request_completed(rq)) {
>>> +             __i915_request_submit(rq);
>>> +             ve->request = NULL;
>>> +     } else {
>>> +             ve->base.execlists.queue_priority_hint = rq_prio(rq);
>>> +             ve->request = i915_request_get(rq);
>>> +             rq->engine = &ve->base; /* fixup from unwind */
>>
>> The last line has me confused. Isn't this the normal veng rq submission
>> path?
> 
> It is also on the normal submission path.
> 
>> In which case rq->engine will already be set to veng.
> 
> Yes
> 
>> But on the
>> unwind path you have removed reset-back of rq->engine to owner. Ah.. the
>> unwind calls veng->submit_request on it, and then we end up in here..
>> Okay, this is outside the normal path, I mean the else block has two
>> functions/paths, and this should be explained in a comment.
> 
> That was the intent of "fixup from unwind?"
> 
> I can squeeze in /* fixup __unwind_incomplete_requests */ is that more
> clueful? Or do you think it needs more?

I thought it does, but then I did not know what I would write so it's 
useful. So I think it's fine after all.

Reviewed-by: Tvrtko Ursulin <tvrtko.ursulin at intel.com>

Regards,

Tvrtko

> -Chris
> 


More information about the Intel-gfx mailing list