[Intel-gfx] [PATCH] drm/i915/guc: Fix revocation of non-persistent contexts

Andrzej Hajda andrzej.hajda at intel.com
Mon Oct 3 09:30:11 UTC 2022


On 03.10.2022 09:59, Tvrtko Ursulin wrote:
> 
> On 30/09/2022 15:52, Andrzej Hajda wrote:
>> On 30.09.2022 11:47, Tvrtko Ursulin wrote:
>>> From: Tvrtko Ursulin <tvrtko.ursulin at intel.com>
>>>
>>> Patch which added graceful exit for non-persistent contexts missed the
>>> fact it is not enough to set the exiting flag on a context and let the
>>> backend handle it from there.
>>>
>>> GuC backend cannot handle it because it runs independently in the
>>> firmware and driver might not see the requests ever again. Patch also
>>> missed the fact some usages of intel_context_is_banned in the GuC 
>>> backend
>>> needed replacing with newly introduced intel_context_is_schedulable.
>>>
>>> Fix the first issue by calling into backend revoke when we know this is
>>> the last chance to do it. Fix the second issue by replacing
>>> intel_context_is_banned with intel_context_is_schedulable, which should
>>> always be safe since latter is a superset of the former.
>>
>> negation of the latter is a ...?
> 
> I did not get what you meant here.
> 
>>> Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin at intel.com>
>>> Fixes: 45c64ecf97ee ("drm/i915: Improve user experience and driver 
>>> robustness under SIGINT or similar")
>>> Cc: Andrzej Hajda <andrzej.hajda at intel.com>
>>> Cc: John Harrison <John.C.Harrison at Intel.com>
>>> Cc: Daniele Ceraolo Spurio <daniele.ceraolospurio at intel.com>
>>> ---
>>>   drivers/gpu/drm/i915/gem/i915_gem_context.c   |  8 +-----
>>>   drivers/gpu/drm/i915/gt/intel_context.c       | 14 +++++++---
>>>   drivers/gpu/drm/i915/gt/intel_context.h       |  8 +-----
>>>   .../gpu/drm/i915/gt/uc/intel_guc_submission.c | 26 +++++++++----------
>>>   4 files changed, 25 insertions(+), 31 deletions(-)
>>>
>>> diff --git a/drivers/gpu/drm/i915/gem/i915_gem_context.c 
>>> b/drivers/gpu/drm/i915/gem/i915_gem_context.c
>>> index 0bcde53c50c6..1e29b1e6d186 100644
>>> --- a/drivers/gpu/drm/i915/gem/i915_gem_context.c
>>> +++ b/drivers/gpu/drm/i915/gem/i915_gem_context.c
>>> @@ -1387,14 +1387,8 @@ kill_engines(struct i915_gem_engines *engines, 
>>> bool exit, bool persistent)
>>>        */
>>>       for_each_gem_engine(ce, engines, it) {
>>>           struct intel_engine_cs *engine;
>>> -        bool skip = false;
>>> -        if (exit)
>>> -            skip = intel_context_set_exiting(ce);
>>> -        else if (!persistent)
>>> -            skip = intel_context_exit_nonpersistent(ce, NULL); > -
>>> -        if (skip)
>>> +        if ((exit || !persistent) && intel_context_revoke(ce))
>>>               continue; /* Already marked. */
>>>           /*
>>> diff --git a/drivers/gpu/drm/i915/gt/intel_context.c 
>>> b/drivers/gpu/drm/i915/gt/intel_context.c
>>> index 654a092ed3d6..398b2a9eed61 100644
>>> --- a/drivers/gpu/drm/i915/gt/intel_context.c
>>> +++ b/drivers/gpu/drm/i915/gt/intel_context.c
>>> @@ -614,13 +614,19 @@ bool intel_context_ban(struct intel_context 
>>> *ce, struct i915_request *rq)
>>>       return ret;
>>>   }
>>> -bool intel_context_exit_nonpersistent(struct intel_context *ce,
>>> -                      struct i915_request *rq)
>>> +bool intel_context_revoke(struct intel_context *ce)
>>>   {
>>>       bool ret = intel_context_set_exiting(ce);
>>> -    if (ce->ops->revoke)
>>> -        ce->ops->revoke(ce, rq, ce->engine->props.preempt_timeout_ms);
>>> +    if (!ret && intel_engine_uses_guc(ce->engine)) {
>>> +        /*
>>> +         * With GuC backend we have to notify it of revocation as soon
>>> +         * as the exiting flag is set.
>>> +         */
>>> +        if (ce->ops->revoke)
>>> +            ce->ops->revoke(ce, NULL,
>>> +                    ce->engine->props.preempt_timeout_ms);
>>> +    }
>>
>> Now revoke is called only with GuC, previously it was called also for 
>> other backends in case non-exiting/non-persistent, is it OK?
> 
> It is okay (execlists has no revoke vfunc, ringbuffer has it but only 
> works if target request is known), but agreed it is a bit ugly. I was in 
> two minds which way to go. Perhaps it would indeed be cleaner to go 
> unconditional. I will resend with that change, copying stable this time 
> round (since 6.0 is out), and can keep your r-b?

Yes, please keep r-b.

Regards
Andrzej

> 
> Regards,
> 
> Tvrtko
> 
>>
>>>       return ret;
>>>   }
>>> diff --git a/drivers/gpu/drm/i915/gt/intel_context.h 
>>> b/drivers/gpu/drm/i915/gt/intel_context.h
>>> index 8e2d70630c49..40f8809d14ea 100644
>>> --- a/drivers/gpu/drm/i915/gt/intel_context.h
>>> +++ b/drivers/gpu/drm/i915/gt/intel_context.h
>>> @@ -319,18 +319,12 @@ static inline bool 
>>> intel_context_is_schedulable(const struct intel_context *ce)
>>>              !test_bit(CONTEXT_BANNED, &ce->flags);
>>>   }
>>> -static inline bool intel_context_is_exiting(const struct 
>>> intel_context *ce)
>>> -{
>>> -    return test_bit(CONTEXT_EXITING, &ce->flags);
>>> -}
>>> -
>>>   static inline bool intel_context_set_exiting(struct intel_context *ce)
>>>   {
>>>       return test_and_set_bit(CONTEXT_EXITING, &ce->flags);
>>>   }
>>> -bool intel_context_exit_nonpersistent(struct intel_context *ce,
>>> -                      struct i915_request *rq);
>>> +bool intel_context_revoke(struct intel_context *ce);
>>>   static inline bool
>>>   intel_context_force_single_submission(const struct intel_context *ce)
>>> diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c 
>>> b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
>>> index 0ef295a94060..88a4476b8e92 100644
>>> --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
>>> +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
>>> @@ -685,7 +685,7 @@ static int __guc_add_request(struct intel_guc 
>>> *guc, struct i915_request *rq)
>>>        * Corner case where requests were sitting in the priority list 
>>> or a
>>>        * request resubmitted after the context was banned.
>>>        */
>>> -    if (unlikely(intel_context_is_banned(ce))) {
>>> +    if (unlikely(!intel_context_is_schedulable(ce))) {
>>>           i915_request_put(i915_request_mark_eio(rq));
>>>           intel_engine_signal_breadcrumbs(ce->engine);
>>>           return 0;
>>> @@ -871,15 +871,15 @@ static int guc_wq_item_append(struct intel_guc 
>>> *guc,
>>>                     struct i915_request *rq)
>>>   {
>>>       struct intel_context *ce = request_to_scheduling_context(rq);
>>> -    int ret = 0;
>>> +    int ret;
>>> -    if (likely(!intel_context_is_banned(ce))) {
>>> -        ret = __guc_wq_item_append(rq);
>>> +    if (unlikely(!intel_context_is_schedulable(ce)))
>>> +        return 0;
>>> -        if (unlikely(ret == -EBUSY)) {
>>> -            guc->stalled_request = rq;
>>> -            guc->submission_stall_reason = STALL_MOVE_LRC_TAIL;
>>> -        }
>>> +    ret = __guc_wq_item_append(rq);
>>> +    if (unlikely(ret == -EBUSY)) {
>>> +        guc->stalled_request = rq;
>>> +        guc->submission_stall_reason = STALL_MOVE_LRC_TAIL;
>>>       }
>>>       return ret;
>>> @@ -898,7 +898,7 @@ static bool multi_lrc_submit(struct i915_request 
>>> *rq)
>>>        * submitting all the requests generated in parallel.
>>>        */
>>>       return test_bit(I915_FENCE_FLAG_SUBMIT_PARALLEL, 
>>> &rq->fence.flags) ||
>>> -        intel_context_is_banned(ce);
>>> +           !intel_context_is_schedulable(ce);
>>>   }
>>>   static int guc_dequeue_one_context(struct intel_guc *guc)
>>> @@ -967,7 +967,7 @@ static int guc_dequeue_one_context(struct 
>>> intel_guc *guc)
>>>           struct intel_context *ce = 
>>> request_to_scheduling_context(last);
>>>           if (unlikely(!ctx_id_mapped(guc, ce->guc_id.id) &&
>>> -                 !intel_context_is_banned(ce))) {
>>> +                 intel_context_is_schedulable(ce))) {
>>>               ret = try_context_registration(ce, false);
>>>               if (unlikely(ret == -EPIPE)) {
>>>                   goto deadlk;
>>> @@ -1577,7 +1577,7 @@ static void guc_reset_state(struct 
>>> intel_context *ce, u32 head, bool scrub)
>>>   {
>>>       struct intel_engine_cs *engine = __context_to_physical_engine(ce);
>>> -    if (intel_context_is_banned(ce))
>>> +    if (!intel_context_is_schedulable(ce))
>>>           return;
>>>       GEM_BUG_ON(!intel_context_is_pinned(ce));
>>> @@ -4518,12 +4518,12 @@ static void guc_handle_context_reset(struct 
>>> intel_guc *guc,
>>>   {
>>>       trace_intel_context_reset(ce);
>>> -    if (likely(!intel_context_is_banned(ce))) {
>>> +    if (likely(intel_context_is_schedulable(ce))) {
>>>           capture_error_state(guc, ce);
>>>           guc_context_replay(ce);
>>>       } else {
>>>           drm_info(&guc_to_gt(guc)->i915->drm,
>>> -             "Ignoring context reset notification of banned context 
>>> 0x%04X on %s",
>>> +             "Ignoring context reset notification of exiting context 
>>> 0x%04X on %s",
>>
>> With small clarifications:
>>
>> Reviewed-by: Andrzej Hajda <andrzej.hajda at intel.com>
>>
>> Regards
>> Andrzej
>>
>>
>>>                ce->guc_id.id, ce->engine->name);
>>>       }
>>>   }
>>



More information about the dri-devel mailing list