[igt-dev] [i-g-t] tests/i915/exec_balancer: Added Skip Guc Submission

Matthew Brost matthew.brost at intel.com
Wed Dec 1 23:42:34 UTC 2021


On Wed, Dec 01, 2021 at 09:46:02AM +0000, Tvrtko Ursulin wrote:
> 
> On 30/11/2021 16:48, Matthew Brost wrote:
> > On Mon, Nov 29, 2021 at 11:15:22AM +0000, Tvrtko Ursulin wrote:
> > > 
> > > On 29/11/2021 10:58, Katragadda, MastanX wrote:
> > > > Hi   Tvrtko Ursulin,
> > > >                Based on following backend information added skip on guc enabled platforms.
> > > > 
> > > > basically there is a quirk in GuC scheduling where if a context reaches the head of the queue and can't be scheduled it blocks the rest of the queue. A queue is an engine class. So in this case if the user submits to VCS0, VCS0, then VCS1 and the first submission to VCS0 is spinner the VCS1 submission is blocked. This more or less is exactly what this test is doing, thus it hangs. We have a request with the GuC firmware team to be able to tweak this head of queue blocking behaviour but don't expect to land anytime soon. Also in the real world this isn't an issue as the user should always be using VEs which should never block the head of the queue unless all the engines within the class are busy.
> > > > 
> > > > Test is expected to fail with GuC submission, skip it in CI.
> > > 
> > > Does "blocks the rest of the queue" mean unrelated contexts submitted
> > > against the same engine class?
> > > 
> > 
> > Yes.
> > 
> > > If so then it would be a DoS vector and the "user _should_ always" would not
> > > be sufficient.
> > > 
> > 
> > This is a DoS vector but there are about a million others and this no
> > worse than most.
> 
> Adding new ones should not be taken lightly.
> 
> > > Or if the blockage is localised to a single context then it might be fine
> > > (but unfortunate) if on top we chose to disallow submission to non-virtual
> > > indices in the engine map (in case of GuC)? If the firmware bug is not
> > 
> > We discussed this we basically land on if the UMDs want to do something
> > stupid, let them as with all the other DoS in the i915. We likely can't
> > disable non-virtual submission as some UMDs want to explictly place some
> > contexts on certain engines (compute). In that case the UMD has to be
> > smart enough to not submit contexts in a way to expose this scheduling
> > quirk.
> 
> "Quirk"... :) Why it happens though? Is the firmware scheduling with GEM
> context granularity, and not intel_context, and that is the disconnect?
> 

The firmware has no idea of GEM contexts or requests only contexts (a LRC).

The scheduling queue is per class, if the HoQ can't be scheduled it
blocks scheduling for that queue until the HoQ can be scheduled. If the
user is using VEs then the HoQ can't block unless all engines within the
class are full. This all goes back to making VEs optional to the user,
once we did that we have exposed ourselves to dump users.

Again this not a bug in the GuC firmware, this is by design and was SoB
by multiple architects. The i915 team doesn't control the GuC firmware
arch BTW. We can ask for changes, but at the end of the day we can't
force any changes without SoB from multiple architects, on multiple
teams.

The current proposal is add a per class KMD controlled policy which
controls the search depth of the queue to find a context which can be
scheduled (e.g. search so many entires past of the HoQ). It is likely to
get implemented.

> > > getting fixed that is. I may be on the wrong track here since I am not 100%
> > 
> > Just because the GuC scheduling doesn't work like execlists doesn't mean
> > that it is bug. Checking for the HoQ only is done for a good reason -
> > performance a uC can do everything and anything like execlists as it is
> > single low power CPU controling the scheduling of an entire GT. We can't
> > protect against everything a user can do that is stupid.
> > 
> > FWIW we do have an open task assigned to the GuC team to allow a KMD
> > configurable search depth when the queue can't be scheduled. No idea
> > when this is going to get implemented.
> 
> This maybe too specific. Question rather could be why is something not
> runnable put at the head of the queue by the firmware. Sounds like a plain
> bug to me.
> 

See above. As far as I'm concerned this is not a bug.

Also see Daniel's comments, this is basically a broken test which
assumes certain scheduler behavior. It is not testing ABI or a POR
arch requirement.

> > > certain I figured out why it exactly gets stuck.
> > > 
> > > Because, looking at the bonded-pair to start with, if the test is emitting a
> > > pair of request on the same context, spinner first, then a another one with
> > > a semaphore dependency I am not sure why it hangs. When the spinner switches
> > > out after time slice expires the second request should run, cancel the
> > > spinner and exit. At which point they are both complete.
> > > 
> > 
> > I think only the MT version of bonded-pair hangs but it has been a while
> > since I looked at this.
> > 
> > IMO this fix is 100% correct as this is a known, tracked issue. It was
> > agreed upon (arch, i915, GuC team) that we just skip these tests with
> > GuC submission.
> 
> I915 team is here on upstream as well.
> 
> Record those acks publicly would be my ask. Unless some security by
> obscurity is happening here? Until then from me it is a soft nack to keep
> disabling tests which show genuine weaknesses in GuC mode. Soft until we get
> a public record of exactly what is broken and in what circumstances, acked
> by architects publicly as you say they acked it somewhere. Commit message
> devoid of detail is not good enough.

I'm fine with not fixing this but you'd have to ask the CI if they can
live with this. If they can't (it crashes CI runs) then we have to skip
this within the test.

Matt

> 
> Regards,
> 
> Tvrtko


More information about the igt-dev mailing list