[Intel-gfx] [PATCH 23/27] drm/i915/guc: Implement no mid batch preemption for multi-lrc

Matthew Brost matthew.brost at intel.com
Wed Sep 29 00:22:57 UTC 2021


On Tue, Sep 28, 2021 at 04:33:24PM -0700, John Harrison wrote:
> On 9/28/2021 15:33, Matthew Brost wrote:
> > On Tue, Sep 28, 2021 at 03:20:42PM -0700, John Harrison wrote:
> > > On 8/20/2021 15:44, Matthew Brost wrote:
> > > > For some users of multi-lrc, e.g. split frame, it isn't safe to preempt
> > > > mid BB. To safely enable preemption at the BB boundary, a handshake
> > > > between to parent and child is needed. This is implemented via custom
> > > > emit_bb_start & emit_fini_breadcrumb functions and enabled via by
> > > via by -> by
> > > 
> > > > default if a context is configured by set parallel extension.
> > > I tend to agree with Tvrtko that this should probably be an opt in change.
> > > Is there a flags word passed in when creating the context?
> > > 
> > I don't disagree but the uAPI in this series is where we landed. It has
> > been acked all by the relevant parties in the RFC, ported to our
> > internal tree, and the media UMD has been updated / posted. Concerns
> > with the uAPI should've been raised in the RFC phase, not now. I really
> > don't feel like changing this uAPI another time.
> The counter argument is that once a UAPI has been merged, it cannot be
> changed. Ever. So it is worth taking the trouble to get it right first time.
> 
> The proposal isn't a major re-write of the interface. It is simply a request
> to set an extra flag when creating the context.
> 

We are basically just talking about the polarity of a flag at this
point. Either by default you can't be preempted mid batch (current GPU /
UMD requirement) or by default you can be preempted mid-batch (no
current GPU / UMD can do this yet but add flags that everyone opts
into). I think Daniel's opinion was just default to what the current GPU
/ UMD wants and if future requirements arise we add flags to the
interface. I understand both points of view for flag / not flag but in
the end it doesn't really matter. Either way the interface works now and
will in the future too.

> 
> > 
> > > Also, it's not just a change in pre-emption behaviour but a change in
> > > synchronisation too, right? Previously, if you had a whole bunch of back to
> > > back submissions then one child could run ahead of another and/or the
> > > parent. After this change, there is a forced regroup at the end of each
> > > batch. So while one could end sooner/later than the others, they can't ever
> > > get an entire batch (or more) ahead or behind. Or was that synchronisation
> > > already in there through other means anyway?
> > > 
> > Yes, each parent / child sync at the of each batch - this is the only
> > way safely insert preemption points. Without this the GuC could attempt
> > a preemption and hang the batches.
> To be clear, I'm not saying that this is wrong. I'm just saying that this
> appears to be new behaviour with this patch but it is not explicitly called
> out in the description of the patch.
> 

Will add some comments explaining this behavior (unless I already have
them).

> 
> > 
> > > > Signed-off-by: Matthew Brost <matthew.brost at intel.com>
> > > > ---
> > > >    drivers/gpu/drm/i915/gt/intel_context.c       |   2 +-
> > > >    drivers/gpu/drm/i915/gt/intel_context_types.h |   3 +
> > > >    drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h   |   2 +-
> > > >    .../gpu/drm/i915/gt/uc/intel_guc_submission.c | 283 +++++++++++++++++-
> > > >    4 files changed, 287 insertions(+), 3 deletions(-)
> > > > 
> > > > diff --git a/drivers/gpu/drm/i915/gt/intel_context.c b/drivers/gpu/drm/i915/gt/intel_context.c
> > > > index 5615be32879c..2de62649e275 100644
> > > > --- a/drivers/gpu/drm/i915/gt/intel_context.c
> > > > +++ b/drivers/gpu/drm/i915/gt/intel_context.c
> > > > @@ -561,7 +561,7 @@ void intel_context_bind_parent_child(struct intel_context *parent,
> > > >    	GEM_BUG_ON(intel_context_is_child(child));
> > > >    	GEM_BUG_ON(intel_context_is_parent(child));
> > > > -	parent->guc_number_children++;
> > > > +	child->guc_child_index = parent->guc_number_children++;
> > > >    	list_add_tail(&child->guc_child_link,
> > > >    		      &parent->guc_child_list);
> > > >    	child->parent = parent;
> > > > diff --git a/drivers/gpu/drm/i915/gt/intel_context_types.h b/drivers/gpu/drm/i915/gt/intel_context_types.h
> > > > index 713d85b0b364..727f91e7f7c2 100644
> > > > --- a/drivers/gpu/drm/i915/gt/intel_context_types.h
> > > > +++ b/drivers/gpu/drm/i915/gt/intel_context_types.h
> > > > @@ -246,6 +246,9 @@ struct intel_context {
> > > >    		/** @guc_number_children: number of children if parent */
> > > >    		u8 guc_number_children;
> > > > +		/** @guc_child_index: index into guc_child_list if child */
> > > > +		u8 guc_child_index;
> > > > +
> > > >    		/**
> > > >    		 * @parent_page: page in context used by parent for work queue,
> > > >    		 * work queue descriptor
> > > > diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h b/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h
> > > > index 6cd26dc060d1..9f61cfa5566a 100644
> > > > --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h
> > > > +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h
> > > > @@ -188,7 +188,7 @@ struct guc_process_desc {
> > > >    	u32 wq_status;
> > > >    	u32 engine_presence;
> > > >    	u32 priority;
> > > > -	u32 reserved[30];
> > > > +	u32 reserved[36];
> > > What is this extra space for? All the extra storage is grabbed from after
> > > the end of this structure, isn't it?
> > > 
> > This is the size of process descriptor in the GuC spec. Even though this
> > is unused space we really don't want the child go / join memory using
> > anything within the process descriptor.
> Okay. So it's more that the code was previously broken and we just hadn't
> hit a problem because of it? Again, worth adding a comment in the
> description to call it out as a bug fix.
>

Sure.
 
> > 
> > > >    } __packed;
> > > >    #define CONTEXT_REGISTRATION_FLAG_KMD	BIT(0)
> > > > diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
> > > > index 91330525330d..1a18f99bf12a 100644
> > > > --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
> > > > +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
> > > > @@ -11,6 +11,7 @@
> > > >    #include "gt/intel_context.h"
> > > >    #include "gt/intel_engine_pm.h"
> > > >    #include "gt/intel_engine_heartbeat.h"
> > > > +#include "gt/intel_gpu_commands.h"
> > > >    #include "gt/intel_gt.h"
> > > >    #include "gt/intel_gt_irq.h"
> > > >    #include "gt/intel_gt_pm.h"
> > > > @@ -366,10 +367,14 @@ static struct i915_priolist *to_priolist(struct rb_node *rb)
> > > >    /*
> > > >     * When using multi-lrc submission an extra page in the context state is
> > > > - * reserved for the process descriptor and work queue.
> > > > + * reserved for the process descriptor, work queue, and preempt BB boundary
> > > > + * handshake between the parent + childlren contexts.
> > > >     *
> > > >     * The layout of this page is below:
> > > >     * 0						guc_process_desc
> > > > + * + sizeof(struct guc_process_desc)		child go
> > > > + * + CACHELINE_BYTES				child join ...
> > > > + * + CACHELINE_BYTES ...
> > > Would be better written as '[num_children]' instead of '...' to make it
> > > clear it is a per child array.
> > > 
> > I think this description is pretty clear.
> Evidently not because it confused me for a moment.
> 

Ok, let me see if I can make this a bit more clear.

> > 
> > > Also, maybe create a struct for this to get rid of the magic '+1's and
> > > 'BYTES / sizeof' constructs in the functions below.
> > > 
> > Let me see if I can create a struct that describes the layout.
> That would definitely make the code a lot clearer.
> 
> > 
> > > >     * ...						unused
> > > >     * PAGE_SIZE / 2				work queue start
> > > >     * ...						work queue
> > > > @@ -1785,6 +1790,30 @@ static int deregister_context(struct intel_context *ce, u32 guc_id, bool loop)
> > > >    	return __guc_action_deregister_context(guc, guc_id, loop);
> > > >    }
> > > > +static inline void clear_children_join_go_memory(struct intel_context *ce)
> > > > +{
> > > > +	u32 *mem = (u32 *)(__get_process_desc(ce) + 1);
> > > > +	u8 i;
> > > > +
> > > > +	for (i = 0; i < ce->guc_number_children + 1; ++i)
> > > > +		mem[i * (CACHELINE_BYTES / sizeof(u32))] = 0;
> > > > +}
> > > > +
> > > > +static inline u32 get_children_go_value(struct intel_context *ce)
> > > > +{
> > > > +	u32 *mem = (u32 *)(__get_process_desc(ce) + 1);
> > > > +
> > > > +	return mem[0];
> > > > +}
> > > > +
> > > > +static inline u32 get_children_join_value(struct intel_context *ce,
> > > > +					  u8 child_index)
> > > > +{
> > > > +	u32 *mem = (u32 *)(__get_process_desc(ce) + 1);
> > > > +
> > > > +	return mem[(child_index + 1) * (CACHELINE_BYTES / sizeof(u32))];
> > > > +}
> > > > +
> > > >    static void guc_context_policy_init(struct intel_engine_cs *engine,
> > > >    				    struct guc_lrc_desc *desc)
> > > >    {
> > > > @@ -1867,6 +1896,8 @@ static int guc_lrc_desc_pin(struct intel_context *ce, bool loop)
> > > >    			desc->context_flags = CONTEXT_REGISTRATION_FLAG_KMD;
> > > >    			guc_context_policy_init(engine, desc);
> > > >    		}
> > > > +
> > > > +		clear_children_join_go_memory(ce);
> > > >    	}
> > > >    	/*
> > > > @@ -2943,6 +2974,31 @@ static const struct intel_context_ops virtual_child_context_ops = {
> > > >    	.get_sibling = guc_virtual_get_sibling,
> > > >    };
> > > > +/*
> > > > + * The below override of the breadcrumbs is enabled when the user configures a
> > > > + * context for parallel submission (multi-lrc, parent-child).
> > > > + *
> > > > + * The overridden breadcrumbs implements an algorithm which allows the GuC to
> > > > + * safely preempt all the hw contexts configured for parallel submission
> > > > + * between each BB. The contract between the i915 and GuC is if the parent
> > > > + * context can be preempted, all the children can be preempted, and the GuC will
> > > > + * always try to preempt the parent before the children. A handshake between the
> > > > + * parent / children breadcrumbs ensures the i915 holds up its end of the deal
> > > > + * creating a window to preempt between each set of BBs.
> > > > + */
> > > > +static int emit_bb_start_parent_no_preempt_mid_batch(struct i915_request *rq,
> > > > +						     u64 offset, u32 len,
> > > > +						     const unsigned int flags);
> > > > +static int emit_bb_start_child_no_preempt_mid_batch(struct i915_request *rq,
> > > > +						    u64 offset, u32 len,
> > > > +						    const unsigned int flags);
> > > > +static u32 *
> > > > +emit_fini_breadcrumb_parent_no_preempt_mid_batch(struct i915_request *rq,
> > > > +						 u32 *cs);
> > > > +static u32 *
> > > > +emit_fini_breadcrumb_child_no_preempt_mid_batch(struct i915_request *rq,
> > > > +						u32 *cs);
> > > > +
> > > >    static struct intel_context *
> > > >    guc_create_parallel(struct intel_engine_cs **engines,
> > > >    		    unsigned int num_siblings,
> > > > @@ -2978,6 +3034,20 @@ guc_create_parallel(struct intel_engine_cs **engines,
> > > >    		}
> > > >    	}
> > > > +	parent->engine->emit_bb_start =
> > > > +		emit_bb_start_parent_no_preempt_mid_batch;
> > > > +	parent->engine->emit_fini_breadcrumb =
> > > > +		emit_fini_breadcrumb_parent_no_preempt_mid_batch;
> > > > +	parent->engine->emit_fini_breadcrumb_dw =
> > > > +		12 + 4 * parent->guc_number_children;
> > > > +	for_each_child(parent, ce) {
> > > > +		ce->engine->emit_bb_start =
> > > > +			emit_bb_start_child_no_preempt_mid_batch;
> > > > +		ce->engine->emit_fini_breadcrumb =
> > > > +			emit_fini_breadcrumb_child_no_preempt_mid_batch;
> > > > +		ce->engine->emit_fini_breadcrumb_dw = 16;
> > > > +	}
> > > > +
> > > >    	kfree(siblings);
> > > >    	return parent;
> > > > @@ -3362,6 +3432,204 @@ void intel_guc_submission_init_early(struct intel_guc *guc)
> > > >    	guc->submission_selected = __guc_submission_selected(guc);
> > > >    }
> > > > +static inline u32 get_children_go_addr(struct intel_context *ce)
> > > > +{
> > > > +	GEM_BUG_ON(!intel_context_is_parent(ce));
> > > > +
> > > > +	return i915_ggtt_offset(ce->state) +
> > > > +		__get_process_desc_offset(ce) +
> > > > +		sizeof(struct guc_process_desc);
> > > > +}
> > > > +
> > > > +static inline u32 get_children_join_addr(struct intel_context *ce,
> > > > +					 u8 child_index)
> > > > +{
> > > > +	GEM_BUG_ON(!intel_context_is_parent(ce));
> > > > +
> > > > +	return get_children_go_addr(ce) + (child_index + 1) * CACHELINE_BYTES;
> > > > +}
> > > > +
> > > > +#define PARENT_GO_BB			1
> > > > +#define PARENT_GO_FINI_BREADCRUMB	0
> > > > +#define CHILD_GO_BB			1
> > > > +#define CHILD_GO_FINI_BREADCRUMB	0
> > > > +static int emit_bb_start_parent_no_preempt_mid_batch(struct i915_request *rq,
> > > > +						     u64 offset, u32 len,
> > > > +						     const unsigned int flags)
> > > > +{
> > > > +	struct intel_context *ce = rq->context;
> > > > +	u32 *cs;
> > > > +	u8 i;
> > > > +
> > > > +	GEM_BUG_ON(!intel_context_is_parent(ce));
> > > > +
> > > > +	cs = intel_ring_begin(rq, 10 + 4 * ce->guc_number_children);
> > > > +	if (IS_ERR(cs))
> > > > +		return PTR_ERR(cs);
> > > > +
> > > > +	/* Wait on chidlren */
> > > chidlren -> children
> > > 
> > Yep.
> > > > +	for (i = 0; i < ce->guc_number_children; ++i) {
> > > > +		*cs++ = (MI_SEMAPHORE_WAIT |
> > > > +			 MI_SEMAPHORE_GLOBAL_GTT |
> > > > +			 MI_SEMAPHORE_POLL |
> > > > +			 MI_SEMAPHORE_SAD_EQ_SDD);
> > > > +		*cs++ = PARENT_GO_BB;
> > > > +		*cs++ = get_children_join_addr(ce, i);
> > > > +		*cs++ = 0;
> > > > +	}
> > > > +
> > > > +	/* Turn off preemption */
> > > > +	*cs++ = MI_ARB_ON_OFF | MI_ARB_DISABLE;
> > > > +	*cs++ = MI_NOOP;
> > > > +
> > > > +	/* Tell children go */
> > > > +	cs = gen8_emit_ggtt_write(cs,
> > > > +				  CHILD_GO_BB,
> > > > +				  get_children_go_addr(ce),
> > > > +				  0);
> > > > +
> > > > +	/* Jump to batch */
> > > > +	*cs++ = MI_BATCH_BUFFER_START_GEN8 |
> > > > +		(flags & I915_DISPATCH_SECURE ? 0 : BIT(8));
> > > > +	*cs++ = lower_32_bits(offset);
> > > > +	*cs++ = upper_32_bits(offset);
> > > > +	*cs++ = MI_NOOP;
> > > > +
> > > > +	intel_ring_advance(rq, cs);
> > > > +
> > > > +	return 0;
> > > > +}
> > > > +
> > > > +static int emit_bb_start_child_no_preempt_mid_batch(struct i915_request *rq,
> > > > +						    u64 offset, u32 len,
> > > > +						    const unsigned int flags)
> > > > +{
> > > > +	struct intel_context *ce = rq->context;
> > > > +	u32 *cs;
> > > > +
> > > > +	GEM_BUG_ON(!intel_context_is_child(ce));
> > > > +
> > > > +	cs = intel_ring_begin(rq, 12);
> > > > +	if (IS_ERR(cs))
> > > > +		return PTR_ERR(cs);
> > > > +
> > > > +	/* Signal parent */
> > > > +	cs = gen8_emit_ggtt_write(cs,
> > > > +				  PARENT_GO_BB,
> > > > +				  get_children_join_addr(ce->parent,
> > > > +							 ce->guc_child_index),
> > > > +				  0);
> > > > +
> > > > +	/* Wait parent on for go */
> > > parent on -> on parent
> > > 
> > Yep.
> > > > +	*cs++ = (MI_SEMAPHORE_WAIT |
> > > > +		 MI_SEMAPHORE_GLOBAL_GTT |
> > > > +		 MI_SEMAPHORE_POLL |
> > > > +		 MI_SEMAPHORE_SAD_EQ_SDD);
> > > > +	*cs++ = CHILD_GO_BB;
> > > > +	*cs++ = get_children_go_addr(ce->parent);
> > > > +	*cs++ = 0;
> > > > +
> > > > +	/* Turn off preemption */
> > > > +	*cs++ = MI_ARB_ON_OFF | MI_ARB_DISABLE;
> > > > +
> > > > +	/* Jump to batch */
> > > > +	*cs++ = MI_BATCH_BUFFER_START_GEN8 |
> > > > +		(flags & I915_DISPATCH_SECURE ? 0 : BIT(8));
> > > > +	*cs++ = lower_32_bits(offset);
> > > > +	*cs++ = upper_32_bits(offset);
> > > > +
> > > > +	intel_ring_advance(rq, cs);
> > > > +
> > > > +	return 0;
> > > > +}
> > > > +
> > > > +static u32 *
> > > > +emit_fini_breadcrumb_parent_no_preempt_mid_batch(struct i915_request *rq,
> > > > +						 u32 *cs)
> > > > +{
> > > > +	struct intel_context *ce = rq->context;
> > > > +	u8 i;
> > > > +
> > > > +	GEM_BUG_ON(!intel_context_is_parent(ce));
> > > > +
> > > > +	/* Wait on children */
> > > > +	for (i = 0; i < ce->guc_number_children; ++i) {
> > > > +		*cs++ = (MI_SEMAPHORE_WAIT |
> > > > +			 MI_SEMAPHORE_GLOBAL_GTT |
> > > > +			 MI_SEMAPHORE_POLL |
> > > > +			 MI_SEMAPHORE_SAD_EQ_SDD);
> > > > +		*cs++ = PARENT_GO_FINI_BREADCRUMB;
> > > > +		*cs++ = get_children_join_addr(ce, i);
> > > > +		*cs++ = 0;
> > > > +	}
> > > > +
> > > > +	/* Turn on preemption */
> > > > +	*cs++ = MI_ARB_ON_OFF | MI_ARB_ENABLE;
> > > > +	*cs++ = MI_NOOP;
> > > > +
> > > > +	/* Tell children go */
> > > > +	cs = gen8_emit_ggtt_write(cs,
> > > > +				  CHILD_GO_FINI_BREADCRUMB,
> > > > +				  get_children_go_addr(ce),
> > > > +				  0);
> > > > +
> > > > +	/* Emit fini breadcrumb */
> > > > +	cs = gen8_emit_ggtt_write(cs,
> > > > +				  rq->fence.seqno,
> > > > +				  i915_request_active_timeline(rq)->hwsp_offset,
> > > > +				  0);
> > > > +
> > > > +	/* User interrupt */
> > > > +	*cs++ = MI_USER_INTERRUPT;
> > > > +	*cs++ = MI_NOOP;
> > > > +
> > > > +	rq->tail = intel_ring_offset(rq, cs);
> > > > +
> > > > +	return cs;
> > > > +}
> > > > +
> > > > +static u32 *
> > > > +emit_fini_breadcrumb_child_no_preempt_mid_batch(struct i915_request *rq, u32 *cs)
> > > > +{
> > > > +	struct intel_context *ce = rq->context;
> > > > +
> > > > +	GEM_BUG_ON(!intel_context_is_child(ce));
> > > > +
> > > > +	/* Turn on preemption */
> > > > +	*cs++ = MI_ARB_ON_OFF | MI_ARB_ENABLE;
> > > > +	*cs++ = MI_NOOP;
> > > > +
> > > > +	/* Signal parent */
> > > > +	cs = gen8_emit_ggtt_write(cs,
> > > > +				  PARENT_GO_FINI_BREADCRUMB,
> > > > +				  get_children_join_addr(ce->parent,
> > > > +							 ce->guc_child_index),
> > > > +				  0);
> > > > +
> > > This is backwards compared to the parent?
> > > 
> > > Parent: wait on children then enable pre-emption
> > > Child: enable pre-emption then signal parent
> > > 
> > > Makes for a window where the parent is waiting in atomic context for a
> > > signal from a child that has been pre-empted and might not get to run for
> > > some time?
> > > 
> > No, this is correct. The rule is if the parent can be preempted all the
> > children can be preempted, thus we can't enable preemption on the parent
> > until all the children have preemption enabled, thus the parent waits
> > for all the children to join before enabling its preemption.
> > 
> > Matt
> But,
> 
> The write to PARENT_GO_FINI can't fail or stall, right? So if it happens
> before the ARB_ON then the child is guaranteed to execute the ARB_ON once it
> has signalled the parent. Indeed, by the time the parent context gets to see
> the update memory value, the child is practically certain to have passed the
> ARB_ON. So, by the time the parent becomes pre-emptible, the children will
> all be pre-emptible. Even if the parent is superfast, the children are
> guaranteed to become pre-emptible immediately - certainly before any
> fail-to-preempt timeout could occur.
>
> Whereas, with the current ordering, it is possible for the child to be
> preempted before it has issued the signal to the parent. So now you have a
> non-preemptible parent hogging the hardware, waiting for a signal that isn't

To be clear the parent is always preempted first by the GuC. The parent
can't be running if the child preempt is attempted.

> going to come for an entire execution quantum. Indeed, it is actually quite
> likely the child would be preempted before it can signal the parent because
> any pre-emption request that was issued at any time during the child's
> execution will take effect immediately on the ARB_ON instruction.
> 

Looking at the code, I do think I have a bug though.

I think I'm missing a MI_ARB_CHECK in the parent after turning on
preemption before releasing the children, right?

This covers the case where the GuC issues a preemption to the parent
while it is waiting on the children, all the children join, the parent
turns on preemption and is preempted with the added MI_ARB_CHECK
instruction, and the children all can be preempted waiting on the parent
go semaphore. Does that sound correct?

Matt

> John.
> 
> 
> > > John.
> > > 
> > > 
> > > > +	/* Wait parent on for go */
> > > > +	*cs++ = (MI_SEMAPHORE_WAIT |
> > > > +		 MI_SEMAPHORE_GLOBAL_GTT |
> > > > +		 MI_SEMAPHORE_POLL |
> > > > +		 MI_SEMAPHORE_SAD_EQ_SDD);
> > > > +	*cs++ = CHILD_GO_FINI_BREADCRUMB;
> > > > +	*cs++ = get_children_go_addr(ce->parent);
> > > > +	*cs++ = 0;
> > > > +
> > > > +	/* Emit fini breadcrumb */
> > > > +	cs = gen8_emit_ggtt_write(cs,
> > > > +				  rq->fence.seqno,
> > > > +				  i915_request_active_timeline(rq)->hwsp_offset,
> > > > +				  0);
> > > > +
> > > > +	/* User interrupt */
> > > > +	*cs++ = MI_USER_INTERRUPT;
> > > > +	*cs++ = MI_NOOP;
> > > > +
> > > > +	rq->tail = intel_ring_offset(rq, cs);
> > > > +
> > > > +	return cs;
> > > > +}
> > > > +
> > > >    static struct intel_context *
> > > >    g2h_context_lookup(struct intel_guc *guc, u32 desc_idx)
> > > >    {
> > > > @@ -3807,6 +4075,19 @@ void intel_guc_submission_print_context_info(struct intel_guc *guc,
> > > >    			drm_printf(p, "\t\tWQI Status: %u\n\n",
> > > >    				   READ_ONCE(desc->wq_status));
> > > > +			drm_printf(p, "\t\tNumber Children: %u\n\n",
> > > > +				   ce->guc_number_children);
> > > > +			if (ce->engine->emit_bb_start ==
> > > > +			    emit_bb_start_parent_no_preempt_mid_batch) {
> > > > +				u8 i;
> > > > +
> > > > +				drm_printf(p, "\t\tChildren Go: %u\n\n",
> > > > +					   get_children_go_value(ce));
> > > > +				for (i = 0; i < ce->guc_number_children; ++i)
> > > > +					drm_printf(p, "\t\tChildren Join: %u\n",
> > > > +						   get_children_join_value(ce, i));
> > > > +			}
> > > > +
> > > >    			for_each_child(ce, child)
> > > >    				guc_log_context(p, child);
> > > >    		}
> 


More information about the Intel-gfx mailing list