[Intel-gfx] [PATCH 15/18] drm/i915/guc: Ensure H2G buffer updates visible before tail update

John Harrison john.c.harrison at intel.com
Fri May 28 01:13:58 UTC 2021


On 5/26/2021 10:58, Matthew Brost wrote:
> On Wed, May 26, 2021 at 02:36:18PM +0200, Michal Wajdeczko wrote:
>> On 26.05.2021 08:42, Matthew Brost wrote:
>>> Ensure H2G buffer updates are visible before descriptor tail updates by
>>> inserting a barrier between the H2G buffer update and the tail. The
>>> barrier is simple wmb() for SMEM and is register write for LMEM. This is
>>> needed if more than 1 H2G can be inflight at once.
>>>
>>> Signed-off-by: Matthew Brost <matthew.brost at intel.com>
>>> Cc: Michal Wajdeczko <michal.wajdeczko at intel.com>
>>> ---
>>>   drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c | 18 ++++++++++++++++++
>>>   1 file changed, 18 insertions(+)
>>>
>>> diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c b/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
>>> index fb875d257536..42063e1c355d 100644
>>> --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
>>> +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
>>> @@ -328,6 +328,18 @@ static u32 ct_get_next_fence(struct intel_guc_ct *ct)
>>>   	return ++ct->requests.last_fence;
>>>   }
>>>   
>>> +static void write_barrier(struct intel_guc_ct *ct) {
>>> +	struct intel_guc *guc = ct_to_guc(ct);
>>> +	struct intel_gt *gt = guc_to_gt(guc);
>>> +
>>> +	if (i915_gem_object_is_lmem(guc->ct.vma->obj)) {
>>> +		GEM_BUG_ON(guc->send_regs.fw_domains);
>>> +		intel_uncore_write_fw(gt->uncore, GEN11_SOFT_SCRATCH(0), 0);
>> hmm, as this is one of the GuC scratch registers used for H2G MMIO
>> communication, writing 0 there might be interpreted by the GuC as new
>> request with action=0 and might results in extra processing/logging on
>> GuC side, and, since from here we don't protect access to this register
>> by send_mutex, we can corrupt other MMIO message being prepared from
>> different thread, ... can't we use other register ?
>>
> Hmm, this code has been internal for a long time and we haven't seen an
> issues. MMIOs are always attempted to be processed each interrupt and
> then CTBs are processed next. A value a 0 in scratch0 results in no MMIOs
> being processed as a value of 0 is a reserved action which translates to
> a NOP.
>
> Also in the current i915 once CTBs are enabled MMIOs are never used.
> That being said, I think once we transition to the new interface +
> enable suspend on a VF MMIOs might be used.
>
> With that I purpose that we merge this as is with a comment saying if we
> ever mix CTBs and MMIOs we need to find another MMIO register. I don't
> changing this now is worth delaying upstreaming this and also any change
> we make now will make us lose confidence in code that has been
> thoroughly tested.
>
> Matt
This was discussed in chat while inspecting the GuC firmware code. 
Writing zero to the scratch does indeed not trigger any extra processing 
of spurious MMIO H2Gs. The register is indeed always checked when the 
host triggers a CTB H2G, but zero counts as invalid and thus will be 
skipped.

So with a comment about not mixing CTB and MMIOs, I think we are good 
for now. It seems unlikely that MMIOs & CTB would be mixed. MMIOs are 
only used for initialisation operations and should not be necessary once 
the CTBs are up and running. If mixing does occur in the future, it 
sounds like something that should be addressed at the GuC architecture 
level!

With the comment added:
Reviewed-by: John Harrison <John.C.Harrison at Intel.com>


>   
>>> +	} else {
>>> +		wmb();
>>> +	}
>>> +}
>>> +
>>>   /**
>>>    * DOC: CTB Host to GuC request
>>>    *
>>> @@ -411,6 +423,12 @@ static int ct_write(struct intel_guc_ct *ct,
>>>   	}
>>>   	GEM_BUG_ON(tail > size);
>>>   
>>> +	/*
>>> +	 * make sure H2G buffer update and LRC tail update (if this triggering a
>>> +	 * submission) are visible before updating the descriptor tail
>>> +	 */
>>> +	write_barrier(ct);
>>> +
>>>   	/* now update desc tail (back in bytes) */
>>>   	desc->tail = tail * 4;
>>>   	return 0;
>>>
> _______________________________________________
> Intel-gfx mailing list
> Intel-gfx at lists.freedesktop.org
> https://lists.freedesktop.org/mailman/listinfo/intel-gfx



More information about the dri-devel mailing list