[Intel-xe] [PATCH v4 1/7] drm/xe: hold mem_access.ref for CT fast-path

Matthew Auld matthew.auld at intel.com
Thu Jul 6 08:29:09 UTC 2023


On 06/07/2023 04:51, Matthew Brost wrote:
> On Wed, Jul 05, 2023 at 05:06:04PM +0100, Matthew Auld wrote:
>> Just checking xe_device_mem_access_ongoing() is not enough, we also need
>> to hold the reference otherwise the ref can transition from 1 -> 0 as we
>> enter g2h_read(), leading to warnings. While we can't do a full rpm sync
>> in the IRQ, we can keep the device awake if the ref is non-zero.
>> Introduce a new helper for this and set it to work in for the CT
>> fast-path.
>>
>> Signed-off-by: Matthew Auld <matthew.auld at intel.com>
>> Cc: Matthew Brost <matthew.brost at intel.com>
>> Cc: José Roberto de Souza <jose.souza at intel.com>
>> ---
>>   drivers/gpu/drm/xe/xe_device.c | 5 +++++
>>   drivers/gpu/drm/xe/xe_device.h | 1 +
>>   drivers/gpu/drm/xe/xe_guc_ct.c | 5 ++++-
>>   3 files changed, 10 insertions(+), 1 deletion(-)
>>
>> diff --git a/drivers/gpu/drm/xe/xe_device.c b/drivers/gpu/drm/xe/xe_device.c
>> index 07ae208af809..94b0089b0dee 100644
>> --- a/drivers/gpu/drm/xe/xe_device.c
>> +++ b/drivers/gpu/drm/xe/xe_device.c
>> @@ -412,6 +412,11 @@ u32 xe_device_ccs_bytes(struct xe_device *xe, u64 size)
>>   		DIV_ROUND_UP(size, NUM_BYTES_PER_CCS_BYTE) : 0;
>>   }
>>   
>> +bool xe_device_mem_access_get_if_ongoing(struct xe_device *xe)
>> +{
>> +	return atomic_inc_not_zero(&xe->mem_access.ref);
>> +}
>> +
>>   void xe_device_mem_access_get(struct xe_device *xe)
>>   {
>>   	bool resumed = xe_pm_runtime_resume_if_suspended(xe);
>> diff --git a/drivers/gpu/drm/xe/xe_device.h b/drivers/gpu/drm/xe/xe_device.h
>> index 779f71d066e6..8e01bbadb149 100644
>> --- a/drivers/gpu/drm/xe/xe_device.h
>> +++ b/drivers/gpu/drm/xe/xe_device.h
>> @@ -138,6 +138,7 @@ static inline struct xe_force_wake * gt_to_fw(struct xe_gt *gt)
>>   }
>>   
>>   void xe_device_mem_access_get(struct xe_device *xe);
>> +bool xe_device_mem_access_get_if_ongoing(struct xe_device *xe);
>>   void xe_device_mem_access_put(struct xe_device *xe);
>>   
>>   static inline bool xe_device_mem_access_ongoing(struct xe_device *xe)
>> diff --git a/drivers/gpu/drm/xe/xe_guc_ct.c b/drivers/gpu/drm/xe/xe_guc_ct.c
>> index 22bc9ce846db..b7aecc480098 100644
>> --- a/drivers/gpu/drm/xe/xe_guc_ct.c
>> +++ b/drivers/gpu/drm/xe/xe_guc_ct.c
>> @@ -1038,7 +1038,8 @@ void xe_guc_ct_fast_path(struct xe_guc_ct *ct)
>>   	struct xe_device *xe = ct_to_xe(ct);
>>   	int len;
>>   
>> -	if (!xe_device_in_fault_mode(xe) || !xe_device_mem_access_ongoing(xe))
>> +	if (!xe_device_in_fault_mode(xe) ||
>> +	    !xe_device_mem_access_get_if_ongoing(xe))
>>   		return;
>>   
>>   	spin_lock(&ct->fast_lock);
>> @@ -1048,6 +1049,8 @@ void xe_guc_ct_fast_path(struct xe_guc_ct *ct)
>>   			g2h_fast_path(ct, ct->fast_msg, len);
>>   	} while (len > 0);
>>   	spin_unlock(&ct->fast_lock);
>> +
>> +	xe_device_mem_access_put(xe);
> 
> Can't this sleep if would go from 1->0, i.e. can't xe_pm_runtime_put sleep?

Thanks for the review. The rpm put() in xe_device_mem_access_put() 
always uses RPM_ASYNC underneath, and that is always safe to use from 
atomic context. The kernel-doc for __pm_runtime_suspend() says:

"This routine may be called in atomic context if the RPM_ASYNC flag is set"

It only really queues the work to run our rpm suspend callback, and 
never runs it directly if using RPM_ASYNC.

> 
> Matt
> 
>>   }
>>   
>>   /* Returns less than zero on error, 0 on done, 1 on more available */
>> -- 
>> 2.41.0
>>


More information about the Intel-xe mailing list