[Intel-xe] [PATCH v9 4/8] drm/xe: tweak lock ordering for freq_lock

Matthew Auld matthew.auld at intel.com
Tue May 23 13:13:29 UTC 2023


On 23/05/2023 14:07, Rodrigo Vivi wrote:
> On Tue, May 23, 2023 at 10:47:23AM +0100, Matthew Auld wrote:
>> Lockdep spits out:
>>
>> Possible unsafe locking scenario:
>>     CPU0				CPU1
>>     ----				----
>>     lock(&xe->mem_access.lock);
>> 				lock(&pc->freq_lock);
>> 				lock(&xe->mem_access.lock);
>>     lock(&pc->freq_lock);
>>
>> It looks like we already mostly take care to ensure that the mem_access
>> ref is taken outside of taking freq_lock, since that is also grabbed
>> from the runtime_pm callbacks, except for a couple of spots in
>> xe_guc_pc and guc_ct_send_recv(). But tt looks like all the callers are
>> already holding the mem_access.ref, so just switch over to use
>> guc_ct_send_locked().
>>
>> v2:
>>   - There are few more cases it seems.
> 
> should we do this to all the ct_send? and kill this function in favor
> of the locked one?

Yeah, my new thinking is that everything calling ct_send or similar 
needs to already be holding the mem_access.ref i.e we drop the 
mem_access_get() and just keep the assert. But need to check all the 
callers...


> 
> or how we can be sure that we are done with all the cases?
> 
>>
>> Signed-off-by: Matthew Auld <matthew.auld at intel.com>
>> Cc: Rodrigo Vivi <rodrigo.vivi at intel.com>
>> ---
>>   drivers/gpu/drm/xe/xe_guc_ct.c | 4 +++-
>>   drivers/gpu/drm/xe/xe_guc_pc.c | 8 ++++++--
>>   2 files changed, 9 insertions(+), 3 deletions(-)
>>
>> diff --git a/drivers/gpu/drm/xe/xe_guc_ct.c b/drivers/gpu/drm/xe/xe_guc_ct.c
>> index 9dc906f2651a..7a8c9d6a03f3 100644
>> --- a/drivers/gpu/drm/xe/xe_guc_ct.c
>> +++ b/drivers/gpu/drm/xe/xe_guc_ct.c
>> @@ -689,7 +689,9 @@ static int guc_ct_send_recv(struct xe_guc_ct *ct, const u32 *action, u32 len,
>>   retry:
>>   	g2h_fence_init(&g2h_fence, response_buffer);
>>   retry_same_fence:
>> -	ret = guc_ct_send(ct, action, len, 0, 0, &g2h_fence);
>> +	mutex_lock(&ct->lock);
>> +	ret = guc_ct_send_locked(ct, action, len, 0, 0, &g2h_fence);
>> +	mutex_unlock(&ct->lock);
>>   	if (unlikely(ret == -ENOMEM)) {
>>   		void *ptr;
>>   
>> diff --git a/drivers/gpu/drm/xe/xe_guc_pc.c b/drivers/gpu/drm/xe/xe_guc_pc.c
>> index e799faa1c6b8..3b56e57d29af 100644
>> --- a/drivers/gpu/drm/xe/xe_guc_pc.c
>> +++ b/drivers/gpu/drm/xe/xe_guc_pc.c
>> @@ -225,7 +225,9 @@ static int pc_action_set_param(struct xe_guc_pc *pc, u8 id, u32 value)
>>   	if (wait_for_pc_state(pc, SLPC_GLOBAL_STATE_RUNNING))
>>   		return -EAGAIN;
>>   
>> -	ret = xe_guc_ct_send(ct, action, ARRAY_SIZE(action), 0, 0);
>> +	mutex_lock(&ct->lock);
>> +	ret = xe_guc_ct_send_locked(ct, action, ARRAY_SIZE(action), 0, 0);
>> +	mutex_unlock(&ct->lock);
>>   	if (ret)
>>   		drm_err(&pc_to_xe(pc)->drm, "GuC PC set param failed: %pe",
>>   			ERR_PTR(ret));
>> @@ -242,7 +244,9 @@ static int pc_action_setup_gucrc(struct xe_guc_pc *pc, u32 mode)
>>   	};
>>   	int ret;
>>   
>> -	ret = xe_guc_ct_send(ct, action, ARRAY_SIZE(action), 0, 0);
>> +	mutex_lock(&ct->lock);
>> +	ret = xe_guc_ct_send_locked(ct, action, ARRAY_SIZE(action), 0, 0);
>> +	mutex_unlock(&ct->lock);
>>   	if (ret)
>>   		drm_err(&pc_to_xe(pc)->drm, "GuC RC enable failed: %pe",
>>   			ERR_PTR(ret));
>> -- 
>> 2.40.1
>>


More information about the Intel-xe mailing list