[Intel-xe] [PATCH] drm/xe: fix xe_device_mem_access_get() race
Thomas Hellström
thomas.hellstrom at linux.intel.com
Thu May 4 13:39:28 UTC 2023
On 5/4/23 07:31, Matthew Brost wrote:
> On Wed, May 03, 2023 at 04:28:02PM +0100, Matthew Auld wrote:
>> It looks like there is at least one race here, given that the
>> pm_runtime_suspended() check looks to return false if we are in the
>> process of suspending the device (RPM_SUSPENDING vs RPM_SUSPENDED). We
>> later also do xe_pm_runtime_get_if_active(), but since the device is
>> suspending or has now suspended, this doesn't do anything either.
>> Following from this we can potentially return from
>> xe_device_mem_access_get() with the device suspended or about to be,
>> leading to broken behaviour.
>>
>> Attempt to fix this by always grabbing the runtime ref when our internal
>> ref transitions from 0 -> 1, and then wrap the whole thing with a lock
>> to ensure callers are serialized.
>>
>> Closes: https://gitlab.freedesktop.org/drm/xe/kernel/-/issues/258
>> Signed-off-by: Matthew Auld <matthew.auld at intel.com>
>> Cc: Rodrigo Vivi <rodrigo.vivi at intel.com>
>> ---
>> drivers/gpu/drm/xe/xe_device.c | 22 ++++++++++++----------
>> drivers/gpu/drm/xe/xe_device_types.h | 5 +++++
>> drivers/gpu/drm/xe/xe_pm.c | 9 ++-------
>> drivers/gpu/drm/xe/xe_pm.h | 2 +-
>> 4 files changed, 20 insertions(+), 18 deletions(-)
>>
>> diff --git a/drivers/gpu/drm/xe/xe_device.c b/drivers/gpu/drm/xe/xe_device.c
>> index 45d6e5ff47fd..5f6554bb34d2 100644
>> --- a/drivers/gpu/drm/xe/xe_device.c
>> +++ b/drivers/gpu/drm/xe/xe_device.c
>> @@ -209,6 +209,8 @@ struct xe_device *xe_device_create(struct pci_dev *pdev,
>>
>> xe->ordered_wq = alloc_ordered_workqueue("xe-ordered-wq", 0);
>>
>> + drmm_mutex_init(&xe->drm, &xe->mem_access.lock);
>> +
>> err = xe_display_create(xe);
>> if (WARN_ON(err))
>> goto err_put;
>> @@ -404,26 +406,26 @@ u32 xe_device_ccs_bytes(struct xe_device *xe, u64 size)
>>
>> void xe_device_mem_access_get(struct xe_device *xe)
>> {
>> - bool resumed = xe_pm_runtime_resume_if_suspended(xe);
>> - int ref = atomic_inc_return(&xe->mem_access.ref);
>> + int ref;
>>
>> + mutex_lock(&xe->mem_access.lock);
>> + ref = atomic_inc_return(&xe->mem_access.ref);
> Drive by comment, if we have a lock then why does this need to be
> atomic?
>
> For the review maybe loop in Maarten because if I recall correctly he
> changed this to an atomic to fix some lockdep splat.
>
> Matt
I also stumbled upon this race some time ago but didn't dig deeper. I
agree we need a mutex here but unless we access the mem_access.ref
somewhere without the mutex, doesn't need to be atomic.
An alternative construct is to grab the mutex only on 0->1 and 1->0
transitions and keep the ref atomic, but I'm not sure how much is gained
assuming nearly all mutex locks are not contended.
/Thomas
>> if (ref == 1)
>> - xe->mem_access.hold_rpm = xe_pm_runtime_get_if_active(xe);
>> -
>> - /* The usage counter increased if device was immediately resumed */
>> - if (resumed)
>> - xe_pm_runtime_put(xe);
>> + xe->mem_access.hold_rpm = xe_pm_runtime_resume_and_get(xe);
>> + mutex_unlock(&xe->mem_access.lock);
>>
>> XE_WARN_ON(ref == S32_MAX);
>> }
>>
>> void xe_device_mem_access_put(struct xe_device *xe)
>> {
>> - bool hold = xe->mem_access.hold_rpm;
>> - int ref = atomic_dec_return(&xe->mem_access.ref);
>> + int ref;
>>
>> - if (!ref && hold)
>> + mutex_lock(&xe->mem_access.lock);
>> + ref = atomic_dec_return(&xe->mem_access.ref);
>> + if (!ref && xe->mem_access.hold_rpm)
>> xe_pm_runtime_put(xe);
>> + mutex_unlock(&xe->mem_access.lock);
>>
>> XE_WARN_ON(ref < 0);
>> }
>> diff --git a/drivers/gpu/drm/xe/xe_device_types.h b/drivers/gpu/drm/xe/xe_device_types.h
>> index 1cb404e48aaa..e8d320f93852 100644
>> --- a/drivers/gpu/drm/xe/xe_device_types.h
>> +++ b/drivers/gpu/drm/xe/xe_device_types.h
>> @@ -257,6 +257,11 @@ struct xe_device {
>> * triggering additional actions when they occur.
>> */
>> struct {
>> + /**
>> + * @lock: Serialize xe_device_mem_access users, and protect the
>> + * below internal state.
>> + */
>> + struct mutex lock;
>> /** @ref: ref count of memory accesses */
>> atomic_t ref;
>> /** @hold_rpm: need to put rpm ref back at the end */
>> diff --git a/drivers/gpu/drm/xe/xe_pm.c b/drivers/gpu/drm/xe/xe_pm.c
>> index b7b57f10ba25..b2ffa001e6f7 100644
>> --- a/drivers/gpu/drm/xe/xe_pm.c
>> +++ b/drivers/gpu/drm/xe/xe_pm.c
>> @@ -210,14 +210,9 @@ int xe_pm_runtime_put(struct xe_device *xe)
>> return pm_runtime_put_autosuspend(xe->drm.dev);
>> }
>>
>> -/* Return true if resume operation happened and usage count was increased */
>> -bool xe_pm_runtime_resume_if_suspended(struct xe_device *xe)
>> +bool xe_pm_runtime_resume_and_get(struct xe_device *xe)
>> {
>> - /* In case we are suspended we need to immediately wake up */
>> - if (pm_runtime_suspended(xe->drm.dev))
>> - return !pm_runtime_resume_and_get(xe->drm.dev);
>> -
>> - return false;
>> + return !pm_runtime_resume_and_get(xe->drm.dev);
>> }
>>
>> int xe_pm_runtime_get_if_active(struct xe_device *xe)
>> diff --git a/drivers/gpu/drm/xe/xe_pm.h b/drivers/gpu/drm/xe/xe_pm.h
>> index 6a885585f653..1b4c15b5e71a 100644
>> --- a/drivers/gpu/drm/xe/xe_pm.h
>> +++ b/drivers/gpu/drm/xe/xe_pm.h
>> @@ -19,7 +19,7 @@ int xe_pm_runtime_suspend(struct xe_device *xe);
>> int xe_pm_runtime_resume(struct xe_device *xe);
>> int xe_pm_runtime_get(struct xe_device *xe);
>> int xe_pm_runtime_put(struct xe_device *xe);
>> -bool xe_pm_runtime_resume_if_suspended(struct xe_device *xe);
>> +bool xe_pm_runtime_resume_and_get(struct xe_device *xe);
>> int xe_pm_runtime_get_if_active(struct xe_device *xe);
>>
>> #endif
>> --
>> 2.40.0
>>
More information about the Intel-xe
mailing list