[Intel-xe] [PATCH] drm/xe: fix xe_device_mem_access_get() race

Rodrigo Vivi rodrigo.vivi at kernel.org
Fri May 5 13:22:43 UTC 2023


On Thu, May 04, 2023 at 05:28:51PM +0100, Matthew Auld wrote:
> On 04/05/2023 16:09, Rodrigo Vivi wrote:
> > On Thu, May 04, 2023 at 03:39:28PM +0200, Thomas Hellström wrote:
> > > 
> > > On 5/4/23 07:31, Matthew Brost wrote:
> > > > On Wed, May 03, 2023 at 04:28:02PM +0100, Matthew Auld wrote:
> > > > > It looks like there is at least one race here, given that the
> > > > > pm_runtime_suspended() check looks to return false if we are in the
> > > > > process of suspending the device (RPM_SUSPENDING vs RPM_SUSPENDED). We
> > > > > later also do xe_pm_runtime_get_if_active(), but since the device is
> > > > > suspending or has now suspended, this doesn't do anything either.
> > > > > Following from this we can potentially return from
> > > > > xe_device_mem_access_get() with the device suspended or about to be,
> > > > > leading to broken behaviour.
> > > > > 
> > > > > Attempt to fix this by always grabbing the runtime ref when our internal
> > > > > ref transitions from 0 -> 1, and then wrap the whole thing with a lock
> > > > > to ensure callers are serialized.
> > 
> > We had the lock and not atomic, but then we had other issue:
> > 
> > https://lore.kernel.org/all/20230228101730.63294-1-maarten.lankhorst@linux.intel.com/
> > 
> > If we are going back to the mutex probably the right way is to revert
> > the last patch. But then we get back to the issue that Maarten solved.
> 
> It sounded like the issue what only with the access_ongoing helper, and
> making the ref atomic fixed that?

Ohh! I see now... Now it makes sense why you are keeping it atomic while
using the mutex.

> As-is we still want that, but in addition
> we wrap the entire operation with a mutex to fix the above race. Should I
> still revert, or perhaps partially revert and keep the lock and add the
> atomic, and then have this patch on top or just squash them together and
> rework the commit message?

I'd say that the best approach is likely a total revert, then a new commit
to add the atomic explaining the removal of the lock on the
xe_device_assert_mem_access...

Or at least some comment explaining why we need the atomic with the lock.

But yeap, I agree that this is needed and it is the simplest we can get
for now.

> 
> > 
> > Probably the serialization there is the only way. And if it is serialized
> > then we just need the atomic, but probably the atomic with the full block
> > of exclusive ownership pattern like the one described in here:
> 
> It doesn't look like using atomics for exclusivity is strictly needed here,
> the dumb mutex approach should also be fine, so long as access_ongoing
> doesn't need to grab it, AFAICT.
> 
> > 
> > https://lwn.net/Articles/698315/
> > 
> > > > > 
> > > > > Closes: https://gitlab.freedesktop.org/drm/xe/kernel/-/issues/258
> > > > > Signed-off-by: Matthew Auld <matthew.auld at intel.com>
> > > > > Cc: Rodrigo Vivi <rodrigo.vivi at intel.com>
> > > > > ---
> > > > >    drivers/gpu/drm/xe/xe_device.c       | 22 ++++++++++++----------
> > > > >    drivers/gpu/drm/xe/xe_device_types.h |  5 +++++
> > > > >    drivers/gpu/drm/xe/xe_pm.c           |  9 ++-------
> > > > >    drivers/gpu/drm/xe/xe_pm.h           |  2 +-
> > > > >    4 files changed, 20 insertions(+), 18 deletions(-)
> > > > > 
> > > > > diff --git a/drivers/gpu/drm/xe/xe_device.c b/drivers/gpu/drm/xe/xe_device.c
> > > > > index 45d6e5ff47fd..5f6554bb34d2 100644
> > > > > --- a/drivers/gpu/drm/xe/xe_device.c
> > > > > +++ b/drivers/gpu/drm/xe/xe_device.c
> > > > > @@ -209,6 +209,8 @@ struct xe_device *xe_device_create(struct pci_dev *pdev,
> > > > >    	xe->ordered_wq = alloc_ordered_workqueue("xe-ordered-wq", 0);
> > > > > +	drmm_mutex_init(&xe->drm, &xe->mem_access.lock);
> > > > > +
> > > > >    	err = xe_display_create(xe);
> > > > >    	if (WARN_ON(err))
> > > > >    		goto err_put;
> > > > > @@ -404,26 +406,26 @@ u32 xe_device_ccs_bytes(struct xe_device *xe, u64 size)
> > > > >    void xe_device_mem_access_get(struct xe_device *xe)
> > > > >    {
> > > > > -	bool resumed = xe_pm_runtime_resume_if_suspended(xe);
> > > > > -	int ref = atomic_inc_return(&xe->mem_access.ref);
> > > > > +	int ref;
> > > > > +	mutex_lock(&xe->mem_access.lock);
> > > > > +	ref = atomic_inc_return(&xe->mem_access.ref);
> > > > Drive by comment, if we have a lock then why does this need to be
> > > > atomic?
> > > > 
> > > > For the review maybe loop in Maarten because if I recall correctly he
> > > > changed this to an atomic to fix some lockdep splat.
> > > > 
> > > > Matt
> > > 
> > > I also stumbled upon this race some time ago but didn't dig deeper. I agree
> > > we need a mutex here but unless we access the mem_access.ref somewhere
> > > without the mutex, doesn't need to be atomic.
> > > 
> > > An alternative construct is to grab the mutex only on 0->1 and 1->0
> > > transitions and keep the ref atomic, but I'm not sure how much is gained
> > > assuming nearly all mutex locks are not contended.
> > > 
> > > /Thomas
> > > 
> > > 
> > > > >    	if (ref == 1)
> > > > > -		xe->mem_access.hold_rpm = xe_pm_runtime_get_if_active(xe);
> > > > > -
> > > > > -	/* The usage counter increased if device was immediately resumed */
> > > > > -	if (resumed)
> > > > > -		xe_pm_runtime_put(xe);
> > > > > +		xe->mem_access.hold_rpm = xe_pm_runtime_resume_and_get(xe);
> > > > > +	mutex_unlock(&xe->mem_access.lock);
> > > > >    	XE_WARN_ON(ref == S32_MAX);
> > > > >    }
> > > > >    void xe_device_mem_access_put(struct xe_device *xe)
> > > > >    {
> > > > > -	bool hold = xe->mem_access.hold_rpm;
> > > > > -	int ref = atomic_dec_return(&xe->mem_access.ref);
> > > > > +	int ref;
> > > > > -	if (!ref && hold)
> > > > > +	mutex_lock(&xe->mem_access.lock);
> > > > > +	ref = atomic_dec_return(&xe->mem_access.ref);
> > > > > +	if (!ref && xe->mem_access.hold_rpm)
> > > > >    		xe_pm_runtime_put(xe);
> > > > > +	mutex_unlock(&xe->mem_access.lock);
> > > > >    	XE_WARN_ON(ref < 0);
> > > > >    }
> > > > > diff --git a/drivers/gpu/drm/xe/xe_device_types.h b/drivers/gpu/drm/xe/xe_device_types.h
> > > > > index 1cb404e48aaa..e8d320f93852 100644
> > > > > --- a/drivers/gpu/drm/xe/xe_device_types.h
> > > > > +++ b/drivers/gpu/drm/xe/xe_device_types.h
> > > > > @@ -257,6 +257,11 @@ struct xe_device {
> > > > >    	 * triggering additional actions when they occur.
> > > > >    	 */
> > > > >    	struct {
> > > > > +		/**
> > > > > +		 * @lock: Serialize xe_device_mem_access users, and protect the
> > > > > +		 * below internal state.
> > > > > +		 */
> > > > > +		struct mutex lock;
> > > > >    		/** @ref: ref count of memory accesses */
> > > > >    		atomic_t ref;
> > > > >    		/** @hold_rpm: need to put rpm ref back at the end */
> > > > > diff --git a/drivers/gpu/drm/xe/xe_pm.c b/drivers/gpu/drm/xe/xe_pm.c
> > > > > index b7b57f10ba25..b2ffa001e6f7 100644
> > > > > --- a/drivers/gpu/drm/xe/xe_pm.c
> > > > > +++ b/drivers/gpu/drm/xe/xe_pm.c
> > > > > @@ -210,14 +210,9 @@ int xe_pm_runtime_put(struct xe_device *xe)
> > > > >    	return pm_runtime_put_autosuspend(xe->drm.dev);
> > > > >    }
> > > > > -/* Return true if resume operation happened and usage count was increased */
> > > > > -bool xe_pm_runtime_resume_if_suspended(struct xe_device *xe)
> > > > > +bool xe_pm_runtime_resume_and_get(struct xe_device *xe)
> > > > >    {
> > > > > -	/* In case we are suspended we need to immediately wake up */
> > > > > -	if (pm_runtime_suspended(xe->drm.dev))
> > > > > -		return !pm_runtime_resume_and_get(xe->drm.dev);
> > > > > -
> > > > > -	return false;
> > > > > +	return !pm_runtime_resume_and_get(xe->drm.dev);
> > > > >    }
> > > > >    int xe_pm_runtime_get_if_active(struct xe_device *xe)
> > > > > diff --git a/drivers/gpu/drm/xe/xe_pm.h b/drivers/gpu/drm/xe/xe_pm.h
> > > > > index 6a885585f653..1b4c15b5e71a 100644
> > > > > --- a/drivers/gpu/drm/xe/xe_pm.h
> > > > > +++ b/drivers/gpu/drm/xe/xe_pm.h
> > > > > @@ -19,7 +19,7 @@ int xe_pm_runtime_suspend(struct xe_device *xe);
> > > > >    int xe_pm_runtime_resume(struct xe_device *xe);
> > > > >    int xe_pm_runtime_get(struct xe_device *xe);
> > > > >    int xe_pm_runtime_put(struct xe_device *xe);
> > > > > -bool xe_pm_runtime_resume_if_suspended(struct xe_device *xe);
> > > > > +bool xe_pm_runtime_resume_and_get(struct xe_device *xe);
> > > > >    int xe_pm_runtime_get_if_active(struct xe_device *xe);
> > > > >    #endif
> > > > > -- 
> > > > > 2.40.0
> > > > > 


More information about the Intel-xe mailing list