[PATCH v2] drm/xe: Defer buffer object shrinker write-backs and GPU waits

Matthew Auld matthew.auld at intel.com
Wed Aug 6 14:45:34 UTC 2025


On 05/08/2025 17:27, Thomas Hellström wrote:
> On Tue, 2025-08-05 at 14:40 +0100, Matthew Auld wrote:
>> On 05/08/2025 08:48, Thomas Hellström wrote:
>>> When the xe buffer-object shrinker allows GPU waits and write-back,
>>> (typically from kswapd), perform multiple passes, skipping
>>> subsequent passes if the shrinker number of scanned objects target
>>> is reached.
>>>
>>> 1) Without GPU waits and write-back
>>> 2) Without write-back
>>> 3) With both GPU-waits and write-back
>>>
>>> This is to avoid stalls and costly write- and readbacks unless they
>>> are really necessary.
>>>
>>> v2:
>>> - Don't test for scan completion twice. (Stuart Summers)
>>> - Update tags.
>>>
>>> Reported-by: melvyn <melvyn2 at dnsense.pub>
>>> Closes: https://gitlab.freedesktop.org/drm/xe/kernel/-/issues/5557
>>> Cc: Summers Stuart <stuart.summers at intel.com>
>>> Fixes: 00c8efc3180f ("drm/xe: Add a shrinker for xe bos")
>>> Cc: <stable at vger.kernel.org> # v6.15+
>>> Signed-off-by: Thomas Hellström <thomas.hellstrom at linux.intel.com>
>>> ---
>>>    drivers/gpu/drm/xe/xe_shrinker.c | 51
>>> +++++++++++++++++++++++++++++---
>>>    1 file changed, 47 insertions(+), 4 deletions(-)
>>>
>>> diff --git a/drivers/gpu/drm/xe/xe_shrinker.c
>>> b/drivers/gpu/drm/xe/xe_shrinker.c
>>> index 1c3c04d52f55..90244fe59b59 100644
>>> --- a/drivers/gpu/drm/xe/xe_shrinker.c
>>> +++ b/drivers/gpu/drm/xe/xe_shrinker.c
>>> @@ -54,10 +54,10 @@ xe_shrinker_mod_pages(struct xe_shrinker
>>> *shrinker, long shrinkable, long purgea
>>>    	write_unlock(&shrinker->lock);
>>>    }
>>>    
>>> -static s64 xe_shrinker_walk(struct xe_device *xe,
>>> -			    struct ttm_operation_ctx *ctx,
>>> -			    const struct xe_bo_shrink_flags flags,
>>> -			    unsigned long to_scan, unsigned long
>>> *scanned)
>>> +static s64 __xe_shrinker_walk(struct xe_device *xe,
>>> +			      struct ttm_operation_ctx *ctx,
>>> +			      const struct xe_bo_shrink_flags
>>> flags,
>>> +			      unsigned long to_scan, unsigned long
>>> *scanned)
>>>    {
>>>    	unsigned int mem_type;
>>>    	s64 freed = 0, lret;
>>> @@ -93,6 +93,48 @@ static s64 xe_shrinker_walk(struct xe_device
>>> *xe,
>>>    	return freed;
>>>    }
>>>    
>>> +/*
>>> + * Try shrinking idle objects without writeback first, then if not
>>> sufficient,
>>> + * try also non-idle objects and finally if that's not sufficient
>>> either,
>>> + * add writeback. This avoids stalls and explicit writebacks with
>>> light or
>>> + * moderate memory pressure.
>>
>> Just one question here, with writeback=false it doesn't really
>> influence
>> which objects are chosen for shrinking, unlike with no_wait_gpu,
>> right?
>> Will having another pass just with writeback=true yield anything
>> different, assuming here that the previous two passes would have
>> already
>> hoovered ~everything up that was a possible candidate, so this pass
>> won't really find anything in practice? If so, does that also mean we
>> never really end up using the writeback=true behaviour any more?
> 
> Good point.
> 
> The assumption is that if allocating shmem backing-store fails during
> shrinking, we'd see an -ENOMEM and fail our target, and the next pass
> with writeback would help avoiding that.

Ah right, since on completion of writeback it would then allow giving 
back the folio/page straight away? So it's hopefully only in obscure 
cases where you might re-trigger ENOMEM on writeback=true, where setting 
writeback earlier could have helped?

> 
> Ofc that requires that a shmem_read_folio() from within reclaim returns
> an ERR_PTR(-ENOMEM) if the kernel reserves are depleted rather than to
> invoke the OOM killer. I should perhaps test that.
> 
> Other options would ofc be to include the writeback in pass 2, which
> would be similar to what the i915 shrinker does.
> 
> Thoughts?

Potentially merging with pass 2 sounds reasonable to me, if with that 
change it still helps the user. But not a blocker or anything, I was 
more just curious from my end.

> 
> Thanks,
> Thomas
> 
> 
> 
>>
>>> + */
>>> +static s64 xe_shrinker_walk(struct xe_device *xe,
>>> +			    struct ttm_operation_ctx *ctx,
>>> +			    const struct xe_bo_shrink_flags flags,
>>> +			    unsigned long to_scan, unsigned long
>>> *scanned)
>>> +{
>>> +	bool no_wait_gpu = true;
>>> +	struct xe_bo_shrink_flags save_flags = flags;
>>> +	s64 lret, freed;
>>> +
>>> +	swap(no_wait_gpu, ctx->no_wait_gpu);
>>> +	save_flags.writeback = false;
>>> +	lret = __xe_shrinker_walk(xe, ctx, save_flags, to_scan,
>>> scanned);
>>> +	swap(no_wait_gpu, ctx->no_wait_gpu);
>>> +	if (lret < 0 || *scanned >= to_scan)
>>> +		return lret;
>>> +
>>> +	freed = lret;
>>> +	if (!ctx->no_wait_gpu) {
>>> +		lret = __xe_shrinker_walk(xe, ctx, save_flags,
>>> to_scan, scanned);
>>> +		if (lret < 0)
>>> +			return lret;
>>> +		freed += lret;
>>> +		if (*scanned >= to_scan)
>>> +			return freed;
>>> +	}
>>> +
>>> +	if (flags.writeback) {
>>> +		lret = __xe_shrinker_walk(xe, ctx, flags, to_scan,
>>> scanned);
>>> +		if (lret < 0)
>>> +			return lret;
>>> +		freed += lret;
>>> +	}
>>> +
>>> +	return freed;
>>> +}
>>> +
>>>    static unsigned long
>>>    xe_shrinker_count(struct shrinker *shrink, struct shrink_control
>>> *sc)
>>>    {
>>> @@ -199,6 +241,7 @@ static unsigned long xe_shrinker_scan(struct
>>> shrinker *shrink, struct shrink_con
>>>    		runtime_pm = xe_shrinker_runtime_pm_get(shrinker,
>>> true, 0, can_backup);
>>>    
>>>    	shrink_flags.purge = false;
>>> +
>>>    	lret = xe_shrinker_walk(shrinker->xe, &ctx, shrink_flags,
>>>    				nr_to_scan, &nr_scanned);
>>>    	if (lret >= 0)
>>
> 



More information about the Intel-xe mailing list