[PATCH v2] drm/xe: Defer buffer object shrinker write-backs and GPU waits

Thomas Hellström thomas.hellstrom at linux.intel.com
Tue Aug 5 16:27:43 UTC 2025


On Tue, 2025-08-05 at 14:40 +0100, Matthew Auld wrote:
> On 05/08/2025 08:48, Thomas Hellström wrote:
> > When the xe buffer-object shrinker allows GPU waits and write-back,
> > (typically from kswapd), perform multiple passes, skipping
> > subsequent passes if the shrinker number of scanned objects target
> > is reached.
> > 
> > 1) Without GPU waits and write-back
> > 2) Without write-back
> > 3) With both GPU-waits and write-back
> > 
> > This is to avoid stalls and costly write- and readbacks unless they
> > are really necessary.
> > 
> > v2:
> > - Don't test for scan completion twice. (Stuart Summers)
> > - Update tags.
> > 
> > Reported-by: melvyn <melvyn2 at dnsense.pub>
> > Closes: https://gitlab.freedesktop.org/drm/xe/kernel/-/issues/5557
> > Cc: Summers Stuart <stuart.summers at intel.com>
> > Fixes: 00c8efc3180f ("drm/xe: Add a shrinker for xe bos")
> > Cc: <stable at vger.kernel.org> # v6.15+
> > Signed-off-by: Thomas Hellström <thomas.hellstrom at linux.intel.com>
> > ---
> >   drivers/gpu/drm/xe/xe_shrinker.c | 51
> > +++++++++++++++++++++++++++++---
> >   1 file changed, 47 insertions(+), 4 deletions(-)
> > 
> > diff --git a/drivers/gpu/drm/xe/xe_shrinker.c
> > b/drivers/gpu/drm/xe/xe_shrinker.c
> > index 1c3c04d52f55..90244fe59b59 100644
> > --- a/drivers/gpu/drm/xe/xe_shrinker.c
> > +++ b/drivers/gpu/drm/xe/xe_shrinker.c
> > @@ -54,10 +54,10 @@ xe_shrinker_mod_pages(struct xe_shrinker
> > *shrinker, long shrinkable, long purgea
> >   	write_unlock(&shrinker->lock);
> >   }
> >   
> > -static s64 xe_shrinker_walk(struct xe_device *xe,
> > -			    struct ttm_operation_ctx *ctx,
> > -			    const struct xe_bo_shrink_flags flags,
> > -			    unsigned long to_scan, unsigned long
> > *scanned)
> > +static s64 __xe_shrinker_walk(struct xe_device *xe,
> > +			      struct ttm_operation_ctx *ctx,
> > +			      const struct xe_bo_shrink_flags
> > flags,
> > +			      unsigned long to_scan, unsigned long
> > *scanned)
> >   {
> >   	unsigned int mem_type;
> >   	s64 freed = 0, lret;
> > @@ -93,6 +93,48 @@ static s64 xe_shrinker_walk(struct xe_device
> > *xe,
> >   	return freed;
> >   }
> >   
> > +/*
> > + * Try shrinking idle objects without writeback first, then if not
> > sufficient,
> > + * try also non-idle objects and finally if that's not sufficient
> > either,
> > + * add writeback. This avoids stalls and explicit writebacks with
> > light or
> > + * moderate memory pressure.
> 
> Just one question here, with writeback=false it doesn't really
> influence 
> which objects are chosen for shrinking, unlike with no_wait_gpu,
> right? 
> Will having another pass just with writeback=true yield anything 
> different, assuming here that the previous two passes would have
> already 
> hoovered ~everything up that was a possible candidate, so this pass 
> won't really find anything in practice? If so, does that also mean we
> never really end up using the writeback=true behaviour any more?

Good point.

The assumption is that if allocating shmem backing-store fails during
shrinking, we'd see an -ENOMEM and fail our target, and the next pass
with writeback would help avoiding that.

Ofc that requires that a shmem_read_folio() from within reclaim returns
an ERR_PTR(-ENOMEM) if the kernel reserves are depleted rather than to
invoke the OOM killer. I should perhaps test that.

Other options would ofc be to include the writeback in pass 2, which
would be similar to what the i915 shrinker does.

Thoughts?

Thanks,
Thomas



> 
> > + */
> > +static s64 xe_shrinker_walk(struct xe_device *xe,
> > +			    struct ttm_operation_ctx *ctx,
> > +			    const struct xe_bo_shrink_flags flags,
> > +			    unsigned long to_scan, unsigned long
> > *scanned)
> > +{
> > +	bool no_wait_gpu = true;
> > +	struct xe_bo_shrink_flags save_flags = flags;
> > +	s64 lret, freed;
> > +
> > +	swap(no_wait_gpu, ctx->no_wait_gpu);
> > +	save_flags.writeback = false;
> > +	lret = __xe_shrinker_walk(xe, ctx, save_flags, to_scan,
> > scanned);
> > +	swap(no_wait_gpu, ctx->no_wait_gpu);
> > +	if (lret < 0 || *scanned >= to_scan)
> > +		return lret;
> > +
> > +	freed = lret;
> > +	if (!ctx->no_wait_gpu) {
> > +		lret = __xe_shrinker_walk(xe, ctx, save_flags,
> > to_scan, scanned);
> > +		if (lret < 0)
> > +			return lret;
> > +		freed += lret;
> > +		if (*scanned >= to_scan)
> > +			return freed;
> > +	}
> > +
> > +	if (flags.writeback) {
> > +		lret = __xe_shrinker_walk(xe, ctx, flags, to_scan,
> > scanned);
> > +		if (lret < 0)
> > +			return lret;
> > +		freed += lret;
> > +	}
> > +
> > +	return freed;
> > +}
> > +
> >   static unsigned long
> >   xe_shrinker_count(struct shrinker *shrink, struct shrink_control
> > *sc)
> >   {
> > @@ -199,6 +241,7 @@ static unsigned long xe_shrinker_scan(struct
> > shrinker *shrink, struct shrink_con
> >   		runtime_pm = xe_shrinker_runtime_pm_get(shrinker,
> > true, 0, can_backup);
> >   
> >   	shrink_flags.purge = false;
> > +
> >   	lret = xe_shrinker_walk(shrinker->xe, &ctx, shrink_flags,
> >   				nr_to_scan, &nr_scanned);
> >   	if (lret >= 0)
> 



More information about the Intel-xe mailing list