[PATCH v3] drm/xe/xe2: Use XE_CACHE_WB pat index
Ghimiray, Himal Prasad
himal.prasad.ghimiray at intel.com
Fri Jan 19 03:49:39 UTC 2024
> -----Original Message-----
> From: Thomas Hellström <thomas.hellstrom at linux.intel.com>
> Sent: 18 January 2024 21:36
> To: Ghimiray, Himal Prasad <himal.prasad.ghimiray at intel.com>; intel-
> xe at lists.freedesktop.org
> Cc: Roper, Matthew D <matthew.d.roper at intel.com>
> Subject: Re: [PATCH v3] drm/xe/xe2: Use XE_CACHE_WB pat index
>
> On Tue, 2024-01-16 at 17:35 +0530, Himal Prasad Ghimiray wrote:
> > The pat table entry associated with XE_CACHE_WB is coherent whereas
> > XE_CACHE_NONE is non coherent. Use the coherent entry XE_CACHE_WB
> for
> > buffers not supporting compression.
>
> "since that is what the migrate code currently expects"?
That is right. Will update it in commit message too.
>
> > For read/write to flat ccs region
> > the issue is not related to coherency with cpu. The hardware expects
> > the pat index associated with GPUVA for indirect access to be
> > compression enabled hence use XE_CACHE_NONE_COMPRESSION.
> >
> > Fixes the broken "xe_migrate_test kunit test" on LNL.
> >
> > v2
> > - Fix the argument to emit_pte, pass the bool directly. (Thomas)
> >
> > v3
> > - Rebase
> > - Update commit message (Matt)
> >
> > Cc: Matt Roper <matthew.d.roper at intel.com>
> > Cc: Thomas Hellström <thomas.hellstrom at linux.intel.com>
>
> There needs to be a Fixes: tag here. Please use "dim fixes" to get it
> right:
>
> https://drm.pages.freedesktop.org/maintainer-tools/dim.html
Sure.
>
>
> > Signed-off-by: Himal Prasad Ghimiray
> > <himal.prasad.ghimiray at intel.com>
>
> Otherwise LGTM.
>
> /Thomas
>
>
> > ---
> > drivers/gpu/drm/xe/xe_migrate.c | 14 +++++++-------
> > 1 file changed, 7 insertions(+), 7 deletions(-)
> >
> > diff --git a/drivers/gpu/drm/xe/xe_migrate.c
> > b/drivers/gpu/drm/xe/xe_migrate.c index 44725f978f3e..5de8ee4245c3
> > 100644
> > --- a/drivers/gpu/drm/xe/xe_migrate.c
> > +++ b/drivers/gpu/drm/xe/xe_migrate.c
> > @@ -481,7 +481,7 @@ static void emit_pte(struct xe_migrate *m,
> > /* Indirect access needs compression enabled uncached PAT index */
> > if (GRAPHICS_VERx100(xe) >= 2000)
> > pat_index = is_comp_pte ? xe-
> > >pat.idx[XE_CACHE_NONE_COMPRESSION] :
> > - xe-
> > >pat.idx[XE_CACHE_NONE];
> > + xe->pat.idx[XE_CACHE_WB];
> > else
> > pat_index = xe->pat.idx[XE_CACHE_WB];
> >
> > @@ -769,14 +769,14 @@ struct dma_fence *xe_migrate_copy(struct
> > xe_migrate *m,
> > if (src_is_vram && xe_migrate_allow_identity(src_L0,
> > &src_it))
> > xe_res_next(&src_it, src_L0);
> > else
> > - emit_pte(m, bb, src_L0_pt, src_is_vram,
> > true, &src_it, src_L0,
> > - src);
> > + emit_pte(m, bb, src_L0_pt, src_is_vram,
> > copy_system_ccs,
> > + &src_it, src_L0, src);
> >
> > if (dst_is_vram && xe_migrate_allow_identity(src_L0,
> > &dst_it))
> > xe_res_next(&dst_it, src_L0);
> > else
> > - emit_pte(m, bb, dst_L0_pt, dst_is_vram,
> > true, &dst_it, src_L0,
> > - dst);
> > + emit_pte(m, bb, dst_L0_pt, dst_is_vram,
> > copy_system_ccs,
> > + &dst_it, src_L0, dst);
> >
> > if (copy_system_ccs)
> > emit_pte(m, bb, ccs_pt, false, false, &ccs_it, ccs_size,
> src); @@
> > -1018,8 +1018,8 @@ struct dma_fence *xe_migrate_clear(struct
> > xe_migrate *m,
> > if (clear_vram &&
> > xe_migrate_allow_identity(clear_L0, &src_it))
> > xe_res_next(&src_it, clear_L0);
> > else
> > - emit_pte(m, bb, clear_L0_pt, clear_vram,
> > true, &src_it, clear_L0,
> > - dst);
> > + emit_pte(m, bb, clear_L0_pt, clear_vram,
> > clear_system_ccs,
> > + &src_it, clear_L0, dst);
> >
> > bb->cs[bb->len++] = MI_BATCH_BUFFER_END;
> > update_idx = bb->len;
More information about the Intel-xe
mailing list