[PATCH 3/4] drm/xe: s/tlb_invalidation/tlb_inval

Summers, Stuart stuart.summers at intel.com
Tue Jul 15 21:24:28 UTC 2025


On Tue, 2025-07-15 at 21:21 +0000, stuartsummers wrote:
> From: Matthew Brost <matthew.brost at intel.com>
> 
> tlb_invalidation is a bit verbose leading to ugly wraps in the code,
> shorten to tlb_inval.
> 
> Signed-off-by: Matthew Brost <matthew.brost at intel.com>
> Signed-off-by: Stuart Summers <stuart.summers at intel.com>

So I did touch this one slightly as part of a rebase, particularly the
pieces around the more recent LMTT changes, but generally I agree with
the changes from Matt that it certainly makes things a little more
manageable and seems obvious with the shortened name anyway:
Reviewed-by: Stuart Summers <stuart.summers at intel.com>

Thanks,
Stuart

> ---
>  drivers/gpu/drm/xe/Makefile                   |   2 +-
>  drivers/gpu/drm/xe/xe_device_types.h          |   4 +-
>  drivers/gpu/drm/xe/xe_ggtt.c                  |   6 +-
>  drivers/gpu/drm/xe/xe_gt.c                    |   8 +-
>  drivers/gpu/drm/xe/xe_gt_pagefault.c          |   1 -
>  ...t_tlb_invalidation.c => xe_gt_tlb_inval.c} | 239 +++++++++-------
> --
>  drivers/gpu/drm/xe/xe_gt_tlb_inval.h          |  40 +++
>  ...dation_types.h => xe_gt_tlb_inval_types.h} |  14 +-
>  drivers/gpu/drm/xe/xe_gt_tlb_invalidation.h   |  40 ---
>  drivers/gpu/drm/xe/xe_gt_types.h              |  18 +-
>  drivers/gpu/drm/xe/xe_guc_ct.c                |   8 +-
>  drivers/gpu/drm/xe/xe_lmtt.c                  |  12 +-
>  drivers/gpu/drm/xe/xe_pci.c                   |   6 +-
>  drivers/gpu/drm/xe/xe_pci_types.h             |   2 +-
>  drivers/gpu/drm/xe/xe_pt.c                    |  55 ++--
>  drivers/gpu/drm/xe/xe_svm.c                   |   4 +-
>  drivers/gpu/drm/xe/xe_trace.h                 |  24 +-
>  drivers/gpu/drm/xe/xe_vm.c                    |  64 +++--
>  drivers/gpu/drm/xe/xe_vm.h                    |   4 +-
>  19 files changed, 270 insertions(+), 281 deletions(-)
>  rename drivers/gpu/drm/xe/{xe_gt_tlb_invalidation.c =>
> xe_gt_tlb_inval.c} (62%)
>  create mode 100644 drivers/gpu/drm/xe/xe_gt_tlb_inval.h
>  rename drivers/gpu/drm/xe/{xe_gt_tlb_invalidation_types.h =>
> xe_gt_tlb_inval_types.h} (55%)
>  delete mode 100644 drivers/gpu/drm/xe/xe_gt_tlb_invalidation.h
> 
> diff --git a/drivers/gpu/drm/xe/Makefile
> b/drivers/gpu/drm/xe/Makefile
> index 07c71a29963d..8ad427f0d6fe 100644
> --- a/drivers/gpu/drm/xe/Makefile
> +++ b/drivers/gpu/drm/xe/Makefile
> @@ -60,7 +60,7 @@ xe-y += xe_bb.o \
>         xe_gt_pagefault.o \
>         xe_gt_sysfs.o \
>         xe_gt_throttle.o \
> -       xe_gt_tlb_invalidation.o \
> +       xe_gt_tlb_inval.o \
>         xe_gt_topology.o \
>         xe_guc.o \
>         xe_guc_ads.o \
> diff --git a/drivers/gpu/drm/xe/xe_device_types.h
> b/drivers/gpu/drm/xe/xe_device_types.h
> index d4d2c6854790..7fba91a5146e 100644
> --- a/drivers/gpu/drm/xe/xe_device_types.h
> +++ b/drivers/gpu/drm/xe/xe_device_types.h
> @@ -336,8 +336,8 @@ struct xe_device {
>                 u8 has_mbx_power_limits:1;
>                 /** @info.has_pxp: Device has PXP support */
>                 u8 has_pxp:1;
> -               /** @info.has_range_tlb_invalidation: Has range based
> TLB invalidations */
> -               u8 has_range_tlb_invalidation:1;
> +               /** @info.has_range_tlb_inval: Has range based TLB
> invalidations */
> +               u8 has_range_tlb_inval:1;
>                 /** @info.has_sriov: Supports SR-IOV */
>                 u8 has_sriov:1;
>                 /** @info.has_usm: Device has unified shared memory
> support */
> diff --git a/drivers/gpu/drm/xe/xe_ggtt.c
> b/drivers/gpu/drm/xe/xe_ggtt.c
> index 29d4d3f51da1..9a06d68946cf 100644
> --- a/drivers/gpu/drm/xe/xe_ggtt.c
> +++ b/drivers/gpu/drm/xe/xe_ggtt.c
> @@ -23,7 +23,7 @@
>  #include "xe_device.h"
>  #include "xe_gt.h"
>  #include "xe_gt_printk.h"
> -#include "xe_gt_tlb_invalidation.h"
> +#include "xe_gt_tlb_inval.h"
>  #include "xe_map.h"
>  #include "xe_mmio.h"
>  #include "xe_pm.h"
> @@ -438,9 +438,9 @@ static void ggtt_invalidate_gt_tlb(struct xe_gt
> *gt)
>         if (!gt)
>                 return;
>  
> -       err = xe_gt_tlb_invalidation_ggtt(gt);
> +       err = xe_gt_tlb_inval_ggtt(gt);
>         if (err)
> -               drm_warn(&gt_to_xe(gt)->drm,
> "xe_gt_tlb_invalidation_ggtt error=%d", err);
> +               drm_warn(&gt_to_xe(gt)->drm, "xe_gt_tlb_inval_ggtt
> error=%d", err);
>  }
>  
>  static void xe_ggtt_invalidate(struct xe_ggtt *ggtt)
> diff --git a/drivers/gpu/drm/xe/xe_gt.c b/drivers/gpu/drm/xe/xe_gt.c
> index c8eda36546d3..a7048e7c7177 100644
> --- a/drivers/gpu/drm/xe/xe_gt.c
> +++ b/drivers/gpu/drm/xe/xe_gt.c
> @@ -37,7 +37,7 @@
>  #include "xe_gt_sriov_pf.h"
>  #include "xe_gt_sriov_vf.h"
>  #include "xe_gt_sysfs.h"
> -#include "xe_gt_tlb_invalidation.h"
> +#include "xe_gt_tlb_inval.h"
>  #include "xe_gt_topology.h"
>  #include "xe_guc_exec_queue_types.h"
>  #include "xe_guc_pc.h"
> @@ -412,7 +412,7 @@ int xe_gt_init_early(struct xe_gt *gt)
>         xe_force_wake_init_gt(gt, gt_to_fw(gt));
>         spin_lock_init(&gt->global_invl_lock);
>  
> -       err = xe_gt_tlb_invalidation_init_early(gt);
> +       err = xe_gt_tlb_inval_init_early(gt);
>         if (err)
>                 return err;
>  
> @@ -842,7 +842,7 @@ static int gt_reset(struct xe_gt *gt)
>  
>         xe_uc_stop(&gt->uc);
>  
> -       xe_gt_tlb_invalidation_reset(gt);
> +       xe_gt_tlb_inval_reset(gt);
>  
>         err = do_gt_reset(gt);
>         if (err)
> @@ -1056,5 +1056,5 @@ void xe_gt_declare_wedged(struct xe_gt *gt)
>         xe_gt_assert(gt, gt_to_xe(gt)->wedged.mode);
>  
>         xe_uc_declare_wedged(&gt->uc);
> -       xe_gt_tlb_invalidation_reset(gt);
> +       xe_gt_tlb_inval_reset(gt);
>  }
> diff --git a/drivers/gpu/drm/xe/xe_gt_pagefault.c
> b/drivers/gpu/drm/xe/xe_gt_pagefault.c
> index 5a75d56d8558..6a24e1eaafa8 100644
> --- a/drivers/gpu/drm/xe/xe_gt_pagefault.c
> +++ b/drivers/gpu/drm/xe/xe_gt_pagefault.c
> @@ -16,7 +16,6 @@
>  #include "xe_gt.h"
>  #include "xe_gt_printk.h"
>  #include "xe_gt_stats.h"
> -#include "xe_gt_tlb_invalidation.h"
>  #include "xe_guc.h"
>  #include "xe_guc_ct.h"
>  #include "xe_migrate.h"
> diff --git a/drivers/gpu/drm/xe/xe_gt_tlb_invalidation.c
> b/drivers/gpu/drm/xe/xe_gt_tlb_inval.c
> similarity index 62%
> rename from drivers/gpu/drm/xe/xe_gt_tlb_invalidation.c
> rename to drivers/gpu/drm/xe/xe_gt_tlb_inval.c
> index 086c12ee3d9d..7ffa9488f947 100644
> --- a/drivers/gpu/drm/xe/xe_gt_tlb_invalidation.c
> +++ b/drivers/gpu/drm/xe/xe_gt_tlb_inval.c
> @@ -3,7 +3,7 @@
>   * Copyright © 2023 Intel Corporation
>   */
>  
> -#include "xe_gt_tlb_invalidation.h"
> +#include "xe_gt_tlb_inval.h"
>  
>  #include "abi/guc_actions_abi.h"
>  #include "xe_device.h"
> @@ -37,7 +37,7 @@ static long tlb_timeout_jiffies(struct xe_gt *gt)
>         return hw_tlb_timeout + 2 * delay;
>  }
>  
> -static void xe_gt_tlb_invalidation_fence_fini(struct
> xe_gt_tlb_invalidation_fence *fence)
> +static void xe_gt_tlb_inval_fence_fini(struct xe_gt_tlb_inval_fence
> *fence)
>  {
>         if (WARN_ON_ONCE(!fence->gt))
>                 return;
> @@ -47,66 +47,66 @@ static void
> xe_gt_tlb_invalidation_fence_fini(struct xe_gt_tlb_invalidation_fenc
>  }
>  
>  static void
> -__invalidation_fence_signal(struct xe_device *xe, struct
> xe_gt_tlb_invalidation_fence *fence)
> +__inval_fence_signal(struct xe_device *xe, struct
> xe_gt_tlb_inval_fence *fence)
>  {
>         bool stack = test_bit(FENCE_STACK_BIT, &fence->base.flags);
>  
> -       trace_xe_gt_tlb_invalidation_fence_signal(xe, fence);
> -       xe_gt_tlb_invalidation_fence_fini(fence);
> +       trace_xe_gt_tlb_inval_fence_signal(xe, fence);
> +       xe_gt_tlb_inval_fence_fini(fence);
>         dma_fence_signal(&fence->base);
>         if (!stack)
>                 dma_fence_put(&fence->base);
>  }
>  
>  static void
> -invalidation_fence_signal(struct xe_device *xe, struct
> xe_gt_tlb_invalidation_fence *fence)
> +inval_fence_signal(struct xe_device *xe, struct
> xe_gt_tlb_inval_fence *fence)
>  {
>         list_del(&fence->link);
> -       __invalidation_fence_signal(xe, fence);
> +       __inval_fence_signal(xe, fence);
>  }
>  
> -void xe_gt_tlb_invalidation_fence_signal(struct
> xe_gt_tlb_invalidation_fence *fence)
> +void xe_gt_tlb_inval_fence_signal(struct xe_gt_tlb_inval_fence
> *fence)
>  {
>         if (WARN_ON_ONCE(!fence->gt))
>                 return;
>  
> -       __invalidation_fence_signal(gt_to_xe(fence->gt), fence);
> +       __inval_fence_signal(gt_to_xe(fence->gt), fence);
>  }
>  
>  static void xe_gt_tlb_fence_timeout(struct work_struct *work)
>  {
>         struct xe_gt *gt = container_of(work, struct xe_gt,
> -
>                                        tlb_invalidation.fence_tdr.work
> );
> +                                       tlb_inval.fence_tdr.work);
>         struct xe_device *xe = gt_to_xe(gt);
> -       struct xe_gt_tlb_invalidation_fence *fence, *next;
> +       struct xe_gt_tlb_inval_fence *fence, *next;
>  
>         LNL_FLUSH_WORK(&gt->uc.guc.ct.g2h_worker);
>  
> -       spin_lock_irq(&gt->tlb_invalidation.pending_lock);
> +       spin_lock_irq(&gt->tlb_inval.pending_lock);
>         list_for_each_entry_safe(fence, next,
> -                                &gt-
> >tlb_invalidation.pending_fences, link) {
> +                                &gt->tlb_inval.pending_fences, link)
> {
>                 s64 since_inval_ms = ktime_ms_delta(ktime_get(),
> -                                                   fence-
> >invalidation_time);
> +                                                   fence-
> >inval_time);
>  
>                 if (msecs_to_jiffies(since_inval_ms) <
> tlb_timeout_jiffies(gt))
>                         break;
>  
> -               trace_xe_gt_tlb_invalidation_fence_timeout(xe,
> fence);
> +               trace_xe_gt_tlb_inval_fence_timeout(xe, fence);
>                 xe_gt_err(gt, "TLB invalidation fence timeout,
> seqno=%d recv=%d",
> -                         fence->seqno, gt-
> >tlb_invalidation.seqno_recv);
> +                         fence->seqno, gt->tlb_inval.seqno_recv);
>  
>                 fence->base.error = -ETIME;
> -               invalidation_fence_signal(xe, fence);
> +               inval_fence_signal(xe, fence);
>         }
> -       if (!list_empty(&gt->tlb_invalidation.pending_fences))
> +       if (!list_empty(&gt->tlb_inval.pending_fences))
>                 queue_delayed_work(system_wq,
> -                                  &gt->tlb_invalidation.fence_tdr,
> +                                  &gt->tlb_inval.fence_tdr,
>                                    tlb_timeout_jiffies(gt));
> -       spin_unlock_irq(&gt->tlb_invalidation.pending_lock);
> +       spin_unlock_irq(&gt->tlb_inval.pending_lock);
>  }
>  
>  /**
> - * xe_gt_tlb_invalidation_init_early - Initialize GT TLB
> invalidation state
> + * xe_gt_tlb_inval_init_early - Initialize GT TLB invalidation state
>   * @gt: GT structure
>   *
>   * Initialize GT TLB invalidation state, purely software
> initialization, should
> @@ -114,27 +114,27 @@ static void xe_gt_tlb_fence_timeout(struct
> work_struct *work)
>   *
>   * Return: 0 on success, negative error code on error.
>   */
> -int xe_gt_tlb_invalidation_init_early(struct xe_gt *gt)
> +int xe_gt_tlb_inval_init_early(struct xe_gt *gt)
>  {
> -       gt->tlb_invalidation.seqno = 1;
> -       INIT_LIST_HEAD(&gt->tlb_invalidation.pending_fences);
> -       spin_lock_init(&gt->tlb_invalidation.pending_lock);
> -       spin_lock_init(&gt->tlb_invalidation.lock);
> -       INIT_DELAYED_WORK(&gt->tlb_invalidation.fence_tdr,
> +       gt->tlb_inval.seqno = 1;
> +       INIT_LIST_HEAD(&gt->tlb_inval.pending_fences);
> +       spin_lock_init(&gt->tlb_inval.pending_lock);
> +       spin_lock_init(&gt->tlb_inval.lock);
> +       INIT_DELAYED_WORK(&gt->tlb_inval.fence_tdr,
>                           xe_gt_tlb_fence_timeout);
>  
>         return 0;
>  }
>  
>  /**
> - * xe_gt_tlb_invalidation_reset - Initialize GT TLB invalidation
> reset
> + * xe_gt_tlb_inval_reset - Initialize GT TLB invalidation reset
>   * @gt: GT structure
>   *
>   * Signal any pending invalidation fences, should be called during a
> GT reset
>   */
> -void xe_gt_tlb_invalidation_reset(struct xe_gt *gt)
> +void xe_gt_tlb_inval_reset(struct xe_gt *gt)
>  {
> -       struct xe_gt_tlb_invalidation_fence *fence, *next;
> +       struct xe_gt_tlb_inval_fence *fence, *next;
>         int pending_seqno;
>  
>         /*
> @@ -151,8 +151,8 @@ void xe_gt_tlb_invalidation_reset(struct xe_gt
> *gt)
>          */
>  
>         mutex_lock(&gt->uc.guc.ct.lock);
> -       spin_lock_irq(&gt->tlb_invalidation.pending_lock);
> -       cancel_delayed_work(&gt->tlb_invalidation.fence_tdr);
> +       spin_lock_irq(&gt->tlb_inval.pending_lock);
> +       cancel_delayed_work(&gt->tlb_inval.fence_tdr);
>         /*
>          * We might have various kworkers waiting for TLB flushes to
> complete
>          * which are not tracked with an explicit TLB fence, however
> at this
> @@ -160,22 +160,22 @@ void xe_gt_tlb_invalidation_reset(struct xe_gt
> *gt)
>          * make sure we signal them here under the assumption that we
> have
>          * completed a full GT reset.
>          */
> -       if (gt->tlb_invalidation.seqno == 1)
> +       if (gt->tlb_inval.seqno == 1)
>                 pending_seqno = TLB_INVALIDATION_SEQNO_MAX - 1;
>         else
> -               pending_seqno = gt->tlb_invalidation.seqno - 1;
> -       WRITE_ONCE(gt->tlb_invalidation.seqno_recv, pending_seqno);
> +               pending_seqno = gt->tlb_inval.seqno - 1;
> +       WRITE_ONCE(gt->tlb_inval.seqno_recv, pending_seqno);
>  
>         list_for_each_entry_safe(fence, next,
> -                                &gt-
> >tlb_invalidation.pending_fences, link)
> -               invalidation_fence_signal(gt_to_xe(gt), fence);
> -       spin_unlock_irq(&gt->tlb_invalidation.pending_lock);
> +                                &gt->tlb_inval.pending_fences, link)
> +               inval_fence_signal(gt_to_xe(gt), fence);
> +       spin_unlock_irq(&gt->tlb_inval.pending_lock);
>         mutex_unlock(&gt->uc.guc.ct.lock);
>  }
>  
> -static bool tlb_invalidation_seqno_past(struct xe_gt *gt, int seqno)
> +static bool tlb_inval_seqno_past(struct xe_gt *gt, int seqno)
>  {
> -       int seqno_recv = READ_ONCE(gt->tlb_invalidation.seqno_recv);
> +       int seqno_recv = READ_ONCE(gt->tlb_inval.seqno_recv);
>  
>         if (seqno - seqno_recv < -(TLB_INVALIDATION_SEQNO_MAX / 2))
>                 return false;
> @@ -186,9 +186,9 @@ static bool tlb_invalidation_seqno_past(struct
> xe_gt *gt, int seqno)
>         return seqno_recv >= seqno;
>  }
>  
> -static int send_tlb_invalidation(struct xe_guc *guc,
> -                                struct xe_gt_tlb_invalidation_fence
> *fence,
> -                                u32 *action, int len)
> +static int send_tlb_inval(struct xe_guc *guc,
> +                         struct xe_gt_tlb_inval_fence *fence,
> +                         u32 *action, int len)
>  {
>         struct xe_gt *gt = guc_to_gt(guc);
>         struct xe_device *xe = gt_to_xe(gt);
> @@ -204,41 +204,41 @@ static int send_tlb_invalidation(struct xe_guc
> *guc,
>          */
>  
>         mutex_lock(&guc->ct.lock);
> -       seqno = gt->tlb_invalidation.seqno;
> +       seqno = gt->tlb_inval.seqno;
>         fence->seqno = seqno;
> -       trace_xe_gt_tlb_invalidation_fence_send(xe, fence);
> +       trace_xe_gt_tlb_inval_fence_send(xe, fence);
>         action[1] = seqno;
>         ret = xe_guc_ct_send_locked(&guc->ct, action, len,
>                                     G2H_LEN_DW_TLB_INVALIDATE, 1);
>         if (!ret) {
> -               spin_lock_irq(&gt->tlb_invalidation.pending_lock);
> +               spin_lock_irq(&gt->tlb_inval.pending_lock);
>                 /*
>                  * We haven't actually published the TLB fence as per
>                  * pending_fences, but in theory our seqno could have
> already
>                  * been written as we acquired the pending_lock. In
> such a case
>                  * we can just go ahead and signal the fence here.
>                  */
> -               if (tlb_invalidation_seqno_past(gt, seqno)) {
> -                       __invalidation_fence_signal(xe, fence);
> +               if (tlb_inval_seqno_past(gt, seqno)) {
> +                       __inval_fence_signal(xe, fence);
>                 } else {
> -                       fence->invalidation_time = ktime_get();
> +                       fence->inval_time = ktime_get();
>                         list_add_tail(&fence->link,
> -                                     &gt-
> >tlb_invalidation.pending_fences);
> +                                     &gt->tlb_inval.pending_fences);
>  
> -                       if (list_is_singular(&gt-
> >tlb_invalidation.pending_fences))
> +                       if (list_is_singular(&gt-
> >tlb_inval.pending_fences))
>                                 queue_delayed_work(system_wq,
> -                                                  &gt-
> >tlb_invalidation.fence_tdr,
> +                                                  &gt-
> >tlb_inval.fence_tdr,
>                                                   
> tlb_timeout_jiffies(gt));
>                 }
> -               spin_unlock_irq(&gt->tlb_invalidation.pending_lock);
> +               spin_unlock_irq(&gt->tlb_inval.pending_lock);
>         } else {
> -               __invalidation_fence_signal(xe, fence);
> +               __inval_fence_signal(xe, fence);
>         }
>         if (!ret) {
> -               gt->tlb_invalidation.seqno = (gt-
> >tlb_invalidation.seqno + 1) %
> +               gt->tlb_inval.seqno = (gt->tlb_inval.seqno + 1) %
>                         TLB_INVALIDATION_SEQNO_MAX;
> -               if (!gt->tlb_invalidation.seqno)
> -                       gt->tlb_invalidation.seqno = 1;
> +               if (!gt->tlb_inval.seqno)
> +                       gt->tlb_inval.seqno = 1;
>         }
>         mutex_unlock(&guc->ct.lock);
>         xe_gt_stats_incr(gt, XE_GT_STATS_ID_TLB_INVAL, 1);
> @@ -251,7 +251,7 @@ static int send_tlb_invalidation(struct xe_guc
> *guc,
>                 XE_GUC_TLB_INVAL_FLUSH_CACHE)
>  
>  /**
> - * xe_gt_tlb_invalidation_guc - Issue a TLB invalidation on this GT
> for the GuC
> + * xe_gt_tlb_inval_guc - Issue a TLB invalidation on this GT for the
> GuC
>   * @gt: GT structure
>   * @fence: invalidation fence which will be signal on TLB
> invalidation
>   * completion
> @@ -261,18 +261,17 @@ static int send_tlb_invalidation(struct xe_guc
> *guc,
>   *
>   * Return: 0 on success, negative error code on error
>   */
> -static int xe_gt_tlb_invalidation_guc(struct xe_gt *gt,
> -                                     struct
> xe_gt_tlb_invalidation_fence *fence)
> +static int xe_gt_tlb_inval_guc(struct xe_gt *gt,
> +                              struct xe_gt_tlb_inval_fence *fence)
>  {
>         u32 action[] = {
>                 XE_GUC_ACTION_TLB_INVALIDATION,
> -               0,  /* seqno, replaced in send_tlb_invalidation */
> +               0,  /* seqno, replaced in send_tlb_inval */
>                 MAKE_INVAL_OP(XE_GUC_TLB_INVAL_GUC),
>         };
>         int ret;
>  
> -       ret = send_tlb_invalidation(&gt->uc.guc, fence, action,
> -                                   ARRAY_SIZE(action));
> +       ret = send_tlb_inval(&gt->uc.guc, fence, action,
> ARRAY_SIZE(action));
>         /*
>          * -ECANCELED indicates the CT is stopped for a GT reset. TLB
> caches
>          *  should be nuked on a GT reset so this error can be
> ignored.
> @@ -284,7 +283,7 @@ static int xe_gt_tlb_invalidation_guc(struct
> xe_gt *gt,
>  }
>  
>  /**
> - * xe_gt_tlb_invalidation_ggtt - Issue a TLB invalidation on this GT
> for the GGTT
> + * xe_gt_tlb_inval_ggtt - Issue a TLB invalidation on this GT for
> the GGTT
>   * @gt: GT structure
>   *
>   * Issue a TLB invalidation for the GGTT. Completion of TLB
> invalidation is
> @@ -292,22 +291,22 @@ static int xe_gt_tlb_invalidation_guc(struct
> xe_gt *gt,
>   *
>   * Return: 0 on success, negative error code on error
>   */
> -int xe_gt_tlb_invalidation_ggtt(struct xe_gt *gt)
> +int xe_gt_tlb_inval_ggtt(struct xe_gt *gt)
>  {
>         struct xe_device *xe = gt_to_xe(gt);
>         unsigned int fw_ref;
>  
>         if (xe_guc_ct_enabled(&gt->uc.guc.ct) &&
>             gt->uc.guc.submission_state.enabled) {
> -               struct xe_gt_tlb_invalidation_fence fence;
> +               struct xe_gt_tlb_inval_fence fence;
>                 int ret;
>  
> -               xe_gt_tlb_invalidation_fence_init(gt, &fence, true);
> -               ret = xe_gt_tlb_invalidation_guc(gt, &fence);
> +               xe_gt_tlb_inval_fence_init(gt, &fence, true);
> +               ret = xe_gt_tlb_inval_guc(gt, &fence);
>                 if (ret)
>                         return ret;
>  
> -               xe_gt_tlb_invalidation_fence_wait(&fence);
> +               xe_gt_tlb_inval_fence_wait(&fence);
>         } else if (xe_device_uc_enabled(xe) && !xe_device_wedged(xe))
> {
>                 struct xe_mmio *mmio = &gt->mmio;
>  
> @@ -330,34 +329,34 @@ int xe_gt_tlb_invalidation_ggtt(struct xe_gt
> *gt)
>         return 0;
>  }
>  
> -static int send_tlb_invalidation_all(struct xe_gt *gt,
> -                                    struct
> xe_gt_tlb_invalidation_fence *fence)
> +static int send_tlb_inval_all(struct xe_gt *gt,
> +                             struct xe_gt_tlb_inval_fence *fence)
>  {
>         u32 action[] = {
>                 XE_GUC_ACTION_TLB_INVALIDATION_ALL,
> -               0,  /* seqno, replaced in send_tlb_invalidation */
> +               0,  /* seqno, replaced in send_tlb_inval */
>                 MAKE_INVAL_OP(XE_GUC_TLB_INVAL_FULL),
>         };
>  
> -       return send_tlb_invalidation(&gt->uc.guc, fence, action,
> ARRAY_SIZE(action));
> +       return send_tlb_inval(&gt->uc.guc, fence, action,
> ARRAY_SIZE(action));
>  }
>  
>  /**
>   * xe_gt_tlb_invalidation_all - Invalidate all TLBs across PF and
> all VFs.
>   * @gt: the &xe_gt structure
> - * @fence: the &xe_gt_tlb_invalidation_fence to be signaled on
> completion
> + * @fence: the &xe_gt_tlb_inval_fence to be signaled on completion
>   *
>   * Send a request to invalidate all TLBs across PF and all VFs.
>   *
>   * Return: 0 on success, negative error code on error
>   */
> -int xe_gt_tlb_invalidation_all(struct xe_gt *gt, struct
> xe_gt_tlb_invalidation_fence *fence)
> +int xe_gt_tlb_inval_all(struct xe_gt *gt, struct
> xe_gt_tlb_inval_fence *fence)
>  {
>         int err;
>  
>         xe_gt_assert(gt, gt == fence->gt);
>  
> -       err = send_tlb_invalidation_all(gt, fence);
> +       err = send_tlb_inval_all(gt, fence);
>         if (err)
>                 xe_gt_err(gt, "TLB invalidation request failed
> (%pe)", ERR_PTR(err));
>  
> @@ -372,8 +371,7 @@ int xe_gt_tlb_invalidation_all(struct xe_gt *gt,
> struct xe_gt_tlb_invalidation_f
>  #define MAX_RANGE_TLB_INVALIDATION_LENGTH
> (rounddown_pow_of_two(ULONG_MAX))
>  
>  /**
> - * xe_gt_tlb_invalidation_range - Issue a TLB invalidation on this
> GT for an
> - * address range
> + * xe_gt_tlb_inval_range - Issue a TLB invalidation on this GT for
> an address range
>   *
>   * @gt: GT structure
>   * @fence: invalidation fence which will be signal on TLB
> invalidation
> @@ -388,9 +386,8 @@ int xe_gt_tlb_invalidation_all(struct xe_gt *gt,
> struct xe_gt_tlb_invalidation_f
>   *
>   * Return: Negative error code on error, 0 on success
>   */
> -int xe_gt_tlb_invalidation_range(struct xe_gt *gt,
> -                                struct xe_gt_tlb_invalidation_fence
> *fence,
> -                                u64 start, u64 end, u32 asid)
> +int xe_gt_tlb_inval_range(struct xe_gt *gt, struct
> xe_gt_tlb_inval_fence *fence,
> +                         u64 start, u64 end, u32 asid)
>  {
>         struct xe_device *xe = gt_to_xe(gt);
>  #define MAX_TLB_INVALIDATION_LEN       7
> @@ -402,13 +399,13 @@ int xe_gt_tlb_invalidation_range(struct xe_gt
> *gt,
>  
>         /* Execlists not supported */
>         if (gt_to_xe(gt)->info.force_execlist) {
> -               __invalidation_fence_signal(xe, fence);
> +               __inval_fence_signal(xe, fence);
>                 return 0;
>         }
>  
>         action[len++] = XE_GUC_ACTION_TLB_INVALIDATION;
> -       action[len++] = 0; /* seqno, replaced in
> send_tlb_invalidation */
> -       if (!xe->info.has_range_tlb_invalidation ||
> +       action[len++] = 0; /* seqno, replaced in send_tlb_inval */
> +       if (!xe->info.has_range_tlb_inval ||
>             length > MAX_RANGE_TLB_INVALIDATION_LENGTH) {
>                 action[len++] = MAKE_INVAL_OP(XE_GUC_TLB_INVAL_FULL);
>         } else {
> @@ -457,33 +454,33 @@ int xe_gt_tlb_invalidation_range(struct xe_gt
> *gt,
>  
>         xe_gt_assert(gt, len <= MAX_TLB_INVALIDATION_LEN);
>  
> -       return send_tlb_invalidation(&gt->uc.guc, fence, action,
> len);
> +       return send_tlb_inval(&gt->uc.guc, fence, action, len);
>  }
>  
>  /**
> - * xe_gt_tlb_invalidation_vm - Issue a TLB invalidation on this GT
> for a VM
> + * xe_gt_tlb_inval_vm - Issue a TLB invalidation on this GT for a VM
>   * @gt: graphics tile
>   * @vm: VM to invalidate
>   *
>   * Invalidate entire VM's address space
>   */
> -void xe_gt_tlb_invalidation_vm(struct xe_gt *gt, struct xe_vm *vm)
> +void xe_gt_tlb_inval_vm(struct xe_gt *gt, struct xe_vm *vm)
>  {
> -       struct xe_gt_tlb_invalidation_fence fence;
> +       struct xe_gt_tlb_inval_fence fence;
>         u64 range = 1ull << vm->xe->info.va_bits;
>         int ret;
>  
> -       xe_gt_tlb_invalidation_fence_init(gt, &fence, true);
> +       xe_gt_tlb_inval_fence_init(gt, &fence, true);
>  
> -       ret = xe_gt_tlb_invalidation_range(gt, &fence, 0, range, vm-
> >usm.asid);
> +       ret = xe_gt_tlb_inval_range(gt, &fence, 0, range, vm-
> >usm.asid);
>         if (ret < 0)
>                 return;
>  
> -       xe_gt_tlb_invalidation_fence_wait(&fence);
> +       xe_gt_tlb_inval_fence_wait(&fence);
>  }
>  
>  /**
> - * xe_guc_tlb_invalidation_done_handler - TLB invalidation done
> handler
> + * xe_guc_tlb_inval_done_handler - TLB invalidation done handler
>   * @guc: guc
>   * @msg: message indicating TLB invalidation done
>   * @len: length of message
> @@ -494,11 +491,11 @@ void xe_gt_tlb_invalidation_vm(struct xe_gt
> *gt, struct xe_vm *vm)
>   *
>   * Return: 0 on success, -EPROTO for malformed messages.
>   */
> -int xe_guc_tlb_invalidation_done_handler(struct xe_guc *guc, u32
> *msg, u32 len)
> +int xe_guc_tlb_inval_done_handler(struct xe_guc *guc, u32 *msg, u32
> len)
>  {
>         struct xe_gt *gt = guc_to_gt(guc);
>         struct xe_device *xe = gt_to_xe(gt);
> -       struct xe_gt_tlb_invalidation_fence *fence, *next;
> +       struct xe_gt_tlb_inval_fence *fence, *next;
>         unsigned long flags;
>  
>         if (unlikely(len != 1))
> @@ -519,74 +516,74 @@ int xe_guc_tlb_invalidation_done_handler(struct
> xe_guc *guc, u32 *msg, u32 len)
>          * officially process the CT message like if racing against
>          * process_g2h_msg().
>          */
> -       spin_lock_irqsave(&gt->tlb_invalidation.pending_lock, flags);
> -       if (tlb_invalidation_seqno_past(gt, msg[0])) {
> -               spin_unlock_irqrestore(&gt-
> >tlb_invalidation.pending_lock, flags);
> +       spin_lock_irqsave(&gt->tlb_inval.pending_lock, flags);
> +       if (tlb_inval_seqno_past(gt, msg[0])) {
> +               spin_unlock_irqrestore(&gt->tlb_inval.pending_lock,
> flags);
>                 return 0;
>         }
>  
> -       WRITE_ONCE(gt->tlb_invalidation.seqno_recv, msg[0]);
> +       WRITE_ONCE(gt->tlb_inval.seqno_recv, msg[0]);
>  
>         list_for_each_entry_safe(fence, next,
> -                                &gt-
> >tlb_invalidation.pending_fences, link) {
> -               trace_xe_gt_tlb_invalidation_fence_recv(xe, fence);
> +                                &gt->tlb_inval.pending_fences, link)
> {
> +               trace_xe_gt_tlb_inval_fence_recv(xe, fence);
>  
> -               if (!tlb_invalidation_seqno_past(gt, fence->seqno))
> +               if (!tlb_inval_seqno_past(gt, fence->seqno))
>                         break;
>  
> -               invalidation_fence_signal(xe, fence);
> +               inval_fence_signal(xe, fence);
>         }
>  
> -       if (!list_empty(&gt->tlb_invalidation.pending_fences))
> +       if (!list_empty(&gt->tlb_inval.pending_fences))
>                 mod_delayed_work(system_wq,
> -                                &gt->tlb_invalidation.fence_tdr,
> +                                &gt->tlb_inval.fence_tdr,
>                                  tlb_timeout_jiffies(gt));
>         else
> -               cancel_delayed_work(&gt->tlb_invalidation.fence_tdr);
> +               cancel_delayed_work(&gt->tlb_inval.fence_tdr);
>  
> -       spin_unlock_irqrestore(&gt->tlb_invalidation.pending_lock,
> flags);
> +       spin_unlock_irqrestore(&gt->tlb_inval.pending_lock, flags);
>  
>         return 0;
>  }
>  
>  static const char *
> -invalidation_fence_get_driver_name(struct dma_fence *dma_fence)
> +inval_fence_get_driver_name(struct dma_fence *dma_fence)
>  {
>         return "xe";
>  }
>  
>  static const char *
> -invalidation_fence_get_timeline_name(struct dma_fence *dma_fence)
> +inval_fence_get_timeline_name(struct dma_fence *dma_fence)
>  {
> -       return "invalidation_fence";
> +       return "inval_fence";
>  }
>  
> -static const struct dma_fence_ops invalidation_fence_ops = {
> -       .get_driver_name = invalidation_fence_get_driver_name,
> -       .get_timeline_name = invalidation_fence_get_timeline_name,
> +static const struct dma_fence_ops inval_fence_ops = {
> +       .get_driver_name = inval_fence_get_driver_name,
> +       .get_timeline_name = inval_fence_get_timeline_name,
>  };
>  
>  /**
> - * xe_gt_tlb_invalidation_fence_init - Initialize TLB invalidation
> fence
> + * xe_gt_tlb_inval_fence_init - Initialize TLB invalidation fence
>   * @gt: GT
>   * @fence: TLB invalidation fence to initialize
>   * @stack: fence is stack variable
>   *
> - * Initialize TLB invalidation fence for use.
> xe_gt_tlb_invalidation_fence_fini
> + * Initialize TLB invalidation fence for use.
> xe_gt_tlb_inval_fence_fini
>   * will be automatically called when fence is signalled (all fences
> must signal),
>   * even on error.
>   */
> -void xe_gt_tlb_invalidation_fence_init(struct xe_gt *gt,
> -                                      struct
> xe_gt_tlb_invalidation_fence *fence,
> -                                      bool stack)
> +void xe_gt_tlb_inval_fence_init(struct xe_gt *gt,
> +                               struct xe_gt_tlb_inval_fence *fence,
> +                               bool stack)
>  {
>         xe_pm_runtime_get_noresume(gt_to_xe(gt));
>  
> -       spin_lock_irq(&gt->tlb_invalidation.lock);
> -       dma_fence_init(&fence->base, &invalidation_fence_ops,
> -                      &gt->tlb_invalidation.lock,
> +       spin_lock_irq(&gt->tlb_inval.lock);
> +       dma_fence_init(&fence->base, &inval_fence_ops,
> +                      &gt->tlb_inval.lock,
>                        dma_fence_context_alloc(1), 1);
> -       spin_unlock_irq(&gt->tlb_invalidation.lock);
> +       spin_unlock_irq(&gt->tlb_inval.lock);
>         INIT_LIST_HEAD(&fence->link);
>         if (stack)
>                 set_bit(FENCE_STACK_BIT, &fence->base.flags);
> diff --git a/drivers/gpu/drm/xe/xe_gt_tlb_inval.h
> b/drivers/gpu/drm/xe/xe_gt_tlb_inval.h
> new file mode 100644
> index 000000000000..801d4ecf88f0
> --- /dev/null
> +++ b/drivers/gpu/drm/xe/xe_gt_tlb_inval.h
> @@ -0,0 +1,40 @@
> +/* SPDX-License-Identifier: MIT */
> +/*
> + * Copyright © 2023 Intel Corporation
> + */
> +
> +#ifndef _XE_GT_TLB_INVAL_H_
> +#define _XE_GT_TLB_INVAL_H_
> +
> +#include <linux/types.h>
> +
> +#include "xe_gt_tlb_inval_types.h"
> +
> +struct xe_gt;
> +struct xe_guc;
> +struct xe_vm;
> +struct xe_vma;
> +
> +int xe_gt_tlb_inval_init_early(struct xe_gt *gt);
> +
> +void xe_gt_tlb_inval_reset(struct xe_gt *gt);
> +int xe_gt_tlb_inval_ggtt(struct xe_gt *gt);
> +void xe_gt_tlb_inval_vm(struct xe_gt *gt, struct xe_vm *vm);
> +int xe_gt_tlb_inval_all(struct xe_gt *gt, struct
> xe_gt_tlb_inval_fence *fence);
> +int xe_gt_tlb_inval_range(struct xe_gt *gt,
> +                         struct xe_gt_tlb_inval_fence *fence,
> +                         u64 start, u64 end, u32 asid);
> +int xe_guc_tlb_inval_done_handler(struct xe_guc *guc, u32 *msg, u32
> len);
> +
> +void xe_gt_tlb_inval_fence_init(struct xe_gt *gt,
> +                               struct xe_gt_tlb_inval_fence *fence,
> +                               bool stack);
> +void xe_gt_tlb_inval_fence_signal(struct xe_gt_tlb_inval_fence
> *fence);
> +
> +static inline void
> +xe_gt_tlb_inval_fence_wait(struct xe_gt_tlb_inval_fence *fence)
> +{
> +       dma_fence_wait(&fence->base, false);
> +}
> +
> +#endif /* _XE_GT_TLB_INVAL_ */
> diff --git a/drivers/gpu/drm/xe/xe_gt_tlb_invalidation_types.h
> b/drivers/gpu/drm/xe/xe_gt_tlb_inval_types.h
> similarity index 55%
> rename from drivers/gpu/drm/xe/xe_gt_tlb_invalidation_types.h
> rename to drivers/gpu/drm/xe/xe_gt_tlb_inval_types.h
> index de6e825e0851..919430359103 100644
> --- a/drivers/gpu/drm/xe/xe_gt_tlb_invalidation_types.h
> +++ b/drivers/gpu/drm/xe/xe_gt_tlb_inval_types.h
> @@ -3,20 +3,20 @@
>   * Copyright © 2023 Intel Corporation
>   */
>  
> -#ifndef _XE_GT_TLB_INVALIDATION_TYPES_H_
> -#define _XE_GT_TLB_INVALIDATION_TYPES_H_
> +#ifndef _XE_GT_TLB_INVAL_TYPES_H_
> +#define _XE_GT_TLB_INVAL_TYPES_H_
>  
>  #include <linux/dma-fence.h>
>  
>  struct xe_gt;
>  
>  /**
> - * struct xe_gt_tlb_invalidation_fence - XE GT TLB invalidation
> fence
> + * struct xe_gt_tlb_inval_fence - XE GT TLB invalidation fence
>   *
> - * Optionally passed to xe_gt_tlb_invalidation and will be signaled
> upon TLB
> + * Optionally passed to xe_gt_tlb_inval and will be signaled upon
> TLB
>   * invalidation completion.
>   */
> -struct xe_gt_tlb_invalidation_fence {
> +struct xe_gt_tlb_inval_fence {
>         /** @base: dma fence base */
>         struct dma_fence base;
>         /** @gt: GT which fence belong to */
> @@ -25,8 +25,8 @@ struct xe_gt_tlb_invalidation_fence {
>         struct list_head link;
>         /** @seqno: seqno of TLB invalidation to signal fence one */
>         int seqno;
> -       /** @invalidation_time: time of TLB invalidation */
> -       ktime_t invalidation_time;
> +       /** @inval_time: time of TLB invalidation */
> +       ktime_t inval_time;
>  };
>  
>  #endif
> diff --git a/drivers/gpu/drm/xe/xe_gt_tlb_invalidation.h
> b/drivers/gpu/drm/xe/xe_gt_tlb_invalidation.h
> deleted file mode 100644
> index f7f0f2eaf4b5..000000000000
> --- a/drivers/gpu/drm/xe/xe_gt_tlb_invalidation.h
> +++ /dev/null
> @@ -1,40 +0,0 @@
> -/* SPDX-License-Identifier: MIT */
> -/*
> - * Copyright © 2023 Intel Corporation
> - */
> -
> -#ifndef _XE_GT_TLB_INVALIDATION_H_
> -#define _XE_GT_TLB_INVALIDATION_H_
> -
> -#include <linux/types.h>
> -
> -#include "xe_gt_tlb_invalidation_types.h"
> -
> -struct xe_gt;
> -struct xe_guc;
> -struct xe_vm;
> -struct xe_vma;
> -
> -int xe_gt_tlb_invalidation_init_early(struct xe_gt *gt);
> -
> -void xe_gt_tlb_invalidation_reset(struct xe_gt *gt);
> -int xe_gt_tlb_invalidation_ggtt(struct xe_gt *gt);
> -void xe_gt_tlb_invalidation_vm(struct xe_gt *gt, struct xe_vm *vm);
> -int xe_gt_tlb_invalidation_all(struct xe_gt *gt, struct
> xe_gt_tlb_invalidation_fence *fence);
> -int xe_gt_tlb_invalidation_range(struct xe_gt *gt,
> -                                struct xe_gt_tlb_invalidation_fence
> *fence,
> -                                u64 start, u64 end, u32 asid);
> -int xe_guc_tlb_invalidation_done_handler(struct xe_guc *guc, u32
> *msg, u32 len);
> -
> -void xe_gt_tlb_invalidation_fence_init(struct xe_gt *gt,
> -                                      struct
> xe_gt_tlb_invalidation_fence *fence,
> -                                      bool stack);
> -void xe_gt_tlb_invalidation_fence_signal(struct
> xe_gt_tlb_invalidation_fence *fence);
> -
> -static inline void
> -xe_gt_tlb_invalidation_fence_wait(struct
> xe_gt_tlb_invalidation_fence *fence)
> -{
> -       dma_fence_wait(&fence->base, false);
> -}
> -
> -#endif /* _XE_GT_TLB_INVALIDATION_ */
> diff --git a/drivers/gpu/drm/xe/xe_gt_types.h
> b/drivers/gpu/drm/xe/xe_gt_types.h
> index 96344c604726..b82795fc0070 100644
> --- a/drivers/gpu/drm/xe/xe_gt_types.h
> +++ b/drivers/gpu/drm/xe/xe_gt_types.h
> @@ -185,34 +185,34 @@ struct xe_gt {
>                 struct work_struct worker;
>         } reset;
>  
> -       /** @tlb_invalidation: TLB invalidation state */
> +       /** @tlb_inval: TLB invalidation state */
>         struct {
> -               /** @tlb_invalidation.seqno: TLB invalidation seqno,
> protected by CT lock */
> +               /** @tlb_inval.seqno: TLB invalidation seqno,
> protected by CT lock */
>  #define TLB_INVALIDATION_SEQNO_MAX     0x100000
>                 int seqno;
>                 /**
> -                * @tlb_invalidation.seqno_recv: last received TLB
> invalidation seqno,
> +                * @tlb_inval.seqno_recv: last received TLB
> invalidation seqno,
>                  * protected by CT lock
>                  */
>                 int seqno_recv;
>                 /**
> -                * @tlb_invalidation.pending_fences: list of pending
> fences waiting TLB
> +                * @tlb_inval.pending_fences: list of pending fences
> waiting TLB
>                  * invaliations, protected by CT lock
>                  */
>                 struct list_head pending_fences;
>                 /**
> -                * @tlb_invalidation.pending_lock: protects
> @tlb_invalidation.pending_fences
> -                * and updating @tlb_invalidation.seqno_recv.
> +                * @tlb_inval.pending_lock: protects
> @tlb_inval.pending_fences
> +                * and updating @tlb_inval.seqno_recv.
>                  */
>                 spinlock_t pending_lock;
>                 /**
> -                * @tlb_invalidation.fence_tdr: schedules a delayed
> call to
> +                * @tlb_inval.fence_tdr: schedules a delayed call to
>                  * xe_gt_tlb_fence_timeout after the timeut interval
> is over.
>                  */
>                 struct delayed_work fence_tdr;
> -               /** @tlb_invalidation.lock: protects TLB invalidation
> fences */
> +               /** @tlb_inval.lock: protects TLB invalidation fences
> */
>                 spinlock_t lock;
> -       } tlb_invalidation;
> +       } tlb_inval;
>  
>         /**
>          * @ccs_mode: Number of compute engines enabled.
> diff --git a/drivers/gpu/drm/xe/xe_guc_ct.c
> b/drivers/gpu/drm/xe/xe_guc_ct.c
> index b6acccfcd351..c213a037b346 100644
> --- a/drivers/gpu/drm/xe/xe_guc_ct.c
> +++ b/drivers/gpu/drm/xe/xe_guc_ct.c
> @@ -26,7 +26,7 @@
>  #include "xe_gt_sriov_pf_control.h"
>  #include "xe_gt_sriov_pf_monitor.h"
>  #include "xe_gt_sriov_printk.h"
> -#include "xe_gt_tlb_invalidation.h"
> +#include "xe_gt_tlb_inval.h"
>  #include "xe_guc.h"
>  #include "xe_guc_log.h"
>  #include "xe_guc_relay.h"
> @@ -1420,8 +1420,7 @@ static int process_g2h_msg(struct xe_guc_ct
> *ct, u32 *msg, u32 len)
>                 ret = xe_guc_pagefault_handler(guc, payload,
> adj_len);
>                 break;
>         case XE_GUC_ACTION_TLB_INVALIDATION_DONE:
> -               ret = xe_guc_tlb_invalidation_done_handler(guc,
> payload,
> -                                                          adj_len);
> +               ret = xe_guc_tlb_inval_done_handler(guc, payload,
> adj_len);
>                 break;
>         case XE_GUC_ACTION_ACCESS_COUNTER_NOTIFY:
>                 ret = xe_guc_access_counter_notify_handler(guc,
> payload,
> @@ -1622,8 +1621,7 @@ static void g2h_fast_path(struct xe_guc_ct *ct,
> u32 *msg, u32 len)
>                 break;
>         case XE_GUC_ACTION_TLB_INVALIDATION_DONE:
>                 __g2h_release_space(ct, len);
> -               ret = xe_guc_tlb_invalidation_done_handler(guc,
> payload,
> -                                                          adj_len);
> +               ret = xe_guc_tlb_inval_done_handler(guc, payload,
> adj_len);
>                 break;
>         default:
>                 xe_gt_warn(gt, "NOT_POSSIBLE");
> diff --git a/drivers/gpu/drm/xe/xe_lmtt.c
> b/drivers/gpu/drm/xe/xe_lmtt.c
> index a2000307d5bf..8869ad491d99 100644
> --- a/drivers/gpu/drm/xe/xe_lmtt.c
> +++ b/drivers/gpu/drm/xe/xe_lmtt.c
> @@ -11,7 +11,7 @@
>  
>  #include "xe_assert.h"
>  #include "xe_bo.h"
> -#include "xe_gt_tlb_invalidation.h"
> +#include "xe_gt_tlb_inval.h"
>  #include "xe_lmtt.h"
>  #include "xe_map.h"
>  #include "xe_mmio.h"
> @@ -225,8 +225,8 @@ void xe_lmtt_init_hw(struct xe_lmtt *lmtt)
>  
>  static int lmtt_invalidate_hw(struct xe_lmtt *lmtt)
>  {
> -       struct xe_gt_tlb_invalidation_fence
> fences[XE_MAX_GT_PER_TILE];
> -       struct xe_gt_tlb_invalidation_fence *fence = fences;
> +       struct xe_gt_tlb_inval_fence fences[XE_MAX_GT_PER_TILE];
> +       struct xe_gt_tlb_inval_fence *fence = fences;
>         struct xe_tile *tile = lmtt_to_tile(lmtt);
>         struct xe_gt *gt;
>         int result = 0;
> @@ -234,8 +234,8 @@ static int lmtt_invalidate_hw(struct xe_lmtt
> *lmtt)
>         u8 id;
>  
>         for_each_gt_on_tile(gt, tile, id) {
> -               xe_gt_tlb_invalidation_fence_init(gt, fence, true);
> -               err = xe_gt_tlb_invalidation_all(gt, fence);
> +               xe_gt_tlb_inval_fence_init(gt, fence, true);
> +               err = xe_gt_tlb_inval_all(gt, fence);
>                 result = result ?: err;
>                 fence++;
>         }
> @@ -249,7 +249,7 @@ static int lmtt_invalidate_hw(struct xe_lmtt
> *lmtt)
>          */
>         fence = fences;
>         for_each_gt_on_tile(gt, tile, id)
> -               xe_gt_tlb_invalidation_fence_wait(fence++);
> +               xe_gt_tlb_inval_fence_wait(fence++);
>  
>         return result;
>  }
> diff --git a/drivers/gpu/drm/xe/xe_pci.c
> b/drivers/gpu/drm/xe/xe_pci.c
> index dc4c8e861a84..2c14e3b10679 100644
> --- a/drivers/gpu/drm/xe/xe_pci.c
> +++ b/drivers/gpu/drm/xe/xe_pci.c
> @@ -55,7 +55,7 @@ static const struct xe_graphics_desc graphics_xelp
> = {
>  };
>  
>  #define XE_HP_FEATURES \
> -       .has_range_tlb_invalidation = true, \
> +       .has_range_tlb_inval = true, \
>         .va_bits = 48, \
>         .vm_max_level = 3
>  
> @@ -103,7 +103,7 @@ static const struct xe_graphics_desc
> graphics_xelpg = {
>         .has_asid = 1, \
>         .has_atomic_enable_pte_bit = 1, \
>         .has_flat_ccs = 1, \
> -       .has_range_tlb_invalidation = 1, \
> +       .has_range_tlb_inval = 1, \
>         .has_usm = 1, \
>         .has_64bit_timestamp = 1, \
>         .va_bits = 48, \
> @@ -673,7 +673,7 @@ static int xe_info_init(struct xe_device *xe,
>         /* Runtime detection may change this later */
>         xe->info.has_flat_ccs = graphics_desc->has_flat_ccs;
>  
> -       xe->info.has_range_tlb_invalidation = graphics_desc-
> >has_range_tlb_invalidation;
> +       xe->info.has_range_tlb_inval = graphics_desc-
> >has_range_tlb_inval;
>         xe->info.has_usm = graphics_desc->has_usm;
>         xe->info.has_64bit_timestamp = graphics_desc-
> >has_64bit_timestamp;
>  
> diff --git a/drivers/gpu/drm/xe/xe_pci_types.h
> b/drivers/gpu/drm/xe/xe_pci_types.h
> index 4de6f69ed975..b63002fc0f67 100644
> --- a/drivers/gpu/drm/xe/xe_pci_types.h
> +++ b/drivers/gpu/drm/xe/xe_pci_types.h
> @@ -60,7 +60,7 @@ struct xe_graphics_desc {
>         u8 has_atomic_enable_pte_bit:1;
>         u8 has_flat_ccs:1;
>         u8 has_indirect_ring_state:1;
> -       u8 has_range_tlb_invalidation:1;
> +       u8 has_range_tlb_inval:1;
>         u8 has_usm:1;
>         u8 has_64bit_timestamp:1;
>  };
> diff --git a/drivers/gpu/drm/xe/xe_pt.c b/drivers/gpu/drm/xe/xe_pt.c
> index 7b441d1a77e9..04c52926828f 100644
> --- a/drivers/gpu/drm/xe/xe_pt.c
> +++ b/drivers/gpu/drm/xe/xe_pt.c
> @@ -13,7 +13,7 @@
>  #include "xe_drm_client.h"
>  #include "xe_exec_queue.h"
>  #include "xe_gt.h"
> -#include "xe_gt_tlb_invalidation.h"
> +#include "xe_gt_tlb_inval.h"
>  #include "xe_migrate.h"
>  #include "xe_pt_types.h"
>  #include "xe_pt_walk.h"
> @@ -1509,8 +1509,8 @@ static int xe_pt_svm_pre_commit(struct
> xe_migrate_pt_update *pt_update)
>  }
>  #endif
>  
> -struct invalidation_fence {
> -       struct xe_gt_tlb_invalidation_fence base;
> +struct inval_fence {
> +       struct xe_gt_tlb_inval_fence base;
>         struct xe_gt *gt;
>         struct dma_fence *fence;
>         struct dma_fence_cb cb;
> @@ -1520,38 +1520,37 @@ struct invalidation_fence {
>         u32 asid;
>  };
>  
> -static void invalidation_fence_cb(struct dma_fence *fence,
> -                                 struct dma_fence_cb *cb)
> +static void inval_fence_cb(struct dma_fence *fence, struct
> dma_fence_cb *cb)
>  {
> -       struct invalidation_fence *ifence =
> -               container_of(cb, struct invalidation_fence, cb);
> +       struct inval_fence *ifence =
> +               container_of(cb, struct inval_fence, cb);
>  
>         if (!ifence->fence->error) {
>                 queue_work(system_wq, &ifence->work);
>         } else {
>                 ifence->base.base.error = ifence->fence->error;
> -               xe_gt_tlb_invalidation_fence_signal(&ifence->base);
> +               xe_gt_tlb_inval_fence_signal(&ifence->base);
>         }
>         dma_fence_put(ifence->fence);
>  }
>  
> -static void invalidation_fence_work_func(struct work_struct *w)
> +static void inval_fence_work_func(struct work_struct *w)
>  {
> -       struct invalidation_fence *ifence =
> -               container_of(w, struct invalidation_fence, work);
> +       struct inval_fence *ifence =
> +               container_of(w, struct inval_fence, work);
>  
> -       xe_gt_tlb_invalidation_range(ifence->gt, &ifence->base,
> ifence->start,
> -                                    ifence->end, ifence->asid);
> +       xe_gt_tlb_inval_range(ifence->gt, &ifence->base, ifence-
> >start,
> +                             ifence->end, ifence->asid);
>  }
>  
> -static void invalidation_fence_init(struct xe_gt *gt,
> -                                   struct invalidation_fence
> *ifence,
> -                                   struct dma_fence *fence,
> -                                   u64 start, u64 end, u32 asid)
> +static void inval_fence_init(struct xe_gt *gt,
> +                            struct inval_fence *ifence,
> +                            struct dma_fence *fence,
> +                            u64 start, u64 end, u32 asid)
>  {
>         int ret;
>  
> -       xe_gt_tlb_invalidation_fence_init(gt, &ifence->base, false);
> +       xe_gt_tlb_inval_fence_init(gt, &ifence->base, false);
>  
>         ifence->fence = fence;
>         ifence->gt = gt;
> @@ -1559,11 +1558,11 @@ static void invalidation_fence_init(struct
> xe_gt *gt,
>         ifence->end = end;
>         ifence->asid = asid;
>  
> -       INIT_WORK(&ifence->work, invalidation_fence_work_func);
> -       ret = dma_fence_add_callback(fence, &ifence->cb,
> invalidation_fence_cb);
> +       INIT_WORK(&ifence->work, inval_fence_work_func);
> +       ret = dma_fence_add_callback(fence, &ifence->cb,
> inval_fence_cb);
>         if (ret == -ENOENT) {
>                 dma_fence_put(ifence->fence);   /* Usually dropped in
> CB */
> -               invalidation_fence_work_func(&ifence->work);
> +               inval_fence_work_func(&ifence->work);
>         } else if (ret) {
>                 dma_fence_put(&ifence->base.base);      /* Caller ref
> */
>                 dma_fence_put(&ifence->base.base);      /* Creation
> ref */
> @@ -2402,7 +2401,7 @@ xe_pt_update_ops_run(struct xe_tile *tile,
> struct xe_vma_ops *vops)
>         struct xe_vm_pgtable_update_ops *pt_update_ops =
>                 &vops->pt_update_ops[tile->id];
>         struct dma_fence *fence;
> -       struct invalidation_fence *ifence = NULL, *mfence = NULL;
> +       struct inval_fence *ifence = NULL, *mfence = NULL;
>         struct dma_fence **fences = NULL;
>         struct dma_fence_array *cf = NULL;
>         struct xe_range_fence *rfence;
> @@ -2489,13 +2488,13 @@ xe_pt_update_ops_run(struct xe_tile *tile,
> struct xe_vma_ops *vops)
>         if (ifence) {
>                 if (mfence)
>                         dma_fence_get(fence);
> -               invalidation_fence_init(tile->primary_gt, ifence,
> fence,
> -                                       pt_update_ops->start,
> -                                       pt_update_ops->last, vm-
> >usm.asid);
> +               inval_fence_init(tile->primary_gt, ifence, fence,
> +                                pt_update_ops->start,
> +                                pt_update_ops->last, vm->usm.asid);
>                 if (mfence) {
> -                       invalidation_fence_init(tile->media_gt,
> mfence, fence,
> -                                               pt_update_ops->start,
> -                                               pt_update_ops->last,
> vm->usm.asid);
> +                       inval_fence_init(tile->media_gt, mfence,
> fence,
> +                                        pt_update_ops->start,
> +                                        pt_update_ops->last, vm-
> >usm.asid);
>                         fences[0] = &ifence->base.base;
>                         fences[1] = &mfence->base.base;
>                         dma_fence_array_init(cf, 2, fences,
> diff --git a/drivers/gpu/drm/xe/xe_svm.c
> b/drivers/gpu/drm/xe/xe_svm.c
> index a7ff5975873f..030412cda6a4 100644
> --- a/drivers/gpu/drm/xe/xe_svm.c
> +++ b/drivers/gpu/drm/xe/xe_svm.c
> @@ -7,7 +7,7 @@
>  
>  #include "xe_bo.h"
>  #include "xe_gt_stats.h"
> -#include "xe_gt_tlb_invalidation.h"
> +#include "xe_gt_tlb_inval.h"
>  #include "xe_migrate.h"
>  #include "xe_module.h"
>  #include "xe_pm.h"
> @@ -224,7 +224,7 @@ static void xe_svm_invalidate(struct drm_gpusvm
> *gpusvm,
>  
>         xe_device_wmb(xe);
>  
> -       err = xe_vm_range_tilemask_tlb_invalidation(vm, adj_start,
> adj_end, tile_mask);
> +       err = xe_vm_range_tilemask_tlb_inval(vm, adj_start, adj_end,
> tile_mask);
>         WARN_ON_ONCE(err);
>  
>  range_notifier_event_end:
> diff --git a/drivers/gpu/drm/xe/xe_trace.h
> b/drivers/gpu/drm/xe/xe_trace.h
> index 21486a6f693a..36538f50d06f 100644
> --- a/drivers/gpu/drm/xe/xe_trace.h
> +++ b/drivers/gpu/drm/xe/xe_trace.h
> @@ -14,7 +14,7 @@
>  
>  #include "xe_exec_queue_types.h"
>  #include "xe_gpu_scheduler_types.h"
> -#include "xe_gt_tlb_invalidation_types.h"
> +#include "xe_gt_tlb_inval_types.h"
>  #include "xe_gt_types.h"
>  #include "xe_guc_exec_queue_types.h"
>  #include "xe_sched_job.h"
> @@ -25,13 +25,13 @@
>  #define __dev_name_gt(gt)      __dev_name_xe(gt_to_xe((gt)))
>  #define __dev_name_eq(q)       __dev_name_gt((q)->gt)
>  
> -DECLARE_EVENT_CLASS(xe_gt_tlb_invalidation_fence,
> -                   TP_PROTO(struct xe_device *xe, struct
> xe_gt_tlb_invalidation_fence *fence),
> +DECLARE_EVENT_CLASS(xe_gt_tlb_inval_fence,
> +                   TP_PROTO(struct xe_device *xe, struct
> xe_gt_tlb_inval_fence *fence),
>                     TP_ARGS(xe, fence),
>  
>                     TP_STRUCT__entry(
>                              __string(dev, __dev_name_xe(xe))
> -                            __field(struct
> xe_gt_tlb_invalidation_fence *, fence)
> +                            __field(struct xe_gt_tlb_inval_fence *,
> fence)
>                              __field(int, seqno)
>                              ),
>  
> @@ -45,23 +45,23 @@ DECLARE_EVENT_CLASS(xe_gt_tlb_invalidation_fence,
>                               __get_str(dev), __entry->fence,
> __entry->seqno)
>  );
>  
> -DEFINE_EVENT(xe_gt_tlb_invalidation_fence,
> xe_gt_tlb_invalidation_fence_send,
> -            TP_PROTO(struct xe_device *xe, struct
> xe_gt_tlb_invalidation_fence *fence),
> +DEFINE_EVENT(xe_gt_tlb_inval_fence, xe_gt_tlb_inval_fence_send,
> +            TP_PROTO(struct xe_device *xe, struct
> xe_gt_tlb_inval_fence *fence),
>              TP_ARGS(xe, fence)
>  );
>  
> -DEFINE_EVENT(xe_gt_tlb_invalidation_fence,
> xe_gt_tlb_invalidation_fence_recv,
> -            TP_PROTO(struct xe_device *xe, struct
> xe_gt_tlb_invalidation_fence *fence),
> +DEFINE_EVENT(xe_gt_tlb_inval_fence, xe_gt_tlb_inval_fence_recv,
> +            TP_PROTO(struct xe_device *xe, struct
> xe_gt_tlb_inval_fence *fence),
>              TP_ARGS(xe, fence)
>  );
>  
> -DEFINE_EVENT(xe_gt_tlb_invalidation_fence,
> xe_gt_tlb_invalidation_fence_signal,
> -            TP_PROTO(struct xe_device *xe, struct
> xe_gt_tlb_invalidation_fence *fence),
> +DEFINE_EVENT(xe_gt_tlb_inval_fence, xe_gt_tlb_inval_fence_signal,
> +            TP_PROTO(struct xe_device *xe, struct
> xe_gt_tlb_inval_fence *fence),
>              TP_ARGS(xe, fence)
>  );
>  
> -DEFINE_EVENT(xe_gt_tlb_invalidation_fence,
> xe_gt_tlb_invalidation_fence_timeout,
> -            TP_PROTO(struct xe_device *xe, struct
> xe_gt_tlb_invalidation_fence *fence),
> +DEFINE_EVENT(xe_gt_tlb_inval_fence, xe_gt_tlb_inval_fence_timeout,
> +            TP_PROTO(struct xe_device *xe, struct
> xe_gt_tlb_inval_fence *fence),
>              TP_ARGS(xe, fence)
>  );
>  
> diff --git a/drivers/gpu/drm/xe/xe_vm.c b/drivers/gpu/drm/xe/xe_vm.c
> index 2035604121e6..a5fe59f63268 100644
> --- a/drivers/gpu/drm/xe/xe_vm.c
> +++ b/drivers/gpu/drm/xe/xe_vm.c
> @@ -28,7 +28,7 @@
>  #include "xe_drm_client.h"
>  #include "xe_exec_queue.h"
>  #include "xe_gt_pagefault.h"
> -#include "xe_gt_tlb_invalidation.h"
> +#include "xe_gt_tlb_inval.h"
>  #include "xe_migrate.h"
>  #include "xe_pat.h"
>  #include "xe_pm.h"
> @@ -1850,7 +1850,7 @@ static void xe_vm_close(struct xe_vm *vm)
>                                         xe_pt_clear(xe, vm-
> >pt_root[id]);
>  
>                         for_each_gt(gt, xe, id)
> -                               xe_gt_tlb_invalidation_vm(gt, vm);
> +                               xe_gt_tlb_inval_vm(gt, vm);
>                 }
>         }
>  
> @@ -3843,7 +3843,7 @@ void xe_vm_unlock(struct xe_vm *vm)
>  }
>  
>  /**
> - * xe_vm_range_tilemask_tlb_invalidation - Issue a TLB invalidation
> on this tilemask for an
> + * xe_vm_range_tilemask_tlb_inval - Issue a TLB invalidation on this
> tilemask for an
>   * address range
>   * @vm: The VM
>   * @start: start address
> @@ -3854,10 +3854,11 @@ void xe_vm_unlock(struct xe_vm *vm)
>   *
>   * Returns 0 for success, negative error code otherwise.
>   */
> -int xe_vm_range_tilemask_tlb_invalidation(struct xe_vm *vm, u64
> start,
> -                                         u64 end, u8 tile_mask)
> +int xe_vm_range_tilemask_tlb_inval(struct xe_vm *vm, u64 start,
> +                                  u64 end, u8 tile_mask)
>  {
> -       struct xe_gt_tlb_invalidation_fence
> fence[XE_MAX_TILES_PER_DEVICE * XE_MAX_GT_PER_TILE];
> +       struct xe_gt_tlb_inval_fence
> +               fence[XE_MAX_TILES_PER_DEVICE * XE_MAX_GT_PER_TILE];
>         struct xe_tile *tile;
>         u32 fence_id = 0;
>         u8 id;
> @@ -3867,39 +3868,34 @@ int
> xe_vm_range_tilemask_tlb_invalidation(struct xe_vm *vm, u64 start,
>                 return 0;
>  
>         for_each_tile(tile, vm->xe, id) {
> -               if (tile_mask & BIT(id)) {
> -                       xe_gt_tlb_invalidation_fence_init(tile-
> >primary_gt,
> -                                                        
> &fence[fence_id], true);
> -
> -                       err = xe_gt_tlb_invalidation_range(tile-
> >primary_gt,
> -                                                         
> &fence[fence_id],
> -                                                          start,
> -                                                          end,
> -                                                          vm-
> >usm.asid);
> -                       if (err)
> -                               goto wait;
> -                       ++fence_id;
> +               if (!(tile_mask & BIT(id)))
> +                       continue;
>  
> -                       if (!tile->media_gt)
> -                               continue;
> +               xe_gt_tlb_inval_fence_init(tile->primary_gt,
> +                                          &fence[fence_id], true);
>  
> -                       xe_gt_tlb_invalidation_fence_init(tile-
> >media_gt,
> -                                                        
> &fence[fence_id], true);
> +               err = xe_gt_tlb_inval_range(tile->primary_gt,
> &fence[fence_id],
> +                                           start, end, vm-
> >usm.asid);
> +               if (err)
> +                       goto wait;
> +               ++fence_id;
>  
> -                       err = xe_gt_tlb_invalidation_range(tile-
> >media_gt,
> -                                                         
> &fence[fence_id],
> -                                                          start,
> -                                                          end,
> -                                                          vm-
> >usm.asid);
> -                       if (err)
> -                               goto wait;
> -                       ++fence_id;
> -               }
> +               if (!tile->media_gt)
> +                       continue;
> +
> +               xe_gt_tlb_inval_fence_init(tile->media_gt,
> +                                          &fence[fence_id], true);
> +
> +               err = xe_gt_tlb_inval_range(tile->media_gt,
> &fence[fence_id],
> +                                           start, end, vm-
> >usm.asid);
> +               if (err)
> +                       goto wait;
> +               ++fence_id;
>         }
>  
>  wait:
>         for (id = 0; id < fence_id; ++id)
> -               xe_gt_tlb_invalidation_fence_wait(&fence[id]);
> +               xe_gt_tlb_inval_fence_wait(&fence[id]);
>  
>         return err;
>  }
> @@ -3958,8 +3954,8 @@ int xe_vm_invalidate_vma(struct xe_vma *vma)
>  
>         xe_device_wmb(xe);
>  
> -       ret = xe_vm_range_tilemask_tlb_invalidation(xe_vma_vm(vma),
> xe_vma_start(vma),
> -                                                   xe_vma_end(vma),
> tile_mask);
> +       ret = xe_vm_range_tilemask_tlb_inval(xe_vma_vm(vma),
> xe_vma_start(vma),
> +                                            xe_vma_end(vma),
> tile_mask);
>  
>         /* WRITE_ONCE pairs with READ_ONCE in
> xe_vm_has_valid_gpu_mapping() */
>         WRITE_ONCE(vma->tile_invalidated, vma->tile_mask);
> diff --git a/drivers/gpu/drm/xe/xe_vm.h b/drivers/gpu/drm/xe/xe_vm.h
> index 3475a118f666..fade8ad92bad 100644
> --- a/drivers/gpu/drm/xe/xe_vm.h
> +++ b/drivers/gpu/drm/xe/xe_vm.h
> @@ -228,8 +228,8 @@ struct dma_fence *xe_vm_range_rebind(struct xe_vm
> *vm,
>  struct dma_fence *xe_vm_range_unbind(struct xe_vm *vm,
>                                      struct xe_svm_range *range);
>  
> -int xe_vm_range_tilemask_tlb_invalidation(struct xe_vm *vm, u64
> start,
> -                                         u64 end, u8 tile_mask);
> +int xe_vm_range_tilemask_tlb_inval(struct xe_vm *vm, u64 start,
> +                                  u64 end, u8 tile_mask);
>  
>  int xe_vm_invalidate_vma(struct xe_vma *vma);
>  



More information about the Intel-xe mailing list