[PATCH] drm/xe: Keep track of TLB inval events
Nirmoy Das
nirmoy.das at intel.com
Thu Jul 25 15:05:45 UTC 2024
Use stats API to keep track of TLB invalidation events per GT.
Signed-off-by: Nirmoy Das <nirmoy.das at intel.com>
---
drivers/gpu/drm/xe/xe_gt_debugfs.c | 10 +++++++
drivers/gpu/drm/xe/xe_gt_tlb_invalidation.c | 31 +++++++++++++++------
drivers/gpu/drm/xe/xe_gt_tlb_invalidation.h | 1 +
drivers/gpu/drm/xe/xe_gt_types.h | 4 +++
drivers/gpu/drm/xe/xe_uc_fw.c | 2 +-
5 files changed, 38 insertions(+), 10 deletions(-)
diff --git a/drivers/gpu/drm/xe/xe_gt_debugfs.c b/drivers/gpu/drm/xe/xe_gt_debugfs.c
index 5e7fd937917a..28a9aa18d3bc 100644
--- a/drivers/gpu/drm/xe/xe_gt_debugfs.c
+++ b/drivers/gpu/drm/xe/xe_gt_debugfs.c
@@ -17,6 +17,7 @@
#include "xe_gt_mcr.h"
#include "xe_gt_sriov_pf_debugfs.h"
#include "xe_gt_sriov_vf_debugfs.h"
+#include "xe_gt_tlb_invalidation.h"
#include "xe_gt_topology.h"
#include "xe_hw_engine.h"
#include "xe_lrc.h"
@@ -27,6 +28,7 @@
#include "xe_reg_sr.h"
#include "xe_reg_whitelist.h"
#include "xe_sriov.h"
+#include "xe_stats.h"
#include "xe_uc_debugfs.h"
#include "xe_wa.h"
@@ -288,6 +290,12 @@ static const struct drm_info_list debugfs_list[] = {
{"default_lrc_vecs", .show = xe_gt_debugfs_simple_show, .data = vecs_default_lrc},
};
+static void xe_gt_debugfs_stats_register(struct xe_gt *gt, struct dentry *root)
+{
+ gt->stats = xe_stats_init(>_to_xe(gt)->drm, root);
+ xe_gt_tlb_invalidation_stats_init(gt);
+}
+
void xe_gt_debugfs_register(struct xe_gt *gt)
{
struct xe_device *xe = gt_to_xe(gt);
@@ -321,4 +329,6 @@ void xe_gt_debugfs_register(struct xe_gt *gt)
xe_gt_sriov_pf_debugfs_register(gt, root);
else if (IS_SRIOV_VF(xe))
xe_gt_sriov_vf_debugfs_register(gt, root);
+
+ xe_gt_debugfs_stats_register(gt, root);
}
diff --git a/drivers/gpu/drm/xe/xe_gt_tlb_invalidation.c b/drivers/gpu/drm/xe/xe_gt_tlb_invalidation.c
index 87cb76a8718c..60f9b8bc4121 100644
--- a/drivers/gpu/drm/xe/xe_gt_tlb_invalidation.c
+++ b/drivers/gpu/drm/xe/xe_gt_tlb_invalidation.c
@@ -37,11 +37,13 @@ static long tlb_timeout_jiffies(struct xe_gt *gt)
}
static void
-__invalidation_fence_signal(struct xe_device *xe, struct xe_gt_tlb_invalidation_fence *fence)
+__invalidation_fence_signal(struct xe_gt *gt, struct xe_gt_tlb_invalidation_fence *fence)
{
+ struct xe_device *xe = gt_to_xe(gt);
bool stack = test_bit(FENCE_STACK_BIT, &fence->base.flags);
trace_xe_gt_tlb_invalidation_fence_signal(xe, fence);
+ xe_stats_decrement_counter(gt->stats, "TLB_INVAL_INFLIGHT");
xe_gt_tlb_invalidation_fence_fini(fence);
dma_fence_signal(&fence->base);
if (!stack)
@@ -49,10 +51,10 @@ __invalidation_fence_signal(struct xe_device *xe, struct xe_gt_tlb_invalidation_
}
static void
-invalidation_fence_signal(struct xe_device *xe, struct xe_gt_tlb_invalidation_fence *fence)
+invalidation_fence_signal(struct xe_gt *gt, struct xe_gt_tlb_invalidation_fence *fence)
{
list_del(&fence->link);
- __invalidation_fence_signal(xe, fence);
+ __invalidation_fence_signal(gt, fence);
}
static void xe_gt_tlb_fence_timeout(struct work_struct *work)
@@ -76,7 +78,7 @@ static void xe_gt_tlb_fence_timeout(struct work_struct *work)
fence->seqno, gt->tlb_invalidation.seqno_recv);
fence->base.error = -ETIME;
- invalidation_fence_signal(xe, fence);
+ invalidation_fence_signal(gt, fence);
}
if (!list_empty(>->tlb_invalidation.pending_fences))
queue_delayed_work(system_wq,
@@ -140,7 +142,7 @@ void xe_gt_tlb_invalidation_reset(struct xe_gt *gt)
list_for_each_entry_safe(fence, next,
>->tlb_invalidation.pending_fences, link)
- invalidation_fence_signal(gt_to_xe(gt), fence);
+ invalidation_fence_signal(gt, fence);
spin_unlock_irq(>->tlb_invalidation.pending_lock);
mutex_unlock(>->uc.guc.ct.lock);
}
@@ -183,6 +185,9 @@ static int send_tlb_invalidation(struct xe_guc *guc,
ret = xe_guc_ct_send_locked(&guc->ct, action, len,
G2H_LEN_DW_TLB_INVALIDATE, 1);
if (!ret) {
+ xe_stats_increment_counter(gt->stats, "TLB_INVAL_INFLIGHT");
+ xe_stats_increment_counter(gt->stats, "TLB_INVAL_TOTAL_SENT");
+
spin_lock_irq(>->tlb_invalidation.pending_lock);
/*
* We haven't actually published the TLB fence as per
@@ -191,7 +196,7 @@ static int send_tlb_invalidation(struct xe_guc *guc,
* we can just go ahead and signal the fence here.
*/
if (tlb_invalidation_seqno_past(gt, seqno)) {
- __invalidation_fence_signal(xe, fence);
+ __invalidation_fence_signal(gt, fence);
} else {
fence->invalidation_time = ktime_get();
list_add_tail(&fence->link,
@@ -204,7 +209,7 @@ static int send_tlb_invalidation(struct xe_guc *guc,
}
spin_unlock_irq(>->tlb_invalidation.pending_lock);
} else if (ret < 0) {
- __invalidation_fence_signal(xe, fence);
+ __invalidation_fence_signal(gt, fence);
}
if (!ret) {
gt->tlb_invalidation.seqno = (gt->tlb_invalidation.seqno + 1) %
@@ -321,7 +326,7 @@ int xe_gt_tlb_invalidation_range(struct xe_gt *gt,
/* Execlists not supported */
if (gt_to_xe(gt)->info.force_execlist) {
- __invalidation_fence_signal(xe, fence);
+ __invalidation_fence_signal(gt, fence);
return 0;
}
@@ -455,7 +460,7 @@ int xe_guc_tlb_invalidation_done_handler(struct xe_guc *guc, u32 *msg, u32 len)
if (!tlb_invalidation_seqno_past(gt, fence->seqno))
break;
- invalidation_fence_signal(xe, fence);
+ invalidation_fence_signal(gt, fence);
}
if (!list_empty(>->tlb_invalidation.pending_fences))
@@ -525,3 +530,11 @@ void xe_gt_tlb_invalidation_fence_fini(struct xe_gt_tlb_invalidation_fence *fenc
{
xe_pm_runtime_put(gt_to_xe(fence->gt));
}
+
+void xe_gt_tlb_invalidation_stats_init(struct xe_gt *gt)
+{
+ xe_stats_add_entry(gt->stats, "TLB_INVAL_TOTAL_SENT",
+ XE_STATS_TYPE_COUNTER);
+ xe_stats_add_entry(gt->stats, "TLB_INVAL_INFLIGHT",
+ XE_STATS_TYPE_COUNTER);
+}
diff --git a/drivers/gpu/drm/xe/xe_gt_tlb_invalidation.h b/drivers/gpu/drm/xe/xe_gt_tlb_invalidation.h
index a84065fa324c..62180379cc40 100644
--- a/drivers/gpu/drm/xe/xe_gt_tlb_invalidation.h
+++ b/drivers/gpu/drm/xe/xe_gt_tlb_invalidation.h
@@ -36,4 +36,5 @@ xe_gt_tlb_invalidation_fence_wait(struct xe_gt_tlb_invalidation_fence *fence)
dma_fence_wait(&fence->base, false);
}
+void xe_gt_tlb_invalidation_stats_init(struct xe_gt *gt);
#endif /* _XE_GT_TLB_INVALIDATION_ */
diff --git a/drivers/gpu/drm/xe/xe_gt_types.h b/drivers/gpu/drm/xe/xe_gt_types.h
index 631928258d71..702b21ac611a 100644
--- a/drivers/gpu/drm/xe/xe_gt_types.h
+++ b/drivers/gpu/drm/xe/xe_gt_types.h
@@ -15,6 +15,7 @@
#include "xe_oa.h"
#include "xe_reg_sr_types.h"
#include "xe_sa_types.h"
+#include "xe_stats.h"
#include "xe_uc_types.h"
struct xe_exec_queue_ops;
@@ -133,6 +134,9 @@ struct xe_gt {
u8 has_indirect_ring_state:1;
} info;
+ /* @stats: Track stats of various events, currently TLB inval */
+ struct xe_stats *stats;
+
/**
* @mmio: mmio info for GT. All GTs within a tile share the same
* register space, but have their own copy of GSI registers at a
diff --git a/drivers/gpu/drm/xe/xe_uc_fw.c b/drivers/gpu/drm/xe/xe_uc_fw.c
index c1dcf58d25d7..0e51d853fa4a 100644
--- a/drivers/gpu/drm/xe/xe_uc_fw.c
+++ b/drivers/gpu/drm/xe/xe_uc_fw.c
@@ -258,7 +258,7 @@ uc_fw_override(struct xe_uc_fw *uc_fw)
path_override = xe_modparam.huc_firmware_path;
break;
case XE_UC_FW_TYPE_GSC:
- path_override = xe_modparam.gsc_firmware_path;
+ path_override = "";
break;
default:
break;
--
2.42.0
More information about the Intel-xe
mailing list