[PATCH v13 7/7] drm/xe/pmu: Add GT C6 events
Rodrigo Vivi
rodrigo.vivi at intel.com
Fri Jan 17 18:44:02 UTC 2025
On Thu, Jan 16, 2025 at 03:07:18PM -0800, Lucas De Marchi wrote:
> From: Vinay Belgaumkar <vinay.belgaumkar at intel.com>
>
> Provide a PMU interface for GT C6 residency counters. The implementation
> is ported over from the i915 PMU code. Residency is provided in units of
> ms(like sysfs entry in - /sys/class/drm/card0/device/tile0/gt0/gtidle).
>
> Sample usage and output:
>
> $ perf list | grep gt-c6
> xe_0000_00_02.0/gt-c6-residency/ [Kernel PMU event]
>
> $ tail /sys/bus/event_source/devices/xe_0000_00_02.0/events/gt-c6-residency*
> ==> /sys/bus/event_source/devices/xe_0000_00_02.0/events/gt-c6-residency <==
> event=0x01
>
> ==> /sys/bus/event_source/devices/xe_0000_00_02.0/events/gt-c6-residency.unit <==
> ms
>
> $ perf stat -e xe_0000_00_02.0/gt-c6-residency,gt=0/ -I1000
> # time counts unit events
> 1.001196056 1,001 ms xe_0000_00_02.0/gt-c6-residency,gt=0/
> 2.005216219 1,003 ms xe_0000_00_02.0/gt-c6-residency,gt=0/
>
> Signed-off-by: Vinay Belgaumkar <vinay.belgaumkar at intel.com>
> Signed-off-by: Lucas De Marchi <lucas.demarchi at intel.com>
> ---
>
> Besides the rebase, that changed a lot how the event was added,
> here is a summary of other changes:
>
> - Use xe_pm_runtime_get_if_active() when reading
> xe_gt_idle_residency_msec() as there's not guarantee it will not be
> suspended anymore by the time it reads the counter
>
> - Drop sample[] from the pmu struct and only use the prev/counter from
> the perf_event struct. This avoids mixing the counter reported to 2
> separate clients.
>
> - Drop time ktime helpers and just use what's provided by
> include/linux/ktime.h
>
> drivers/gpu/drm/xe/xe_pmu.c | 56 +++++++++++++++++++++++++++++++------
> 1 file changed, 48 insertions(+), 8 deletions(-)
>
> diff --git a/drivers/gpu/drm/xe/xe_pmu.c b/drivers/gpu/drm/xe/xe_pmu.c
> index c2af82ec3f793..37df9d3cc110c 100644
> --- a/drivers/gpu/drm/xe/xe_pmu.c
> +++ b/drivers/gpu/drm/xe/xe_pmu.c
> @@ -11,6 +11,7 @@
> #include "xe_device.h"
> #include "xe_force_wake.h"
> #include "xe_gt_clock.h"
> +#include "xe_gt_idle.h"
> #include "xe_gt_printk.h"
> #include "xe_mmio.h"
> #include "xe_macros.h"
> @@ -117,16 +118,50 @@ static int xe_pmu_event_init(struct perf_event *event)
> return 0;
> }
>
> -static u64 __xe_pmu_event_read(struct perf_event *event)
> +static u64 read_gt_c6_residency(struct xe_pmu *pmu, struct xe_gt *gt, u64 prev)
> {
> - struct xe_device *xe = container_of(event->pmu, typeof(*xe), pmu.base);
> + struct xe_device *xe = gt_to_xe(gt);
> + unsigned long flags;
> + ktime_t t0;
> + s64 delta;
> +
> + if (xe_pm_runtime_get_if_active(xe)) {
> + u64 val = xe_gt_idle_residency_msec(>->gtidle);
> +
> + xe_pm_runtime_put(xe);
> +
> + return val;
> + }
> +
> + /*
> + * Estimate the idle residency by looking at the time the device was
> + * suspended: should be good enough as long as the sampling frequency is
> + * 2x or more than the suspend frequency.
> + */
> + raw_spin_lock_irqsave(&pmu->lock, flags);
> + t0 = pmu->suspend_timestamp[gt->info.id];
> + raw_spin_unlock_irqrestore(&pmu->lock, flags);
I believe we should simply avoid the patch 4 and this block and use
xe_pm_runtime_get(xe) instead of the if_active variant.
I'm afraid this won't be a good estimative since this mix 2 different wakes
of 2 different IP blocks... the runtime pm is for the entire device, while
in runtime active the GTs can be sleeping in GT-C6.
or did we get some deadlock risk on that scenario?
perhaps we should get runtime pm in upper layers when using PMU?
> +
> + delta = ktime_ms_delta(ktime_get(), t0);
> +
> + return prev + delta;
> +}
> +
> +static u64 __xe_pmu_event_read(struct perf_event *event, u64 prev)
> +{
> + struct xe_pmu *pmu = container_of(event->pmu, typeof(*pmu), base);
> + struct xe_device *xe = container_of(pmu, typeof(*xe), pmu);
> struct xe_gt *gt = event_to_gt(event);
> - u64 val = 0;
>
> if (!gt)
> - return 0;
> + return prev;
> +
> + switch (config_to_event_id(event->attr.config)) {
> + case XE_PMU_EVENT_GT_C6_RESIDENCY:
> + return read_gt_c6_residency(pmu, gt, prev);
> + }
>
> - return val;
> + return prev;
> }
>
> static void xe_pmu_event_update(struct perf_event *event)
> @@ -136,10 +171,11 @@ static void xe_pmu_event_update(struct perf_event *event)
>
> prev = local64_read(&hwc->prev_count);
> do {
> - new = __xe_pmu_event_read(event);
> + new = __xe_pmu_event_read(event, prev);
> } while (!local64_try_cmpxchg(&hwc->prev_count, &prev, new));
>
> - local64_add(new - prev, &event->count);
> + if (new > prev)
> + local64_add(new - prev, &event->count);
> }
>
> static void xe_pmu_event_read(struct perf_event *event)
> @@ -162,7 +198,7 @@ static void xe_pmu_enable(struct perf_event *event)
> * for all listeners. Even when the event was already enabled and has
> * an existing non-zero value.
> */
> - local64_set(&event->hw.prev_count, __xe_pmu_event_read(event));
> + local64_set(&event->hw.prev_count, __xe_pmu_event_read(event, 0));
> }
>
> static void xe_pmu_event_start(struct perf_event *event, int flags)
> @@ -267,6 +303,10 @@ static const struct attribute_group pmu_events_attr_group = {
>
> static void set_supported_events(struct xe_pmu *pmu)
> {
> + struct xe_device *xe = container_of(pmu, typeof(*xe), pmu);
> +
> + if (!xe->info.skip_guc_pc)
> + pmu->supported_events |= BIT_ULL(XE_PMU_EVENT_GT_C6_RESIDENCY);
> }
>
> /**
> --
> 2.48.0
>
More information about the Intel-xe
mailing list