[igt-dev] [PATCH v2 1/1] tests/xe/perf_pmu: Tests for the XE pmu interface
Belgaumkar, Vinay
vinay.belgaumkar at intel.com
Wed Sep 20 23:10:44 UTC 2023
On 7/4/2023 6:55 AM, Venkata Ramana Nayana wrote:
> There are set of engine group busyness counters provided by HW which are
> exposed via PMU events. Adding a basic unit tests to read those counters.
>
> v2: Added idle condition checks while reading the counters. (Rahul)
This series needs to be re-based and re-compiled. Xe tests have moved to
within tests/intel now.
Thanks,
Vinay.
>
> Cc: Janga Rahul Kumar <janga.rahul.kumar at intel.com>
> Signed-off-by: Venkata Ramana Nayana <venkata.ramana.nayana at intel.com>
> ---
> include/drm-uapi/xe_drm.h | 22 +++
> lib/igt_perf.c | 36 +++++
> lib/igt_perf.h | 5 +
> tests/meson.build | 1 +
> tests/xe/xe_perf_pmu.c | 331 ++++++++++++++++++++++++++++++++++++++
> 5 files changed, 395 insertions(+)
> create mode 100644 tests/xe/xe_perf_pmu.c
>
> diff --git a/include/drm-uapi/xe_drm.h b/include/drm-uapi/xe_drm.h
> index 432bd87ca..81dae10de 100644
> --- a/include/drm-uapi/xe_drm.h
> +++ b/include/drm-uapi/xe_drm.h
> @@ -732,6 +732,28 @@ struct drm_xe_engine_create {
> __u64 reserved[2];
> };
>
> +/**
> + * DOC: perf_events exposed by xe through /sys/bus/event_sources/drivers/xe
> + *
> + */
> +
> +
> +/* PMU event config IDs */
> +
> +/*
> + * Top 4 bits of every counter are GT id.
> + */
> +#define __XE_PMU_GT_SHIFT (60)
> +
> +#define ___XE_PMU_OTHER(gt, x) \
> + (((__u64)(x)) | ((__u64)(gt) << __XE_PMU_GT_SHIFT))
> +
> +#define XE_PMU_INTERRUPTS(gt) ___XE_PMU_OTHER(gt, 0)
> +#define XE_PMU_RENDER_GROUP_BUSY(gt) ___XE_PMU_OTHER(gt, 1)
> +#define XE_PMU_COPY_GROUP_BUSY(gt) ___XE_PMU_OTHER(gt, 2)
> +#define XE_PMU_MEDIA_GROUP_BUSY(gt) ___XE_PMU_OTHER(gt, 3)
> +#define XE_PMU_ANY_ENGINE_GROUP_BUSY(gt) ___XE_PMU_OTHER(gt, 4)
> +
> struct drm_xe_engine_get_property {
> /** @extensions: Pointer to the first extension struct, if any */
> __u64 extensions;
> diff --git a/lib/igt_perf.c b/lib/igt_perf.c
> index ffe078adc..3866c6d77 100644
> --- a/lib/igt_perf.c
> +++ b/lib/igt_perf.c
> @@ -69,6 +69,36 @@ const char *i915_perf_device(int i915, char *buf, int buflen)
> return buf;
> }
>
> +const char *xe_perf_device(int xe, char *buf, int buflen)
> +{
> + char *s;
> + char pref[] = "xe_";
> + int len = strlen(pref);
> +
> +
> + if (!buf || buflen < len)
> + return "xe";
> +
> + memcpy(buf, pref, len);
> +
> + if (!bus_address(xe, buf + len, buflen - len))
> + buf[len - 1] = '\0';
> +
> + /* Convert all colons in the address to '_', thanks perf! */
> + for (s = buf; *s; s++)
> + if (*s == ':')
> + *s = '_';
> +
> + return buf;
> +}
> +
> +uint64_t xe_perf_type_id(int xe)
> +{
> + char buf[80];
> +
> + return igt_perf_type_id(xe_perf_device(xe, buf, sizeof(buf)));
> +}
> +
> uint64_t i915_perf_type_id(int i915)
> {
> char buf[80];
> @@ -147,6 +177,12 @@ int perf_igfx_open_group(uint64_t config, int group)
> PERF_FORMAT_TOTAL_TIME_ENABLED | PERF_FORMAT_GROUP);
> }
>
> +int perf_xe_open(int xe, uint64_t config)
> +{
> + return _perf_open(xe_perf_type_id(xe), config, -1,
> + PERF_FORMAT_TOTAL_TIME_ENABLED);
> +}
> +
> int perf_i915_open(int i915, uint64_t config)
> {
> return _perf_open(i915_perf_type_id(i915), config, -1,
> diff --git a/lib/igt_perf.h b/lib/igt_perf.h
> index 4d86e31ae..3d9ba2917 100644
> --- a/lib/igt_perf.h
> +++ b/lib/igt_perf.h
> @@ -61,10 +61,15 @@ int igt_perf_open_group(uint64_t type, uint64_t config, int group);
> const char *i915_perf_device(int i915, char *buf, int buflen);
> uint64_t i915_perf_type_id(int i915);
>
> +const char *xe_perf_device(int xe, char *buf, int buflen);
> +uint64_t xe_perf_type_id(int);
> +
> int perf_igfx_open(uint64_t config);
> int perf_igfx_open_group(uint64_t config, int group);
>
> int perf_i915_open(int i915, uint64_t config);
> int perf_i915_open_group(int i915, uint64_t config, int group);
>
> +int perf_xe_open(int xe, uint64_t config);
> +
> #endif /* I915_PERF_H */
> diff --git a/tests/meson.build b/tests/meson.build
> index ee066b849..115bcf3ff 100644
> --- a/tests/meson.build
> +++ b/tests/meson.build
> @@ -267,6 +267,7 @@ xe_progs = [
> 'xe_noexec_ping_pong',
> 'xe_pm',
> 'xe_prime_self_import',
> + 'xe_perf_pmu',
> 'xe_query',
> 'xe_vm',
> 'xe_waitfence',
> diff --git a/tests/xe/xe_perf_pmu.c b/tests/xe/xe_perf_pmu.c
> new file mode 100644
> index 000000000..1fc940338
> --- /dev/null
> +++ b/tests/xe/xe_perf_pmu.c
> @@ -0,0 +1,331 @@
> +// SPDX-License-Identifier: MIT
> +/*
> + * Copyright © 2021 Intel Corporation
> + */
> +
> +/**
> + * TEST: Basic tests for verify pmu perf interface
> + * Category: Hardware building block
> + * Sub-category: pmu interface
> + * Functionality: pmu
> + * Test category: functionality test
> + */
> +
> +#include <fcntl.h>
> +#include <string.h>
> +
> +#include "igt.h"
> +#include "lib/igt_syncobj.h"
> +#include "lib/intel_reg.h"
> +#include "lib/igt_perf.h"
> +#include "xe_drm.h"
> +#include "xe/xe_ioctl.h"
> +#include "xe/xe_query.h"
> +#include "xe/xe_spin.h"
> +
> +#define MAX_INSTANCE 9
> +
> +static uint64_t pmu_read(int fd)
> +{
> + uint64_t data[2];
> +
> + igt_assert_eq(read(fd, data, sizeof(data)), sizeof(data));
> +
> + return data[0];
> +}
> +
> +static int open_pmu(int fd, uint64_t config)
> +{
> + int perf_fd;
> +
> + perf_fd = perf_xe_open(fd, config);
> + igt_skip_on(perf_fd < 0 && errno == ENODEV);
> + igt_assert(perf_fd >= 0);
> +
> + return perf_fd;
> +}
> +
> +static uint64_t engine_group_get_config(int gt, int class)
> +{
> + uint64_t config;
> +
> + switch (class) {
> + case DRM_XE_ENGINE_CLASS_COPY:
> + config = XE_PMU_COPY_GROUP_BUSY(gt);
> + break;
> + case DRM_XE_ENGINE_CLASS_RENDER:
> + case DRM_XE_ENGINE_CLASS_COMPUTE:
> + config = XE_PMU_RENDER_GROUP_BUSY(gt);
> + break;
> + case DRM_XE_ENGINE_CLASS_VIDEO_DECODE:
> + case DRM_XE_ENGINE_CLASS_VIDEO_ENHANCE:
> + config = XE_PMU_MEDIA_GROUP_BUSY(gt);
> + break;
> + }
> +
> + return config;
> +}
> +
> +/**
> + * Test: Basic test for measure the active time when engine of any class active
> + *
> + * SUBTEST: any-engine-group-busy
> + * Description:
> + * Run a test to measure the global activity time by submitting
> + * the WL to all existing engines.
> + * Run type: FULL
> + *
> + */
> +static void test_any_engine_busyness(int fd, struct drm_xe_engine_class_instance *eci)
> +{
> + uint32_t vm;
> + uint64_t addr = 0x1a0000;
> + struct drm_xe_sync sync[2] = {
> + { .flags = DRM_XE_SYNC_SYNCOBJ | DRM_XE_SYNC_SIGNAL, },
> + { .flags = DRM_XE_SYNC_SYNCOBJ | DRM_XE_SYNC_SIGNAL, },
> + };
> + struct drm_xe_exec exec = {
> + .num_batch_buffer = 1,
> + .num_syncs = 2,
> + .syncs = to_user_pointer(sync),
> + };
> + uint32_t engine;
> + uint32_t syncobj;
> + size_t bo_size;
> + uint32_t bo = 0;
> + struct xe_spin *spin;
> + uint32_t pmu_fd;
> + uint64_t count, idle;
> +
> + vm = xe_vm_create(fd, DRM_XE_VM_CREATE_ASYNC_BIND_OPS, 0);
> + bo_size = sizeof(*spin);
> + bo_size = ALIGN(bo_size + xe_cs_prefetch_size(fd),
> + xe_get_default_alignment(fd));
> +
> + bo = xe_bo_create(fd, eci->gt_id, vm, bo_size);
> + spin = xe_bo_map(fd, bo, bo_size);
> +
> + engine = xe_engine_create(fd, vm, eci, 0);
> + syncobj = syncobj_create(fd, 0);
> +
> + sync[0].handle = syncobj_create(fd, 0);
> + xe_vm_bind_async(fd, vm, 0, bo, 0, addr, bo_size, sync, 1);
> +
> + pmu_fd = open_pmu(fd, XE_PMU_ANY_ENGINE_GROUP_BUSY(eci->gt_id));
> + idle = pmu_read(pmu_fd);
> + igt_assert(!idle);
> +
> + xe_spin_init(spin, addr, false);
> +
> + sync[0].flags &= ~DRM_XE_SYNC_SIGNAL;
> + sync[1].flags |= DRM_XE_SYNC_SIGNAL;
> + sync[1].handle = syncobj;
> +
> + exec.engine_id = engine;
> + exec.address = addr;
> + xe_exec(fd, &exec);
> +
> + xe_spin_wait_started(spin);
> + usleep(50000);
> +
> + igt_assert(!syncobj_wait(fd, &syncobj, 1, 1, 0, NULL));
> + xe_spin_end(spin);
> +
> + igt_assert(syncobj_wait(fd, &syncobj, 1, INT64_MAX, 0, NULL));
> + igt_assert(syncobj_wait(fd, &sync[0].handle, 1, INT64_MAX, 0, NULL));
> +
> + sync[0].flags |= DRM_XE_SYNC_SIGNAL;
> + xe_vm_unbind_async(fd, vm, 0, 0, addr, bo_size, sync, 1);
> + igt_assert(syncobj_wait(fd, &sync[0].handle, 1, INT64_MAX, 0, NULL));
> +
> + syncobj_destroy(fd, sync[0].handle);
> + syncobj_destroy(fd, syncobj);
> +
> + count = pmu_read(pmu_fd);
> + igt_assert_lt_u64(idle, count);
> + igt_debug("Incrementing counter all-busy-group %ld ns\n", count);
> +
> + xe_engine_destroy(fd, engine);
> + munmap(spin, bo_size);
> + gem_close(fd, bo);
> + xe_vm_destroy(fd, vm);
> + close(pmu_fd);
> +}
> +
> +/**
> + * Test: Basic test for measure the active time across engine class
> + *
> + * SUBTEST: render-busy
> + * Description:
> + * Run a test to measure the active engine class time by submitting the
> + * WL to all instances of a class
> + * Run type: FULL
> + *
> + * SUBTEST: compute-busy
> + * Description: Run copy-group-busy test
> + * Run type: FULL
> + *
> + * SUBTEST: copy-busy
> + * Description: Run copy-group-busy test
> + * Run type: FULL
> + *
> + * SUBTEST: vcs-busy
> + * Description: Run copy-group-busy test
> + * Run type: FULL
> + *
> + * SUBTEST: vecs-busy
> + * Description: Run copy-group-busy test
> + * Run type: FULL
> + *
> + */
> +
> +static void test_engine_group_busyness(int fd, int gt, int class, const char *name)
> +{
> + uint32_t vm;
> + uint64_t addr = 0x1a0000;
> + struct drm_xe_sync sync[2] = {
> + { .flags = DRM_XE_SYNC_SYNCOBJ | DRM_XE_SYNC_SIGNAL, },
> + { .flags = DRM_XE_SYNC_SYNCOBJ | DRM_XE_SYNC_SIGNAL, },
> + };
> + struct drm_xe_exec exec = {
> + .num_batch_buffer = 1,
> + .num_syncs = 2,
> + .syncs = to_user_pointer(sync),
> + };
> + uint32_t engines[MAX_INSTANCE];
> + uint32_t syncobjs[MAX_INSTANCE];
> + int pmu_fd;
> + size_t bo_size;
> + uint32_t bo = 0, i = 0;
> + struct {
> + struct xe_spin spin;
> + } *data;
> + struct drm_xe_engine_class_instance *hwe;
> + struct drm_xe_engine_class_instance eci[MAX_INSTANCE];
> + int num_placements = 0;
> + uint64_t config, count, idle;
> +
> + config = engine_group_get_config(gt, class);
> +
> + xe_for_each_hw_engine(fd, hwe) {
> + if (hwe->engine_class != class || hwe->gt_id != gt)
> + continue;
> +
> + eci[num_placements++] = *hwe;
> + }
> +
> + igt_skip_on_f(!num_placements, "Engine class:%d gt:%d not enabled on this platform\n",
> + class, gt);
> +
> + vm = xe_vm_create(fd, DRM_XE_VM_CREATE_ASYNC_BIND_OPS, 0);
> + bo_size = sizeof(*data) * num_placements;
> + bo_size = ALIGN(bo_size + xe_cs_prefetch_size(fd), xe_get_default_alignment(fd));
> +
> + bo = xe_bo_create(fd, gt, vm, bo_size);
> + data = xe_bo_map(fd, bo, bo_size);
> +
> + for (i = 0; i < num_placements; i++) {
> + struct drm_xe_engine_create create = {
> + .vm_id = vm,
> + .width = 1,
> + .num_placements = num_placements,
> + .instances = to_user_pointer(eci),
> + };
> +
> + igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_XE_ENGINE_CREATE,
> + &create), 0);
> + engines[i] = create.engine_id;
> + syncobjs[i] = syncobj_create(fd, 0);
> + };
> +
> + sync[0].handle = syncobj_create(fd, 0);
> + xe_vm_bind_async(fd, vm, 0, bo, 0, addr, bo_size, sync, 1);
> +
> + pmu_fd = open_pmu(fd, config);
> + idle = pmu_read(pmu_fd);
> + igt_assert(!idle);
> +
> + for (i = 0; i < num_placements; i++) {
> + uint64_t spin_offset = (char *)&data[i].spin - (char *)data;
> + uint64_t spin_addr = addr + spin_offset;
> +
> + xe_spin_init(&data[i].spin, spin_addr, false);
> + sync[0].flags &= ~DRM_XE_SYNC_SIGNAL;
> + sync[1].flags |= DRM_XE_SYNC_SIGNAL;
> + sync[1].handle = syncobjs[i];
> +
> + exec.engine_id = engines[i];
> + exec.address = spin_addr;
> + xe_exec(fd, &exec);
> + xe_spin_wait_started(&data[i].spin);
> + }
> +
> + for (i = 0; i < num_placements; i++) {
> + xe_spin_end(&data[i].spin);
> + igt_assert(syncobj_wait(fd, &syncobjs[i], 1, INT64_MAX, 0,
> + NULL));
> + }
> +
> + igt_assert(syncobj_wait(fd, &sync[0].handle, 1, INT64_MAX, 0, NULL));
> +
> + sync[0].flags |= DRM_XE_SYNC_SIGNAL;
> + xe_vm_unbind_async(fd, vm, 0, 0, addr, bo_size, sync, 1);
> + igt_assert(syncobj_wait(fd, &sync[0].handle, 1, INT64_MAX, 0, NULL));
> +
> +
> + syncobj_destroy(fd, sync[0].handle);
> + for (i = 0; i < num_placements; i++) {
> + syncobj_destroy(fd, syncobjs[i]);
> + xe_engine_destroy(fd, engines[i]);
> + }
> +
> + count = pmu_read(pmu_fd);
> + igt_assert_lt_u64(idle, count);
> + igt_debug("Incrementing counter %s-gt-%d %ld ns\n", name, gt, count);
> +
> + munmap(data, bo_size);
> + gem_close(fd, bo);
> + xe_vm_destroy(fd, vm);
> + close(pmu_fd);
> +}
> +
> +igt_main
> +{
> + struct drm_xe_engine_class_instance *hwe;
> + const struct section {
> + const char *name;
> + int class;
> + } sections[] = {
> + { "render-busy", DRM_XE_ENGINE_CLASS_RENDER },
> + { "compute-busy", DRM_XE_ENGINE_CLASS_COMPUTE },
> + { "copy-busy", DRM_XE_ENGINE_CLASS_COPY },
> + { "vcs-busy", DRM_XE_ENGINE_CLASS_VIDEO_DECODE },
> + { "vecs-busy", DRM_XE_ENGINE_CLASS_VIDEO_ENHANCE },
> + { NULL },
> + };
> + int gt;
> + int class;
> + int fd;
> +
> + igt_fixture {
> + fd = drm_open_driver(DRIVER_XE);
> + xe_device_get(fd);
> + }
> +
> + for (const struct section *s = sections; s->name; s++) {
> + igt_subtest_f("%s", s->name)
> + xe_for_each_gt(fd, gt)
> + xe_for_each_hw_engine_class(class)
> + if (class == s->class)
> + test_engine_group_busyness(fd, gt, class, s->name);
> + }
> +
> + igt_subtest("any-engine-group-busy")
> + xe_for_each_hw_engine(fd, hwe)
> + test_any_engine_busyness(fd, hwe);
> +
> + igt_fixture {
> + xe_device_put(fd);
> + close(fd);
> + }
> +}
More information about the igt-dev
mailing list