[PATCH i-g-t] tests/xe: Remove xe_perf_pmu
Umesh Nerlige Ramappa
umesh.nerlige.ramappa at intel.com
Mon Dec 11 22:45:15 UTC 2023
On Mon, Dec 11, 2023 at 08:25:36AM -0800, Ashutosh Dixit wrote:
>Group busyness uapi implemented in PMU in the kernel has been discontinued
>and will be replaced by a different group busyness uapi and a different
>implementation. Because of this, we are proposing to remove the PMU code
>from the kernel, till the correct uapi is implemented. Therefore drop the
>corresponding IGT too, since it caters to the old uapi.
>
>Signed-off-by: Ashutosh Dixit <ashutosh.dixit at intel.com>
Since the corresponding xe implementation will be removed, this is a
required change.
Reviewed-by: Umesh Nerlige Ramappa <umesh.nerlige.ramappa at intel.com>
Thanks,
Umesh
>---
> tests/intel/xe_perf_pmu.c | 331 --------------------------------------
> tests/meson.build | 1 -
> 2 files changed, 332 deletions(-)
> delete mode 100644 tests/intel/xe_perf_pmu.c
>
>diff --git a/tests/intel/xe_perf_pmu.c b/tests/intel/xe_perf_pmu.c
>deleted file mode 100644
>index 42cf62729142..000000000000
>--- a/tests/intel/xe_perf_pmu.c
>+++ /dev/null
>@@ -1,331 +0,0 @@
>-// SPDX-License-Identifier: MIT
>-/*
>- * Copyright © 2021 Intel Corporation
>- */
>-
>-/**
>- * TEST: Basic tests for verify pmu perf interface
>- * Category: Hardware building block
>- * Sub-category: pmu interface
>- * Functionality: pmu
>- * Test category: functionality test
>- */
>-
>-#include <fcntl.h>
>-#include <string.h>
>-
>-#include "igt.h"
>-#include "lib/igt_syncobj.h"
>-#include "lib/intel_reg.h"
>-#include "lib/igt_perf.h"
>-#include "xe_drm.h"
>-#include "xe/xe_ioctl.h"
>-#include "xe/xe_query.h"
>-#include "xe/xe_spin.h"
>-
>-#define MAX_INSTANCE 9
>-
>-static uint64_t pmu_read(int fd)
>-{
>- uint64_t data[2];
>-
>- igt_assert_eq(read(fd, data, sizeof(data)), sizeof(data));
>-
>- return data[0];
>-}
>-
>-static int open_pmu(int fd, uint64_t config)
>-{
>- int perf_fd;
>-
>- perf_fd = perf_xe_open(fd, config);
>- igt_skip_on(perf_fd < 0 && errno == ENODEV);
>- igt_assert(perf_fd >= 0);
>-
>- return perf_fd;
>-}
>-
>-static uint64_t engine_group_get_config(int gt, int class)
>-{
>- uint64_t config;
>-
>- switch (class) {
>- case DRM_XE_ENGINE_CLASS_COPY:
>- config = DRM_XE_PMU_COPY_GROUP_BUSY(gt);
>- break;
>- case DRM_XE_ENGINE_CLASS_RENDER:
>- case DRM_XE_ENGINE_CLASS_COMPUTE:
>- config = DRM_XE_PMU_RENDER_GROUP_BUSY(gt);
>- break;
>- case DRM_XE_ENGINE_CLASS_VIDEO_DECODE:
>- case DRM_XE_ENGINE_CLASS_VIDEO_ENHANCE:
>- config = DRM_XE_PMU_MEDIA_GROUP_BUSY(gt);
>- break;
>- }
>-
>- return config;
>-}
>-
>-/**
>- * Test: Basic test for measure the active time when engine of any class active
>- *
>- * SUBTEST: any-engine-group-busy
>- * Description:
>- * Run a test to measure the global activity time by submitting
>- * the WL to all existing engines.
>- * Run type: FULL
>- *
>- */
>-static void test_any_engine_busyness(int fd, struct drm_xe_engine_class_instance *eci)
>-{
>- uint32_t vm;
>- uint64_t addr = 0x1a0000;
>- struct drm_xe_sync sync[2] = {
>- { .type = DRM_XE_SYNC_TYPE_SYNCOBJ, .flags = DRM_XE_SYNC_FLAG_SIGNAL, },
>- { .type = DRM_XE_SYNC_TYPE_SYNCOBJ, .flags = DRM_XE_SYNC_FLAG_SIGNAL, },
>- };
>- struct drm_xe_exec exec = {
>- .num_batch_buffer = 1,
>- .num_syncs = 2,
>- .syncs = to_user_pointer(sync),
>- };
>- uint32_t exec_queue;
>- uint32_t syncobj;
>- size_t bo_size;
>- uint32_t bo = 0;
>- struct xe_spin *spin;
>- struct xe_spin_opts spin_opts = { .addr = addr, .preempt = false };
>- uint32_t pmu_fd;
>- uint64_t count, idle;
>-
>- vm = xe_vm_create(fd, DRM_XE_VM_CREATE_FLAG_ASYNC_DEFAULT, 0);
>- bo_size = sizeof(*spin);
>- bo_size = ALIGN(bo_size + xe_cs_prefetch_size(fd),
>- xe_get_default_alignment(fd));
>-
>- bo = xe_bo_create(fd, vm, bo_size, vram_if_possible(fd, eci->gt_id), 0);
>- spin = xe_bo_map(fd, bo, bo_size);
>-
>- exec_queue = xe_exec_queue_create(fd, vm, eci, 0);
>- syncobj = syncobj_create(fd, 0);
>-
>- sync[0].handle = syncobj_create(fd, 0);
>- xe_vm_bind_async(fd, vm, 0, bo, 0, addr, bo_size, sync, 1);
>-
>- pmu_fd = open_pmu(fd, DRM_XE_PMU_ANY_ENGINE_GROUP_BUSY(eci->gt_id));
>- idle = pmu_read(pmu_fd);
>- igt_assert(!idle);
>-
>- xe_spin_init(spin, &spin_opts);
>-
>- sync[0].flags &= ~DRM_XE_SYNC_FLAG_SIGNAL;
>- sync[1].flags |= DRM_XE_SYNC_FLAG_SIGNAL;
>- sync[1].handle = syncobj;
>-
>- exec.exec_queue_id = exec_queue;
>- exec.address = addr;
>- xe_exec(fd, &exec);
>-
>- xe_spin_wait_started(spin);
>- usleep(50000);
>-
>- igt_assert(!syncobj_wait(fd, &syncobj, 1, 1, 0, NULL));
>- xe_spin_end(spin);
>-
>- igt_assert(syncobj_wait(fd, &syncobj, 1, INT64_MAX, 0, NULL));
>- igt_assert(syncobj_wait(fd, &sync[0].handle, 1, INT64_MAX, 0, NULL));
>-
>- sync[0].flags |= DRM_XE_SYNC_FLAG_SIGNAL;
>- xe_vm_unbind_async(fd, vm, 0, 0, addr, bo_size, sync, 1);
>- igt_assert(syncobj_wait(fd, &sync[0].handle, 1, INT64_MAX, 0, NULL));
>-
>- syncobj_destroy(fd, sync[0].handle);
>- syncobj_destroy(fd, syncobj);
>-
>- count = pmu_read(pmu_fd);
>- igt_assert_lt_u64(idle, count);
>- igt_debug("Incrementing counter all-busy-group %ld ns\n", count);
>-
>- xe_exec_queue_destroy(fd, exec_queue);
>- munmap(spin, bo_size);
>- gem_close(fd, bo);
>- xe_vm_destroy(fd, vm);
>- close(pmu_fd);
>-}
>-
>-/**
>- * Test: Basic test for measure the active time across engine class
>- *
>- * SUBTEST: render-busy
>- * Description:
>- * Run a test to measure the active engine class time by submitting the
>- * WL to all instances of a class
>- * Run type: FULL
>- *
>- * SUBTEST: compute-busy
>- * Description: Run copy-group-busy test
>- * Run type: FULL
>- *
>- * SUBTEST: copy-busy
>- * Description: Run copy-group-busy test
>- * Run type: FULL
>- *
>- * SUBTEST: vcs-busy
>- * Description: Run copy-group-busy test
>- * Run type: FULL
>- *
>- * SUBTEST: vecs-busy
>- * Description: Run copy-group-busy test
>- * Run type: FULL
>- *
>- */
>-
>-static void test_engine_group_busyness(int fd, int gt, int class, const char *name)
>-{
>- uint32_t vm;
>- uint64_t addr = 0x1a0000;
>- struct drm_xe_sync sync[2] = {
>- { .type = DRM_XE_SYNC_TYPE_SYNCOBJ, .flags = DRM_XE_SYNC_FLAG_SIGNAL, },
>- { .type = DRM_XE_SYNC_TYPE_SYNCOBJ, .flags = DRM_XE_SYNC_FLAG_SIGNAL, },
>- };
>- struct drm_xe_exec exec = {
>- .num_batch_buffer = 1,
>- .num_syncs = 2,
>- .syncs = to_user_pointer(sync),
>- };
>- uint32_t exec_queues[MAX_INSTANCE];
>- uint32_t syncobjs[MAX_INSTANCE];
>- int pmu_fd;
>- size_t bo_size;
>- uint32_t bo = 0, i = 0;
>- struct {
>- struct xe_spin spin;
>- } *data;
>- struct xe_spin_opts spin_opts = { .addr = addr, .preempt = false };
>- struct drm_xe_engine_class_instance *hwe;
>- struct drm_xe_engine_class_instance eci[MAX_INSTANCE];
>- int num_placements = 0;
>- uint64_t config, count, idle;
>-
>- config = engine_group_get_config(gt, class);
>-
>- xe_for_each_engine(fd, hwe) {
>- if (hwe->engine_class != class || hwe->gt_id != gt)
>- continue;
>-
>- eci[num_placements++] = *hwe;
>- }
>-
>- igt_skip_on_f(!num_placements, "Engine class:%d gt:%d not enabled on this platform\n",
>- class, gt);
>-
>- vm = xe_vm_create(fd, DRM_XE_VM_CREATE_FLAG_ASYNC_DEFAULT, 0);
>- bo_size = sizeof(*data) * num_placements;
>- bo_size = ALIGN(bo_size + xe_cs_prefetch_size(fd), xe_get_default_alignment(fd));
>-
>- bo = xe_bo_create(fd, vm, bo_size, vram_if_possible(fd, gt), 0);
>- data = xe_bo_map(fd, bo, bo_size);
>-
>- for (i = 0; i < num_placements; i++) {
>- struct drm_xe_exec_queue_create create = {
>- .vm_id = vm,
>- .width = 1,
>- .num_placements = num_placements,
>- .instances = to_user_pointer(eci),
>- };
>-
>- igt_assert_eq(igt_ioctl(fd, DRM_IOCTL_XE_EXEC_QUEUE_CREATE,
>- &create), 0);
>- exec_queues[i] = create.exec_queue_id;
>- syncobjs[i] = syncobj_create(fd, 0);
>- };
>-
>- sync[0].handle = syncobj_create(fd, 0);
>- xe_vm_bind_async(fd, vm, 0, bo, 0, addr, bo_size, sync, 1);
>-
>- pmu_fd = open_pmu(fd, config);
>- idle = pmu_read(pmu_fd);
>- igt_assert(!idle);
>-
>- for (i = 0; i < num_placements; i++) {
>- spin_opts.addr = addr + (char *)&data[i].spin - (char *)data;
>- xe_spin_init(&data[i].spin, &spin_opts);
>- sync[0].flags &= ~DRM_XE_SYNC_FLAG_SIGNAL;
>- sync[1].flags |= DRM_XE_SYNC_FLAG_SIGNAL;
>- sync[1].handle = syncobjs[i];
>-
>- exec.exec_queue_id = exec_queues[i];
>- exec.address = spin_opts.addr;
>- xe_exec(fd, &exec);
>- xe_spin_wait_started(&data[i].spin);
>- }
>-
>- for (i = 0; i < num_placements; i++) {
>- xe_spin_end(&data[i].spin);
>- igt_assert(syncobj_wait(fd, &syncobjs[i], 1, INT64_MAX, 0,
>- NULL));
>- }
>-
>- igt_assert(syncobj_wait(fd, &sync[0].handle, 1, INT64_MAX, 0, NULL));
>-
>- sync[0].flags |= DRM_XE_SYNC_FLAG_SIGNAL;
>- xe_vm_unbind_async(fd, vm, 0, 0, addr, bo_size, sync, 1);
>- igt_assert(syncobj_wait(fd, &sync[0].handle, 1, INT64_MAX, 0, NULL));
>-
>-
>- syncobj_destroy(fd, sync[0].handle);
>- for (i = 0; i < num_placements; i++) {
>- syncobj_destroy(fd, syncobjs[i]);
>- xe_exec_queue_destroy(fd, exec_queues[i]);
>- }
>-
>- count = pmu_read(pmu_fd);
>- igt_assert_lt_u64(idle, count);
>- igt_debug("Incrementing counter %s-gt-%d %ld ns\n", name, gt, count);
>-
>- munmap(data, bo_size);
>- gem_close(fd, bo);
>- xe_vm_destroy(fd, vm);
>- close(pmu_fd);
>-}
>-
>-igt_main
>-{
>- struct drm_xe_engine_class_instance *hwe;
>- const struct section {
>- const char *name;
>- int class;
>- } sections[] = {
>- { "render-busy", DRM_XE_ENGINE_CLASS_RENDER },
>- { "compute-busy", DRM_XE_ENGINE_CLASS_COMPUTE },
>- { "copy-busy", DRM_XE_ENGINE_CLASS_COPY },
>- { "vcs-busy", DRM_XE_ENGINE_CLASS_VIDEO_DECODE },
>- { "vecs-busy", DRM_XE_ENGINE_CLASS_VIDEO_ENHANCE },
>- { NULL },
>- };
>- int gt;
>- int class;
>- int fd;
>-
>- igt_fixture {
>- fd = drm_open_driver(DRIVER_XE);
>- xe_device_get(fd);
>- }
>-
>- for (const struct section *s = sections; s->name; s++) {
>- igt_subtest_f("%s", s->name)
>- xe_for_each_gt(fd, gt)
>- xe_for_each_engine_class(class)
>- if (class == s->class)
>- test_engine_group_busyness(fd, gt, class, s->name);
>- }
>-
>- igt_subtest("any-engine-group-busy")
>- xe_for_each_engine(fd, hwe)
>- test_any_engine_busyness(fd, hwe);
>-
>- igt_fixture {
>- xe_device_put(fd);
>- close(fd);
>- }
>-}
>diff --git a/tests/meson.build b/tests/meson.build
>index f6cfbcf5ef9f..4430ca739275 100644
>--- a/tests/meson.build
>+++ b/tests/meson.build
>@@ -305,7 +305,6 @@ intel_xe_progs = [
> 'xe_pm',
> 'xe_pm_residency',
> 'xe_prime_self_import',
>- 'xe_perf_pmu',
> 'xe_query',
> 'xe_sysfs_tile',
> 'xe_uevent',
>--
>2.41.0
>
More information about the igt-dev
mailing list