[PATCH i-g-t] tests/intel/xe_exec_compute_mode: Add non-blocking subtest
Dandamudi, Priyanka
priyanka.dandamudi at intel.com
Mon Jan 8 07:56:13 UTC 2024
> -----Original Message-----
> From: Dandamudi, Priyanka
> Sent: Monday, January 8, 2024 1:12 PM
> To: sai.gowtham.ch at intel.com; igt-dev at lists.freedesktop.org; Brost, Matthew
> <matthew.brost at intel.com>; Ch, Sai Gowtham <sai.gowtham.ch at intel.com>
> Subject: RE: [PATCH i-g-t] tests/intel/xe_exec_compute_mode: Add non-
> blocking subtest
>
>
>
> > -----Original Message-----
> > From: igt-dev <igt-dev-bounces at lists.freedesktop.org> On Behalf Of
> > sai.gowtham.ch at intel.com
> > Sent: Saturday, January 6, 2024 2:12 AM
> > To: igt-dev at lists.freedesktop.org; Brost, Matthew
> > <matthew.brost at intel.com>; Ch, Sai Gowtham
> <sai.gowtham.ch at intel.com>
> > Subject: [PATCH i-g-t] tests/intel/xe_exec_compute_mode: Add
> > non-blocking subtest
> >
> > From: Sai Gowtham Ch <sai.gowtham.ch at intel.com>
> >
> > Fill the ring with maximum workload and expecte kernel to return -
> > EWOULDBLOCK error.
> >
> > v2: Fixing batch buffer submission, updating sync mapping addresses
> > of VM, SPIN & EXEC for better readability. (Matthew Brost)
> >
> > Signed-off-by: Matthew Brost <matthew.brost at intel.com>
> > Signed-off-by: Sai Gowtham Ch <sai.gowtham.ch at intel.com>
> > ---
> > tests/intel/xe_exec_compute_mode.c | 111
> > +++++++++++++++++++++++++++++
> > 1 file changed, 111 insertions(+)
> >
> > diff --git a/tests/intel/xe_exec_compute_mode.c
> > b/tests/intel/xe_exec_compute_mode.c
> > index b935ee0bc..6ebfa1a86 100644
> > --- a/tests/intel/xe_exec_compute_mode.c
> > +++ b/tests/intel/xe_exec_compute_mode.c
> > @@ -19,6 +19,7 @@
> >
> > #include "xe/xe_ioctl.h"
> > #include "xe/xe_query.h"
> > +#include "xe/xe_spin.h"
> > #include <string.h>
> >
> > #define MAX_N_EXECQUEUES 16
> > @@ -289,6 +290,112 @@ test_exec(int fd, struct
> > drm_xe_engine_class_instance *eci,
> > close(map_fd);
> > }
> >
> > +/**
> > + * SUBTEST: non-blocking
> > + * Description: Fill the ring and check we get expected errors.
> > + * Test category: functionality test
> > + */
> > +#define ONE_SEC MS_TO_NS(1000)
> > +#define VM_DATA 0
> > +#define SPIN_DATA 1
> > +#define EXEC_DATA 2
> > +#define DATA_COUNT 3
> > +static void non_block(int fd, int expect) {
> > + uint64_t addr = 0x1a0000;
> > + struct drm_xe_sync sync[1] = {
> > + { .type = DRM_XE_SYNC_TYPE_USER_FENCE,
> > + .flags = DRM_XE_SYNC_FLAG_SIGNAL,
> > + .timeline_value = USER_FENCE_VALUE},
> > + };
> > + struct drm_xe_exec exec = {
> > + .num_batch_buffer = 1,
> > + .num_syncs = 1,
> > + .syncs = to_user_pointer(&sync),
> > + };
> > + struct {
> > + struct xe_spin spin;
> > + uint32_t batch[16];
> > + uint64_t vm_sync;
> > + uint32_t data;
> > + uint64_t exec_sync;
> > + } *data = NULL;
> > + struct xe_spin_opts spin_opts = { .addr = addr, .preempt = false };
> > + struct drm_xe_engine *engine;
> > + uint64_t batch_offset = (char *)&data[EXEC_DATA].batch - (char
> > *)data;
> > + uint64_t batch_addr = addr + batch_offset;
> > + uint64_t sdi_offset = (char *)&data[EXEC_DATA].data - (char *)data;
> > + uint64_t sdi_addr = addr + sdi_offset;
> > + size_t bo_size;
> > + uint32_t vm;
> > + uint32_t exec_queue;
> > + int value = 0x123456;
> > + uint32_t bo = 0;
> > + int b, err;
> > +
> > + vm = xe_vm_create(fd, DRM_XE_VM_CREATE_FLAG_LR_MODE, 0);
> > + bo_size = sizeof(*data) * DATA_COUNT;
> > + bo_size = ALIGN(bo_size + xe_cs_prefetch_size(fd),
> > +xe_get_default_alignment(fd));
> > +
> > + engine = xe_engine(fd, 1);
> > + bo = xe_bo_create(fd, vm, bo_size, vram_if_possible(fd,
> > +engine->instance.gt_id), 0);
> > +
> > + data = xe_bo_map(fd, bo, bo_size);
> > + memset(data, 0, bo_size);
> > + exec_queue = xe_exec_queue_create(fd, vm, &engine->instance, 0);
> > +
> > + sync[0].addr = to_user_pointer(&data[VM_DATA].vm_sync);
> > + xe_vm_bind_async(fd, vm, engine->instance.gt_id, bo, 0,
> > + addr, bo_size, sync, 1);
> > + xe_wait_ufence(fd, &data[VM_DATA].vm_sync, USER_FENCE_VALUE,
> > 0, ONE_SEC);
> > + data[VM_DATA].vm_sync = 0;
> > +
> > + xe_spin_init(&data[SPIN_DATA].spin, &spin_opts);
> > + sync[0].addr = addr + (char *)&data[SPIN_DATA].exec_sync - (char
> > *)data;
> > + exec.exec_queue_id = exec_queue;
> > + exec.address = spin_opts.addr;
> > + xe_exec(fd, &exec);
> > + xe_spin_wait_started(&data[SPIN_DATA].spin);
> > +
> > + b = 0;
> > + data[0].batch[b++] = MI_STORE_DWORD_IMM_GEN4;
> > + data[0].batch[b++] = sdi_addr;
> > + data[0].batch[b++] = sdi_addr >> 32;
> > + data[0].batch[b++] = value;
> > + data[0].batch[b++] = MI_BATCH_BUFFER_END;
> > + igt_assert(b <= ARRAY_SIZE(data->batch));
> > +
> > + sync[0].addr = addr + (char *)&data[EXEC_DATA].exec_sync - (char
> > *)data;
> > + exec.num_syncs = 0;
> > + exec.exec_queue_id = exec_queue;
> > + exec.address = batch_addr;
> > +
> > + while (1) {
> > + err = __xe_exec(fd, &exec);
> > +
> > + if (err == -EWOULDBLOCK)
> > + break;
> > + }
> > + igt_assert_eq(err, expect);
> > +
> > + xe_spin_end(&data[SPIN_DATA].spin);
> > + xe_wait_ufence(fd, &data[SPIN_DATA].exec_sync,
> > USER_FENCE_VALUE, 0, ONE_SEC);
> > + usleep(50000);
> > +
> > + exec.num_syncs = 1;
> > + xe_exec(fd, &exec);
> > + xe_wait_ufence(fd, &data[EXEC_DATA].exec_sync,
> > USER_FENCE_VALUE, 0,
> > +ONE_SEC);
> > +
> > + sync[0].addr = to_user_pointer(&data[VM_DATA].vm_sync);
> > + xe_vm_unbind_async(fd, vm, 0, 0, addr, bo_size, sync, 1);
> > + xe_wait_ufence(fd, &data[VM_DATA].vm_sync, USER_FENCE_VALUE,
> > 0, ONE_SEC);
> > + munmap(data, bo_size);
> > + gem_close(fd, bo);
> > +
> > + xe_exec_queue_destroy(fd, exec_queue);
> > + xe_vm_destroy(fd, vm);
> > +}
> > +
> > igt_main
> > {
> > struct drm_xe_engine_class_instance *hwe; @@ -346,6 +453,10 @@
> > igt_main
> > s->flags);
> > }
> >
> > + igt_subtest("non-blocking")
> > + non_block(fd, -EWOULDBLOCK);
> > +
> LGTM.
> Reviewed-by: Priyanka Dandamudi <priyanka.dandamudi at intel.com>
> > +
> > igt_fixture
> > drm_close_driver(fd);
> > }
> > --
> > 2.39.1
More information about the igt-dev
mailing list