[PATCH v2 2/2] tests/xe/xe_exec_basic: Add section to test zero number of BB in exec IOCTL

Matthew Brost matthew.brost at intel.com
Tue Jul 23 18:39:35 UTC 2024


On Tue, Jul 23, 2024 at 02:27:33PM +0200, Kamil Konieczny wrote:
> Hi Matthew,
> On 2024-07-19 at 12:41:07 -0700, Matthew Brost wrote:
> 
> in subject you use 'tests/xe/':
> [PATCH v2 2/2] tests/xe/xe_exec_basic: Add section to test zero number of BB in exec IOCTL
> 
> imho this should be:
> [PATCH v2 2/2] tests/intel/xe_exec_basic: Add test for zero number of BB
> 

Will change.

Matt

> Regards,
> Kamil
> 
> > Verify the exec queue ordering works wrt out-sync signaling when zero
> > number of BB passed to exec IOCTL.
> > 
> > v2;
> >  - Fix build error
> > 
> > Signed-off-by: Matthew Brost <matthew.brost at intel.com>
> > ---
> >  tests/intel/xe_exec_basic.c | 108 ++++++++++++++++++++++++++++++++++++
> >  1 file changed, 108 insertions(+)
> > 
> > diff --git a/tests/intel/xe_exec_basic.c b/tests/intel/xe_exec_basic.c
> > index 0fd1ae062c..02b78d55f6 100644
> > --- a/tests/intel/xe_exec_basic.c
> > +++ b/tests/intel/xe_exec_basic.c
> > @@ -19,6 +19,7 @@
> >  
> >  #include "xe/xe_ioctl.h"
> >  #include "xe/xe_query.h"
> > +#include "xe/xe_spin.h"
> >  #include <string.h>
> >  
> >  #define MAX_N_EXEC_QUEUES 16
> > @@ -314,6 +315,109 @@ test_exec(int fd, struct drm_xe_engine_class_instance *eci,
> >  	}
> >  }
> >  
> > +/**
> > + * SUBTEST: zero-execs
> > + * Description: Test zero execs in IOCTL
> > + * Functionality: exec IOCTL
> > + * Run type: BAT
> > + */
> > +
> > +static void test_zero_execs(int fd, struct drm_xe_engine_class_instance *eci,
> > +			    int n_execs)
> > +{
> > +	uint32_t vm;
> > +	uint64_t addr = 0x1a0000;
> > +	struct drm_xe_sync sync[2] = {
> > +		{ .flags = DRM_XE_SYNC_TYPE_SYNCOBJ | DRM_XE_SYNC_FLAG_SIGNAL, },
> > +		{ .flags = DRM_XE_SYNC_TYPE_SYNCOBJ | DRM_XE_SYNC_FLAG_SIGNAL, },
> > +	};
> > +	struct drm_xe_exec exec = {
> > +		.num_batch_buffer = 1,
> > +		.num_syncs = 1,
> > +		.syncs = to_user_pointer(sync),
> > +	};
> > +	size_t bo_size;
> > +	uint32_t bo = 0;
> > +	uint32_t syncobj;
> > +	uint32_t exec_queue;
> > +	struct xe_cork cork;
> > +	struct {
> > +		uint32_t batch[16];
> > +		uint64_t pad;
> > +		uint32_t data;
> > +	} *data;
> > +	int i, b;
> > +
> > +	vm = xe_vm_create(fd, 0, 0);
> > +	bo_size = sizeof(*data) * n_execs;
> > +	bo_size = ALIGN(bo_size + xe_cs_prefetch_size(fd),
> > +			xe_get_default_alignment(fd));
> > +	bo = xe_bo_create(fd, vm, bo_size,
> > +			  vram_if_possible(fd, eci->gt_id),
> > +			  DRM_XE_GEM_CREATE_FLAG_NEEDS_VISIBLE_VRAM);
> > +	data = xe_bo_map(fd, bo, bo_size);
> > +	exec_queue = xe_exec_queue_create(fd, vm, eci, 0);
> > +
> > +	xe_cork_init(fd, eci, &cork);
> > +	xe_cork_wait_started(&cork);
> > +
> > +	/* Initial bind behind cork */
> > +	sync[0].handle = syncobj = syncobj_create(fd, 0);
> > +	sync[0].flags |= DRM_XE_SYNC_FLAG_SIGNAL;
> > +	sync[1].handle = cork.syncobj;
> > +	sync[1].flags &= ~DRM_XE_SYNC_FLAG_SIGNAL;
> > +	xe_vm_bind_async(fd, vm, 0, bo, 0, addr, bo_size, sync, 2);
> > +
> > +	/* Exec behind bind */
> > +	for (i = 0; i < n_execs; i++) {
> > +		uint64_t batch_offset = (char *)&data[i].batch - (char *)data;
> > +		uint64_t batch_addr = addr + batch_offset;
> > +		uint64_t sdi_offset = (char *)&data[i].data - (char *)data;
> > +		uint64_t sdi_addr = addr + sdi_offset;
> > +
> > +		b = 0;
> > +		data[i].batch[b++] = MI_STORE_DWORD_IMM_GEN4;
> > +		data[i].batch[b++] = sdi_addr;
> > +		data[i].batch[b++] = sdi_addr >> 32;
> > +		data[i].batch[b++] = 0xc0ffee;
> > +		data[i].batch[b++] = MI_BATCH_BUFFER_END;
> > +		igt_assert(b <= ARRAY_SIZE(data[i].batch));
> > +
> > +		sync[0].flags &= ~DRM_XE_SYNC_FLAG_SIGNAL;
> > +
> > +		exec.exec_queue_id = exec_queue;
> > +		exec.address = batch_addr;
> > +		xe_exec(fd, &exec);
> > +	}
> > +
> > +	/* Exec with no batch buffer */
> > +	sync[0].handle = syncobj_create(fd, 0);
> > +	sync[0].flags |= DRM_XE_SYNC_FLAG_SIGNAL;
> > +	exec.num_batch_buffer = 0;
> > +	exec.address = 0;
> > +	xe_exec(fd, &exec);
> > +
> > +	/* Let jobs runs for a bit */
> > +	usleep(100000);
> > +
> > +	/* both bind and execs are waiting */
> > +	igt_assert(!syncobj_wait(fd, &syncobj, 1, 0, 0, NULL));
> > +	igt_assert(!syncobj_wait(fd, &sync[0].handle, 1, 0, 0, NULL));
> > +
> > +	/* Release cork */
> > +	xe_cork_end(&cork);
> > +	xe_cork_wait_done(&cork);
> > +	xe_cork_fini(&cork);
> > +
> > +	/* both binds are done */
> > +	igt_assert(syncobj_wait(fd, &syncobj, 1, INT64_MAX, 0, NULL));
> > +	igt_assert(syncobj_wait(fd, &sync[0].handle, 1, INT64_MAX, 0, NULL));
> > +
> > +	syncobj_destroy(fd, sync[0].handle);
> > +	gem_close(fd, bo);
> > +	xe_vm_destroy(fd, vm);
> > +}
> > +
> >  igt_main
> >  {
> >  	struct drm_xe_engine_class_instance *hwe;
> > @@ -383,6 +487,10 @@ igt_main
> >  				test_exec(fd, hwe, 1, 0, 1, s->flags);
> >  	}
> >  
> > +	igt_subtest("zero-execs")
> > +		xe_for_each_engine(fd, hwe)
> > +			test_zero_execs(fd, hwe, 1);
> > +
> >  	igt_fixture
> >  		drm_close_driver(fd);
> >  
> > -- 
> > 2.34.1
> > 


More information about the igt-dev mailing list