[igt-dev] [PATCH i-g-t] tests/xe: Add a test for corner cases of eviction.

Maarten Lankhorst maarten.lankhorst at linux.intel.com
Tue May 30 14:55:40 UTC 2023


Cc

On 2023-05-29 14:48, Maarten Lankhorst wrote:
> Gitlab issues 194 and 239 show some corner cases. In particular,
> signal handling, pt eviction and losing track of external objects
> on the list.
>
> The patch series to fix those was sent as
> https://patchwork.freedesktop.org/series/118428/
>
> But because it's tricky to reproduce those issues, I had to add some
> custom igt's to handle those.
>
> Signed-off-by: Maarten Lankhorst <maarten.lankhorst at linux.intel.com>
> ---
>  tests/meson.build     |   1 +
>  tests/xe/xe_evicted.c | 313 ++++++++++++++++++++++++++++++++++++++++++
>  2 files changed, 314 insertions(+)
>  create mode 100644 tests/xe/xe_evicted.c
>
> diff --git a/tests/meson.build b/tests/meson.build
> index 6551194fe..166937b6c 100644
> --- a/tests/meson.build
> +++ b/tests/meson.build
> @@ -248,6 +248,7 @@ xe_progs = [
>  	'xe_dma_buf_sync',
>  	'xe_debugfs',
>  	'xe_evict',
> +	'xe_evicted',
>  	'xe_exec_balancer',
>  	'xe_exec_basic',
>  	'xe_exec_compute_mode',
> diff --git a/tests/xe/xe_evicted.c b/tests/xe/xe_evicted.c
> new file mode 100644
> index 000000000..3f5e6a82c
> --- /dev/null
> +++ b/tests/xe/xe_evicted.c
> @@ -0,0 +1,313 @@
> +// SPDX-License-Identifier: MIT
> +/*
> + * Copyright © 2023 Intel Corporation
> + */
> +
> +/**
> + * TEST: Test various eviction and ENOMEM corner cases.
> + * Category: Software building block
> + * Sub-category: VM_BIND / eviction
> + * Test category: functionality test
> + */
> +
> +#include "igt.h"
> +#include "lib/igt_syncobj.h"
> +#include "lib/intel_reg.h"
> +#include "xe_drm.h"
> +
> +#include "xe/xe_ioctl.h"
> +#include "xe/xe_query.h"
> +#include "xe/xe_spin.h"
> +#include <string.h>
> +
> +static void evict_mem(struct xe_device *xe)
> +{
> +	if (xe->has_vram)
> +		igt_debugfs_write(xe->fd, "vram0_evict", "1");
> +	else
> +		igt_debugfs_write(xe->fd, "gtt_evict", "1");
> +}
> +
> +static void execstress(struct xe_device *xe, uint32_t engine, uint32_t addr)
> +{
> +	struct timespec start, end;
> +
> +	igt_fork(child, 2) {
> +		struct drm_xe_sync sync = {
> +			.flags = DRM_XE_SYNC_SYNCOBJ | DRM_XE_SYNC_SIGNAL,
> +			.handle = syncobj_create(xe->fd, 0),
> +		};
> +
> +		clock_gettime(CLOCK_MONOTONIC, &start);
> +
> +		do {
> +			if (!child)
> +				evict_mem(xe);
> +			else
> +				xe_exec_sync(xe->fd, engine, addr, &sync, 1);
> +
> +			clock_gettime(CLOCK_MONOTONIC, &end);
> +		} while (end.tv_sec - start.tv_sec <= 3);
> +
> +		if (child)
> +			igt_assert(syncobj_wait(xe->fd, &sync.handle, 1, INT64_MAX, 0, NULL));
> +		syncobj_destroy(xe->fd, sync.handle);
> +	}
> +
> +	igt_waitchildren();
> +
> +	/* Leave function in a known state */
> +	evict_mem(xe);
> +}
> +
> +/* See issue 194, unbinding first VMA may cause BO not to be locked correctly */
> +static void test_multimap(struct xe_device *xe, bool external)
> +{
> +	uint64_t size = xe->default_alignment;
> +	uint32_t vm = xe_vm_create(xe->fd, DRM_XE_VM_CREATE_SCRATCH_PAGE, 0);
> +	uint32_t bo = xe_bo_create(xe->fd, 0, external ? 0 : vm, size);
> +	uint32_t engine = xe_engine_create_class(xe->fd, vm, DRM_XE_ENGINE_CLASS_COPY);
> +	uint32_t *map = xe_bo_map(xe->fd, bo, size);
> +
> +	*map = MI_BATCH_BUFFER_END;
> +	munmap(map, size);
> +
> +	/* Create 2 mappings */
> +	xe_vm_bind_sync(xe->fd, vm, bo, 0, 0, size);
> +	xe_vm_bind_sync(xe->fd, vm, bo, 0, size, size);
> +
> +	execstress(xe, engine, size);
> +
> +	/* Unbind the first one, to possibly confuse stuff */
> +	xe_vm_unbind_sync(xe->fd, vm, 0, 0, size);
> +
> +	execstress(xe, engine, size);
> +
> +	/* Teardown.. */
> +	xe_engine_destroy(xe->fd, engine);
> +	xe_vm_destroy(xe->fd, vm);
> +	gem_close(xe->fd, bo);
> +}
> +
> +static uint64_t avail_vram(struct xe_device **xe)
> +{
> +	int fd = (*xe)->fd;
> +	int region_idx;
> +
> +	/* Refresh to get up-to-date values */
> +	xe_device_put(fd);
> +	*xe = xe_device_get(fd);
> +
> +	region_idx = ffs((*xe)->gts->gts[0].native_mem_regions) - 1;
> +
> +	return (*xe)->mem_usage->regions[region_idx].total_size -
> +	       (*xe)->mem_usage->regions[region_idx].used;
> +}
> +
> +static void test_pt_eviction(struct xe_device *xe, bool external)
> +{
> +	uint32_t bo, i;
> +	uint64_t size, full_size, full_addr, addr, avail;
> +	uint32_t vm;
> +	uint32_t engine;
> +	uint32_t *map;
> +	struct drm_xe_exec exec = {};
> +	igt_require(xe->has_vram);
> +
> +	vm = xe_vm_create(xe->fd, 0, 0);
> +	engine = xe_engine_create_class(xe->fd, vm, DRM_XE_ENGINE_CLASS_COPY);
> +
> +	/* Refresh xe_device size after creating vm + engine */
> +	evict_mem(xe);
> +	full_size = (avail_vram(&xe) - (2 << 20ULL)) & ~((2 << 20ULL) - 1);
> +
> +	bo = xe_bo_create(xe->fd, 0, external ? 0 : vm, full_size);
> +
> +	/* Only map what we need.. */
> +	map = xe_bo_map(xe->fd, bo, xe->default_alignment);
> +	*map = MI_BATCH_BUFFER_END;
> +	munmap(map, xe->default_alignment);
> +
> +	/* Stuff bo at the end, so we can use start for partial bo's */
> +	full_addr = (1ULL << (uint64_t)xe_va_bits(xe->fd)) - full_size;
> +	xe_vm_bind_sync(xe->fd, vm, bo, 0, full_addr, full_size);
> +
> +	exec.engine_id = engine;
> +	exec.num_batch_buffer = 1;
> +	exec.address = full_addr;
> +
> +	/* Now map partially at 0, and see how many pagetables we can still fill.. */
> +	addr = 0;
> +	size = xe->default_alignment;
> +	xe_vm_bind_sync(xe->fd, vm, bo, 0, addr, size);
> +
> +	xe_exec_sync(xe->fd, engine, full_addr, NULL, 0);
> +
> +	avail = full_size >> 12;
> +	while (1) {
> +		uint64_t new_avail = avail_vram(&xe) >> 12;
> +
> +		if (new_avail >= avail)
> +			igt_info("TTM delayed destroy may have freed some memory\n");
> +		avail = new_avail;
> +
> +		igt_info("%"PRIu64" available 4K pages left\n", avail);
> +		if (!avail)
> +			break;
> +
> +		if (avail == 1) {
> +			xe_vm_bind_sync(xe->fd, vm, bo, 0, 2 << 20ULL, size);
> +			break;
> +		}
> +
> +		addr += 1 << 30ULL;
> +
> +		/* Remove 1 for second level PT */
> +		avail--;
> +
> +		for (i = 0; i < min((uint64_t)512, avail); i++) {
> +			uint64_t this_addr = addr + (i << 21);
> +
> +			xe_vm_bind_sync(xe->fd, vm, bo, 0, this_addr, size);
> +		}
> +	}
> +
> +	/* Verify that with 0 memory available, exec still completes */
> +	xe_exec_sync(xe->fd, engine, full_addr, NULL, 0);
> +
> +	/* Bind 1 more to go over the edge, this should never succeed.. */
> +	if (__xe_vm_bind(xe->fd, vm, 0, bo, 0, 4 << 20ULL, size,
> +			 XE_VM_BIND_OP_MAP, NULL, 0, 0, 0) == 0) {
> +
> +		/* Boom! */
> +		igt_assert(igt_ioctl(xe->fd, DRM_IOCTL_XE_EXEC, &exec) < 0 &&
> +			   errno == ENOMEM);
> +
> +		xe_vm_unbind_sync(xe->fd, vm, 0, 4 << 20ULL, size);
> +	} else {
> +		igt_assert(errno == ENOMEM);
> +	}
> +
> +	/*
> +	 * After evicting, we may end up creating new page tables,
> +	 * so this should fail..
> +	 */
> +	evict_mem(xe);
> +	igt_assert(igt_ioctl(xe->fd, DRM_IOCTL_XE_EXEC, &exec) < 0 &&
> +		   errno == ENOMEM);
> +
> +	/* Cleanup */
> +	xe_engine_destroy(xe->fd, engine);
> +	xe_vm_destroy(xe->fd, vm);
> +	gem_close(xe->fd, bo);
> +}
> +
> +/* Test OOM with signal handler enabled, to mimic X.org (issue #239) */
> +static void test_signal_oom(struct xe_device *xe, bool external)
> +{
> +	uint32_t bo[13], num_bo = 0, i;
> +	uint64_t binds, size = xe->has_vram ? xe->vram_size[0] : igt_get_avail_ram_mb();
> +	uint32_t vm = xe_vm_create(xe->fd, DRM_XE_VM_CREATE_SCRATCH_PAGE, 0);
> +	uint32_t engine = xe_engine_create_class(xe->fd, vm, DRM_XE_ENGINE_CLASS_COPY);
> +	int err;
> +	struct drm_xe_gem_create create = {
> +		.flags = vram_if_possible(xe->fd, 0),
> +	};
> +	struct drm_xe_sync sync[2] = {
> +		{
> +			.flags = DRM_XE_SYNC_SYNCOBJ | DRM_XE_SYNC_SIGNAL,
> +			.handle = syncobj_create(xe->fd, 0),
> +		},
> +		{
> +			.flags = DRM_XE_SYNC_SYNCOBJ | DRM_XE_SYNC_SIGNAL,
> +			.handle = syncobj_create(xe->fd, 0),
> +		},
> +	};
> +	struct drm_xe_exec exec = {
> +		.engine_id = engine,
> +		.syncs = (uintptr_t)sync,
> +		.num_syncs = ARRAY_SIZE(sync),
> +		.address = 0,
> +		.num_batch_buffer = 1,
> +	};
> +
> +	size /= 12;
> +	size &= ~(xe->default_alignment - 1);
> +
> +	create.size = size;
> +	if (!external)
> +		create.vm_id = vm;
> +
> +	/* Create as many bo's as we can without OOMs */
> +	while (num_bo < ARRAY_SIZE(bo)) {
> +		igt_while_interruptible(true) {
> +			if (num_bo >= ARRAY_SIZE(bo))
> +				continue;
> +
> +			create.handle = 0;
> +			err = igt_ioctl(xe->fd, DRM_IOCTL_XE_GEM_CREATE, &create);
> +			if (!err)
> +				bo[num_bo++] = create.handle;
> +		}
> +
> +		igt_assert(err == 0 || errno == ENOMEM);
> +		if (err) {
> +			igt_assert(!xe->has_vram || num_bo > 9);
> +			break;
> +		}
> +	}
> +
> +	igt_info("Created %u bo's (total: %"PRIu64" MB)\n",
> +		 num_bo, (num_bo * size) >> 20ULL);
> +
> +	binds = 0;
> +	for (i = 0; i < num_bo; i++) {
> +		evict_mem(xe);
> +		igt_debug("Binding bo %i, total attempts %"PRIi64"\n", i, binds);
> +
> +		igt_while_interruptible(true)
> +			xe_vm_bind(xe->fd, vm, bo[i], 0, ((binds++) << 32ULL), size, sync, 1);
> +	}
> +
> +	sync[0].flags &= ~DRM_XE_SYNC_SIGNAL;
> +
> +	/* Try a few times, why not? */
> +	for (i = 0; i < 4; i++) {
> +		igt_while_interruptible(true) {
> +			err = igt_ioctl(xe->fd, DRM_IOCTL_XE_EXEC, &exec);
> +			igt_assert(err < 0 && errno == ENOMEM);
> +		}
> +	}
> +
> +	xe_vm_destroy(xe->fd, vm);
> +	while (num_bo--)
> +		gem_close(xe->fd, bo[num_bo]);
> +	syncobj_destroy(xe->fd, sync[1].handle);
> +	syncobj_destroy(xe->fd, sync[0].handle);
> +}
> +
> +igt_main
> +{
> +	int fd;
> +
> +	igt_fixture
> +		fd = drm_open_driver(DRIVER_XE);
> +
> +	igt_subtest("multimap-internal")
> +		test_multimap(xe_device_get(fd), false);
> +
> +	igt_subtest("multimap-external")
> +		test_multimap(xe_device_get(fd), true);
> +
> +	igt_subtest("pt-eviction-internal")
> +		test_pt_eviction(xe_device_get(fd), false);
> +
> +	igt_subtest("pt-eviction-external")
> +		test_pt_eviction(xe_device_get(fd), false);
> +
> +	igt_subtest("signal-oom-internal")
> +		test_signal_oom(xe_device_get(fd), false);
> +
> +	igt_subtest("signal-oom-external")
> +		test_signal_oom(xe_device_get(fd), false);
> +}


More information about the igt-dev mailing list