[PATCH i-g-t v2] tests/intel/xe_pat: limit number of dynamic tests
Kamil Konieczny
kamil.konieczny at linux.intel.com
Mon Dec 11 12:30:25 UTC 2023
Hi Matthew,
On 2023-12-11 at 10:12:51 +0000, Matthew Auld wrote:
> Due to of the all the combinations here, we generate hundreds of dynamic
> testcases. This does allow running any individual testcase quite easily,
> but the sheer number of tests is a bit unwieldy for CI to handle. Trim
> this back to the following tests:
>
> pat-index-{xelp, xehpc, xelpg, xe2}@{dw, blt, render}
>
> v2 (Kamil): Tweak the pat-index-* test description in the commit
> message.
>
> Signed-off-by: Matthew Auld <matthew.auld at intel.com>
> Cc: Kamil Konieczny <kamil.konieczny at linux.intel.com>
> Cc: Himanshu Girotra <himanshu.girotra at intel.com>
Reviewed-by: Kamil Konieczny <kamil.konieczny at linux.intel.com>
> ---
> tests/intel/xe_pat.c | 128 +++++++++++++++++++++----------------------
> 1 file changed, 64 insertions(+), 64 deletions(-)
>
> diff --git a/tests/intel/xe_pat.c b/tests/intel/xe_pat.c
> index 8189390a8..7273242b0 100644
> --- a/tests/intel/xe_pat.c
> +++ b/tests/intel/xe_pat.c
> @@ -369,13 +369,15 @@ static void pat_index_render(struct xe_pat_param *p)
> int bpp = 32;
> int i;
>
> - bops = buf_ops_create(fd);
> -
> render_copy = igt_get_render_copyfunc(devid);
> igt_require(render_copy);
> - igt_require(!p->r2_compressed); /* XXX */
> igt_require(xe_has_engine_class(fd, DRM_XE_ENGINE_CLASS_RENDER));
>
> + if (p->r2_compressed) /* XXX */
> + return;
> +
> + bops = buf_ops_create(fd);
> +
> ibb = intel_bb_create_full(fd, 0, 0, NULL, xe_get_default_alignment(fd),
> 0, 0, p->size->alignment,
> INTEL_ALLOCATOR_SIMPLE,
> @@ -456,8 +458,8 @@ static void pat_index_dw(struct xe_pat_param *p)
> int bpp = 32, i, n_engines;
> uint32_t ctx, vm;
>
> - igt_require(!p->r1_compressed);
> - igt_require(!p->r2_compressed);
> + if (p->r1_compressed || p->r2_compressed)
> + return;
>
> bops = buf_ops_create(fd);
>
> @@ -901,84 +903,82 @@ static void subtest_pat_index_modes_with_regions(int fd,
>
> copy_mode = copy_modes[igt_collection_get_value(copies, 0)];
>
> - for_each_variation_r(regions, 2, regions_set) {
> - struct igt_collection *pat_modes;
> - uint32_t r1, r2;
> - char *reg_str;
> + igt_dynamic_f("%s", copy_mode.name) {
> + for_each_variation_r(regions, 2, regions_set) {
> + struct igt_collection *pat_modes;
> + uint32_t r1, r2;
> + char *reg_str;
>
> - r1 = igt_collection_get_value(regions, 0);
> - r2 = igt_collection_get_value(regions, 1);
> + r1 = igt_collection_get_value(regions, 0);
> + r2 = igt_collection_get_value(regions, 1);
>
> - reg_str = xe_memregion_dynamic_subtest_name(fd, regions);
> + reg_str = xe_memregion_dynamic_subtest_name(fd, regions);
>
> - for_each_variation_r(pat_modes, 2, pat_index_set) {
> - struct igt_collection *sizes;
> - struct pat_index_entry r1_entry, r2_entry;
> - int r1_idx, r2_idx;
> + for_each_variation_r(pat_modes, 2, pat_index_set) {
> + struct igt_collection *sizes;
> + struct pat_index_entry r1_entry, r2_entry;
> + int r1_idx, r2_idx;
>
> - r1_idx = igt_collection_get_value(pat_modes, 0);
> - r2_idx = igt_collection_get_value(pat_modes, 1);
> + r1_idx = igt_collection_get_value(pat_modes, 0);
> + r2_idx = igt_collection_get_value(pat_modes, 1);
>
> - r1_entry = modes_arr[r1_idx];
> - r2_entry = modes_arr[r2_idx];
> + r1_entry = modes_arr[r1_idx];
> + r2_entry = modes_arr[r2_idx];
>
> - if (r1_entry.get_pat_index) {
> - p.r1_pat_index = r1_entry.get_pat_index(fd, &p.r1_compressed);
> - } else {
> - p.r1_pat_index = r1_entry.pat_index;
> - p.r1_compressed = r1_entry.compressed;
> - }
> + if (r1_entry.get_pat_index) {
> + p.r1_pat_index = r1_entry.get_pat_index(fd, &p.r1_compressed);
> + } else {
> + p.r1_pat_index = r1_entry.pat_index;
> + p.r1_compressed = r1_entry.compressed;
> + }
>
> - if (r2_entry.get_pat_index)
> - p.r2_pat_index = r2_entry.get_pat_index(fd, &p.r2_compressed);
> - else {
> - p.r2_pat_index = r2_entry.pat_index;
> - p.r2_compressed = r2_entry.compressed;
> - }
> + if (r2_entry.get_pat_index)
> + p.r2_pat_index = r2_entry.get_pat_index(fd, &p.r2_compressed);
> + else {
> + p.r2_pat_index = r2_entry.pat_index;
> + p.r2_compressed = r2_entry.compressed;
> + }
>
> - p.r1 = r1;
> - p.r2 = r2;
> + p.r1 = r1;
> + p.r2 = r2;
>
> - for_each_variation_r(sizes, 1, sizes_set) {
> - int size_mode_idx = igt_collection_get_value(sizes, 0);
> - int bpp = 32;
> - int size;
> + for_each_variation_r(sizes, 1, sizes_set) {
> + int size_mode_idx = igt_collection_get_value(sizes, 0);
> + int bpp = 32;
> + int size;
>
> - p.size = &size_modes[size_mode_idx];
> + p.size = &size_modes[size_mode_idx];
>
> - size = p.size->width * p.size->height * bpp / 8;
> - p.r1_bo = create_object(fd, p.r1, size, r1_entry.coh_mode,
> - r1_entry.force_cpu_wc);
> - p.r1_map = xe_bo_map(fd, p.r1_bo, size);
> + size = p.size->width * p.size->height * bpp / 8;
> + p.r1_bo = create_object(fd, p.r1, size, r1_entry.coh_mode,
> + r1_entry.force_cpu_wc);
> + p.r1_map = xe_bo_map(fd, p.r1_bo, size);
>
> - p.r2_bo = create_object(fd, p.r2, size, r2_entry.coh_mode,
> - r2_entry.force_cpu_wc);
> - p.r2_map = xe_bo_map(fd, p.r2_bo, size);
> + p.r2_bo = create_object(fd, p.r2, size, r2_entry.coh_mode,
> + r2_entry.force_cpu_wc);
> + p.r2_map = xe_bo_map(fd, p.r2_bo, size);
>
> - igt_debug("[r1]: r: %u, idx: %u, coh: %u, wc: %d, comp: %d\n",
> - p.r1, p.r1_pat_index, r1_entry.coh_mode,
> - r1_entry.force_cpu_wc, p.r1_compressed);
> - igt_debug("[r2]: r: %u, idx: %u, coh: %u, wc: %d, comp: %d, w: %u, h: %u, a: %u\n",
> - p.r2, p.r2_pat_index, r2_entry.coh_mode,
> - r1_entry.force_cpu_wc, p.r2_compressed,
> - p.size->width, p.size->height,
> - p.size->alignment);
> + igt_debug("[r1]: r: %u, idx: %u (%s), coh: %u, wc: %d, comp: %d\n",
> + p.r1, p.r1_pat_index, r1_entry.name, r1_entry.coh_mode,
> + r1_entry.force_cpu_wc, p.r1_compressed);
> + igt_debug("[r2]: r: %u, idx: %u (%s), coh: %u, wc: %d, comp: %d, w: %u, h: %u, a: %u\n",
> + p.r2, p.r2_pat_index, r2_entry.name, r2_entry.coh_mode,
> + r1_entry.force_cpu_wc, p.r2_compressed,
> + p.size->width, p.size->height,
> + p.size->alignment);
>
> - igt_dynamic_f("%s-%s-%s-%s-%s",
> - copy_mode.name,
> - reg_str, r1_entry.name,
> - r2_entry.name, p.size->name)
> copy_mode.fn(&p);
>
> - munmap(p.r1_map, size);
> - munmap(p.r2_map, size);
> + munmap(p.r1_map, size);
> + munmap(p.r2_map, size);
>
> - gem_close(fd, p.r1_bo);
> - gem_close(fd, p.r2_bo);
> + gem_close(fd, p.r1_bo);
> + gem_close(fd, p.r2_bo);
> + }
> }
> +
> + free(reg_str);
> }
> -
> - free(reg_str);
> }
> }
> }
> --
> 2.43.0
>
More information about the igt-dev
mailing list