[igt-dev] [RFC v3] tests/gem_watchdog: Initial set of tests for GPU watchdog
Tvrtko Ursulin
tvrtko.ursulin at linux.intel.com
Tue Apr 23 16:25:39 UTC 2019
Hi,
How is the test doing against the code in progress? Any issues
discovered etc?
In general it looks a bit untidy. Coding style is off, alignment is
random etc.. so a bit difficult to read. But I'll make a pass and see if
I can make sense of it. I won't leave comments on coding style and
formatting since there are too many. But please do fix that before
re-posting.
On 23/04/2019 01:05, Carlos Santa wrote:
> This test adds basic set of tests to reset the different
> GPU engines through the gpu watchdog timer.
>
> Credits to Antonio for the original codebase this is based on.
>
> v2: remove gem_context_get_param() during set (Antonio)
> remove clearing of the engines_threshold[] in the default case
> inside context_set_watchdog(). (Antonio)
> fix indexing when creating low/high priority contexts
> get rid of 2 threads idea (Antonio)
> fix context prio bug due to wrong indexing (Antonio)
> v3: no need to force clear drm_i915_gem_watchdog_timeout struct
> when setting the watchdog timeouts (Antonio)
> -use gem_context_set_priority() instead to use error checking
> -always use a looper to test all engines (Antonio)
> -modify test gpu_watchodg_hang_long_batch_single_engine()
> to use a couple of batches and set the timeout in between them (Antonio)
> -make use of for_each_engine_class_instance() (Antonio)
>
> Cc: Ursulin Tvrtko <tvrtko.ursulin at intel.com>
> Cc: Antonio Argenziano <antonio.argenziano at intel.com>
> Cc: Chris Wilson <chris at chris-wilson.co.uk>
> Signed-off-by: Carlos Santa <carlos.santa at intel.com>
> ---
> tests/Makefile.sources | 3 +
> tests/i915/gem_watchdog.c | 517 ++++++++++++++++++++++++++++++++++++++++++++++
> tests/meson.build | 1 +
> 3 files changed, 521 insertions(+)
> create mode 100644 tests/i915/gem_watchdog.c
>
> diff --git a/tests/Makefile.sources b/tests/Makefile.sources
> index 214698d..7f17f20 100644
> --- a/tests/Makefile.sources
> +++ b/tests/Makefile.sources
> @@ -444,6 +444,9 @@ gem_userptr_blits_SOURCES = i915/gem_userptr_blits.c
> TESTS_progs += gem_wait
> gem_wait_SOURCES = i915/gem_wait.c
>
> +TESTS_progs += gem_watchdog
> +gem_watchdog_SOURCES = i915/gem_watchdog.c
> +
> TESTS_progs += gem_workarounds
> gem_workarounds_SOURCES = i915/gem_workarounds.c
>
> diff --git a/tests/i915/gem_watchdog.c b/tests/i915/gem_watchdog.c
> new file mode 100644
> index 0000000..8f789ab
> --- /dev/null
> +++ b/tests/i915/gem_watchdog.c
> @@ -0,0 +1,517 @@
> +/*
> + * Copyright © 2016 Intel Corporation
> + *
> + * Permission is hereby granted, free of charge, to any person obtaining a
> + * copy of this software and associated documentation files (the "Software"),
> + * to deal in the Software without restriction, including without limitation
> + * the rights to use, copy, modify, merge, publish, distribute, sublicense,
> + * and/or sell copies of the Software, and to permit persons to whom the
> + * Software is furnished to do so, subject to the following conditions:
> + *
> + * The above copyright notice and this permission notice (including the next
> + * paragraph) shall be included in all copies or substantial portions of the
> + * Software.
> + *
> + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
> + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
> + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
> + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
> + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
> + * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
> + * IN THE SOFTWARE.
> + */
> +#include "igt.h"
> +#include "igt_sysfs.h"
> +#include "sw_sync.h"
> +
> +#include <pthread.h>
> +#include <fcntl.h>
> +
> +#include <sys/ioctl.h>
> +#include <sys/poll.h>
> +#include <sys/signal.h>
> +#include "i915/gem_ring.h"
> +
> +#define LOCAL_I915_EXEC_BSD_SHIFT (13)
> +#define LOCAL_I915_EXEC_BSD_RING1 (1 << LOCAL_I915_EXEC_BSD_SHIFT)
> +#define LOCAL_I915_EXEC_BSD_RING2 (2 << LOCAL_I915_EXEC_BSD_SHIFT)
These are in old so no need for local definitions.
> +
> +#define MAX_PRIO LOCAL_I915_CONTEXT_MAX_USER_PRIORITY
> +#define MIN_PRIO LOCAL_I915_CONTEXT_MIN_USER_PRIORITY
> +#define HIGH 1
> +#define LOW 0
> +#define TRUE 1
> +#define FALSE 0
true and false are fair game in igt.
> +#define WATCHDOG_THRESHOLD (100)
> +#define MAX_ENGINES 5
> +#define RENDER_CLASS 0
> +#define VIDEO_DECODE_CLASS 1
> +#define VIDEO_ENHANCEMENT_CLASS 2
> +#define COPY_ENGINE_CLASS 3
Classes are in i915_drm.h as well.
> +#define LOCAL_I915_CONTEXT_PARAM_WATCHDOG 0x10
> +
> +#define RS_NO_ERROR 0
> +#define RS_BATCH_ACTIVE (1 << 0)
> +#define RS_BATCH_PENDING (1 << 1)
> +#define RS_UNKNOWN (1 << 2)
What are these? Look unused in code.
> +
> +#define GET_RESET_STATS_IOCTL DRM_IOWR(DRM_COMMAND_BASE + 0x32, struct local_drm_i915_reset_stats)
> +struct local_drm_i915_reset_stats {
> + __u32 ctx_id;
> + __u32 flags;
> + __u32 reset_count;
> + __u32 batch_active;
> + __u32 batch_pending;
> + __u32 pad;
> +};
> +
> +const uint64_t timeout_100ms = 100000000LL;
> +int num;
> +
> +struct drm_i915_gem_watchdog_timeout {
> + union {
> + struct {
> + /*
> + * Engine class & instance to be configured or queried.
> + */
> + __u16 engine_class;
> + __u16 engine_instance;
> + };
> + /* Index based addressing mode */
> + __u32 index;
> + };
> + /* GPU Engine watchdog resets timeout in us */
> + __u32 timeout_us;
> +};
> +
> +static void clear_error_state(int fd)
> +{
> + int dir;
> +
> + dir = igt_sysfs_open(fd);
> +
> + if (dir < 0)
> + return;
igt_require(dir >= 0) if test relies on this being present?
> +
> + /* Any write to the error state clears it */
> + igt_sysfs_set(dir, "error", "");
> + close(dir);
> +}
> +
> +static unsigned int e2ring(int fd, const struct intel_execution_engine2 *e)
> +{
> + return gem_class_instance_to_eb_flags(fd, e->class, e->instance);
> +}
> +
> +static void context_set_watchdog(int fd, unsigned engine_id,
> + unsigned ctx_id, unsigned threshold)
> +{
> + struct drm_i915_gem_watchdog_timeout engines_threshold[MAX_ENGINES];
> + struct drm_i915_gem_context_param arg = {
> + .param = LOCAL_I915_CONTEXT_PARAM_WATCHDOG,
> + .ctx_id = ctx_id,
> + .size = sizeof(engines_threshold),
> + .value = (uint64_t)&engines_threshold
> + };
> +
> + memset(&engines_threshold, 0, sizeof(engines_threshold));
> +
> + switch (engine_id & I915_EXEC_RING_MASK) {
> + case I915_EXEC_RENDER:
> + engines_threshold[RENDER_CLASS].timeout_us = threshold;
Looks like uAPI is in some interim state? I was expecting to see here
something like:
engines_threshold.engine_class = RENDER_CLASS;
engines_threshold.engine_instance = 0;
engines_threshold.timeout_us = threshold;
But you seem to be passing an array of structs which carry
class/instance data, don't use it but instead infer the class from the
position in the array.
> + break;
> + case I915_EXEC_BSD:
> + engines_threshold[VIDEO_DECODE_CLASS].timeout_us = threshold;
> + break;
> + case I915_EXEC_VEBOX:
> + engines_threshold[VIDEO_ENHANCEMENT_CLASS].timeout_us = threshold;
> + break;
> + default:
> + break;
> + }
> +
> + gem_context_set_param(fd, &arg);
> +}
> +
> +static void batch_buffer_factory(uint32_t fd, uint32_t ctx_id, unsigned exec_id, uint32_t target, uint32_t offset, uint32_t *handle, uint64_t timeout, int *fence, int fence_index)
> +{
> + struct drm_i915_gem_exec_object2 obj[2];
> + struct drm_i915_gem_relocation_entry reloc;
> + struct drm_i915_gem_execbuffer2 execbuf;
> + igt_spin_t *spin = NULL;
> + const uint32_t bbe = MI_BATCH_BUFFER_END;
> + int i = 0;
> +
> + gem_quiescent_gpu(fd);
> +
> + memset(&execbuf, 0, sizeof(execbuf));
> + memset(&obj, 0, sizeof(obj));
> + memset(&reloc, 0, sizeof(reloc));
> +
> + execbuf.buffers_ptr = to_user_pointer(obj);
> +
> + execbuf.buffer_count = 2;
> + execbuf.flags = exec_id | I915_EXEC_FENCE_OUT ;
> +
> + obj[0].handle = target;
> + obj[1].handle = gem_create(fd, 4096);
> +
> + obj[1].relocation_count = 1;
> + obj[1].relocs_ptr = to_user_pointer(&reloc);
> +
> + reloc.target_handle = obj[0].handle;
> + reloc.read_domains = I915_GEM_DOMAIN_COMMAND;
> + reloc.write_domain = I915_GEM_DOMAIN_COMMAND;
> + reloc.delta = offset * sizeof(uint32_t);
> +
> + reloc.offset = i * sizeof(uint32_t);
> + gem_write(fd, obj[1].handle, 0, &bbe, sizeof(bbe));
> +
> + __sync_synchronize();
> +
> + if (handle) {
> + *handle = obj[1].handle;
> + return;
Is this path used by the callers?
> + }
> +
> + gem_sync(fd, obj[1].handle);
> + execbuf.rsvd1 = ctx_id;
> + execbuf.rsvd2 = -1;
> +
> + spin = igt_spin_batch_new(fd, .dependency = obj[0].handle);
> + igt_spin_batch_set_timeout(spin, timeout);
> + igt_assert(gem_bo_busy(fd, obj[0].handle));
> +
> + gem_execbuf_wr(fd, &execbuf);
> + igt_spin_batch_free(fd, spin);
> +
> + fence[fence_index] = execbuf.rsvd2 >> 32;
> +
> + gem_close(fd, obj[1].handle);
> + gem_quiescent_gpu(fd);
> +}
> +
> +static uint32_t create_ctx_with_priority(int fd, int ctx_prio)
> +{
> + uint32_t ctx = gem_context_create(fd);
> +
> + switch (ctx_prio) {
> + case HIGH:
> + gem_context_set_priority(fd, ctx, MAX_PRIO);
> + igt_info("Setting MAX priority %d\n", ctx_prio);
> + break;
> + case LOW:
> + gem_context_set_priority(fd, ctx, MIN_PRIO);
> + igt_info("Setting MIN priority %d\n", ctx_prio);
> + break;
> + default:
> + igt_info("Ignoring context priority %d\n", ctx_prio);
> + break;
> + }
> +
> + return ctx;
Why do you need this helper which translates an integer to an integer?
Couldn't you just pass in the real priority to start with?
> +}
> +
> +static int gem_reset_stats(int fd, int ctx_id,
> + struct local_drm_i915_reset_stats *rs)
> +{
> + memset(rs, 0, sizeof(*rs));
> + rs->ctx_id = ctx_id;
> + rs->reset_count = -1;
> +
> + if (drmIoctl(fd, GET_RESET_STATS_IOCTL, rs)) {
> + printf("error:%d\n",-errno);
> + return -errno;
Make it assert success here since all callers do that anyway and it
matches the convention for gem prefixed ioctl helpers?
> + }
> + igt_assert(rs->reset_count != -1);
> + return 0;
> +}
> +
> +static unsigned index_to_engine_id(int fd, int index)
> +{
> + unsigned engine_id = I915_EXEC_RENDER;
> +
> + switch (index) {
> + case 1:
> + engine_id = I915_EXEC_RENDER;
> + break;
> + case 2:
> + if (gem_has_bsd2(fd))
> + engine_id = I915_EXEC_BSD | I915_EXEC_BSD_RING1;
> + else
> + engine_id = I915_EXEC_BSD;
> + break;
> + case 3:
> + if (gem_has_bsd2(fd))
> + engine_id = I915_EXEC_BSD | I915_EXEC_BSD_RING2;
> + else
> + engine_id = I915_EXEC_BSD;
> + break;
> + case 4:
> + engine_id = I915_EXEC_VEBOX;
> + break;
> + default:
> + break;
> + }
I hope we can avoid this maintenance burden, will see later in patch how.
> +
> + return engine_id;
> +}
> +
> +static void inject_hang(uint32_t fd, unsigned engine_id, uint32_t ctx_id, unsigned flags)
> +{
> + igt_hang_t hang;
> + hang = igt_hang_ctx(fd, ctx_id, engine_id, flags);
> + gem_sync(fd, hang.spin->handle);
> +}
> +
> +static void gpu_watchdog_long_batch_2_contexts(int fd, int nengine, int prio_ctx1, int reset_ctx1, int prio_ctx2, int reset_ctx2)
> +{
> + uint32_t ctx[48];
> + uint32_t scratch[48];
> + unsigned flags = HANG_ALLOW_CAPTURE;
> + const uint64_t batch_timeout_ms = timeout_100ms * 3;
ms suffix usually means the unit stored is ms. While here it seems to be
something else.
const unsigned long batch_timeout_ms = 300;
const unsigned long batch_timeout_us = 300e3;
const unsigned long batch_timeout_ns = 300e6;
Not sure which one you want. Probably use ms and scale up at the use site.
> + const struct intel_execution_engine2 *e;
> + const struct intel_execution_engine2 *e2;
> + int i = 0, index = 0, engine_id = 0, ctx_cnt = 0;
> + int *fence = 0, active_count = 0;
> + struct local_drm_i915_reset_stats rs;
> +
> + igt_require(nengine);
> +
> + fence = (int *)malloc(sizeof(int) * 48);
> +
> + if (!fence) {
> + igt_info("Out of memory\n");
> + exit(1);
> + }
Replace with igt_assert(fence).
> +
> + /* Intializes random number generator */
> + srand((unsigned) time(NULL));
> +
> + /* The outer looper is ctx1 and the inner looper is ctx2
> + * Both contexts can set their priority, create work, and cancel themselves
> + */
> + for_each_engine_class_instance(fd, e) {
> + if ( strcmp(e->name, "bcs0") == 0 )
> + continue;
Use -ENODEV or -EINVAL from context_set_watchdog to skip.
> +
> + scratch[ctx_cnt] = gem_create(fd, 4096);
> + index = rand() % 4 + 1;
> + engine_id = index_to_engine_id(fd, index);
> +
> + /* get current batch_active count */
> + igt_assert_eq(gem_reset_stats(fd, 0, &rs), 0);
> + active_count = rs.batch_active;
> + igt_assert_eq(rs.batch_active, active_count);
> +
> + ctx[ctx_cnt] = create_ctx_with_priority(fd, prio_ctx1);
> + batch_buffer_factory(fd, ctx[ctx_cnt], engine_id, scratch[ctx_cnt], 0, NULL, batch_timeout_ms, fence, ctx_cnt);
> + context_set_watchdog(fd, engine_id, ctx[ctx_cnt], WATCHDOG_THRESHOLD);
> + sleep(1);
What's the sleep for? If unavoidable at least needs a comment.
> +
> + if (reset_ctx1) {
> + clear_error_state(fd);
> + inject_hang(fd, engine_id, ctx[ctx_cnt], flags);
> + active_count++;
> + }
> +
> + /* check batch_active count once again after the hang */
> + igt_assert_eq(gem_reset_stats(fd, 0, &rs), 0);
> + igt_assert_eq(rs.batch_active, active_count);
> +
> + ctx_cnt++;
> + i = 1;
> +
> + for_each_engine_class_instance(fd, e2) {
> + if ( strcmp(e->name, "bcs0") == 0 )
e2, but remove it altogether as said above.
I don't get what is the purpose of two for_each_engine_class_instance
loops when neither e or e2 appear to be used?
Since either I am missing something critical or the test really is very
confused, until clarified, I am skipping a bit lower down.
> + continue;
> +
> + scratch[ctx_cnt] = gem_create(fd, 4096);
> + engine_id = index_to_engine_id(fd, i);
> +
> + /* get current batch_active count */
> + igt_assert_eq(gem_reset_stats(fd, 0, &rs), 0);
> + active_count = rs.batch_active;
> + igt_assert_eq(rs.batch_active, active_count);
> +
> + ctx[ctx_cnt] = create_ctx_with_priority(fd, prio_ctx2);
> + batch_buffer_factory(fd, ctx[ctx_cnt], engine_id, scratch[ctx_cnt], 0, NULL, batch_timeout_ms, fence, ctx_cnt);
> + context_set_watchdog(fd, engine_id, ctx[ctx_cnt], WATCHDOG_THRESHOLD);
> +
> + if (reset_ctx2) {
> + clear_error_state(fd);
> + inject_hang(fd, engine_id, ctx[ctx_cnt], flags);
> + active_count++;
> + }
> +
> + /* check batch_active count once again after the hang */
> + igt_assert_eq(gem_reset_stats(fd, 0, &rs), 0);
> + igt_assert_eq(rs.batch_active, active_count);
> +
> + ctx_cnt++;
> + i++;
> + }
> + }
> +
> + for (i = 0; i < ctx_cnt; i++) {
> + close(fence[i]);
> + gem_context_destroy(fd, ctx[i]);
> + gem_close(fd, scratch[i]);
> + }
> +}
> +
> +static void gpu_watchodg_hang_long_batch_single_engine(int fd, const char *name)
> +{
> + uint32_t ctx[16];
> + uint32_t scratch[16];
> + int *fence;
> + const struct intel_execution_engine2 *e;
> + unsigned nengine = 0;
> + unsigned engine_id = 0;
> + int active_count = 0;
> + struct local_drm_i915_reset_stats rs;
> + int i;
> + unsigned flags = HANG_ALLOW_CAPTURE;
> + const uint64_t batch_timeout_ms = timeout_100ms * 4;
> +
> + fence = (int *)malloc(sizeof(int)*16);
> +
> + if (!fence) {
> + igt_info("Out of memory\n");
> + exit(1);
> + }
> +
> + for_each_engine_class_instance(fd, e) {
> + if ( strcmp(e->name, "bcs0") == 0 )
> + continue;
> +
> + engine_id = e2ring(fd, e);
> +
> + scratch[nengine] = gem_create(fd, 4096);
> + ctx[nengine] = create_ctx_with_priority(fd, -1);
> +
> + /* get current batch_active count */
> + igt_assert_eq(gem_reset_stats(fd, 0, &rs), 0);
> + active_count = rs.batch_active;
> + igt_assert_eq(rs.batch_active, active_count);
> +
> + /* create some work on this engine using the above ctx*/
> + batch_buffer_factory(fd, ctx[nengine], engine_id, scratch[nengine], 0, NULL, batch_timeout_ms, fence, nengine);
> +
> + /* set the gpu watchdog timeout */
> + context_set_watchdog(fd, engine_id, ctx[nengine], WATCHDOG_THRESHOLD);
> + clear_error_state(fd);
> +
> + /*
> + * give enough time for the batch buffer to run
> + */
> + nanosleep(&(struct timespec){ .tv_sec = 0,
> + .tv_nsec = batch_timeout_ms * 2 }, NULL);
> +
> + /* create some work on this engine using the above ctx*/
> + batch_buffer_factory(fd, ctx[nengine], engine_id, scratch[nengine], 0, NULL, batch_timeout_ms, fence, nengine);
> + context_set_watchdog(fd, engine_id, ctx[nengine], WATCHDOG_THRESHOLD);
> + clear_error_state(fd);
> +
> + /* cancel only the batch requested */
> + if ( strcmp(e->name, name) == 0 ) {
> + inject_hang(fd, engine_id, ctx[nengine], flags);
> + active_count++;
> + }
> +
> + /* check batch_active count once again after the hang */
> + igt_assert_eq(gem_reset_stats(fd, 0, &rs), 0);
> + active_count = rs.batch_active;
> + igt_assert_eq(rs.batch_active, active_count);
> +
> + nengine++;
> + }
> +
> + for (i = 0; i < nengine; i++) {
> + close(fence[i]);
> + gem_context_destroy(fd, ctx[i]);
> + gem_close(fd, scratch[i]);
> + }
> +}
> +
> +igt_main
> +{
> + int fd;
> + unsigned int nengine = 0;
> + unsigned int engine;
> + const struct intel_execution_engine2 *e;
> +
> + igt_skip_on_simulation();
> +
> + igt_fixture {
> + fd = drm_open_driver(DRIVER_INTEL);
> + igt_require_gem(fd);
> +
> + for_each_physical_engine(fd, engine)
> + nengine++;
> + igt_require(nengine);
Do we have GPUs with no engines?
> + }
> +
> + igt_subtest_group {
> +
> + igt_subtest_f("gpu-watchdog-long-batch-low-prio-ctx1-reset-ctx2-high-prio-execute") {
> + int prio_ctx1 = LOW;
> + int prio_ctx2 = HIGH;
> + int reset_ctx1 = TRUE;
> + int reset_ctx2 = FALSE;
> + gpu_watchdog_long_batch_2_contexts(fd, nengine, prio_ctx1, reset_ctx1, prio_ctx2, reset_ctx2);
> + }
This is kind of verbose. I recommend out usual pattern along the lines of:
struct {
char *name;
int prio[2];
bool reset[2];
} tests[] = {
{ "change me", {0, 0}, {false, false} },
{ "change me", {-1, 0}, {true, false} },
...
};
for (i = 0; i < ARRAY_SIZE(tests); i++) {
igt_subtest_f("long-batch-%s", tests[i]->name)
long_batch(fd, tests[i].prio[0], tests[i].ctx[0], ...);
}
So you run all subtests from a single loop.
Regards,
Tvrtko
> +
> + igt_subtest_f("gpu-watchdog-long-batch-high-prio-ctx1-reset-ctx2-low-prio-execute") {
> + int prio_ctx1 = HIGH;
> + int prio_ctx2 = LOW;
> + int reset_ctx1 = TRUE;
> + int reset_ctx2 = FALSE;
> + gpu_watchdog_long_batch_2_contexts(fd, nengine, prio_ctx1, reset_ctx1, prio_ctx2, reset_ctx2);
> + }
> +
> + igt_subtest_f("gpu-watchdog-long-batch-ctx1-reset-ctx2-execute") {
> + int prio_ctx1 = -1;
> + int prio_ctx2 = -1;
> + int reset_ctx1 = TRUE;
> + int reset_ctx2 = FALSE;
> + gpu_watchdog_long_batch_2_contexts(fd, nengine, prio_ctx1, reset_ctx1, prio_ctx2, reset_ctx2);
> + }
> +
> + igt_subtest_f("gpu-watchdog-long-batch-ctx1-execute-ctx2-reset") {
> + int prio_ctx1 = -1;
> + int prio_ctx2 = -1;
> + int reset_ctx1 = FALSE;
> + int reset_ctx2 = TRUE;
> + gpu_watchdog_long_batch_2_contexts(fd, nengine, prio_ctx1, reset_ctx1, prio_ctx2, reset_ctx2);
> + }
> +
> + igt_subtest_f("gpu-watchdog-long-batch-ctx1-execute-ctx2-execute") {
> + int prio_ctx1 = -1;
> + int prio_ctx2 = -1;
> + int reset_ctx1 = FALSE;
> + int reset_ctx2 = FALSE;
> + gpu_watchdog_long_batch_2_contexts(fd, nengine, prio_ctx1, reset_ctx1, prio_ctx2, reset_ctx2);
> + }
> +
> + __for_each_engine_class_instance(e) {
> + igt_subtest_group {
> + igt_fixture {
> + gem_require_engine(fd,
> + e->class,
> + e->instance);
> + }
> + /* no support for gpu watchdog on bcs0 */
> + if (strcmp(e->name, "bcs0") == 0)
> + continue;
Software implementation can support it. So don't skip here but skip in
the subtest if you get -ENODEV when you try to configure it.
> +
> + igt_subtest_f("gpu-watchdog-long-batch-single-engine-reset-%s", e->name)
Don't prefix all tests with gpu-watchdog, it is very verbose to read and
the test itself is called gem_watchdog already.
> + gpu_watchodg_hang_long_batch_single_engine(fd, e->name);
> + }
> + }
> + }
> +
> + igt_fixture {
> + close(fd);
> + }
> +}
> diff --git a/tests/meson.build b/tests/meson.build
> index 5167a6c..b281b75 100644
> --- a/tests/meson.build
> +++ b/tests/meson.build
> @@ -210,6 +210,7 @@ i915_progs = [
> 'gem_unref_active_buffers',
> 'gem_userptr_blits',
> 'gem_wait',
> + 'gem_watchdog',
> 'gem_workarounds',
> 'gem_write_read_ring_switch',
> 'i915_fb_tiling',
>
More information about the igt-dev
mailing list