[RFC 2/5] drm/scheduler: Add scheduler unit testing infrastructure and some basic tests
Tvrtko Ursulin
tvrtko.ursulin at igalia.com
Thu Feb 6 10:42:00 UTC 2025
On 06/02/2025 09:51, Philipp Stanner wrote:
> On Mon, 2025-02-03 at 15:30 +0000, Tvrtko Ursulin wrote:
>> Implement a mock scheduler backend and add some basic test to
>> exercise the
>> core scheduler code paths.
>>
>> Mock backend (kind of like a very simple mock GPU) can either process
>> jobs
>> by tests manually advancing the "timeline" job at a time, or
>> alternatively
>> jobs can be configured with a time duration in which case they get
>> completed asynchronously from the unit test code.
>>
>> Core scheduler classes are subclassed to support this mock
>> implementation.
>>
>> The tests added are just a few simple submission patterns.
>>
>> Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin at igalia.com>
>
> Could add a Suggested-by: Philipp :)
Will do.
>> Cc: Christian König <christian.koenig at amd.com>
>> Cc: Danilo Krummrich <dakr at kernel.org>
>> Cc: Matthew Brost <matthew.brost at intel.com>
>> Cc: Philipp Stanner <phasta at kernel.org>
>> ---
>> drivers/gpu/drm/Kconfig.debug | 12 +
>> drivers/gpu/drm/scheduler/.kunitconfig | 12 +
>> drivers/gpu/drm/scheduler/Makefile | 1 +
>> drivers/gpu/drm/scheduler/tests/Makefile | 4 +
>> .../gpu/drm/scheduler/tests/drm_mock_entity.c | 29 ++
>> .../gpu/drm/scheduler/tests/drm_mock_job.c | 3 +
>> .../drm/scheduler/tests/drm_mock_scheduler.c | 254
>> ++++++++++++++++++
>> .../gpu/drm/scheduler/tests/drm_sched_tests.h | 124 +++++++++
>> .../scheduler/tests/drm_sched_tests_basic.c | 247
>> +++++++++++++++++
>> 9 files changed, 686 insertions(+)
>> create mode 100644 drivers/gpu/drm/scheduler/.kunitconfig
>> create mode 100644 drivers/gpu/drm/scheduler/tests/Makefile
>> create mode 100644 drivers/gpu/drm/scheduler/tests/drm_mock_entity.c
>> create mode 100644 drivers/gpu/drm/scheduler/tests/drm_mock_job.c
>> create mode 100644
>> drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c
>> create mode 100644 drivers/gpu/drm/scheduler/tests/drm_sched_tests.h
>> create mode 100644
>> drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c
>>
>> diff --git a/drivers/gpu/drm/Kconfig.debug
>> b/drivers/gpu/drm/Kconfig.debug
>> index a35d74171b7b..89f777f21e95 100644
>> --- a/drivers/gpu/drm/Kconfig.debug
>> +++ b/drivers/gpu/drm/Kconfig.debug
>> @@ -88,5 +88,17 @@ config DRM_TTM_KUNIT_TEST
>>
>> If in doubt, say "N".
>>
>> +config DRM_SCHED_KUNIT_TEST
>> + tristate "KUnit tests for the DRM scheduler" if
>> !KUNIT_ALL_TESTS
>> + select DRM_SCHED
>> + depends on DRM && KUNIT
>> + default KUNIT_ALL_TESTS
>> + help
>> + Choose this option to build unit tests for the DRM
>> scheduler.
>
> nit: Might say "DRM GPU scheduler" so readers not familiar with all
> that get a better idea of what it's about
>
>> +
>> + Recommended for driver developers only.
>
> nit: s/driver developers/DRM developers
> ?
>
>> +
>> + If in doubt, say "N".
>> +
>> config DRM_EXPORT_FOR_TESTS
>> bool
>> diff --git a/drivers/gpu/drm/scheduler/.kunitconfig
>> b/drivers/gpu/drm/scheduler/.kunitconfig
>> new file mode 100644
>> index 000000000000..cece53609fcf
>> --- /dev/null
>> +++ b/drivers/gpu/drm/scheduler/.kunitconfig
>> @@ -0,0 +1,12 @@
>> +CONFIG_KUNIT=y
>> +CONFIG_DRM=y
>> +CONFIG_DRM_SCHED_KUNIT_TEST=y
>> +CONFIG_EXPERT=y
>> +CONFIG_DEBUG_SPINLOCK=y
>> +CONFIG_DEBUG_MUTEXES=y
>> +CONFIG_DEBUG_ATOMIC_SLEEP=y
>> +CONFIG_LOCK_DEBUGGING_SUPPORT=y
>> +CONFIG_PROVE_LOCKING=y
>> +CONFIG_LOCKDEP=y
>> +CONFIG_DEBUG_LOCKDEP=y
>> +CONFIG_DEBUG_LIST=y
>> diff --git a/drivers/gpu/drm/scheduler/Makefile
>> b/drivers/gpu/drm/scheduler/Makefile
>> index 53863621829f..46dfdca0758a 100644
>> --- a/drivers/gpu/drm/scheduler/Makefile
>> +++ b/drivers/gpu/drm/scheduler/Makefile
>> @@ -23,3 +23,4 @@
>> gpu-sched-y := sched_main.o sched_fence.o sched_entity.o
>>
>> obj-$(CONFIG_DRM_SCHED) += gpu-sched.o
>> +obj-$(CONFIG_DRM_SCHED_KUNIT_TEST) += tests/
>> diff --git a/drivers/gpu/drm/scheduler/tests/Makefile
>> b/drivers/gpu/drm/scheduler/tests/Makefile
>> new file mode 100644
>> index 000000000000..d69eab6a2e9b
>> --- /dev/null
>> +++ b/drivers/gpu/drm/scheduler/tests/Makefile
>> @@ -0,0 +1,4 @@
>> +
>> +obj-$(CONFIG_DRM_SCHED_KUNIT_TEST) += \
>> + drm_mock_scheduler.o \
>> + drm_sched_tests_basic.o
>> diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c
>> b/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c
>> new file mode 100644
>> index 000000000000..c9205f9ff524
>> --- /dev/null
>> +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c
>> @@ -0,0 +1,29 @@
>> +
>> +#include "drm_sched_tests.h"
>> +
>> +struct drm_mock_sched_entity *
>> +drm_mock_sched_entity_new(struct kunit *test,
>> + enum drm_sched_priority priority,
>> + struct drm_mock_scheduler *sched)
>
>
> I personally do like this function head style and
>
>
>> +{
>> + struct drm_sched_mock_entity *entity;
>> + int ret;
>> +
>> + entity = kunit_kmalloc(test, sizeof(*entity), GFP_KERNEL);
>> + KUNIT_ASSERT_NOT_NULL(test, entity);
>> +
>> + ret = drm_sched_entity_init(&entity->base,
>> + priority,
>> + &sched->base, 1,
>> + NULL);
>> + KUNIT_ASSERT_EQ(test, ret, 0);
>> +
>> + entity->test = test;
>> +
>> + return entity;
>> +}
>> +
>> +void drm_mock_sched_entity_free(struct drm_mock_sched_entity
>> *entity)
>
> do believe it should then be used consistently everywhere, regardless
> of length.
>
>
>> +{
>> + drm_sched_entity_fini(&entity->base);
>> +}
>> diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_job.c
>> b/drivers/gpu/drm/scheduler/tests/drm_mock_job.c
>> new file mode 100644
>> index 000000000000..d94568cf3da9
>> --- /dev/null
>> +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_job.c
>> @@ -0,0 +1,3 @@
>> +
>> +#include "drm_sched_tests.h"
>> +
>> diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c
>> b/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c
>> new file mode 100644
>> index 000000000000..f1985900a6ba
>> --- /dev/null
>> +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c
>> @@ -0,0 +1,254 @@
>> +
>> +#include "drm_sched_tests.h"
>> +
>> +struct drm_mock_sched_entity *
>> +drm_mock_new_sched_entity(struct kunit *test,
>> + enum drm_sched_priority priority,
>> + struct drm_mock_scheduler *sched)
>> +{
>> + struct drm_mock_sched_entity *entity;
>> + struct drm_gpu_scheduler *drm_sched;
>> + int ret;
>> +
>> + entity = kunit_kzalloc(test, sizeof(*entity), GFP_KERNEL);
>> + KUNIT_ASSERT_NOT_NULL(test, entity);
>> +
>> + drm_sched = &sched->base;
>> + ret = drm_sched_entity_init(&entity->base,
>> + priority,
>> + &drm_sched, 1,
>> + NULL);
>> + KUNIT_ASSERT_EQ(test, ret, 0);
>> +
>> + entity->test = test;
>> +
>> + return entity;
>> +}
>> +
>> +void drm_mock_sched_entity_free(struct drm_mock_sched_entity
>> *entity)
>> +{
>> + drm_sched_entity_destroy(&entity->base);
>> +}
>> +
>> +static enum hrtimer_restart
>> +drm_sched_mock_job_signal_timer(struct hrtimer *hrtimer)
>
> I think we should get up with a consistent naming convention. Some
> things are called drm_mock_sched_, some others drm_sched_mock_.
> As far as I saw, drm_mock_* does not yet exist.
> So do you want to introduce drm_mock as something generic for drm? or
> just for drm/sched?
This one is literally the only when where I tranposed the two. But it is
also local and I am not too bothered about naming conventions of a
local functions. If it were to me I would favour brevity and emit the
long drm_.. prefixes which IMO do not help readability. If anything,
seeing something with a long drm_.. prefix can only be confusing since
one can assume it is some sort of exported interface.
I will change this instance.
>> +{
>> + struct drm_mock_sched_job *upto =
>> + container_of(hrtimer, typeof(*upto), timer);
>> + struct drm_mock_scheduler *sched =
>> + drm_sched_to_mock_sched(upto->base.sched);
>> + struct drm_mock_sched_job *job, *next;
>> + ktime_t now = ktime_get();
>> + unsigned long flags;
>> + LIST_HEAD(signal);
>> +
>> + spin_lock_irqsave(&sched->lock, flags);
>> + list_for_each_entry_safe(job, next, &sched->job_list, link)
>> {
>> + if (!job->duration_us)
>> + break;
>> +
>> + if (ktime_before(now, job->finish_at))
>> + break;
>> +
>> + list_move_tail(&job->link, &signal);
>> + sched->seqno = job->hw_fence.seqno;
>> + }
>> + spin_unlock_irqrestore(&sched->lock, flags);
>> +
>> + list_for_each_entry(job, &signal, link) {
>> + dma_fence_signal(&job->hw_fence);
>> + dma_fence_put(&job->hw_fence);
>> + }
>> +
>> + return HRTIMER_NORESTART;
>> +}
>> +
>> +struct drm_mock_sched_job *
>> +drm_mock_new_sched_job(struct kunit *test,
>> + struct drm_mock_sched_entity *entity)
>> +{
>> + struct drm_mock_sched_job *job;
>> + int ret;
>> +
>> + job = kunit_kzalloc(test, sizeof(*job), GFP_KERNEL);
>> + KUNIT_ASSERT_NOT_NULL(test, job);
>> +
>> + ret = drm_sched_job_init(&job->base,
>> + &entity->base,
>> + 1,
>> + NULL);
>> + KUNIT_ASSERT_EQ(test, ret, 0);
>> +
>> + job->test = test;
>> +
>> + spin_lock_init(&job->lock);
>> + INIT_LIST_HEAD(&job->link);
>> + hrtimer_init(&job->timer, CLOCK_MONOTONIC,
>> HRTIMER_MODE_ABS);
>> + job->timer.function = drm_sched_mock_job_signal_timer;
>> +
>> + return job;
>> +}
>> +
>> +static const char *drm_mock_sched_hw_fence_driver_name(struct
>> dma_fence *fence)
>> +{
>> + return "drm_mock_sched";
>> +}
>> +
>> +static const char *
>> +drm_mock_sched_hw_fence_timeline_name(struct dma_fence *fence)
>> +{
>> + struct drm_mock_sched_job *job =
>> + container_of(fence, typeof(*job), hw_fence);
>> +
>> + return (const char *)job->base.sched->name;
>> +}
>> +static void drm_mock_sched_hw_fence_release(struct dma_fence *fence)
>> +{
>> + struct drm_mock_sched_job *job =
>> + container_of(fence, typeof(*job), hw_fence);
>> +
>> + hrtimer_cancel(&job->timer);
>> +
>> + /* Freed by the kunit framework */
>> +}
>> +
>> +static const struct dma_fence_ops drm_mock_sched_hw_fence_ops = {
>> + .get_driver_name = drm_mock_sched_hw_fence_driver_name,
>> + .get_timeline_name = drm_mock_sched_hw_fence_timeline_name,
>> + .release = drm_mock_sched_hw_fence_release,
>> +};
>> +
>> +static struct dma_fence *mock_sched_run_job(struct drm_sched_job
>> *sched_job)
>> +{
>> + struct drm_mock_scheduler *sched =
>> + drm_sched_to_mock_sched(sched_job->sched);
>> + struct drm_mock_sched_job *job =
>> drm_sched_job_to_mock_job(sched_job);
>> +
>> + dma_fence_init(&job->hw_fence,
>> + &drm_mock_sched_hw_fence_ops,
>> + &job->lock,
>> + sched->hw_fence.context,
>> + atomic_inc_return(&sched->hw_fence.seqno));
>> +
>> + dma_fence_get(&job->hw_fence); /* Reference for the job_list
>> */
>> +
>> + spin_lock_irq(&sched->lock);
>> + if (job->duration_us) {
>> + ktime_t prev_finish_at = 0;
>> +
>> + if (!list_empty(&sched->job_list)) {
>> + struct drm_mock_sched_job *prev =
>> + list_last_entry(&sched->job_list,
>> typeof(*prev),
>> + link);
>> +
>> + prev_finish_at = prev->finish_at;
>> + }
>> +
>> + if (!prev_finish_at)
>> + prev_finish_at = ktime_get();
>> +
>> + job->finish_at = ktime_add_us(prev_finish_at, job-
>>> duration_us);
>> + }
>> + list_add_tail(&job->link, &sched->job_list);
>> + if (job->finish_at)
>> + hrtimer_start(&job->timer, job->finish_at,
>> HRTIMER_MODE_ABS);
>> + spin_unlock_irq(&sched->lock);
>> +
>> + return &job->hw_fence;
>> +}
>> +
>> +static enum drm_gpu_sched_stat
>> +mock_sched_timedout_job(struct drm_sched_job *sched_job)
>> +{
>> + return DRM_GPU_SCHED_STAT_ENODEV;
>> +}
>> +
>> +static void mock_sched_free_job(struct drm_sched_job *sched_job)
>> +{
>> + drm_sched_job_cleanup(sched_job);
>> +}
>> +
>> +static const struct drm_sched_backend_ops drm_mock_scheduler_ops = {
>> + .run_job = mock_sched_run_job,
>> + .timedout_job = mock_sched_timedout_job,
>> + .free_job = mock_sched_free_job
>> +};
>> +
>> +struct drm_mock_scheduler *drm_mock_new_scheduler(struct kunit
>> *test)
>> +{
>> + struct drm_mock_scheduler *sched;
>> + int ret;
>> +
>> + sched = kunit_kzalloc(test, sizeof(*sched), GFP_KERNEL);
>> + KUNIT_ASSERT_NOT_NULL(test, sched);
>> +
>> + ret = drm_sched_init(&sched->base,
>> + &drm_mock_scheduler_ops,
>> + NULL, /* wq */
>> + DRM_SCHED_PRIORITY_COUNT,
>> + U32_MAX, /* max credits */
>> + UINT_MAX, /* hang limit */
>> + MAX_SCHEDULE_TIMEOUT, /* timeout */
>> + NULL, /* timeout wq */
>> + NULL, /* score */
>> + "drm-mock-scheduler",
>> + NULL /* dev */);
>> + KUNIT_ASSERT_EQ(test, ret, 0);
>> +
>> + sched->test = test;
>> + sched->hw_fence.context = dma_fence_context_alloc(1);
>> + atomic_set(&sched->hw_fence.seqno, 0);
>> + INIT_LIST_HEAD(&sched->job_list);
>> + spin_lock_init(&sched->lock);
>> +
>> + return sched;
>> +}
>> +
>> +void drm_mock_scheduler_fini(struct drm_mock_scheduler *sched)
>> +{
>> + struct drm_mock_sched_job *job, *next;
>> + unsigned long flags;
>> + LIST_HEAD(signal);
>> +
>> + spin_lock_irqsave(&sched->lock, flags);
>> + list_for_each_entry_safe(job, next, &sched->job_list, link)
>> + list_move_tail(&job->link, &signal);
>> + spin_unlock_irqrestore(&sched->lock, flags);
>
> So maybe you can help me get up to speed here a bit. What is the
> purpose behind job->link? Is the general idea documented somewhere?
List versus link suffix convention I use to distinguish struct list_head
which is a list versus struct list_head which is used to put something
on the list.
In this case job->link is for the mock GPU driver to keep track of jobs
which have been submitted to the "hardware" for "execution".
I will of course document these things once the high level design is
settled.
>> +
>> + list_for_each_entry(job, &signal, link) {
>> + hrtimer_cancel(&job->timer);
>> + dma_fence_put(&job->hw_fence);
>> + }
>> +
>> + drm_sched_fini(&sched->base);
>> +}
>> +
>> +unsigned int drm_mock_sched_advance(struct drm_mock_scheduler
>> *sched,
>> + unsigned int num)
>> +{
>> + struct drm_mock_sched_job *job, *next;
>> + unsigned int found = 0;
>> + unsigned long flags;
>> + LIST_HEAD(signal);
>> +
>> + spin_lock_irqsave(&sched->lock, flags);
>> + if (WARN_ON_ONCE(sched->seqno + num < sched->seqno))
>> + goto unlock;
>> + sched->seqno += num;
>> + list_for_each_entry_safe(job, next, &sched->job_list, link)
>> {
>> + if (sched->seqno < job->hw_fence.seqno)
>> + break;
>> +
>> + list_move_tail(&job->link, &signal);
>> + found++;
>> + }
>> +unlock:
>> + spin_unlock_irqrestore(&sched->lock, flags);
>> +
>> + list_for_each_entry(job, &signal, link) {
>> + dma_fence_signal(&job->hw_fence);
>> + dma_fence_put(&job->hw_fence);
>> + }
>> +
>> + return found;
>> +}
>> diff --git a/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h
>> b/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h
>> new file mode 100644
>> index 000000000000..421ee2712985
>> --- /dev/null
>> +++ b/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h
>> @@ -0,0 +1,124 @@
>> +
>> +#include <kunit/test.h>
>> +#include <linux/atomic.h>
>> +#include <linux/dma-fence.h>
>> +#include <linux/hrtimer.h>
>> +#include <linux/ktime.h>
>> +#include <linux/list.h>
>> +#include <linux/atomic.h>
>> +#include <linux/mutex.h>
>> +#include <linux/types.h>
>> +
>> +#include <drm/gpu_scheduler.h>
>> +
>> +struct drm_mock_scheduler {
>> + struct drm_gpu_scheduler base;
>> +
>> + struct kunit *test;
>> +
>> + struct {
>> + u64 context;
>> + atomic_t seqno;
>> + } hw_fence;
>
> Hm, well, so this is confusing. drm_mock_sched_job below contains an
> actual struct dma_fence that is also called hw_fence, whereas this here
> seems to be some pseudo-hw_fence?
>
> What is its purpose?
>
> I believe we agreed that "Hardware fence" should always mean a fence
> per job that is signaled by the hardware (driver interrupt) once the
> job is done.
>
> If this hw_fence here is the same, why is it per scheduler? That's
> confusing.
Mock_job->hw_fence is what the mock GPU driver returns from the
sched->run_job().
Mock_sched->hw_fence is what the mock GPU driver uses to track execution
of jobs submitted to it for execution. If Irename this to hw_timeline
will it make it clearer?
>> +
>> + spinlock_t lock;
>
> Maybe document the lock's purpose
>
>
>> + unsigned int seqno;
>> + struct list_head job_list;
>> +};
>> +
>> +struct drm_mock_sched_entity {
>> + struct drm_sched_entity base;
>> +
>> + struct kunit *test;
>> +};
>> +
>> +struct drm_mock_sched_job {
>> + struct drm_sched_job base;
>> +
>> + struct list_head link;
>> + struct hrtimer timer;
>> +
>> + unsigned int duration_us;
>> + ktime_t finish_at;
>> +
>> + spinlock_t lock;
>
> Same
>
>> + struct dma_fence hw_fence;
>> +
>> + struct kunit *test;
>> +};
>> +
>> +static inline struct drm_mock_scheduler *
>> +drm_sched_to_mock_sched(struct drm_gpu_scheduler *sched)
>> +{
>> + return container_of(sched, struct drm_mock_scheduler, base);
>> +};
>> +
>> +static inline struct drm_mock_sched_entity *
>> +drm_sched_entity_to_mock_entity(struct drm_sched_entity
>> *sched_entity)
>> +{
>> + return container_of(sched_entity, struct
>> drm_mock_sched_entity, base);
>> +};
>> +
>> +static inline struct drm_mock_sched_job *
>> +drm_sched_job_to_mock_job(struct drm_sched_job *sched_job)
>> +{
>> + return container_of(sched_job, struct drm_mock_sched_job,
>> base);
>> +};
>> +
>> +struct drm_mock_scheduler *drm_mock_new_scheduler(struct kunit
>> *test);
>> +void drm_mock_scheduler_fini(struct drm_mock_scheduler *sched);
>> +unsigned int drm_mock_sched_advance(struct drm_mock_scheduler
>> *sched,
>> + unsigned int num);
>> +
>> +struct drm_mock_sched_entity *
>> +drm_mock_new_sched_entity(struct kunit *test,
>> + enum drm_sched_priority priority,
>> + struct drm_mock_scheduler *sched);
>> +void drm_mock_sched_entity_free(struct drm_mock_sched_entity
>> *entity);
>> +
>> +struct drm_mock_sched_job *
>> +drm_mock_new_sched_job(struct kunit *test,
>> + struct drm_mock_sched_entity *entity);
>> +
>> +static inline void drm_mock_sched_job_submit(struct
>> drm_mock_sched_job *job)
>> +{
>> + drm_sched_job_arm(&job->base);
>> + drm_sched_entity_push_job(&job->base);
>> +}
>> +
>> +static inline void
>> +drm_mock_sched_job_set_duration_us(struct drm_mock_sched_job *job,
>> + unsigned int duration_us)
>> +{
>> + job->duration_us = duration_us;
>> +}
>> +
>> +static inline bool
>> +drm_mock_sched_job_is_finished(struct drm_mock_sched_job *job)
>> +{
>> + return dma_fence_is_signaled(&job->base.s_fence->finished);
>> +}
>> +
>> +static inline bool
>> +drm_mock_sched_job_wait_finished(struct drm_mock_sched_job *job,
>> long timeout)
>> +{
>> + long ret;
>> +
>> + ret = dma_fence_wait_timeout(&job->base.s_fence->finished,
>> + false,
>> + timeout);
>> +
>> + return ret != 0;
>> +}
>> +
>> +static inline long
>> +drm_mock_sched_job_wait_scheduled(struct drm_mock_sched_job *job,
>> long timeout)
>> +{
>> + long ret;
>> +
>> + ret = dma_fence_wait_timeout(&job->base.s_fence->scheduled,
>> + false,
>> + timeout);
>> +
>> + return ret != 0;
>> +}
>> diff --git a/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c
>> b/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c
>> new file mode 100644
>> index 000000000000..6fd39bea95b1
>> --- /dev/null
>> +++ b/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c
>> @@ -0,0 +1,247 @@
>> +
>> +#include "drm_sched_tests.h"
>> +
>> +static int drm_sched_basic_init(struct kunit *test)
>> +{
>> + test->priv = drm_mock_new_scheduler(test);
>> +
>> + return 0;
>> +}
>> +
>> +static void drm_sched_basic_exit(struct kunit *test)
>> +{
>> + struct drm_mock_scheduler *sched = test->priv;
>> +
>> + drm_mock_scheduler_fini(sched);
>> +}
>> +
>> +static void drm_sched_basic_submit(struct kunit *test)
>> +{
>> + struct drm_mock_scheduler *sched = test->priv;
>> + struct drm_mock_sched_entity *entity;
>> + struct drm_mock_sched_job *job;
>> + unsigned int i;
>> + bool done;
>> +
>> + /*
>> + * Submit one job to the scheduler and verify that it gets
>> scheduled
>> + * and completed only when the mock hw backend processes it.
>> + */
>> +
>> + entity = drm_mock_new_sched_entity(test,
>> +
>> DRM_SCHED_PRIORITY_NORMAL,
>> + sched);
>> + job = drm_mock_new_sched_job(test, entity);
>> +
>> + drm_mock_sched_job_submit(job);
>> +
>> + done = drm_mock_sched_job_wait_scheduled(job, HZ);
>> + KUNIT_ASSERT_EQ(test, done, true);
>> +
>> + done = drm_mock_sched_job_wait_finished(job, HZ / 2);
>> + KUNIT_ASSERT_EQ(test, done, false);
>> +
>> + i = drm_mock_sched_advance(sched, 1);
>> + KUNIT_ASSERT_EQ(test, i, 1);
>> +
>> + done = drm_mock_sched_job_wait_finished(job, HZ);
>> + KUNIT_ASSERT_EQ(test, done, true);
>> +
>> + drm_mock_sched_entity_free(entity);
>> +}
>> +
>> +static void drm_sched_basic_queue(struct kunit *test)
>> +{
>> + struct drm_mock_scheduler *sched = test->priv;
>> + struct drm_mock_sched_entity *entity;
>> + struct drm_mock_sched_job *job;
>> + const unsigned int qd = 100;
>
> Not the best variable name – is this "queue depth"? Why is it 100? ->
> global define & document
I wouldn't promote this to global and TBH if you look how small this
test function is it feels pretty self documenting.
>
>
>> + unsigned int i;
>> + bool done;
>> +
>> + /*
>> + * Submit a queue of jobs on the same entity, let them be
>> completed by
>> + * the mock hw backend and check the status of the last job.
>> + */
>> +
>> + entity = drm_mock_new_sched_entity(test,
>> +
>> DRM_SCHED_PRIORITY_NORMAL,
>> + sched);
>> +
>> + for (i = 0; i < qd; i++) {
>> + job = drm_mock_new_sched_job(test, entity);
>> + drm_mock_sched_job_set_duration_us(job, 1000);
>> + drm_mock_sched_job_submit(job);
>> + }
>> +
>> + done = drm_mock_sched_job_wait_finished(job, HZ);
>> + KUNIT_ASSERT_EQ(test, done, true);
>> +
>> + drm_mock_sched_entity_free(entity);
>> +}
>> +
>> +static void drm_sched_basic_chain(struct kunit *test)
>> +{
>> + struct drm_mock_scheduler *sched = test->priv;
>> + struct drm_mock_sched_job *job, *prev = NULL;
>> + struct drm_mock_sched_entity *entity;
>> + const unsigned int qd = 100;
>> + unsigned int i;
>> + bool done;
>> +
>> + /*
>> + * Submit a queue of jobs on the same entity but with an
>> explicit
>> + * chain of dependencies between them. Let the jobs be
>> completed by
>> + * the mock hw backend and check the status of the last job.
>> + */
>> +
>> + entity = drm_mock_new_sched_entity(test,
>> +
>> DRM_SCHED_PRIORITY_NORMAL,
>> + sched);
>> +
>> + for (i = 0; i < qd; i++) {
>> + job = drm_mock_new_sched_job(test, entity);
>> + drm_mock_sched_job_set_duration_us(job, 1000);
>> + if (prev)
>> + drm_sched_job_add_dependency(&job->base,
>> +
>> dma_fence_get(&prev->base.s_fence->finished));
>> + drm_mock_sched_job_submit(job);
>> + prev = job;
>> + }
>> +
>> + done = drm_mock_sched_job_wait_finished(job, HZ);
>> + KUNIT_ASSERT_EQ(test, done, true);
>> +
>> + drm_mock_sched_entity_free(entity);
>> +}
>> +
>> +static void drm_sched_basic_entities(struct kunit *test)
>> +{
>> + struct drm_mock_scheduler *sched = test->priv;
>> + struct drm_mock_sched_entity *entity[4];
>> + struct drm_mock_sched_job *job;
>> + const unsigned int qd = 100;
>> + unsigned int i, cur_ent = 0;
>> + bool done;
>> +
>> + /*
>> + * Submit a queue of jobs across different entities, let
>> them be
>> + * completed by the mock hw backend and check the status of
>> the last
>> + * job.
>> + */
>> +
>> + for (i = 0; i < ARRAY_SIZE(entity); i++)
>> + entity[i] = drm_mock_new_sched_entity(test,
>> +
>> DRM_SCHED_PRIORITY_NORMAL,
>> + sched);
>> +
>> + for (i = 0; i < qd; i++) {
>> + job = drm_mock_new_sched_job(test,
>> entity[cur_ent++]);
>> + cur_ent %= ARRAY_SIZE(entity);
>> + drm_mock_sched_job_set_duration_us(job, 1000);
>> + drm_mock_sched_job_submit(job);
>> + }
>> +
>> + done = drm_mock_sched_job_wait_finished(job, HZ);
>> + KUNIT_ASSERT_EQ(test, done, true);
>> +
>> + for (i = 0; i < ARRAY_SIZE(entity); i++)
>> + drm_mock_sched_entity_free(entity[i]);
>> +}
>> +
>> +static void drm_sched_basic_entities_chain(struct kunit *test)
>> +{
>> + struct drm_mock_scheduler *sched = test->priv;
>> + struct drm_mock_sched_job *job, *prev = NULL;
>> + struct drm_mock_sched_entity *entity[4];
>> + const unsigned int qd = 100;
>> + unsigned int i, cur_ent = 0;
>> + bool done;
>> +
>> + /*
>> + * Submit a queue of jobs across different entities with an
>> explicit
>> + * chain of dependencies between them. Let the jobs be
>> completed by
>> + * the mock hw backend and check the status of the last job.
>> + */
>> +
>> + for (i = 0; i < ARRAY_SIZE(entity); i++)
>> + entity[i] = drm_mock_new_sched_entity(test,
>> +
>> DRM_SCHED_PRIORITY_NORMAL,
>> + sched);
>> +
>> + for (i = 0; i < qd; i++) {
>> + job = drm_mock_new_sched_job(test,
>> entity[cur_ent++]);
>> + cur_ent %= ARRAY_SIZE(entity);
>> + drm_mock_sched_job_set_duration_us(job, 1000);
>> + if (prev)
>> + drm_sched_job_add_dependency(&job->base,
>> +
>> dma_fence_get(&prev->base.s_fence->finished));
>> + drm_mock_sched_job_submit(job);
>> + prev = job;
>> + }
>> +
>> + done = drm_mock_sched_job_wait_finished(job, HZ);
>> + KUNIT_ASSERT_EQ(test, done, true);
>> +
>> + for (i = 0; i < ARRAY_SIZE(entity); i++)
>> + drm_mock_sched_entity_free(entity[i]);
>> +}
>> +
>> +static void drm_sched_basic_entity_cleanup(struct kunit *test)
>> +{
>> + struct drm_mock_sched_job *job, *mid, *prev = NULL;
>> + struct drm_mock_scheduler *sched = test->priv;
>> + struct drm_mock_sched_entity *entity[4];
>
> 4 is used in several places, so could be defined globally. In case
> there's a special reason for why it's 4, that could be mentioned, too
It's completely arbitrary. In this test we just need a bunch of entities
to be active on the scheduler as we trigger a client exit in the middle
of it. IMO the comment few lines below should be good enough to explain
the idea. I fear that promoting this to a global define would just give
it more weight that what it has. Then some test would want a different
number etc.
Regards,
Tvrtko
>> + const unsigned int qd = 100;
>> + unsigned int i, cur_ent = 0;
>> + bool done;
>> +
>> + /*
>> + * Submit a queue of jobs across different entities with an
>> explicit
>> + * chain of dependencies between them and trigger entity
>> cleanup while
>> + * the queue is still being processed.
>> + */
>> +
>> + for (i = 0; i < ARRAY_SIZE(entity); i++)
>> + entity[i] = drm_mock_new_sched_entity(test,
>> +
>> DRM_SCHED_PRIORITY_NORMAL,
>> + sched);
>> +
>> + for (i = 0; i < qd; i++) {
>> + job = drm_mock_new_sched_job(test,
>> entity[cur_ent++]);
>> + cur_ent %= ARRAY_SIZE(entity);
>> + drm_mock_sched_job_set_duration_us(job, 1000);
>> + if (prev)
>> + drm_sched_job_add_dependency(&job->base,
>> +
>> dma_fence_get(&prev->base.s_fence->finished));
>> + drm_mock_sched_job_submit(job);
>> + if (i == qd / 2)
>> + mid = job;
>> + prev = job;
>> + }
>> +
>> + done = drm_mock_sched_job_wait_finished(mid, HZ);
>> + KUNIT_ASSERT_EQ(test, done, true);
>> +
>> + /* Exit with half of the queue still pending to be executed.
>> */
>> + for (i = 0; i < ARRAY_SIZE(entity); i++)
>> + drm_mock_sched_entity_free(entity[i]);}
>> +
>> +static struct kunit_case drm_sched_basic_tests[] = {
>> + KUNIT_CASE(drm_sched_basic_submit),
>> + KUNIT_CASE(drm_sched_basic_queue),
>> + KUNIT_CASE(drm_sched_basic_chain),
>> + KUNIT_CASE(drm_sched_basic_entities),
>> + KUNIT_CASE(drm_sched_basic_entities_chain),
>> + KUNIT_CASE(drm_sched_basic_entity_cleanup),
>> + {}
>> +};
>> +
>> +static struct kunit_suite drm_sched_basic = {
>> + .name = "drm_sched_basic_tests",
>> + .init = drm_sched_basic_init,
>> + .exit = drm_sched_basic_exit,
>> + .test_cases = drm_sched_basic_tests,
>> +};
>> +
>> +kunit_test_suite(drm_sched_basic);
>
More information about the dri-devel
mailing list