[PATCH] panfrost: Fix job timeout handling
Steven Price
steven.price at arm.com
Thu Oct 1 14:49:39 UTC 2020
On 01/10/2020 15:01, Boris Brezillon wrote:
> If more than two or more jobs end up timeout-ing concurrently, only one
> of them (the one attached to the scheduler acquiring the lock) is fully
> handled. The other one remains in a dangling state where it's no longer
> part of the scheduling queue, but still blocks something in scheduler
> thus leading to repetitive timeouts when new jobs are queued.
>
> Let's make sure all bad jobs are properly handled by the thread acquiring
> the lock.
>
> Signed-off-by: Boris Brezillon <boris.brezillon at collabora.com>
> Fixes: f3ba91228e8e ("drm/panfrost: Add initial panfrost driver")
> Cc: <stable at vger.kernel.org>
> ---
> drivers/gpu/drm/panfrost/panfrost_job.c | 18 ++++++++++++++----
> 1 file changed, 14 insertions(+), 4 deletions(-)
>
> diff --git a/drivers/gpu/drm/panfrost/panfrost_job.c b/drivers/gpu/drm/panfrost/panfrost_job.c
> index 30e7b7196dab..e87edca51d84 100644
> --- a/drivers/gpu/drm/panfrost/panfrost_job.c
> +++ b/drivers/gpu/drm/panfrost/panfrost_job.c
> @@ -25,7 +25,7 @@
>
> struct panfrost_queue_state {
> struct drm_gpu_scheduler sched;
> -
> + struct drm_sched_job *bad;
> u64 fence_context;
> u64 emit_seqno;
> };
> @@ -392,19 +392,29 @@ static void panfrost_job_timedout(struct drm_sched_job *sched_job)
> job_read(pfdev, JS_TAIL_LO(js)),
> sched_job);
>
> + /*
> + * Collect the bad job here so it can be processed by the thread
> + * acquiring the reset lock.
> + */
> + pfdev->js->queue[js].bad = sched_job;
> +
> if (!mutex_trylock(&pfdev->reset_lock))
> return;
>
> for (i = 0; i < NUM_JOB_SLOTS; i++) {
> struct drm_gpu_scheduler *sched = &pfdev->js->queue[i].sched;
>
> - drm_sched_stop(sched, sched_job);
> if (js != i)
> /* Ensure any timeouts on other slots have finished */
> cancel_delayed_work_sync(&sched->work_tdr);
> - }
>
> - drm_sched_increase_karma(sched_job);
> + drm_sched_stop(sched, pfdev->js->queue[i].bad);
So I can see that the call to drm_sched_stop() needs to move below the
cancel_delayed_work_sync() to ensure that the update to queue->bad is
synchronised. What I'm not so sure about is whether it's possible for
the scheduler to make progress between the 'cancel' and the 'stop' -
there is a reason I wrote it the other way round...
The hole for things to go round is clearly much smaller with this
change, but I'm not sure it's completely plugged. Am I missing something?
> +
> + if (pfdev->js->queue[i].bad)
> + drm_sched_increase_karma(pfdev->js->queue[i].bad);
> +
> + pfdev->js->queue[i].bad = NULL;
> + }
>
> spin_lock_irqsave(&pfdev->js->job_lock, flags);
> for (i = 0; i < NUM_JOB_SLOTS; i++) {
>
While we're on potential holes... some more context:
> if (pfdev->jobs[i]) {
> pm_runtime_put_noidle(pfdev->dev);
> panfrost_devfreq_record_idle(pfdev);
> pfdev->jobs[i] = NULL;
> }
> }
> spin_unlock_irqrestore(&pfdev->js->job_lock, flags);
>
> panfrost_device_reset(pfdev);
>
> for (i = 0; i < NUM_JOB_SLOTS; i++)
> drm_sched_resubmit_jobs(&pfdev->js->queue[i].sched);
>
> /* restart scheduler after GPU is usable again */
> for (i = 0; i < NUM_JOB_SLOTS; i++)
> drm_sched_start(&pfdev->js->queue[i].sched, true);
>
> mutex_unlock(&pfdev->reset_lock);
I'm wondering whether the mutex_unlock() should actually happen before
the drm_sched_start() - in the (admittedly very unlikely) case where a
timeout occurs before all the drm_sched_start() calls have completed
it's possible for the timeout to be completely missed because the mutex
is still held.
Thanks,
Steve
More information about the dri-devel
mailing list