[PATCH] drm/vkms: avoid race-condition between flushing and destroying
Louis Chauvet
louis.chauvet at bootlin.com
Tue Nov 5 13:59:44 UTC 2024
On 29/07/23 - 19:49, Maíra Canal wrote:
> After we flush the workqueue at the commit tale, we need to make sure
> that no work is queued until we destroy the state. Currently, new work
> can be queued in the workqueue, even after the commit tale, as the
> vblank thread is still running.
>
> Therefore, to avoid a race-condition that will lead to the trigger of a
> WARN_ON() at the function vkms_atomic_crtc_destroy_state(), add a mutex
> to protect the sections where the queue is manipulated.
>
> This way we can make sure that no work will be added to the workqueue
> between flushing the queue (at the commit tail) and destroying the
> state.
>
> Signed-off-by: Maíra Canal <mairacanal at riseup.net>
Hi Maìra,
Thanks for pointing to this patch, it does not apply on drm-misc-next, but
it was simple to manually rebase (see [0]).
I think it should solve the issue, and the CI seems to agree.
But it seems to be imperfect, as it introduce a warning on mutex unlock
imbalance [1] (not always reproducable). It is better than a kernel crash
already.
Do you want/have time to continue this fix?
[0]:https://gitlab.freedesktop.org/louischauvet/kernel/-/commit/017210f48c809730296d1f562e615b666fdbcfdc
[1]:https://gitlab.freedesktop.org/louischauvet/kernel/-/jobs/66118565/viewer#L803
Thanks,
Louis Chauvet
> ---
> drivers/gpu/drm/vkms/vkms_crtc.c | 10 +++++++++-
> drivers/gpu/drm/vkms/vkms_drv.c | 1 +
> drivers/gpu/drm/vkms/vkms_drv.h | 8 ++++++++
> 3 files changed, 18 insertions(+), 1 deletion(-)
>
> diff --git a/drivers/gpu/drm/vkms/vkms_crtc.c b/drivers/gpu/drm/vkms/vkms_crtc.c
> index 3c5ebf106b66..e5ec21a0da05 100644
> --- a/drivers/gpu/drm/vkms/vkms_crtc.c
> +++ b/drivers/gpu/drm/vkms/vkms_crtc.c
> @@ -49,7 +49,9 @@ static enum hrtimer_restart vkms_vblank_simulate(struct hrtimer *timer)
> state->crc_pending = true;
> spin_unlock(&output->composer_lock);
>
> + mutex_lock(&state->queue_lock);
> ret = queue_work(output->composer_workq, &state->composer_work);
> + mutex_unlock(&state->queue_lock);
> if (!ret)
> DRM_DEBUG_DRIVER("Composer worker already queued\n");
> }
> @@ -129,6 +131,7 @@ vkms_atomic_crtc_duplicate_state(struct drm_crtc *crtc)
>
> __drm_atomic_helper_crtc_duplicate_state(crtc, &vkms_state->base);
>
> + mutex_init(&vkms_state->queue_lock);
> INIT_WORK(&vkms_state->composer_work, vkms_composer_worker);
>
> return &vkms_state->base;
> @@ -142,6 +145,9 @@ static void vkms_atomic_crtc_destroy_state(struct drm_crtc *crtc,
> __drm_atomic_helper_crtc_destroy_state(state);
>
> WARN_ON(work_pending(&vkms_state->composer_work));
> + mutex_unlock(&vkms_state->queue_lock);
> +
> + mutex_destroy(&vkms_state->queue_lock);
> kfree(vkms_state->active_planes);
> kfree(vkms_state);
> }
> @@ -155,8 +161,10 @@ static void vkms_atomic_crtc_reset(struct drm_crtc *crtc)
> vkms_atomic_crtc_destroy_state(crtc, crtc->state);
>
> __drm_atomic_helper_crtc_reset(crtc, &vkms_state->base);
> - if (vkms_state)
> + if (vkms_state) {
> + mutex_init(&vkms_state->queue_lock);
> INIT_WORK(&vkms_state->composer_work, vkms_composer_worker);
> + }
> }
>
> static const struct drm_crtc_funcs vkms_crtc_funcs = {
> diff --git a/drivers/gpu/drm/vkms/vkms_drv.c b/drivers/gpu/drm/vkms/vkms_drv.c
> index dd0af086e7fa..9212686ca88a 100644
> --- a/drivers/gpu/drm/vkms/vkms_drv.c
> +++ b/drivers/gpu/drm/vkms/vkms_drv.c
> @@ -84,6 +84,7 @@ static void vkms_atomic_commit_tail(struct drm_atomic_state *old_state)
> struct vkms_crtc_state *vkms_state =
> to_vkms_crtc_state(old_crtc_state);
>
> + mutex_lock(&vkms_state->queue_lock);
> flush_work(&vkms_state->composer_work);
> }
>
> diff --git a/drivers/gpu/drm/vkms/vkms_drv.h b/drivers/gpu/drm/vkms/vkms_drv.h
> index c7ae6c2ba1df..83767692469a 100644
> --- a/drivers/gpu/drm/vkms/vkms_drv.h
> +++ b/drivers/gpu/drm/vkms/vkms_drv.h
> @@ -89,6 +89,14 @@ struct vkms_crtc_state {
> struct vkms_writeback_job *active_writeback;
> struct vkms_color_lut gamma_lut;
>
> + /* protects the access to the workqueue
> + *
> + * we need to hold this lock between flushing the workqueue and
> + * destroying the state to avoid work to be queued by the worker
> + * thread
> + */
> + struct mutex queue_lock;
> +
> /* below four are protected by vkms_output.composer_lock */
> bool crc_pending;
> bool wb_pending;
> --
> 2.41.0
>
More information about the dri-devel
mailing list