[PATCH] drm/xe: Clean up VM / exec queue file lock usage.
Matthew Auld
matthew.auld at intel.com
Mon Sep 23 13:10:00 UTC 2024
On 21/09/2024 02:17, Matthew Brost wrote:
> Both the VM / exec queue file lock protect the lookup and reference to
> the object, nothing more. These locks are not intended anything else
> underneath them. XA have their own locking too, so no need to take the
> VM / exec queue file lock aside from when doing a lookup and reference
> get.
>
> Add some kernel doc to make this clear and cleanup a few typos too.
>
> Signed-off-by: Matthew Brost <matthew.brost at intel.com>
> ---
> drivers/gpu/drm/xe/xe_device.c | 2 --
> drivers/gpu/drm/xe/xe_device_types.h | 14 +++++++++++---
> drivers/gpu/drm/xe/xe_drm_client.c | 9 ++++++++-
> drivers/gpu/drm/xe/xe_exec_queue.c | 2 --
> drivers/gpu/drm/xe/xe_vm.c | 4 ----
> 5 files changed, 19 insertions(+), 12 deletions(-)
>
> diff --git a/drivers/gpu/drm/xe/xe_device.c b/drivers/gpu/drm/xe/xe_device.c
> index cb5a9fd820cf..53dcece40fc5 100644
> --- a/drivers/gpu/drm/xe/xe_device.c
> +++ b/drivers/gpu/drm/xe/xe_device.c
> @@ -171,10 +171,8 @@ static void xe_file_close(struct drm_device *dev, struct drm_file *file)
> xe_exec_queue_kill(q);
> xe_exec_queue_put(q);
> }
> - mutex_lock(&xef->vm.lock);
> xa_for_each(&xef->vm.xa, idx, vm)
> xe_vm_close_and_put(vm);
> - mutex_unlock(&xef->vm.lock);
>
> xe_file_put(xef);
>
> diff --git a/drivers/gpu/drm/xe/xe_device_types.h b/drivers/gpu/drm/xe/xe_device_types.h
> index 5ad96d283a71..484fb34dde98 100644
> --- a/drivers/gpu/drm/xe/xe_device_types.h
> +++ b/drivers/gpu/drm/xe/xe_device_types.h
> @@ -593,15 +593,23 @@ struct xe_file {
> struct {
> /** @vm.xe: xarray to store VMs */
> struct xarray xa;
> - /** @vm.lock: protects file VM state */
> + /**
> + * @vm.lock: Protects VM lookup + reference and removal a from
> + * file xarray. Not an intended to be an outer lock which does
> + * thing while being held.
> + */
> struct mutex lock;
> } vm;
>
> /** @exec_queue: Submission exec queue state for file */
> struct {
> - /** @exec_queue.xe: xarray to store engines */
> + /** @exec_queue.xa: xarray to store exece queues */
> struct xarray xa;
> - /** @exec_queue.lock: protects file engine state */
> + /**
> + * @exec_queue.lock: Protects exec queue lookup + reference and
> + * removal a frommfile xarray. Not an intended to be an outer
> + * lock which does thing while being held.
> + */
> struct mutex lock;
> } exec_queue;
>
> diff --git a/drivers/gpu/drm/xe/xe_drm_client.c b/drivers/gpu/drm/xe/xe_drm_client.c
> index c4add8b38bbd..fb52a23e28f8 100644
> --- a/drivers/gpu/drm/xe/xe_drm_client.c
> +++ b/drivers/gpu/drm/xe/xe_drm_client.c
> @@ -283,8 +283,15 @@ static void show_run_ticks(struct drm_printer *p, struct drm_file *file)
>
> /* Accumulate all the exec queues from this client */
> mutex_lock(&xef->exec_queue.lock);
> - xa_for_each(&xef->exec_queue.xa, i, q)
> + xa_for_each(&xef->exec_queue.xa, i, q) {
> + xe_exec_queue_get(q);
> + mutex_unlock(&xef->exec_queue.lock);
> +
> xe_exec_queue_update_run_ticks(q);
> +
> + mutex_lock(&xef->exec_queue.lock);
> + xe_exec_queue_put(q);
> + }
> mutex_unlock(&xef->exec_queue.lock);
>
> /* Get the total GPU cycles */
> diff --git a/drivers/gpu/drm/xe/xe_exec_queue.c b/drivers/gpu/drm/xe/xe_exec_queue.c
> index 7f28b7fc68d5..7743ebdcbf4b 100644
> --- a/drivers/gpu/drm/xe/xe_exec_queue.c
> +++ b/drivers/gpu/drm/xe/xe_exec_queue.c
> @@ -635,9 +635,7 @@ int xe_exec_queue_create_ioctl(struct drm_device *dev, void *data,
> }
> }
>
> - mutex_lock(&xef->exec_queue.lock);
> err = xa_alloc(&xef->exec_queue.xa, &id, q, xa_limit_32b, GFP_KERNEL);
> - mutex_unlock(&xef->exec_queue.lock);
> if (err)
> goto kill_exec_queue;
>
> diff --git a/drivers/gpu/drm/xe/xe_vm.c b/drivers/gpu/drm/xe/xe_vm.c
> index a3d7cb7cfd22..31fe31db3fdc 100644
> --- a/drivers/gpu/drm/xe/xe_vm.c
> +++ b/drivers/gpu/drm/xe/xe_vm.c
> @@ -1765,9 +1765,7 @@ int xe_vm_create_ioctl(struct drm_device *dev, void *data,
> if (IS_ERR(vm))
> return PTR_ERR(vm);
>
> - mutex_lock(&xef->vm.lock);
> err = xa_alloc(&xef->vm.xa, &id, vm, xa_limit_32b, GFP_KERNEL);
> - mutex_unlock(&xef->vm.lock);
While reviewing this also spotted some potential uaf here and for queue:
https://patchwork.freedesktop.org/series/138995/
Reviewed-by: Matthew Auld <matthew.auld at intel.com>
> if (err)
> goto err_close_and_put;
>
> @@ -1799,9 +1797,7 @@ int xe_vm_create_ioctl(struct drm_device *dev, void *data,
> return 0;
>
> err_free_id:
> - mutex_lock(&xef->vm.lock);
> xa_erase(&xef->vm.xa, id);
> - mutex_unlock(&xef->vm.lock);
> err_close_and_put:
> xe_vm_close_and_put(vm);
>
More information about the Intel-xe
mailing list