[Intel-xe] [PATCH] drm/xe: add lockdep annotation for xe_device_mem_access_put()
Rodrigo Vivi
rodrigo.vivi at intel.com
Fri Jul 21 18:59:22 UTC 2023
On Fri, Jul 21, 2023 at 01:34:25PM +0100, Matthew Auld wrote:
> The main motivation is with d3cold which will make the suspend and
> resume callbacks even more scary, but is useful regardless. We already
> have the needed annotation on the acquire side with
> xe_device_mem_access_get(), and by adding the annotation on the release
> side we should have a lot more confidence that our locking hierarchy is
> correct.
good idea!
Reviewed-by: Rodrigo Vivi <rodrigo.vivi at intel.com>
>
> Signed-off-by: Matthew Auld <matthew.auld at intel.com>
> Cc: Thomas Hellström <thomas.hellstrom at linux.intel.com>
> Cc: Anshuman Gupta <anshuman.gupta at intel.com>
> Cc: Rodrigo Vivi <rodrigo.vivi at intel.com>
> ---
> drivers/gpu/drm/xe/xe_device.c | 2 +-
> drivers/gpu/drm/xe/xe_device.h | 4 ++++
> drivers/gpu/drm/xe/xe_pm.c | 24 ++++++++++++++++++++++++
> 3 files changed, 29 insertions(+), 1 deletion(-)
>
> diff --git a/drivers/gpu/drm/xe/xe_device.c b/drivers/gpu/drm/xe/xe_device.c
> index 1c57944014e0..fed3015e2d96 100644
> --- a/drivers/gpu/drm/xe/xe_device.c
> +++ b/drivers/gpu/drm/xe/xe_device.c
> @@ -36,7 +36,7 @@
> #include "xe_wait_user_fence.h"
>
> #ifdef CONFIG_LOCKDEP
> -static struct lockdep_map xe_device_mem_access_lockdep_map = {
> +struct lockdep_map xe_device_mem_access_lockdep_map = {
> .name = "xe_device_mem_access_lockdep_map"
> };
> #endif
> diff --git a/drivers/gpu/drm/xe/xe_device.h b/drivers/gpu/drm/xe/xe_device.h
> index 8b085ffdc5f8..593accb68281 100644
> --- a/drivers/gpu/drm/xe/xe_device.h
> +++ b/drivers/gpu/drm/xe/xe_device.h
> @@ -16,6 +16,10 @@ struct xe_file;
> #include "xe_force_wake.h"
> #include "xe_macros.h"
>
> +#ifdef CONFIG_LOCKDEP
> +extern struct lockdep_map xe_device_mem_access_lockdep_map;
> +#endif
> +
> static inline struct xe_device *to_xe_device(const struct drm_device *dev)
> {
> return container_of(dev, struct xe_device, drm);
> diff --git a/drivers/gpu/drm/xe/xe_pm.c b/drivers/gpu/drm/xe/xe_pm.c
> index 17a69b7af155..d1b2aa52ea03 100644
> --- a/drivers/gpu/drm/xe/xe_pm.c
> +++ b/drivers/gpu/drm/xe/xe_pm.c
> @@ -199,6 +199,29 @@ int xe_pm_runtime_suspend(struct xe_device *xe)
> /* Disable access_ongoing asserts and prevent recursive pm calls */
> xe_pm_write_callback_task(xe, current);
>
> + /*
> + * The actual xe_device_mem_access_put() is always async underneath, so
> + * exactly where that is called should makes no difference to us. However
> + * we still need to be very careful with the locks that this callback
> + * acquires and the locks that are acquired and held by any callers of
> + * xe_device_mem_access_get(). We already have the matching annotation
> + * on that side, but we also need it here. For example lockdep should be
> + * able to tell us if the following scenario is in theory possible:
> + *
> + * CPU0 | CPU1 (kworker)
> + * lock(A) |
> + * | xe_pm_runtime_suspend()
> + * | lock(A)
> + * xe_device_mem_access_get() |
> + *
> + * This will clearly deadlock since rpm core needs to wait for
> + * xe_pm_runtime_suspend() to complete, but here we are holding lock(A)
> + * on CPU0 which prevents CPU1 making forward progress. With the
> + * annotation here and in xe_device_mem_access_get() lockdep will see
> + * the potential lock inversion and give us a nice splat.
> + */
> + lock_map_acquire(&xe_device_mem_access_lockdep_map);
> +
> if (xe->d3cold.allowed) {
> err = xe_bo_evict_all(xe);
> if (err)
> @@ -213,6 +236,7 @@ int xe_pm_runtime_suspend(struct xe_device *xe)
>
> xe_irq_suspend(xe);
> out:
> + lock_map_release(&xe_device_mem_access_lockdep_map);
> xe_pm_write_callback_task(xe, NULL);
> return err;
> }
> --
> 2.41.0
>
More information about the Intel-xe
mailing list