[PATCH 3/7] drm/xe: Relax runtime pm protection during execution
Thomas Hellström
thomas.hellstrom at linux.intel.com
Mon May 6 11:57:22 UTC 2024
On Fri, 2024-05-03 at 15:13 -0400, Rodrigo Vivi wrote:
> In the regular use case scenario, user space will create an
> exec queue, and keep it alive to reuse that until it is done
> with that kind of workload.
>
> For the regular desktop cases, it means that the exec_queue
> is alive even on idle scenarios where display goes off. This
> is unacceptable since this would entirely block runtime PM
> indefinitely, blocking deeper Package-C state. This would be
> a waste drainage of power.
>
> So, let's limit the protection for the moments where we are
> really sending jobs for execution.
>
> This patch also introduces a protection to guc submit fini,
> which is not protected anymore by the exec_queue life.
Nit: Imperative language ^^^
>
> Cc: Matthew Brost <matthew.brost at intel.com>
> Cc: Francois Dugast <francois.dugast at intel.com>
> Signed-off-by: Rodrigo Vivi <rodrigo.vivi at intel.com>
Otherwise LGTM.
Reviewed-by: Thomas Hellström <thomas.hellstrom at linux.intel.com>
> ---
> drivers/gpu/drm/xe/xe_exec_queue.c | 14 --------------
> drivers/gpu/drm/xe/xe_guc_submit.c | 3 +++
> drivers/gpu/drm/xe/xe_sched_job.c | 10 +++-------
> 3 files changed, 6 insertions(+), 21 deletions(-)
>
> diff --git a/drivers/gpu/drm/xe/xe_exec_queue.c
> b/drivers/gpu/drm/xe/xe_exec_queue.c
> index 395de93579fa..33b03605a1d1 100644
> --- a/drivers/gpu/drm/xe/xe_exec_queue.c
> +++ b/drivers/gpu/drm/xe/xe_exec_queue.c
> @@ -106,7 +106,6 @@ static struct xe_exec_queue
> *__xe_exec_queue_alloc(struct xe_device *xe,
>
> static int __xe_exec_queue_init(struct xe_exec_queue *q)
> {
> - struct xe_device *xe = gt_to_xe(q->gt);
> int i, err;
>
> for (i = 0; i < q->width; ++i) {
> @@ -119,17 +118,6 @@ static int __xe_exec_queue_init(struct
> xe_exec_queue *q)
> if (err)
> goto err_lrc;
>
> - /*
> - * Normally the user vm holds an rpm ref to keep the device
> - * awake, and the context holds a ref for the vm, however
> for
> - * some engines we use the kernels migrate vm underneath
> which offers no
> - * such rpm ref, or we lack a vm. Make sure we keep a ref
> here, so we
> - * can perform GuC CT actions when needed. Caller is
> expected to have
> - * already grabbed the rpm ref outside any sensitive locks.
> - */
> - if (!(q->flags & EXEC_QUEUE_FLAG_PERMANENT) && (q->flags &
> EXEC_QUEUE_FLAG_VM || !q->vm))
> - xe_pm_runtime_get_noresume(xe);
> -
> return 0;
>
> err_lrc:
> @@ -216,8 +204,6 @@ void xe_exec_queue_fini(struct xe_exec_queue *q)
>
> for (i = 0; i < q->width; ++i)
> xe_lrc_finish(q->lrc + i);
> - if (!(q->flags & EXEC_QUEUE_FLAG_PERMANENT) && (q->flags &
> EXEC_QUEUE_FLAG_VM || !q->vm))
> - xe_pm_runtime_put(gt_to_xe(q->gt));
> __xe_exec_queue_free(q);
> }
>
> diff --git a/drivers/gpu/drm/xe/xe_guc_submit.c
> b/drivers/gpu/drm/xe/xe_guc_submit.c
> index d274a139010b..86baebd49c19 100644
> --- a/drivers/gpu/drm/xe/xe_guc_submit.c
> +++ b/drivers/gpu/drm/xe/xe_guc_submit.c
> @@ -35,6 +35,7 @@
> #include "xe_macros.h"
> #include "xe_map.h"
> #include "xe_mocs.h"
> +#include "xe_pm.h"
> #include "xe_ring_ops_types.h"
> #include "xe_sched_job.h"
> #include "xe_trace.h"
> @@ -1071,6 +1072,7 @@ static void __guc_exec_queue_fini_async(struct
> work_struct *w)
> struct xe_exec_queue *q = ge->q;
> struct xe_guc *guc = exec_queue_to_guc(q);
>
> + xe_pm_runtime_get(guc_to_xe(guc));
> trace_xe_exec_queue_destroy(q);
>
> if (xe_exec_queue_is_lr(q))
> @@ -1081,6 +1083,7 @@ static void __guc_exec_queue_fini_async(struct
> work_struct *w)
>
> kfree(ge);
> xe_exec_queue_fini(q);
> + xe_pm_runtime_put(guc_to_xe(guc));
> }
>
> static void guc_exec_queue_fini_async(struct xe_exec_queue *q)
> diff --git a/drivers/gpu/drm/xe/xe_sched_job.c
> b/drivers/gpu/drm/xe/xe_sched_job.c
> index cd8a2fba5438..a4e030f5e019 100644
> --- a/drivers/gpu/drm/xe/xe_sched_job.c
> +++ b/drivers/gpu/drm/xe/xe_sched_job.c
> @@ -158,11 +158,7 @@ struct xe_sched_job *xe_sched_job_create(struct
> xe_exec_queue *q,
> for (i = 0; i < width; ++i)
> job->batch_addr[i] = batch_addr[i];
>
> - /* All other jobs require a VM to be open which has a ref */
> - if (unlikely(q->flags & EXEC_QUEUE_FLAG_KERNEL))
> - xe_pm_runtime_get_noresume(job_to_xe(job));
> - xe_device_assert_mem_access(job_to_xe(job));
> -
> + xe_pm_runtime_get_noresume(job_to_xe(job));
> trace_xe_sched_job_create(job);
> return job;
>
> @@ -191,13 +187,13 @@ void xe_sched_job_destroy(struct kref *ref)
> {
> struct xe_sched_job *job =
> container_of(ref, struct xe_sched_job, refcount);
> + struct xe_device *xe = job_to_xe(job);
>
> - if (unlikely(job->q->flags & EXEC_QUEUE_FLAG_KERNEL))
> - xe_pm_runtime_put(job_to_xe(job));
> xe_exec_queue_put(job->q);
> dma_fence_put(job->fence);
> drm_sched_job_cleanup(&job->drm);
> job_free(job);
> + xe_pm_runtime_put(xe);
> }
>
> void xe_sched_job_set_error(struct xe_sched_job *job, int error)
More information about the Intel-xe
mailing list