[Intel-xe] [PATCH v4] drm/xe: Fix pagefault and access counter worker functions
Matthew Brost
matthew.brost at intel.com
Fri Nov 3 22:47:47 UTC 2023
On Fri, Nov 03, 2023 at 12:46:38PM -0700, Brian Welty wrote:
> When processing G2H messages for pagefault or access counters, we queue a
> work item and call queue_work(). This fails if the worker thread is already
> queued to run.
> The expectation is that the worker function will do more than process a
> single item and return. It needs to either process all pending items or
> requeue itself if items are pending. But requeuing will add latency and
> potential context switch can occur.
>
> We don't want to add unnecessary latency and so the worker should process
> as many faults as it can within a reasonable duration of time.
> We also do not want to hog the cpu core, so here we execute in a loop
> and requeue if still running after more than 20 ms.
> This seems reasonable framework and easy to tune this futher if needed.
>
> This resolves issues seen with several igt at xe_exec_fault_mode subtests
> where the GPU will hang when KMD ignores a pending pagefault.
>
> v2: requeue the worker instead of having an internal processing loop.
> v3: implement hybrid model of v1 and v2
> now, run for 20 msec before we will requeue if still running
> v4: only requeue in worker if queue is non-empty (Matt B)
>
> Signed-off-by: Brian Welty <brian.welty at intel.com>
With CI results:
Reviewed-by: Matthew Brost <matthew.brost at intel.com>
> ---
> drivers/gpu/drm/xe/xe_gt_pagefault.c | 82 ++++++++++++++++------------
> 1 file changed, 48 insertions(+), 34 deletions(-)
>
> diff --git a/drivers/gpu/drm/xe/xe_gt_pagefault.c b/drivers/gpu/drm/xe/xe_gt_pagefault.c
> index ab6daebbd77c..5b3585e2c125 100644
> --- a/drivers/gpu/drm/xe/xe_gt_pagefault.c
> +++ b/drivers/gpu/drm/xe/xe_gt_pagefault.c
> @@ -276,10 +276,10 @@ static void print_pagefault(struct xe_device *xe, struct pagefault *pf)
>
> #define PF_MSG_LEN_DW 4
>
> -static int get_pagefault(struct pf_queue *pf_queue, struct pagefault *pf)
> +static bool get_pagefault(struct pf_queue *pf_queue, struct pagefault *pf)
> {
> const struct xe_guc_pagefault_desc *desc;
> - int ret = 0;
> + bool ret = false;
>
> spin_lock_irq(&pf_queue->lock);
> if (pf_queue->head != pf_queue->tail) {
> @@ -303,8 +303,7 @@ static int get_pagefault(struct pf_queue *pf_queue, struct pagefault *pf)
>
> pf_queue->head = (pf_queue->head + PF_MSG_LEN_DW) %
> PF_QUEUE_NUM_DW;
> - } else {
> - ret = -1;
> + ret = true;
> }
> spin_unlock_irq(&pf_queue->lock);
>
> @@ -348,6 +347,8 @@ int xe_guc_pagefault_handler(struct xe_guc *guc, u32 *msg, u32 len)
> return full ? -ENOSPC : 0;
> }
>
> +#define USM_QUEUE_MAX_RUNTIME_MS 20
> +
> static void pf_queue_work_func(struct work_struct *w)
> {
> struct pf_queue *pf_queue = container_of(w, struct pf_queue, worker);
> @@ -355,31 +356,38 @@ static void pf_queue_work_func(struct work_struct *w)
> struct xe_device *xe = gt_to_xe(gt);
> struct xe_guc_pagefault_reply reply = {};
> struct pagefault pf = {};
> + unsigned long threshold;
> int ret;
>
> - ret = get_pagefault(pf_queue, &pf);
> - if (ret)
> - return;
> + threshold = jiffies + msecs_to_jiffies(USM_QUEUE_MAX_RUNTIME_MS);
>
> - ret = handle_pagefault(gt, &pf);
> - if (unlikely(ret)) {
> - print_pagefault(xe, &pf);
> - pf.fault_unsuccessful = 1;
> - drm_dbg(&xe->drm, "Fault response: Unsuccessful %d\n", ret);
> - }
> + while (get_pagefault(pf_queue, &pf)) {
> + ret = handle_pagefault(gt, &pf);
> + if (unlikely(ret)) {
> + print_pagefault(xe, &pf);
> + pf.fault_unsuccessful = 1;
> + drm_dbg(&xe->drm, "Fault response: Unsuccessful %d\n", ret);
> + }
> +
> + reply.dw0 = FIELD_PREP(PFR_VALID, 1) |
> + FIELD_PREP(PFR_SUCCESS, pf.fault_unsuccessful) |
> + FIELD_PREP(PFR_REPLY, PFR_ACCESS) |
> + FIELD_PREP(PFR_DESC_TYPE, FAULT_RESPONSE_DESC) |
> + FIELD_PREP(PFR_ASID, pf.asid);
>
> - reply.dw0 = FIELD_PREP(PFR_VALID, 1) |
> - FIELD_PREP(PFR_SUCCESS, pf.fault_unsuccessful) |
> - FIELD_PREP(PFR_REPLY, PFR_ACCESS) |
> - FIELD_PREP(PFR_DESC_TYPE, FAULT_RESPONSE_DESC) |
> - FIELD_PREP(PFR_ASID, pf.asid);
> + reply.dw1 = FIELD_PREP(PFR_VFID, pf.vfid) |
> + FIELD_PREP(PFR_ENG_INSTANCE, pf.engine_instance) |
> + FIELD_PREP(PFR_ENG_CLASS, pf.engine_class) |
> + FIELD_PREP(PFR_PDATA, pf.pdata);
>
> - reply.dw1 = FIELD_PREP(PFR_VFID, pf.vfid) |
> - FIELD_PREP(PFR_ENG_INSTANCE, pf.engine_instance) |
> - FIELD_PREP(PFR_ENG_CLASS, pf.engine_class) |
> - FIELD_PREP(PFR_PDATA, pf.pdata);
> + send_pagefault_reply(>->uc.guc, &reply);
>
> - send_pagefault_reply(>->uc.guc, &reply);
> + if (time_after(jiffies, threshold) &&
> + pf_queue->head != pf_queue->tail) {
> + queue_work(gt->usm.pf_wq, w);
> + break;
> + }
> + }
> }
>
> static void acc_queue_work_func(struct work_struct *w);
> @@ -544,10 +552,10 @@ static int handle_acc(struct xe_gt *gt, struct acc *acc)
>
> #define ACC_MSG_LEN_DW 4
>
> -static int get_acc(struct acc_queue *acc_queue, struct acc *acc)
> +static bool get_acc(struct acc_queue *acc_queue, struct acc *acc)
> {
> const struct xe_guc_acc_desc *desc;
> - int ret = 0;
> + bool ret = false;
>
> spin_lock(&acc_queue->lock);
> if (acc_queue->head != acc_queue->tail) {
> @@ -567,8 +575,7 @@ static int get_acc(struct acc_queue *acc_queue, struct acc *acc)
>
> acc_queue->head = (acc_queue->head + ACC_MSG_LEN_DW) %
> ACC_QUEUE_NUM_DW;
> - } else {
> - ret = -1;
> + ret = true;
> }
> spin_unlock(&acc_queue->lock);
>
> @@ -581,16 +588,23 @@ static void acc_queue_work_func(struct work_struct *w)
> struct xe_gt *gt = acc_queue->gt;
> struct xe_device *xe = gt_to_xe(gt);
> struct acc acc = {};
> + unsigned long threshold;
> int ret;
>
> - ret = get_acc(acc_queue, &acc);
> - if (ret)
> - return;
> + threshold = jiffies + msecs_to_jiffies(USM_QUEUE_MAX_RUNTIME_MS);
>
> - ret = handle_acc(gt, &acc);
> - if (unlikely(ret)) {
> - print_acc(xe, &acc);
> - drm_warn(&xe->drm, "ACC: Unsuccessful %d\n", ret);
> + while (get_acc(acc_queue, &acc)) {
> + ret = handle_acc(gt, &acc);
> + if (unlikely(ret)) {
> + print_acc(xe, &acc);
> + drm_warn(&xe->drm, "ACC: Unsuccessful %d\n", ret);
> + }
> +
> + if (time_after(jiffies, threshold) &&
> + acc_queue->head != acc_queue->tail) {
> + queue_work(gt->usm.acc_wq, w);
> + break;
> + }
> }
> }
>
> --
> 2.38.0
>
More information about the Intel-xe
mailing list