[PATCH] drm/xe/guc_pc: Retry and wait longer for GuC PC start
Rodrigo Vivi
rodrigo.vivi at intel.com
Fri Mar 7 16:00:41 UTC 2025
On Fri, Mar 07, 2025 at 03:35:39PM +0000, Cavitt, Jonathan wrote:
> -----Original Message-----
> From: Vivi, Rodrigo <rodrigo.vivi at intel.com>
> Sent: Thursday, March 6, 2025 3:39 PM
> To: intel-xe at lists.freedesktop.org
> Cc: Vivi, Rodrigo <rodrigo.vivi at intel.com>; Belgaumkar, Vinay <vinay.belgaumkar at intel.com>; Cavitt, Jonathan <jonathan.cavitt at intel.com>; Harrison, John C <john.c.harrison at intel.com>
> Subject: [PATCH] drm/xe/guc_pc: Retry and wait longer for GuC PC start
> >
> > In a rare situation of thermal limit during resume, GuC can
> > be slow and run into delays like this:
> >
> > xe 0000:00:02.0: [drm] GT1: excessive init time: 667ms! \
> > [status = 0x8002F034, timeouts = 0]
> > xe 0000:00:02.0: [drm] GT1: excessive init time: \
> > [freq = 100MHz (req = 800MHz), before = 100MHz, \
> > perf_limit_reasons = 0x1C001000]
> > xe 0000:00:02.0: [drm] *ERROR* GT1: GuC PC Start failed
> > ------------[ cut here ]------------
> > xe 0000:00:02.0: [drm] GT1: Failed to start GuC PC: -EIO
> >
> > When this happens, it will block entirely the GPU to be used.
> > So, let's try and with a huge timeout in the hope it comes back.
> >
> > Also, let's collect some information on how long it is usually
> > taking on situations like this, so perhaps the time can be tuned
> > later.
> >
> > Cc: Vinay Belgaumkar <vinay.belgaumkar at intel.com>
> > Cc: Jonathan Cavitt <jonathan.cavitt at intel.com>
> > Cc: John Harrison <John.C.Harrison at Intel.com>
> > Signed-off-by: Rodrigo Vivi <rodrigo.vivi at intel.com>
>
> s/rought/roughly for the comment next to SLPC_RESET_TIMEOUT_MS.
> There are also two nits below, but nothing blocking.
will change, thank you
> Otherwise:
> Reviewed-by: Jonathan Cavitt <jonathan.cavitt at intel.com>
>
> > ---
> > drivers/gpu/drm/xe/xe_guc_pc.c | 53 +++++++++++++++++++++++++---------
> > 1 file changed, 40 insertions(+), 13 deletions(-)
> >
> > diff --git a/drivers/gpu/drm/xe/xe_guc_pc.c b/drivers/gpu/drm/xe/xe_guc_pc.c
> > index 25040efa043f..ec39a8f2781f 100644
> > --- a/drivers/gpu/drm/xe/xe_guc_pc.c
> > +++ b/drivers/gpu/drm/xe/xe_guc_pc.c
> > @@ -6,6 +6,7 @@
> > #include "xe_guc_pc.h"
> >
> > #include <linux/delay.h>
> > +#include <linux/ktime.h>
> >
> > #include <drm/drm_managed.h>
> > #include <drm/drm_print.h>
> > @@ -20,6 +21,7 @@
> > #include "xe_gt.h"
> > #include "xe_gt_idle.h"
> > #include "xe_gt_printk.h"
> > +#include "xe_gt_throttle.h"
> > #include "xe_gt_types.h"
> > #include "xe_guc.h"
> > #include "xe_guc_ct.h"
> > @@ -50,6 +52,9 @@
> > #define LNL_MERT_FREQ_CAP 800
> > #define BMG_MERT_FREQ_CAP 2133
> >
> > +#define SLPC_RESET_TIMEOUT_MS 5 /* rought 5ms, but no need for precision */
> > +#define SLPC_RESET_EXTENDED_TIMEOUT_MS 1000 /* To be used only at pc_start */
> > +
> > /**
> > * DOC: GuC Power Conservation (PC)
> > *
> > @@ -114,9 +119,10 @@ static struct iosys_map *pc_to_maps(struct xe_guc_pc *pc)
> > FIELD_PREP(HOST2GUC_PC_SLPC_REQUEST_MSG_1_EVENT_ARGC, count))
> >
> > static int wait_for_pc_state(struct xe_guc_pc *pc,
> > - enum slpc_global_state state)
> > + enum slpc_global_state state,
> > + int timeout_ms)
> > {
>
> NIT:
> If we're only planning on having two timeout times, and one of them is only
> used once, maybe we should instead provide information about which case
> to use, rather than providing the full timeout to the wait_for_pc_state
> function? I suspect that there isn't enough information in the xe_guc_pc
> at the time to determine which case to use, so maybe we can just provide
> a flag here to select the timeout instead. Something like:
>
> """
> #define SLPC_EXTENDED_TIMEOUT BIT(1)
> ...
> static int wait_for_pc_state(struct xe_guc_pc *pc,
> enum slpc_global_state state,
> int flags)
> {
> /**
> * Wait 5 ms in the base case, or 1 second on extended timeout.
> * No need to be precise here, though.
> */
> int timeout_us = 1000;
> timeout_us *= flag & SLPC_EXTENDED_TIMEOUT ?
> 5 : 1000;
> ...
> if (wait_for_pc_state(pc, SLPC_GLOBAL_STATE_RUNNING, 0))
> ...
> if (wait_for_pc_state(pc, SLPC_GLOBAL_STATE_RUNNING,
> SLPC_EXTENDED_TIMEOUT)) {
> """
> Either way works, though using flags instead might be more extensible. On the
> other hand, it's rather unlikely that we'll need to pass more options to this
> function in the near to distant future, so keeping it as is would also be fine.
I thought about that, but the problem with that is that most of the callers
are calling with '0', then you need to read the function definition to
see the meaning, while in the current way it is obvious to any reader at
first sight. A similar problem with the bool paramenters.
>
> > - int timeout_us = 5000; /* rought 5ms, but no need for precision */
> > + int timeout_us = 1000 * timeout_ms;
> > int slept, wait = 10;
> >
> > xe_device_assert_mem_access(pc_to_xe(pc));
> > @@ -165,7 +171,8 @@ static int pc_action_query_task_state(struct xe_guc_pc *pc)
> > };
> > int ret;
> >
> > - if (wait_for_pc_state(pc, SLPC_GLOBAL_STATE_RUNNING))
> > + if (wait_for_pc_state(pc, SLPC_GLOBAL_STATE_RUNNING,
> > + SLPC_RESET_TIMEOUT_MS))
> > return -EAGAIN;
> >
> > /* Blocking here to ensure the results are ready before reading them */
> > @@ -188,7 +195,8 @@ static int pc_action_set_param(struct xe_guc_pc *pc, u8 id, u32 value)
> > };
> > int ret;
> >
> > - if (wait_for_pc_state(pc, SLPC_GLOBAL_STATE_RUNNING))
> > + if (wait_for_pc_state(pc, SLPC_GLOBAL_STATE_RUNNING,
> > + SLPC_RESET_TIMEOUT_MS))
> > return -EAGAIN;
> >
> > ret = xe_guc_ct_send(ct, action, ARRAY_SIZE(action), 0, 0);
> > @@ -209,7 +217,8 @@ static int pc_action_unset_param(struct xe_guc_pc *pc, u8 id)
> > struct xe_guc_ct *ct = &pc_to_guc(pc)->ct;
> > int ret;
> >
> > - if (wait_for_pc_state(pc, SLPC_GLOBAL_STATE_RUNNING))
> > + if (wait_for_pc_state(pc, SLPC_GLOBAL_STATE_RUNNING,
> > + SLPC_RESET_TIMEOUT_MS))
> > return -EAGAIN;
> >
> > ret = xe_guc_ct_send(ct, action, ARRAY_SIZE(action), 0, 0);
> > @@ -443,6 +452,15 @@ u32 xe_guc_pc_get_act_freq(struct xe_guc_pc *pc)
> > return freq;
> > }
> >
> > +static u32 get_cur_freq(struct xe_gt *gt)
> > +{
> > + u32 freq;
> > +
> > + freq = xe_mmio_read32(>->mmio, RPNSWREQ);
> > + freq = REG_FIELD_GET(REQ_RATIO_MASK, freq);
> > + return decode_freq(freq);
> > +}
>
> NIT:
> Creating this helper function seems unrelated to this patch
> and should probably be split into a separate one. If you
> decide to split this into a separate patch without any
> additional changes, you can preemptively add my
> Reviewed-by to that patch as well.
I also had the same feeling when I was reading this patch again
before sending it out. But then I realized again that it is like
this so the forcewake doesn't need to be duplicated. And the
only second user of this function comes in this patch below, so
if split in another patch it would be a silly patch moving just
the 3 lines out to another function that is used once, so I kept it.
Thank you so much,
Rodrigo.
> -Jonathan Cavitt
>
> > +
> > /**
> > * xe_guc_pc_get_cur_freq - Get Current requested frequency
> > * @pc: The GuC PC
> > @@ -466,10 +484,7 @@ int xe_guc_pc_get_cur_freq(struct xe_guc_pc *pc, u32 *freq)
> > return -ETIMEDOUT;
> > }
> >
> > - *freq = xe_mmio_read32(>->mmio, RPNSWREQ);
> > -
> > - *freq = REG_FIELD_GET(REQ_RATIO_MASK, *freq);
> > - *freq = decode_freq(*freq);
> > + *freq = get_cur_freq(gt);
> >
> > xe_force_wake_put(gt_to_fw(gt), fw_ref);
> > return 0;
> > @@ -1016,6 +1031,7 @@ int xe_guc_pc_start(struct xe_guc_pc *pc)
> > struct xe_gt *gt = pc_to_gt(pc);
> > u32 size = PAGE_ALIGN(sizeof(struct slpc_shared_data));
> > unsigned int fw_ref;
> > + ktime_t earlier;
> > int ret;
> >
> > xe_gt_assert(gt, xe_device_uc_enabled(xe));
> > @@ -1040,14 +1056,25 @@ int xe_guc_pc_start(struct xe_guc_pc *pc)
> > memset(pc->bo->vmap.vaddr, 0, size);
> > slpc_shared_data_write(pc, header.size, size);
> >
> > + earlier = ktime_get();
> > ret = pc_action_reset(pc);
> > if (ret)
> > goto out;
> >
> > - if (wait_for_pc_state(pc, SLPC_GLOBAL_STATE_RUNNING)) {
> > - xe_gt_err(gt, "GuC PC Start failed\n");
> > - ret = -EIO;
> > - goto out;
> > + if (wait_for_pc_state(pc, SLPC_GLOBAL_STATE_RUNNING,
> > + SLPC_RESET_TIMEOUT_MS)) {
> > + xe_gt_warn(gt, "GuC PC start taking longer than normal [freq = %dMHz (req = %dMHz), perf_limit_reasons = 0x%08X]\n",
> > + xe_guc_pc_get_act_freq(pc), get_cur_freq(gt),
> > + xe_gt_throttle_get_limit_reasons(gt));
> > +
> > + if (wait_for_pc_state(pc, SLPC_GLOBAL_STATE_RUNNING,
> > + SLPC_RESET_EXTENDED_TIMEOUT_MS)) {
> > + xe_gt_err(gt, "GuC PC Start failed: Dynamic GT frequency control and GT sleep states are now disabled.\n");
> > + goto out;
> > + }
> > +
> > + xe_gt_warn(gt, "GuC PC excessive start time: %lldms",
> > + ktime_ms_delta(ktime_get(), earlier));
> > }
> >
> > ret = pc_init_freqs(pc);
> > --
> > 2.48.1
> >
> >
More information about the Intel-xe
mailing list