[Intel-gfx] [PATCH v8 3/7] drm/i915: Define and use GuC and CTB TLB invalidation routines
Nirmoy Das
nirmoy.das at intel.com
Mon Oct 9 12:12:02 UTC 2023
On 10/6/2023 8:20 PM, Jonathan Cavitt wrote:
> From: Prathap Kumar Valsan<prathap.kumar.valsan at intel.com>
>
> The GuC firmware had defined the interface for Translation Look-Aside
> Buffer (TLB) invalidation. We should use this interface when
> invalidating the engine and GuC TLBs.
> Add additional functionality to intel_gt_invalidate_tlb, invalidating
> the GuC TLBs and falling back to GT invalidation when the GuC is
> disabled.
> The invalidation is done by sending a request directly to the GuC
> tlb_lookup that invalidates the table. The invalidation is submitted as
> a wait request and is performed in the CT event handler. This means we
> cannot perform this TLB invalidation path if the CT is not enabled.
> If the request isn't fulfilled in two seconds, this would constitute
> an error in the invalidation as that would constitute either a lost
> request or a severe GuC overload.
>
> With this new invalidation routine, we can perform GuC-based GGTT
> invalidations. GuC-based GGTT invalidation is incompatible with
> MMIO invalidation so we should not perform MMIO invalidation when
> GuC-based GGTT invalidation is expected.
>
> Purpose of xarray:
> The tlb_lookup table is allocated as an xarray because the set of
> pending TLB invalidations may have no upper bound. The consequence of
> this is that all actions interfacing with this table need to use the
> xarray functions, such as xa_alloc_cyclic_irq for array insertion.
>
> Purpose of must_wait_woken:
> Our wait for the G2H ack for the completion of a TLB invalidation is
> mandatory; we must wait for the HW to confirm that the physical
> addresses are no longer accessible before we return those to the system.
>
> On switching to using the wait_woken() convenience routine, we
> introduced ourselves to an issue where wait_woken() may complete early
> under a kthread that is stopped. Since we send a TLB invalidation when
> we try to release pages from the shrinker, we can be called from any
> process; including kthreads.
>
> Using wait_woken() from any process context causes another issue. The
> use of is_kthread_should_stop() assumes that any task with PF_KTHREAD
> set was made by kthread_create() and has called set_kthread_struct().
> This is not true for the raw kernel_thread():
>
> BUG: kernel NULL pointer dereference, address: 0000000000000000
> [ 3089.759660] Call Trace:
> [ 3089.762110] wait_woken+0x4f/0x80
> [ 3089.765496] guc_send_invalidate_tlb+0x1fe/0x310 [i915]
> [ 3089.770725] ? syscall_return_via_sysret+0xf/0x7f
> [ 3089.775426] ? do_wait_intr_irq+0xb0/0xb0
> [ 3089.779430] ? __switch_to_asm+0x40/0x70
> [ 3089.783349] ? __switch_to_asm+0x34/0x70
> [ 3089.787273] ? __switch_to+0x7a/0x3e0
> [ 3089.790930] ? __switch_to_asm+0x34/0x70
> [ 3089.794883] intel_guc_invalidate_tlb_full+0x92/0xa0 [i915]
> [ 3089.800487] intel_invalidate_tlb_full+0x94/0x190 [i915]
> [ 3089.805824] intel_invalidate_tlb_full_sync+0x1b/0x30 [i915]
> [ 3089.811508] __i915_gem_object_unset_pages+0x138/0x150 [i915]
> [ 3089.817279] __i915_gem_object_put_pages+0x25/0x90 [i915]
> [ 3089.822706] i915_gem_shrink+0x532/0x7e0 [i915]
> [ 3089.827264] i915_gem_shrinker_scan+0x3d/0xd0 [i915]
> [ 3089.832230] do_shrink_slab+0x12c/0x2a0
> [ 3089.836065] shrink_slab+0xad/0x2b0
> [ 3089.839550] shrink_node+0xcc/0x410
> [ 3089.843035] do_try_to_free_pages+0xc6/0x380
> [ 3089.847306] try_to_free_pages+0xec/0x1c0
> [ 3089.851312] __alloc_pages_slowpath+0x3ad/0xd10
> [ 3089.855845] ? update_sd_lb_stats+0x636/0x710
> [ 3089.860204] __alloc_pages_nodemask+0x2d5/0x310
> [ 3089.864737] new_slab+0x265/0xa80
> [ 3089.868053] ___slab_alloc+0y_to_free_pages+0xec/0x1c0
> [ 3089.871798] ? copy_process+0x1e5/0x1a00
> [ 3089.875717] ? load_balance+0x165/0xb20
> [ 3089.879555] __slab_alloc+0x1c/0x30
> [ 3089.883047] kmem_cache_alloc_node+0x9f/0x240
> [ 3089.887397] ? copy_process+0x1e5/0x1a00
> [ 3089.891314] copy_process+0x1e5/0x1a00
> [ 3089.895058] ? __switch_to_asm+0x40/0x70
> [ 3089.879555] __slab_alloc+0x1c/0x30
> [ 3089.883047] kmem_cache_alloc_node+0x9f/0x240
> [ 3089.887397] ? copy_process+0x1e5/0x1a00
> [ 3089.891314] copy_process+0x1e5/0x1a00
> [ 3089.895058] ? __switch_to_asm+0x40/0x70
> [ 3089.898977] ? __switch_to_asm+0x34/0x70
> [ 3089.902903] ? __switch_to_asm+0x40/0x70
> [ 3089.906828] ? __switch_to_asm+0x34/0x70
> [ 3089.910745] _do_fork+0x83/0x350
> [ 3089.913969] ? __switch_to+0x7a/0x3e0
> [ 3089.917626] ? __switch_to_asm+0x34/0x70
> [ 3089.921545] kernel_thread+0x58/0x80
> [ 3089.925124] ? kthread_park+0x80/0x80
> [ 3089.928788] kthreadd+0x162/0x1b0
> [ 3089.932098] ? kthread_create_on_cpu+0xa0/0xa0
> [ 3089.936538] ret_from_fork+0x1f/0x40
>
> Signed-off-by: Prathap Kumar Valsan<prathap.kumar.valsan at intel.com>
> Signed-off-by: Bruce Chang<yu.bruce.chang at intel.com>
> Signed-off-by: Chris Wilson<chris.p.wilson at intel.com>
> Signed-off-by: Umesh Nerlige Ramappa<umesh.nerlige.ramappa at intel.com>
> Signed-off-by: Jonathan Cavitt<jonathan.cavitt at intel.com>
> Signed-off-by: Aravind Iddamsetty<aravind.iddamsetty at intel.com>
> Signed-off-by: Fei Yang<fei.yang at intel.com>
> CC: Andi Shyti<andi.shyti at linux.intel.com>
> ---
> drivers/gpu/drm/i915/gt/intel_ggtt.c | 34 ++-
> drivers/gpu/drm/i915/gt/intel_tlb.c | 15 +-
> .../gpu/drm/i915/gt/uc/abi/guc_actions_abi.h | 33 +++
> drivers/gpu/drm/i915/gt/uc/intel_guc.h | 22 ++
> drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c | 4 +
> drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h | 1 +
> .../gpu/drm/i915/gt/uc/intel_guc_submission.c | 200 +++++++++++++++++-
> 7 files changed, 297 insertions(+), 12 deletions(-)
>
> diff --git a/drivers/gpu/drm/i915/gt/intel_ggtt.c b/drivers/gpu/drm/i915/gt/intel_ggtt.c
> index 4d7d88b92632b..a1f7bdc602996 100644
> --- a/drivers/gpu/drm/i915/gt/intel_ggtt.c
> +++ b/drivers/gpu/drm/i915/gt/intel_ggtt.c
> @@ -206,22 +206,38 @@ static void gen8_ggtt_invalidate(struct i915_ggtt *ggtt)
> intel_uncore_write_fw(uncore, GFX_FLSH_CNTL_GEN6, GFX_FLSH_CNTL_EN);
> }
>
> +static void guc_ggtt_ct_invalidate(struct intel_gt *gt)
> +{
> + struct intel_uncore *uncore = gt->uncore;
> + intel_wakeref_t wakeref;
> +
> + with_intel_runtime_pm_if_active(uncore->rpm, wakeref) {
> + struct intel_guc *guc = >->uc.guc;
> +
> + intel_guc_invalidate_tlb_guc(guc);
> + }
> +}
> +
> static void guc_ggtt_invalidate(struct i915_ggtt *ggtt)
> {
> struct drm_i915_private *i915 = ggtt->vm.i915;
> + struct intel_gt *gt;
>
> - gen8_ggtt_invalidate(ggtt);
> -
> - if (GRAPHICS_VER(i915) >= 12) {
> - struct intel_gt *gt;
> + if (!HAS_GUC_TLB_INVALIDATION(i915))
> + gen8_ggtt_invalidate(ggtt);
>
> - list_for_each_entry(gt, &ggtt->gt_list, ggtt_link)
> + list_for_each_entry(gt, &ggtt->gt_list, ggtt_link) {
> + if (HAS_GUC_TLB_INVALIDATION(i915) &&
> + intel_guc_is_ready(>->uc.guc)) {
> + guc_ggtt_ct_invalidate(gt);
> + } else if (GRAPHICS_VER(i915) >= 12) {
> intel_uncore_write_fw(gt->uncore,
> GEN12_GUC_TLB_INV_CR,
> GEN12_GUC_TLB_INV_CR_INVALIDATE);
> - } else {
> - intel_uncore_write_fw(ggtt->vm.gt->uncore,
> - GEN8_GTCR, GEN8_GTCR_INVALIDATE);
> + } else {
> + intel_uncore_write_fw(gt->uncore,
> + GEN8_GTCR, GEN8_GTCR_INVALIDATE);
> + }
> }
> }
>
> @@ -1243,7 +1259,7 @@ static int gen8_gmch_probe(struct i915_ggtt *ggtt)
> ggtt->vm.raw_insert_page = gen8_ggtt_insert_page;
> }
>
> - if (intel_uc_wants_guc(&ggtt->vm.gt->uc))
> + if (intel_uc_wants_guc_submission(&ggtt->vm.gt->uc))
> ggtt->invalidate = guc_ggtt_invalidate;
> else
> ggtt->invalidate = gen8_ggtt_invalidate;
> diff --git a/drivers/gpu/drm/i915/gt/intel_tlb.c b/drivers/gpu/drm/i915/gt/intel_tlb.c
> index 139608c30d978..1278622b9554d 100644
> --- a/drivers/gpu/drm/i915/gt/intel_tlb.c
> +++ b/drivers/gpu/drm/i915/gt/intel_tlb.c
> @@ -12,6 +12,7 @@
> #include "intel_gt_print.h"
> #include "intel_gt_regs.h"
> #include "intel_tlb.h"
> +#include "uc/intel_guc.h"
>
> /*
> * HW architecture suggest typical invalidation time at 40us,
> @@ -131,11 +132,23 @@ void intel_gt_invalidate_tlb_full(struct intel_gt *gt, u32 seqno)
> return;
>
> with_intel_gt_pm_if_awake(gt, wakeref) {
> + struct intel_guc *guc = >->uc.guc;
> +
> mutex_lock(>->tlb.invalidate_lock);
> if (tlb_seqno_passed(gt, seqno))
> goto unlock;
>
> - mmio_invalidate_full(gt);
> + if (HAS_GUC_TLB_INVALIDATION(gt->i915)) {
> + /*
> + * Only perform GuC TLB invalidation if GuC is ready.
> + * If GuC is not ready, then there are no TLBs to
> + * invalidate. Ergo, skip invalidation.
> + */
> + if (intel_guc_is_ready(guc))
> + intel_guc_invalidate_tlb_engines(guc);
> + } else {
> + mmio_invalidate_full(gt);
> + }
>
> write_seqcount_invalidate(>->tlb.seqno);
> unlock:
> diff --git a/drivers/gpu/drm/i915/gt/uc/abi/guc_actions_abi.h b/drivers/gpu/drm/i915/gt/uc/abi/guc_actions_abi.h
> index f359bef046e0b..33f253410d0c8 100644
> --- a/drivers/gpu/drm/i915/gt/uc/abi/guc_actions_abi.h
> +++ b/drivers/gpu/drm/i915/gt/uc/abi/guc_actions_abi.h
> @@ -138,6 +138,8 @@ enum intel_guc_action {
> INTEL_GUC_ACTION_REGISTER_CONTEXT_MULTI_LRC = 0x4601,
> INTEL_GUC_ACTION_CLIENT_SOFT_RESET = 0x5507,
> INTEL_GUC_ACTION_SET_ENG_UTIL_BUFF = 0x550A,
> + INTEL_GUC_ACTION_TLB_INVALIDATION = 0x7000,
> + INTEL_GUC_ACTION_TLB_INVALIDATION_DONE = 0x7001,
> INTEL_GUC_ACTION_STATE_CAPTURE_NOTIFICATION = 0x8002,
> INTEL_GUC_ACTION_NOTIFY_FLUSH_LOG_BUFFER_TO_FILE = 0x8003,
> INTEL_GUC_ACTION_NOTIFY_CRASH_DUMP_POSTED = 0x8004,
> @@ -181,4 +183,35 @@ enum intel_guc_state_capture_event_status {
>
> #define INTEL_GUC_STATE_CAPTURE_EVENT_STATUS_MASK 0x000000FF
>
> +#define INTEL_GUC_TLB_INVAL_TYPE_MASK REG_GENMASK(7, 0)
> +#define INTEL_GUC_TLB_INVAL_MODE_MASK REG_GENMASK(11, 8)
> +#define INTEL_GUC_TLB_INVAL_FLUSH_CACHE REG_BIT(31)
> +
> +enum intel_guc_tlb_invalidation_type {
> + INTEL_GUC_TLB_INVAL_ENGINES = 0x0,
> + INTEL_GUC_TLB_INVAL_GUC = 0x3,
> +};
> +
> +/*
> + * 0: Heavy mode of Invalidation:
> + * The pipeline of the engine(s) for which the invalidation is targeted to is
> + * blocked, and all the in-flight transactions are guaranteed to be Globally
> + * Observed before completing the TLB invalidation
> + * 1: Lite mode of Invalidation:
> + * TLBs of the targeted engine(s) are immediately invalidated.
> + * In-flight transactions are NOT guaranteed to be Globally Observed before
> + * completing TLB invalidation.
> + * Light Invalidation Mode is to be used only when
> + * it can be guaranteed (by SW) that the address translations remain invariant
> + * for the in-flight transactions across the TLB invalidation. In other words,
> + * this mode can be used when the TLB invalidation is intended to clear out the
> + * stale cached translations that are no longer in use. Light Invalidation Mode
> + * is much faster than the Heavy Invalidation Mode, as it does not wait for the
> + * in-flight transactions to be GOd.
> + */
> +enum intel_guc_tlb_inval_mode {
> + INTEL_GUC_TLB_INVAL_MODE_HEAVY = 0x0,
> + INTEL_GUC_TLB_INVAL_MODE_LITE = 0x1,
> +};
> +
> #endif /* _ABI_GUC_ACTIONS_ABI_H */
> diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc.h b/drivers/gpu/drm/i915/gt/uc/intel_guc.h
> index 6c392bad29c19..88deb43bbdc48 100644
> --- a/drivers/gpu/drm/i915/gt/uc/intel_guc.h
> +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc.h
> @@ -79,6 +79,18 @@ struct intel_guc {
> */
> atomic_t outstanding_submission_g2h;
>
> + /** @tlb_lookup: xarray to store all pending TLB invalidation requests */
> + struct xarray tlb_lookup;
> +
> + /**
> + * @serial_slot: id to the initial waiter created in tlb_lookup,
> + * which is used only when failed to allocate new waiter.
> + */
> + u32 serial_slot;
> +
> + /** @next_seqno: the next id (sequence no.) to allocate. */
> + u32 next_seqno;
> +
> /** @interrupts: pointers to GuC interrupt-managing functions. */
> struct {
> bool enabled;
> @@ -296,6 +308,11 @@ struct intel_guc {
> #define MAKE_GUC_VER_STRUCT(ver) MAKE_GUC_VER((ver).major, (ver).minor, (ver).patch)
> #define GUC_SUBMIT_VER(guc) MAKE_GUC_VER_STRUCT((guc)->submission_version)
>
> +struct intel_guc_tlb_wait {
> + struct wait_queue_head wq;
> + bool busy;
> +};
> +
> static inline struct intel_guc *log_to_guc(struct intel_guc_log *log)
> {
> return container_of(log, struct intel_guc, log);
> @@ -418,6 +435,11 @@ static inline bool intel_guc_is_supported(struct intel_guc *guc)
> return intel_uc_fw_is_supported(&guc->fw);
> }
>
> +int intel_guc_invalidate_tlb_engines(struct intel_guc *guc);
> +int intel_guc_invalidate_tlb_guc(struct intel_guc *guc);
> +int intel_guc_tlb_invalidation_done(struct intel_guc *guc, u32 size,
> + u32 len, u32 seqno);
> +
> static inline bool intel_guc_is_wanted(struct intel_guc *guc)
> {
> return intel_uc_fw_is_enabled(&guc->fw);
> diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c b/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
> index 6e22af31513a5..e7ac8402245b7 100644
> --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
> +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
> @@ -1115,6 +1115,9 @@ static int ct_process_request(struct intel_guc_ct *ct, struct ct_incoming_msg *r
> case INTEL_GUC_ACTION_NOTIFY_EXCEPTION:
> ret = intel_guc_crash_process_msg(guc, action);
> break;
> + case INTEL_GUC_ACTION_TLB_INVALIDATION_DONE:
> + ret = intel_guc_tlb_invalidation_done(guc, request->size, len, payload[0]);
> + break;
> default:
> ret = -EOPNOTSUPP;
> break;
> @@ -1186,6 +1189,7 @@ static int ct_handle_event(struct intel_guc_ct *ct, struct ct_incoming_msg *requ
> switch (action) {
> case INTEL_GUC_ACTION_SCHED_CONTEXT_MODE_DONE:
> case INTEL_GUC_ACTION_DEREGISTER_CONTEXT_DONE:
> + case INTEL_GUC_ACTION_TLB_INVALIDATION_DONE:
> g2h_release_space(ct, request->size);
> }
>
> diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h b/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h
> index b4d56eccfb1f0..a7c9874e122a3 100644
> --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h
> +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h
> @@ -22,6 +22,7 @@
> /* Payload length only i.e. don't include G2H header length */
> #define G2H_LEN_DW_SCHED_CONTEXT_MODE_SET 2
> #define G2H_LEN_DW_DEREGISTER_CONTEXT 1
> +#define G2H_LEN_DW_INVALIDATE_TLB 1
>
> #define GUC_CONTEXT_DISABLE 0
> #define GUC_CONTEXT_ENABLE 1
> diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
> index 2cce5ec1ff00d..0c5ccda1b3e87 100644
> --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
> +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
> @@ -1798,9 +1798,11 @@ static void __guc_reset_context(struct intel_context *ce, intel_engine_mask_t st
>
> void intel_guc_submission_reset(struct intel_guc *guc, intel_engine_mask_t stalled)
> {
> + struct intel_guc_tlb_wait *wait;
> struct intel_context *ce;
> unsigned long index;
> unsigned long flags;
> + unsigned long i;
>
> if (unlikely(!guc_submission_initialized(guc))) {
> /* Reset called during driver load? GuC not yet initialised! */
> @@ -1826,6 +1828,18 @@ void intel_guc_submission_reset(struct intel_guc *guc, intel_engine_mask_t stall
>
> /* GuC is blown away, drop all references to contexts */
> xa_destroy(&guc->context_lookup);
> +
> + /*
> + * The full GT reset will have cleared the TLB caches and flushed the
> + * G2H message queue; we can release all the blocked waiters.
> + *
> + * This is safe to do unlocked because the xarray is not dependent
> + * on the GT reset, and there's a separate execution path for TLB
> + * invalidations on GT reset, and there's a large window of time
> + * between the GT reset and GuC becoming available.
> + */
> + xa_for_each(&guc->tlb_lookup, i, wait)
> + wake_up(&wait->wq);
> }
>
> static void guc_cancel_context_requests(struct intel_context *ce)
> @@ -1948,6 +1962,50 @@ void intel_guc_submission_reset_finish(struct intel_guc *guc)
> static void destroyed_worker_func(struct work_struct *w);
> static void reset_fail_worker_func(struct work_struct *w);
>
> +static int init_tlb_lookup(struct intel_guc *guc)
> +{
> + struct intel_guc_tlb_wait *wait;
> + int err;
> +
> + if (!HAS_GUC_TLB_INVALIDATION(guc_to_gt(guc)->i915))
> + return 0;
> +
> + xa_init_flags(&guc->tlb_lookup, XA_FLAGS_ALLOC);
> +
> + wait = kzalloc(sizeof(*wait), GFP_KERNEL);
> + if (!wait)
> + return -ENOMEM;
> +
> + init_waitqueue_head(&wait->wq);
> +
> + /* Preallocate a shared id for use under memory pressure. */
> + err = xa_alloc_cyclic_irq(&guc->tlb_lookup, &guc->serial_slot, wait,
> + xa_limit_32b, &guc->next_seqno, GFP_KERNEL);
> + /* Only error if out of memory, not when busy (list full)*/
> + if (err == -ENOMEM) {
> + kfree(wait);
> + return err;
> + }
> +
> + return 0;
> +}
> +
> +static void fini_tlb_lookup(struct intel_guc *guc)
> +{
> + struct intel_guc_tlb_wait *wait;
> +
> + if (!HAS_GUC_TLB_INVALIDATION(guc_to_gt(guc)->i915))
> + return;
> +
> + wait = xa_load(&guc->tlb_lookup, guc->serial_slot);
> + if (wait) {
> + guc_dbg(guc, "fini_tlb_lookup: Unexpected item in tlb_lookup\n");
> + kfree(wait);
> + }
> +
> + xa_destroy(&guc->tlb_lookup);
> +}
> +
> /*
> * Set up the memory resources to be shared with the GuC (via the GGTT)
> * at firmware loading time.
> @@ -1966,11 +2024,15 @@ int intel_guc_submission_init(struct intel_guc *guc)
> return ret;
> }
>
> + ret = init_tlb_lookup(guc);
> + if (ret)
> + goto destroy_pool;
> +
> guc->submission_state.guc_ids_bitmap =
> bitmap_zalloc(NUMBER_MULTI_LRC_GUC_ID(guc), GFP_KERNEL);
> if (!guc->submission_state.guc_ids_bitmap) {
> ret = -ENOMEM;
> - goto destroy_pool;
> + goto destroy_tlb;
> }
>
> guc->timestamp.ping_delay = (POLL_TIME_CLKS / gt->clock_frequency + 1) * HZ;
> @@ -1979,9 +2041,10 @@ int intel_guc_submission_init(struct intel_guc *guc)
>
> return 0;
>
> +destroy_tlb:
> + fini_tlb_lookup(guc);
> destroy_pool:
> guc_lrc_desc_pool_destroy_v69(guc);
> -
> return ret;
> }
>
> @@ -1994,6 +2057,7 @@ void intel_guc_submission_fini(struct intel_guc *guc)
> guc_lrc_desc_pool_destroy_v69(guc);
> i915_sched_engine_put(guc->sched_engine);
> bitmap_free(guc->submission_state.guc_ids_bitmap);
> + fini_tlb_lookup(guc);
> guc->submission_initialized = false;
> }
>
> @@ -4624,6 +4688,138 @@ g2h_context_lookup(struct intel_guc *guc, u32 ctx_id)
> return ce;
> }
>
> +static void wait_wake_outstanding_tlb_g2h(struct intel_guc *guc, u32 seqno)
> +{
> + struct intel_guc_tlb_wait *wait;
> + unsigned long flags;
> +
> + xa_lock_irqsave(&guc->tlb_lookup, flags);
> + wait = xa_load(&guc->tlb_lookup, seqno);
> +
> + if (wait)
> + wake_up(&wait->wq);
> + else
> + guc_dbg(guc,
> + "Stale TLB invalidation response with seqno %d\n", seqno);
> +
> + xa_unlock_irqrestore(&guc->tlb_lookup, flags);
> +}
> +
> +int intel_guc_tlb_invalidation_done(struct intel_guc *guc, u32 size, u32 len, u32 seqno)
> +{
> + /* Check for underflow */
> + if (unlikely(len < 1 || len > size))
> + return -EPROTO;
> +
> + wait_wake_outstanding_tlb_g2h(guc, seqno);
> + return 0;
> +}
> +
> +static long must_wait_woken(struct wait_queue_entry *wq_entry, long timeout)
> +{
> + /*
> + * This is equivalent to wait_woken() with the exception that
> + * we do not wake up early if the kthread task has been completed.
> + * As we are called from page reclaim in any task context,
> + * we may be invoked from stopped kthreads, but we *must*
> + * complete the wait from the HW .
> + *
> + * A second problem is that since we are called under reclaim
> + * and wait_woken() inspected the thread state, it makes an invalid
> + * assumption that all PF_KTHREAD tasks have set_kthread_struct()
> + * called upon them, and will trigger a GPF
As discussed internally, the GPF issue is resolved with
https://lore.kernel.org/all/20230602212350.535358-1-jstultz@google.com/
<https://lore.kernel.org/all/20230602212350.535358-1-jstultz@google.com/>
> in is_kthread_should_stop().
> + */
> + do {
> + set_current_state(TASK_UNINTERRUPTIBLE);
> + if (wq_entry->flags & WQ_FLAG_WOKEN)
> + break;
> +
> + timeout = schedule_timeout(timeout);
> + } while (timeout);
> + __set_current_state(TASK_RUNNING);
> +
> + /* See wait_woken() and woken_wake_function() */
> + smp_store_mb(wq_entry->flags, wq_entry->flags & ~WQ_FLAG_WOKEN);
> +
> + return timeout;
> +}
> +
> +static int guc_send_invalidate_tlb(struct intel_guc *guc, enum intel_guc_tlb_inval_mode type)
2nd param should be intel_guc_tlb_invalidation_type not
intel_guc_tlb_inval_mod.
Not sure why didn't CI complained.
Regards,
Nirmoy
> +{
> + struct intel_guc_tlb_wait _wq, *wq = &_wq;
> + DEFINE_WAIT_FUNC(wait, woken_wake_function);
> + int err;
> + u32 seqno;
> + u32 action[] = {
> + INTEL_GUC_ACTION_TLB_INVALIDATION,
> + 0,
> + REG_FIELD_PREP(INTEL_GUC_TLB_INVAL_TYPE_MASK, type) |
> + REG_FIELD_PREP(INTEL_GUC_TLB_INVAL_MODE_MASK,
> + INTEL_GUC_TLB_INVAL_MODE_HEAVY) |
> + INTEL_GUC_TLB_INVAL_FLUSH_CACHE,
> + };
> + u32 size = ARRAY_SIZE(action);
> +
> + init_waitqueue_head(&_wq.wq);
> +
> + if (xa_alloc_cyclic_irq(&guc->tlb_lookup, &seqno, wq,
> + xa_limit_32b, &guc->next_seqno,
> + GFP_ATOMIC | __GFP_NOWARN) < 0) {
> + /* Under severe memory pressure? Serialise TLB allocations */
> + xa_lock_irq(&guc->tlb_lookup);
> + wq = xa_load(&guc->tlb_lookup, guc->serial_slot);
> + wait_event_lock_irq(wq->wq,
> + !READ_ONCE(wq->busy),
> + guc->tlb_lookup.xa_lock);
> + /*
> + * Update wq->busy under lock to ensure only one waiter can
> + * issue the TLB invalidation command using the serial slot at a
> + * time. The condition is set to true before releasing the lock
> + * so that other caller continue to wait until woken up again.
> + */
> + wq->busy = true;
> + xa_unlock_irq(&guc->tlb_lookup);
> +
> + seqno = guc->serial_slot;
> + }
> +
> + action[1] = seqno;
> +
> + add_wait_queue(&wq->wq, &wait);
> +
> + /*
> + * This is a critical reclaim path and thus we must loop here:
> + * We cannot block for anything that is on the GPU.
> + */
> + err = intel_guc_send_busy_loop(guc, action, size, G2H_LEN_DW_INVALIDATE_TLB, true);
> + if (err)
> + goto out;
> +
> + if (!must_wait_woken(&wait, intel_guc_ct_expected_delay(&guc->ct))) {
> + guc_err(guc,
> + "TLB invalidation response timed out for seqno %u\n", seqno);
> + err = -ETIME;
> + }
> +out:
> + remove_wait_queue(&wq->wq, &wait);
> + if (seqno != guc->serial_slot)
> + xa_erase_irq(&guc->tlb_lookup, seqno);
> +
> + return err;
> +}
> +
> +/* Full TLB invalidation */
> +int intel_guc_invalidate_tlb_engines(struct intel_guc *guc)
> +{
> + return guc_send_invalidate_tlb(guc, INTEL_GUC_TLB_INVAL_ENGINES);
> +}
> +
> +/* GuC TLB Invalidation: Invalidate the TLB's of GuC itself. */
> +int intel_guc_invalidate_tlb_guc(struct intel_guc *guc)
> +{
> + return guc_send_invalidate_tlb(guc, INTEL_GUC_TLB_INVAL_GUC);
> +}
> +
> int intel_guc_deregister_done_process_msg(struct intel_guc *guc,
> const u32 *msg,
> u32 len)
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.freedesktop.org/archives/intel-gfx/attachments/20231009/732b5982/attachment-0001.htm>
More information about the Intel-gfx
mailing list