[PATCH v4 5/5] drm/msm: Temporarily disable stall-on-fault after a page fault
Rob Clark
robdclark at gmail.com
Wed Mar 5 19:07:00 UTC 2025
On Tue, Mar 4, 2025 at 8:57 AM Connor Abbott <cwabbott0 at gmail.com> wrote:
>
> When things go wrong, the GPU is capable of quickly generating millions
> of faulting translation requests per second. When that happens, in the
> stall-on-fault model each access will stall until it wins the race to
> signal the fault and then the RESUME register is written. This slows
> processing page faults to a crawl as the GPU can generate faults much
> faster than the CPU can acknowledge them. It also means that all
> available resources in the SMMU are saturated waiting for the stalled
> transactions, so that other transactions such as transactions generated
> by the GMU, which shares a context bank with the GPU, cannot proceed.
Nit, the GMU does not actually share a cb.. looking on x1e80100.dtsi,
the GMU has cb 5 and gpu has 0 and 1. (Currently we just use the
first, but I guess the 2nd would be used if we supported protected
content?)
fwiw, you can read this from dtsi, ie. in the GMU node, "iommus =
<&adreno_smmu 5 0x0>;"
> This causes a GMU watchdog timeout, which leads to a failed reset
> because GX cannot collapse when there is a transaction pending and a
> permanently hung GPU.
>
> On older platforms with qcom,smmu-v2, it seems that when one transaction
> is stalled subsequent faulting transactions are terminated, which avoids
> this problem, but the MMU-500 follows the spec here.
>
> To work around these problem, disable stall-on-fault as soon as we get a
> page fault until a cooldown period after pagefaults stop. This allows
> the GMU some guaranteed time to continue working. We only use
> stall-on-fault to halt the GPU while we collect a devcoredump and we
> always terminate the transaction afterward, so it's fine to miss some
> subsequent page faults. We also keep it disabled so long as the current
> devcoredump hasn't been deleted, because in that case we likely won't
> capture another one if there's a fault.
>
> After this commit HFI messages still occasionally time out, because the
> crashdump handler doesn't run fast enough to let the GMU resume, but the
> driver seems to recover from it. This will probably go away after the
> HFI timeout is increased.
>
> Signed-off-by: Connor Abbott <cwabbott0 at gmail.com>
> ---
> drivers/gpu/drm/msm/adreno/a5xx_gpu.c | 2 ++
> drivers/gpu/drm/msm/adreno/a6xx_gpu.c | 4 ++++
> drivers/gpu/drm/msm/adreno/adreno_gpu.c | 42 ++++++++++++++++++++++++++++++++-
> drivers/gpu/drm/msm/adreno/adreno_gpu.h | 24 +++++++++++++++++++
> drivers/gpu/drm/msm/msm_iommu.c | 9 +++++++
> drivers/gpu/drm/msm/msm_mmu.h | 1 +
> 6 files changed, 81 insertions(+), 1 deletion(-)
>
> diff --git a/drivers/gpu/drm/msm/adreno/a5xx_gpu.c b/drivers/gpu/drm/msm/adreno/a5xx_gpu.c
> index 71dca78cd7a5324e9ff5b14f173e2209fa42e196..670141531112c9d29cef8ef1fd51b74759fdd6d2 100644
> --- a/drivers/gpu/drm/msm/adreno/a5xx_gpu.c
> +++ b/drivers/gpu/drm/msm/adreno/a5xx_gpu.c
> @@ -131,6 +131,8 @@ static void a5xx_submit(struct msm_gpu *gpu, struct msm_gem_submit *submit)
> struct msm_ringbuffer *ring = submit->ring;
> unsigned int i, ibs = 0;
>
> + adreno_check_and_reenable_stall(adreno_gpu);
> +
> if (IS_ENABLED(CONFIG_DRM_MSM_GPU_SUDO) && submit->in_rb) {
> ring->cur_ctx_seqno = 0;
> a5xx_submit_in_rb(gpu, submit);
> diff --git a/drivers/gpu/drm/msm/adreno/a6xx_gpu.c b/drivers/gpu/drm/msm/adreno/a6xx_gpu.c
> index 0ae29a7c8a4d3f74236a35cc919f69d5c0a384a0..5a34cd2109a2d74c92841448a61ccb0d4f34e264 100644
> --- a/drivers/gpu/drm/msm/adreno/a6xx_gpu.c
> +++ b/drivers/gpu/drm/msm/adreno/a6xx_gpu.c
> @@ -212,6 +212,8 @@ static void a6xx_submit(struct msm_gpu *gpu, struct msm_gem_submit *submit)
> struct msm_ringbuffer *ring = submit->ring;
> unsigned int i, ibs = 0;
>
> + adreno_check_and_reenable_stall(adreno_gpu);
> +
> a6xx_set_pagetable(a6xx_gpu, ring, submit);
>
> get_stats_counter(ring, REG_A6XX_RBBM_PERFCTR_CP(0),
> @@ -335,6 +337,8 @@ static void a7xx_submit(struct msm_gpu *gpu, struct msm_gem_submit *submit)
> struct msm_ringbuffer *ring = submit->ring;
> unsigned int i, ibs = 0;
>
> + adreno_check_and_reenable_stall(adreno_gpu);
> +
> /*
> * Toggle concurrent binning for pagetable switch and set the thread to
> * BR since only it can execute the pagetable switch packets.
> diff --git a/drivers/gpu/drm/msm/adreno/adreno_gpu.c b/drivers/gpu/drm/msm/adreno/adreno_gpu.c
> index 1238f326597808eb28b4c6822cbd41a26e555eb9..bac586101dc0494f46b069a8440a45825dfe9b5e 100644
> --- a/drivers/gpu/drm/msm/adreno/adreno_gpu.c
> +++ b/drivers/gpu/drm/msm/adreno/adreno_gpu.c
> @@ -246,16 +246,53 @@ u64 adreno_private_address_space_size(struct msm_gpu *gpu)
> return SZ_4G;
> }
>
> +void adreno_check_and_reenable_stall(struct adreno_gpu *adreno_gpu)
> +{
> + struct msm_gpu *gpu = &adreno_gpu->base;
> + unsigned long flags;
> +
> + /*
> + * Wait until the cooldown period has passed and we would actually
> + * collect a crashdump to re-enable stall-on-fault.
> + */
> + spin_lock_irqsave(&adreno_gpu->fault_stall_lock, flags);
> + if (!adreno_gpu->stall_enabled &&
> + ktime_after(ktime_get(), adreno_gpu->stall_reenable_time) &&
> + !READ_ONCE(gpu->crashstate)) {
> + adreno_gpu->stall_enabled = true;
> +
> + gpu->aspace->mmu->funcs->set_stall(gpu->aspace->mmu, true);
> + }
> + spin_unlock_irqrestore(&adreno_gpu->fault_stall_lock, flags);
> +}
> +
> #define ARM_SMMU_FSR_TF BIT(1)
> #define ARM_SMMU_FSR_PF BIT(3)
> #define ARM_SMMU_FSR_EF BIT(4)
> +#define ARM_SMMU_FSR_SS BIT(30)
>
> int adreno_fault_handler(struct msm_gpu *gpu, unsigned long iova, int flags,
> struct adreno_smmu_fault_info *info, const char *block,
> u32 scratch[4])
> {
> + struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu);
> const char *type = "UNKNOWN";
> - bool do_devcoredump = info && !READ_ONCE(gpu->crashstate);
> + bool do_devcoredump = info && (info->fsr & ARM_SMMU_FSR_SS) &&
> + !READ_ONCE(gpu->crashstate);
> + unsigned long irq_flags;
> +
> + /*
> + * In case there is a subsequent storm of pagefaults, disable
> + * stall-on-fault for at least half a second.
> + */
> + spin_lock_irqsave(&adreno_gpu->fault_stall_lock, irq_flags);
> + if (adreno_gpu->stall_enabled) {
> + adreno_gpu->stall_enabled = false;
> +
> + gpu->aspace->mmu->funcs->set_stall(gpu->aspace->mmu, false);
> + }
> + adreno_gpu->stall_reenable_time = ktime_add_ms(ktime_get(), 500);
> + spin_unlock_irqrestore(&adreno_gpu->fault_stall_lock, irq_flags);
>
> /*
> * If we aren't going to be resuming later from fault_worker, then do
> @@ -1143,6 +1180,9 @@ int adreno_gpu_init(struct drm_device *drm, struct platform_device *pdev,
> adreno_gpu->info->inactive_period);
> pm_runtime_use_autosuspend(dev);
>
> + spin_lock_init(&adreno_gpu->fault_stall_lock);
> + adreno_gpu->stall_enabled = true;
> +
> return msm_gpu_init(drm, pdev, &adreno_gpu->base, &funcs->base,
> gpu_name, &adreno_gpu_config);
> }
> diff --git a/drivers/gpu/drm/msm/adreno/adreno_gpu.h b/drivers/gpu/drm/msm/adreno/adreno_gpu.h
> index dcf454629ce037b2a8274a6699674ad754ce1f07..a528036b46216bd898f6d48c5fb0555c4c4b053b 100644
> --- a/drivers/gpu/drm/msm/adreno/adreno_gpu.h
> +++ b/drivers/gpu/drm/msm/adreno/adreno_gpu.h
> @@ -205,6 +205,28 @@ struct adreno_gpu {
> /* firmware: */
> const struct firmware *fw[ADRENO_FW_MAX];
>
> + /**
> + * fault_stall_lock:
nit, @fault_stall_lock: And for
fault_stall_reenable_time/stall_enabled, it wouldn't hurt to add
something along the lines of "Protected by @fault_stall_lock". I've
been slowly trying to improve the comment docs over time, I have some
of that in my vmbind patchset.
Anyways, with those nits addressed, r-b
BR,
-R
> + *
> + * Serialize changes to stall-on-fault state.
> + */
> + spinlock_t fault_stall_lock;
> +
> + /**
> + * fault_stall_reenable_time:
> + *
> + * if stall_enabled is false, when to reenable stall-on-fault.
> + */
> + ktime_t stall_reenable_time;
> +
> + /**
> + * stall_enabled:
> + *
> + * Whether stall-on-fault is currently enabled.
> + */
> + bool stall_enabled;
> +
> +
> struct {
> /**
> * @rgb565_predicator: Unknown, introduced with A650 family,
> @@ -629,6 +651,8 @@ int adreno_fault_handler(struct msm_gpu *gpu, unsigned long iova, int flags,
> struct adreno_smmu_fault_info *info, const char *block,
> u32 scratch[4]);
>
> +void adreno_check_and_reenable_stall(struct adreno_gpu *gpu);
> +
> int adreno_read_speedbin(struct device *dev, u32 *speedbin);
>
> /*
> diff --git a/drivers/gpu/drm/msm/msm_iommu.c b/drivers/gpu/drm/msm/msm_iommu.c
> index 2a94e82316f95c5f9dcc37ef0a4664a29e3492b2..8d5380e6dcc217c7c209b51527bf15748b3ada71 100644
> --- a/drivers/gpu/drm/msm/msm_iommu.c
> +++ b/drivers/gpu/drm/msm/msm_iommu.c
> @@ -351,6 +351,14 @@ static void msm_iommu_resume_translation(struct msm_mmu *mmu)
> adreno_smmu->resume_translation(adreno_smmu->cookie, true);
> }
>
> +static void msm_iommu_set_stall(struct msm_mmu *mmu, bool enable)
> +{
> + struct adreno_smmu_priv *adreno_smmu = dev_get_drvdata(mmu->dev);
> +
> + if (adreno_smmu->set_stall)
> + adreno_smmu->set_stall(adreno_smmu->cookie, enable);
> +}
> +
> static void msm_iommu_detach(struct msm_mmu *mmu)
> {
> struct msm_iommu *iommu = to_msm_iommu(mmu);
> @@ -399,6 +407,7 @@ static const struct msm_mmu_funcs funcs = {
> .unmap = msm_iommu_unmap,
> .destroy = msm_iommu_destroy,
> .resume_translation = msm_iommu_resume_translation,
> + .set_stall = msm_iommu_set_stall,
> };
>
> struct msm_mmu *msm_iommu_new(struct device *dev, unsigned long quirks)
> diff --git a/drivers/gpu/drm/msm/msm_mmu.h b/drivers/gpu/drm/msm/msm_mmu.h
> index 88af4f490881f2a6789ae2d03e1c02d10046331a..2694a356a17904e7572b767b16ed0cee806406cf 100644
> --- a/drivers/gpu/drm/msm/msm_mmu.h
> +++ b/drivers/gpu/drm/msm/msm_mmu.h
> @@ -16,6 +16,7 @@ struct msm_mmu_funcs {
> int (*unmap)(struct msm_mmu *mmu, uint64_t iova, size_t len);
> void (*destroy)(struct msm_mmu *mmu);
> void (*resume_translation)(struct msm_mmu *mmu);
> + void (*set_stall)(struct msm_mmu *mmu, bool enable);
> };
>
> enum msm_mmu_type {
>
> --
> 2.47.1
>
More information about the Freedreno
mailing list