[PATCH 2/4] drm/msm/adreno: use a single register offset for gpu_read64/gpu_write64

Rob Clark robdclark at gmail.com
Sat Apr 2 01:39:16 UTC 2022


On Sun, Mar 27, 2022 at 1:27 PM Jonathan Marek <jonathan at marek.ca> wrote:
>
> The high half of 64-bit registers is always at +1 offset, so change these
> helpers to be more convenient by removing the unnecessary argument.
>
> Signed-off-by: Jonathan Marek <jonathan at marek.ca>

I'd been meaning to do this for a while.. so I think I'll cherry-pick
this ahead of the rest of the series

Reviewed-by: Rob Clark <robdclark at gmail.com>

> ---
>  drivers/gpu/drm/msm/adreno/a4xx_gpu.c       |  3 +--
>  drivers/gpu/drm/msm/adreno/a5xx_gpu.c       | 27 ++++++++-------------
>  drivers/gpu/drm/msm/adreno/a5xx_preempt.c   |  4 +--
>  drivers/gpu/drm/msm/adreno/a6xx_gpu.c       | 25 ++++++-------------
>  drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c |  3 +--
>  drivers/gpu/drm/msm/msm_gpu.h               | 12 ++++-----
>  6 files changed, 27 insertions(+), 47 deletions(-)
>
> diff --git a/drivers/gpu/drm/msm/adreno/a4xx_gpu.c b/drivers/gpu/drm/msm/adreno/a4xx_gpu.c
> index 0c6b2a6d0b4c9..da5e18bd74a45 100644
> --- a/drivers/gpu/drm/msm/adreno/a4xx_gpu.c
> +++ b/drivers/gpu/drm/msm/adreno/a4xx_gpu.c
> @@ -606,8 +606,7 @@ static int a4xx_pm_suspend(struct msm_gpu *gpu) {
>
>  static int a4xx_get_timestamp(struct msm_gpu *gpu, uint64_t *value)
>  {
> -       *value = gpu_read64(gpu, REG_A4XX_RBBM_PERFCTR_CP_0_LO,
> -               REG_A4XX_RBBM_PERFCTR_CP_0_HI);
> +       *value = gpu_read64(gpu, REG_A4XX_RBBM_PERFCTR_CP_0_LO);
>
>         return 0;
>  }
> diff --git a/drivers/gpu/drm/msm/adreno/a5xx_gpu.c b/drivers/gpu/drm/msm/adreno/a5xx_gpu.c
> index 407f50a15faa4..1916cb759cd5c 100644
> --- a/drivers/gpu/drm/msm/adreno/a5xx_gpu.c
> +++ b/drivers/gpu/drm/msm/adreno/a5xx_gpu.c
> @@ -605,11 +605,9 @@ static int a5xx_ucode_init(struct msm_gpu *gpu)
>                 a5xx_ucode_check_version(a5xx_gpu, a5xx_gpu->pfp_bo);
>         }
>
> -       gpu_write64(gpu, REG_A5XX_CP_ME_INSTR_BASE_LO,
> -               REG_A5XX_CP_ME_INSTR_BASE_HI, a5xx_gpu->pm4_iova);
> +       gpu_write64(gpu, REG_A5XX_CP_ME_INSTR_BASE_LO, a5xx_gpu->pm4_iova);
>
> -       gpu_write64(gpu, REG_A5XX_CP_PFP_INSTR_BASE_LO,
> -               REG_A5XX_CP_PFP_INSTR_BASE_HI, a5xx_gpu->pfp_iova);
> +       gpu_write64(gpu, REG_A5XX_CP_PFP_INSTR_BASE_LO, a5xx_gpu->pfp_iova);
>
>         return 0;
>  }
> @@ -868,8 +866,7 @@ static int a5xx_hw_init(struct msm_gpu *gpu)
>          * memory rendering at this point in time and we don't want to block off
>          * part of the virtual memory space.
>          */
> -       gpu_write64(gpu, REG_A5XX_RBBM_SECVID_TSB_TRUSTED_BASE_LO,
> -               REG_A5XX_RBBM_SECVID_TSB_TRUSTED_BASE_HI, 0x00000000);
> +       gpu_write64(gpu, REG_A5XX_RBBM_SECVID_TSB_TRUSTED_BASE_LO, 0x00000000);
>         gpu_write(gpu, REG_A5XX_RBBM_SECVID_TSB_TRUSTED_SIZE, 0x00000000);
>
>         /* Put the GPU into 64 bit by default */
> @@ -908,8 +905,7 @@ static int a5xx_hw_init(struct msm_gpu *gpu)
>                 return ret;
>
>         /* Set the ringbuffer address */
> -       gpu_write64(gpu, REG_A5XX_CP_RB_BASE, REG_A5XX_CP_RB_BASE_HI,
> -               gpu->rb[0]->iova);
> +       gpu_write64(gpu, REG_A5XX_CP_RB_BASE, gpu->rb[0]->iova);
>
>         /*
>          * If the microcode supports the WHERE_AM_I opcode then we can use that
> @@ -936,7 +932,7 @@ static int a5xx_hw_init(struct msm_gpu *gpu)
>                 }
>
>                 gpu_write64(gpu, REG_A5XX_CP_RB_RPTR_ADDR,
> -                       REG_A5XX_CP_RB_RPTR_ADDR_HI, shadowptr(a5xx_gpu, gpu->rb[0]));
> +                       shadowptr(a5xx_gpu, gpu->rb[0]));
>         } else if (gpu->nr_rings > 1) {
>                 /* Disable preemption if WHERE_AM_I isn't available */
>                 a5xx_preempt_fini(gpu);
> @@ -1239,9 +1235,9 @@ static void a5xx_fault_detect_irq(struct msm_gpu *gpu)
>                 gpu_read(gpu, REG_A5XX_RBBM_STATUS),
>                 gpu_read(gpu, REG_A5XX_CP_RB_RPTR),
>                 gpu_read(gpu, REG_A5XX_CP_RB_WPTR),
> -               gpu_read64(gpu, REG_A5XX_CP_IB1_BASE, REG_A5XX_CP_IB1_BASE_HI),
> +               gpu_read64(gpu, REG_A5XX_CP_IB1_BASE),
>                 gpu_read(gpu, REG_A5XX_CP_IB1_BUFSZ),
> -               gpu_read64(gpu, REG_A5XX_CP_IB2_BASE, REG_A5XX_CP_IB2_BASE_HI),
> +               gpu_read64(gpu, REG_A5XX_CP_IB2_BASE),
>                 gpu_read(gpu, REG_A5XX_CP_IB2_BUFSZ));
>
>         /* Turn off the hangcheck timer to keep it from bothering us */
> @@ -1427,8 +1423,7 @@ static int a5xx_pm_suspend(struct msm_gpu *gpu)
>
>  static int a5xx_get_timestamp(struct msm_gpu *gpu, uint64_t *value)
>  {
> -       *value = gpu_read64(gpu, REG_A5XX_RBBM_ALWAYSON_COUNTER_LO,
> -               REG_A5XX_RBBM_ALWAYSON_COUNTER_HI);
> +       *value = gpu_read64(gpu, REG_A5XX_RBBM_ALWAYSON_COUNTER_LO);
>
>         return 0;
>  }
> @@ -1465,8 +1460,7 @@ static int a5xx_crashdumper_run(struct msm_gpu *gpu,
>         if (IS_ERR_OR_NULL(dumper->ptr))
>                 return -EINVAL;
>
> -       gpu_write64(gpu, REG_A5XX_CP_CRASH_SCRIPT_BASE_LO,
> -               REG_A5XX_CP_CRASH_SCRIPT_BASE_HI, dumper->iova);
> +       gpu_write64(gpu, REG_A5XX_CP_CRASH_SCRIPT_BASE_LO, dumper->iova);
>
>         gpu_write(gpu, REG_A5XX_CP_CRASH_DUMP_CNTL, 1);
>
> @@ -1670,8 +1664,7 @@ static unsigned long a5xx_gpu_busy(struct msm_gpu *gpu)
>         if (pm_runtime_get_if_in_use(&gpu->pdev->dev) == 0)
>                 return 0;
>
> -       busy_cycles = gpu_read64(gpu, REG_A5XX_RBBM_PERFCTR_RBBM_0_LO,
> -                       REG_A5XX_RBBM_PERFCTR_RBBM_0_HI);
> +       busy_cycles = gpu_read64(gpu, REG_A5XX_RBBM_PERFCTR_RBBM_0_LO);
>
>         busy_time = busy_cycles - gpu->devfreq.busy_cycles;
>         do_div(busy_time, clk_get_rate(gpu->core_clk) / 1000000);
> diff --git a/drivers/gpu/drm/msm/adreno/a5xx_preempt.c b/drivers/gpu/drm/msm/adreno/a5xx_preempt.c
> index 8abc9a2b114a2..7658e89844b46 100644
> --- a/drivers/gpu/drm/msm/adreno/a5xx_preempt.c
> +++ b/drivers/gpu/drm/msm/adreno/a5xx_preempt.c
> @@ -137,7 +137,6 @@ void a5xx_preempt_trigger(struct msm_gpu *gpu)
>
>         /* Set the address of the incoming preemption record */
>         gpu_write64(gpu, REG_A5XX_CP_CONTEXT_SWITCH_RESTORE_ADDR_LO,
> -               REG_A5XX_CP_CONTEXT_SWITCH_RESTORE_ADDR_HI,
>                 a5xx_gpu->preempt_iova[ring->id]);
>
>         a5xx_gpu->next_ring = ring;
> @@ -211,8 +210,7 @@ void a5xx_preempt_hw_init(struct msm_gpu *gpu)
>         }
>
>         /* Write a 0 to signal that we aren't switching pagetables */
> -       gpu_write64(gpu, REG_A5XX_CP_CONTEXT_SWITCH_SMMU_INFO_LO,
> -               REG_A5XX_CP_CONTEXT_SWITCH_SMMU_INFO_HI, 0);
> +       gpu_write64(gpu, REG_A5XX_CP_CONTEXT_SWITCH_SMMU_INFO_LO, 0);
>
>         /* Reset the preemption state */
>         set_preempt_state(a5xx_gpu, PREEMPT_NONE);
> diff --git a/drivers/gpu/drm/msm/adreno/a6xx_gpu.c b/drivers/gpu/drm/msm/adreno/a6xx_gpu.c
> index 83c31b2ad865b..a624cb2df233b 100644
> --- a/drivers/gpu/drm/msm/adreno/a6xx_gpu.c
> +++ b/drivers/gpu/drm/msm/adreno/a6xx_gpu.c
> @@ -246,8 +246,7 @@ static void a6xx_submit(struct msm_gpu *gpu, struct msm_gem_submit *submit)
>         OUT_RING(ring, submit->seqno);
>
>         trace_msm_gpu_submit_flush(submit,
> -               gpu_read64(gpu, REG_A6XX_CP_ALWAYS_ON_COUNTER_LO,
> -                       REG_A6XX_CP_ALWAYS_ON_COUNTER_HI));
> +               gpu_read64(gpu, REG_A6XX_CP_ALWAYS_ON_COUNTER_LO));
>
>         a6xx_flush(gpu, ring);
>  }
> @@ -878,8 +877,7 @@ static int a6xx_ucode_init(struct msm_gpu *gpu)
>                 }
>         }
>
> -       gpu_write64(gpu, REG_A6XX_CP_SQE_INSTR_BASE,
> -               REG_A6XX_CP_SQE_INSTR_BASE+1, a6xx_gpu->sqe_iova);
> +       gpu_write64(gpu, REG_A6XX_CP_SQE_INSTR_BASE, a6xx_gpu->sqe_iova);
>
>         return 0;
>  }
> @@ -926,8 +924,7 @@ static int hw_init(struct msm_gpu *gpu)
>          * memory rendering at this point in time and we don't want to block off
>          * part of the virtual memory space.
>          */
> -       gpu_write64(gpu, REG_A6XX_RBBM_SECVID_TSB_TRUSTED_BASE_LO,
> -               REG_A6XX_RBBM_SECVID_TSB_TRUSTED_BASE_HI, 0x00000000);
> +       gpu_write64(gpu, REG_A6XX_RBBM_SECVID_TSB_TRUSTED_BASE_LO, 0x00000000);
>         gpu_write(gpu, REG_A6XX_RBBM_SECVID_TSB_TRUSTED_SIZE, 0x00000000);
>
>         /* Turn on 64 bit addressing for all blocks */
> @@ -976,11 +973,8 @@ static int hw_init(struct msm_gpu *gpu)
>
>         if (!adreno_is_a650_family(adreno_gpu)) {
>                 /* Set the GMEM VA range [0x100000:0x100000 + gpu->gmem - 1] */
> -               gpu_write64(gpu, REG_A6XX_UCHE_GMEM_RANGE_MIN_LO,
> -                       REG_A6XX_UCHE_GMEM_RANGE_MIN_HI, 0x00100000);
> -
> +               gpu_write64(gpu, REG_A6XX_UCHE_GMEM_RANGE_MIN_LO, 0x00100000);
>                 gpu_write64(gpu, REG_A6XX_UCHE_GMEM_RANGE_MAX_LO,
> -                       REG_A6XX_UCHE_GMEM_RANGE_MAX_HI,
>                         0x00100000 + adreno_gpu->gmem - 1);
>         }
>
> @@ -1072,8 +1066,7 @@ static int hw_init(struct msm_gpu *gpu)
>                 goto out;
>
>         /* Set the ringbuffer address */
> -       gpu_write64(gpu, REG_A6XX_CP_RB_BASE, REG_A6XX_CP_RB_BASE_HI,
> -               gpu->rb[0]->iova);
> +       gpu_write64(gpu, REG_A6XX_CP_RB_BASE, gpu->rb[0]->iova);
>
>         /* Targets that support extended APRIV can use the RPTR shadow from
>          * hardware but all the other ones need to disable the feature. Targets
> @@ -1105,7 +1098,6 @@ static int hw_init(struct msm_gpu *gpu)
>                 }
>
>                 gpu_write64(gpu, REG_A6XX_CP_RB_RPTR_ADDR_LO,
> -                       REG_A6XX_CP_RB_RPTR_ADDR_HI,
>                         shadowptr(a6xx_gpu, gpu->rb[0]));
>         }
>
> @@ -1394,9 +1386,9 @@ static void a6xx_fault_detect_irq(struct msm_gpu *gpu)
>                 gpu_read(gpu, REG_A6XX_RBBM_STATUS),
>                 gpu_read(gpu, REG_A6XX_CP_RB_RPTR),
>                 gpu_read(gpu, REG_A6XX_CP_RB_WPTR),
> -               gpu_read64(gpu, REG_A6XX_CP_IB1_BASE, REG_A6XX_CP_IB1_BASE_HI),
> +               gpu_read64(gpu, REG_A6XX_CP_IB1_BASE),
>                 gpu_read(gpu, REG_A6XX_CP_IB1_REM_SIZE),
> -               gpu_read64(gpu, REG_A6XX_CP_IB2_BASE, REG_A6XX_CP_IB2_BASE_HI),
> +               gpu_read64(gpu, REG_A6XX_CP_IB2_BASE),
>                 gpu_read(gpu, REG_A6XX_CP_IB2_REM_SIZE));
>
>         /* Turn off the hangcheck timer to keep it from bothering us */
> @@ -1607,8 +1599,7 @@ static int a6xx_get_timestamp(struct msm_gpu *gpu, uint64_t *value)
>         /* Force the GPU power on so we can read this register */
>         a6xx_gmu_set_oob(&a6xx_gpu->gmu, GMU_OOB_PERFCOUNTER_SET);
>
> -       *value = gpu_read64(gpu, REG_A6XX_CP_ALWAYS_ON_COUNTER_LO,
> -                           REG_A6XX_CP_ALWAYS_ON_COUNTER_HI);
> +       *value = gpu_read64(gpu, REG_A6XX_CP_ALWAYS_ON_COUNTER_LO);
>
>         a6xx_gmu_clear_oob(&a6xx_gpu->gmu, GMU_OOB_PERFCOUNTER_SET);
>
> diff --git a/drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c b/drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
> index 55f443328d8e7..c61b233aff09b 100644
> --- a/drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
> +++ b/drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
> @@ -147,8 +147,7 @@ static int a6xx_crashdumper_run(struct msm_gpu *gpu,
>         /* Make sure all pending memory writes are posted */
>         wmb();
>
> -       gpu_write64(gpu, REG_A6XX_CP_CRASH_SCRIPT_BASE_LO,
> -               REG_A6XX_CP_CRASH_SCRIPT_BASE_HI, dumper->iova);
> +       gpu_write64(gpu, REG_A6XX_CP_CRASH_SCRIPT_BASE_LO, dumper->iova);
>
>         gpu_write(gpu, REG_A6XX_CP_CRASH_DUMP_CNTL, 1);
>
> diff --git a/drivers/gpu/drm/msm/msm_gpu.h b/drivers/gpu/drm/msm/msm_gpu.h
> index 02419f2ca2bc5..f7fca687d45de 100644
> --- a/drivers/gpu/drm/msm/msm_gpu.h
> +++ b/drivers/gpu/drm/msm/msm_gpu.h
> @@ -503,7 +503,7 @@ static inline void gpu_rmw(struct msm_gpu *gpu, u32 reg, u32 mask, u32 or)
>         msm_rmw(gpu->mmio + (reg << 2), mask, or);
>  }
>
> -static inline u64 gpu_read64(struct msm_gpu *gpu, u32 lo, u32 hi)
> +static inline u64 gpu_read64(struct msm_gpu *gpu, u32 reg)
>  {
>         u64 val;
>
> @@ -521,17 +521,17 @@ static inline u64 gpu_read64(struct msm_gpu *gpu, u32 lo, u32 hi)
>          * when the lo is read, so make sure to read the lo first to trigger
>          * that
>          */
> -       val = (u64) msm_readl(gpu->mmio + (lo << 2));
> -       val |= ((u64) msm_readl(gpu->mmio + (hi << 2)) << 32);
> +       val = (u64) msm_readl(gpu->mmio + (reg << 2));
> +       val |= ((u64) msm_readl(gpu->mmio + ((reg + 1) << 2)) << 32);
>
>         return val;
>  }
>
> -static inline void gpu_write64(struct msm_gpu *gpu, u32 lo, u32 hi, u64 val)
> +static inline void gpu_write64(struct msm_gpu *gpu, u32 reg, u64 val)
>  {
>         /* Why not a writeq here? Read the screed above */
> -       msm_writel(lower_32_bits(val), gpu->mmio + (lo << 2));
> -       msm_writel(upper_32_bits(val), gpu->mmio + (hi << 2));
> +       msm_writel(lower_32_bits(val), gpu->mmio + (reg << 2));
> +       msm_writel(upper_32_bits(val), gpu->mmio + ((reg + 1) << 2));
>  }
>
>  int msm_gpu_pm_suspend(struct msm_gpu *gpu);
> --
> 2.26.1
>


More information about the dri-devel mailing list