[PATCH v1 3/3] drm/amdgpu: add cp queue registers for gfx9 ipdump
Alex Deucher
alexdeucher at gmail.com
Fri May 31 04:04:44 UTC 2024
On Wed, May 29, 2024 at 5:50 AM Sunil Khatri <sunil.khatri at amd.com> wrote:
>
> Add gfx9 support of CP queue registers for all queues
> to be used by devcoredump.
>
> Signed-off-by: Sunil Khatri <sunil.khatri at amd.com>
> ---
> drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c | 110 +++++++++++++++++++++++++-
> 1 file changed, 108 insertions(+), 2 deletions(-)
>
> diff --git a/drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c b/drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
> index f3d281de2a34..f1a7b60e73ce 100644
> --- a/drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
> +++ b/drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
> @@ -237,6 +237,47 @@ static const struct amdgpu_hwip_reg_entry gc_reg_list_9[] = {
> SOC15_REG_ENTRY_STR(GC, 0, mmGRBM_STATUS_SE3)
> };
>
> +static const struct amdgpu_hwip_reg_entry gc_cp_reg_list_9[] = {
> + /* compute queue registers */
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_VMID),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_ACTIVE),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_PERSISTENT_STATE),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_PIPE_PRIORITY),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_QUEUE_PRIORITY),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_QUANTUM),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_PQ_BASE),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_PQ_BASE_HI),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_PQ_RPTR),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_PQ_WPTR_POLL_ADDR),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_PQ_WPTR_POLL_ADDR_HI),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_PQ_DOORBELL_CONTROL),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_PQ_CONTROL),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_IB_BASE_ADDR),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_IB_BASE_ADDR_HI),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_IB_RPTR),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_IB_CONTROL),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_DEQUEUE_REQUEST),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_EOP_BASE_ADDR),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_EOP_BASE_ADDR_HI),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_EOP_CONTROL),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_EOP_RPTR),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_EOP_WPTR),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_EOP_EVENTS),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_CTX_SAVE_BASE_ADDR_LO),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_CTX_SAVE_BASE_ADDR_HI),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_CTX_SAVE_CONTROL),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_CNTL_STACK_OFFSET),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_CNTL_STACK_SIZE),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_WG_STATE_OFFSET),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_CTX_SAVE_SIZE),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_GDS_RESOURCE_STATE),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_ERROR),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_EOP_WPTR_MEM),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_PQ_WPTR_LO),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_PQ_WPTR_HI),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_GFX_STATUS),
> +};
> +
> enum ta_ras_gfx_subblock {
> /*CPC*/
> TA_RAS_BLOCK__GFX_CPC_INDEX_START = 0,
> @@ -2086,6 +2127,7 @@ static void gfx_v9_0_alloc_ip_dump(struct amdgpu_device *adev)
> {
> uint32_t reg_count = ARRAY_SIZE(gc_reg_list_9);
> uint32_t *ptr;
> + uint32_t inst;
>
> ptr = kcalloc(reg_count, sizeof(uint32_t), GFP_KERNEL);
> if (ptr == NULL) {
> @@ -2094,6 +2136,19 @@ static void gfx_v9_0_alloc_ip_dump(struct amdgpu_device *adev)
> } else {
> adev->gfx.ip_dump_core = ptr;
> }
> +
> + /* Allocate memory for compute queue registers for all the instances */
> + reg_count = ARRAY_SIZE(gc_cp_reg_list_9);
> + inst = adev->gfx.mec.num_mec * adev->gfx.mec.num_pipe_per_mec *
> + adev->gfx.mec.num_queue_per_pipe;
> +
> + ptr = kcalloc(reg_count * inst, sizeof(uint32_t), GFP_KERNEL);
> + if (ptr == NULL) {
> + DRM_ERROR("Failed to allocate memory for GFX CP IP Dump\n");
> + adev->gfx.ip_dump_cp_queues = NULL;
> + } else {
> + adev->gfx.ip_dump_cp_queues = ptr;
As a follow on patch, you may want to rename this to
ip_dump_compute_queues since these are compute queue registers.
Other than that, the series is:
Reviewed-by: Alex Deucher <alexander.deucher at amd.com>
> + }
> }
>
> static int gfx_v9_0_sw_init(void *handle)
> @@ -2311,6 +2366,7 @@ static int gfx_v9_0_sw_fini(void *handle)
> gfx_v9_0_free_microcode(adev);
>
> kfree(adev->gfx.ip_dump_core);
> + kfree(adev->gfx.ip_dump_cp_queues);
>
> return 0;
> }
> @@ -6949,7 +7005,7 @@ static void gfx_v9_0_emit_wave_limit(struct amdgpu_ring *ring, bool enable)
> static void gfx_v9_ip_print(void *handle, struct drm_printer *p)
> {
> struct amdgpu_device *adev = (struct amdgpu_device *)handle;
> - uint32_t i;
> + uint32_t i, j, k, reg, index = 0;
> uint32_t reg_count = ARRAY_SIZE(gc_reg_list_9);
>
> if (!adev->gfx.ip_dump_core)
> @@ -6960,12 +7016,36 @@ static void gfx_v9_ip_print(void *handle, struct drm_printer *p)
> gc_reg_list_9[i].reg_name,
> adev->gfx.ip_dump_core[i]);
>
> + /* print compute queue registers for all instances */
> + if (!adev->gfx.ip_dump_cp_queues)
> + return;
> +
> + reg_count = ARRAY_SIZE(gc_cp_reg_list_9);
> + drm_printf(p, "\nnum_mec: %d num_pipe: %d num_queue: %d\n",
> + adev->gfx.mec.num_mec,
> + adev->gfx.mec.num_pipe_per_mec,
> + adev->gfx.mec.num_queue_per_pipe);
> +
> + for (i = 0; i < adev->gfx.mec.num_mec; i++) {
> + for (j = 0; j < adev->gfx.mec.num_pipe_per_mec; j++) {
> + for (k = 0; k < adev->gfx.mec.num_queue_per_pipe; k++) {
> + drm_printf(p, "\nmec %d, pipe %d, queue %d\n", i, j, k);
> + for (reg = 0; reg < reg_count; reg++) {
> + drm_printf(p, "%-50s \t 0x%08x\n",
> + gc_cp_reg_list_9[reg].reg_name,
> + adev->gfx.ip_dump_cp_queues[index + reg]);
> + }
> + index += reg_count;
> + }
> + }
> + }
> +
> }
>
> static void gfx_v9_ip_dump(void *handle)
> {
> struct amdgpu_device *adev = (struct amdgpu_device *)handle;
> - uint32_t i;
> + uint32_t i, j, k, reg, index = 0;
> uint32_t reg_count = ARRAY_SIZE(gc_reg_list_9);
>
> if (!adev->gfx.ip_dump_core)
> @@ -6976,6 +7056,32 @@ static void gfx_v9_ip_dump(void *handle)
> adev->gfx.ip_dump_core[i] = RREG32(SOC15_REG_ENTRY_OFFSET(gc_reg_list_9[i]));
> amdgpu_gfx_off_ctrl(adev, true);
>
> + /* dump compute queue registers for all instances */
> + if (!adev->gfx.ip_dump_cp_queues)
> + return;
> +
> + reg_count = ARRAY_SIZE(gc_cp_reg_list_9);
> + amdgpu_gfx_off_ctrl(adev, false);
> + mutex_lock(&adev->srbm_mutex);
> + for (i = 0; i < adev->gfx.mec.num_mec; i++) {
> + for (j = 0; j < adev->gfx.mec.num_pipe_per_mec; j++) {
> + for (k = 0; k < adev->gfx.mec.num_queue_per_pipe; k++) {
> + /* ME0 is for GFX so start from 1 for CP */
> + soc15_grbm_select(adev, 1 + i, j, k, 0, 0);
> +
> + for (reg = 0; reg < reg_count; reg++) {
> + adev->gfx.ip_dump_cp_queues[index + reg] =
> + RREG32(SOC15_REG_ENTRY_OFFSET(
> + gc_cp_reg_list_9[reg]));
> + }
> + index += reg_count;
> + }
> + }
> + }
> + soc15_grbm_select(adev, 0, 0, 0, 0, 0);
> + mutex_unlock(&adev->srbm_mutex);
> + amdgpu_gfx_off_ctrl(adev, true);
> +
> }
>
> static const struct amd_ip_funcs gfx_v9_0_ip_funcs = {
> --
> 2.34.1
>
More information about the amd-gfx
mailing list