[PATCH] drm/amd/pm: conditionally disable pcie lane switching for some sienna_cichlid SKUs
Limonciello, Mario
Mario.Limonciello at amd.com
Fri Apr 21 13:39:42 UTC 2023
[Public]
> -----Original Message-----
> From: Quan, Evan <Evan.Quan at amd.com>
> Sent: Friday, April 21, 2023 02:29
> To: amd-gfx at lists.freedesktop.org
> Cc: Deucher, Alexander <Alexander.Deucher at amd.com>; Limonciello, Mario
> <Mario.Limonciello at amd.com>; Quan, Evan <Evan.Quan at amd.com>
> Subject: [PATCH] drm/amd/pm: conditionally disable pcie lane switching for
> some sienna_cichlid SKUs
>
> Disable the pcie lane switching for some sienna_cichlid SKUs since it
> might not work well on some platforms.
>
> Signed-off-by: Evan Quan <evan.quan at amd.com>
> Change-Id: Iea9ceaa146c8706768ee077c10e5d33bce9bc1c2
You can drop the Gerrit Change-Id here
> ---
> .../amd/pm/swsmu/smu11/sienna_cichlid_ppt.c | 92 +++++++++++++++----
> 1 file changed, 74 insertions(+), 18 deletions(-)
>
> diff --git a/drivers/gpu/drm/amd/pm/swsmu/smu11/sienna_cichlid_ppt.c
> b/drivers/gpu/drm/amd/pm/swsmu/smu11/sienna_cichlid_ppt.c
> index 4b91cdc3eaa0..e7223513e384 100644
> --- a/drivers/gpu/drm/amd/pm/swsmu/smu11/sienna_cichlid_ppt.c
> +++ b/drivers/gpu/drm/amd/pm/swsmu/smu11/sienna_cichlid_ppt.c
> @@ -2067,33 +2067,94 @@ static int
> sienna_cichlid_display_disable_memory_clock_switch(struct smu_context
> return ret;
> }
>
> +static void sienna_cichlid_get_override_pcie_settings(struct smu_context
> *smu,
> + uint32_t
> *gen_speed_override,
> + uint32_t
> *lane_width_override)
> +{
> + struct amdgpu_device *adev = smu->adev;
> +
> + *gen_speed_override = 0xff;
> + *lane_width_override = 0xff;
> +
> + switch (adev->pdev->device) {
> + case 0x73A0:
> + case 0x73A1:
> + case 0x73A2:
> + case 0x73A3:
> + case 0x73AB:
> + case 0x73AE:
> + /* Bit 7:0: PCIE lane width, 1 to 7 corresponds is x1 to x32 */
> + *lane_width_override = 6;
> + break;
> + case 0x73E0:
> + case 0x73E1:
> + case 0x73E3:
> + *lane_width_override = 4;
> + break;
> + case 0x7420:
> + case 0x7421:
> + case 0x7422:
> + case 0x7423:
> + case 0x7424:
> + *lane_width_override = 3;
> + break;
> + default:
> + break;
> + }
> +}
> +
> +#define MAX(a, b) ((a) > (b) ? (a) : (b))
> +
> static int sienna_cichlid_update_pcie_parameters(struct smu_context *smu,
> uint32_t pcie_gen_cap,
> uint32_t pcie_width_cap)
> {
> struct smu_11_0_dpm_context *dpm_context = smu-
> >smu_dpm.dpm_context;
> -
> - uint32_t smu_pcie_arg;
> + struct smu_11_0_pcie_table *pcie_table = &dpm_context-
> >dpm_tables.pcie_table;
> + uint32_t gen_speed_override, lane_width_override;
> uint8_t *table_member1, *table_member2;
> + uint32_t min_gen_speed, max_gen_speed;
> + uint32_t min_lane_width, max_lane_width;
> + uint32_t smu_pcie_arg;
> int ret, i;
>
> GET_PPTABLE_MEMBER(PcieGenSpeed, &table_member1);
> GET_PPTABLE_MEMBER(PcieLaneCount, &table_member2);
>
> - /* lclk dpm table setup */
> - for (i = 0; i < MAX_PCIE_CONF; i++) {
> - dpm_context->dpm_tables.pcie_table.pcie_gen[i] =
> table_member1[i];
> - dpm_context->dpm_tables.pcie_table.pcie_lane[i] =
> table_member2[i];
> + sienna_cichlid_get_override_pcie_settings(smu,
> + &gen_speed_override,
> + &lane_width_override);
> +
> + /* PCIE gen speed override */
> + if (gen_speed_override != 0xff) {
> + min_gen_speed = MIN(pcie_gen_cap, gen_speed_override);
> + max_gen_speed = MIN(pcie_gen_cap, gen_speed_override);
> + } else {
> + min_gen_speed = MAX(0, table_member1[0]);
> + max_gen_speed = MIN(pcie_gen_cap, table_member1[1]);
> + min_gen_speed = min_gen_speed > max_gen_speed ?
> + max_gen_speed : min_gen_speed;
> }
> + pcie_table->pcie_gen[0] = min_gen_speed;
> + pcie_table->pcie_gen[1] = max_gen_speed;
> +
> + /* PCIE lane width override */
> + if (lane_width_override != 0xff) {
> + min_lane_width = MIN(pcie_width_cap, lane_width_override);
> + max_lane_width = MIN(pcie_width_cap, lane_width_override);
> + } else {
> + min_lane_width = MAX(1, table_member2[0]);
> + max_lane_width = MIN(pcie_width_cap, table_member2[1]);
> + min_lane_width = min_lane_width > max_lane_width ?
> + max_lane_width : min_lane_width;
> + }
> + pcie_table->pcie_lane[0] = min_lane_width;
> + pcie_table->pcie_lane[1] = max_lane_width;
>
> for (i = 0; i < NUM_LINK_LEVELS; i++) {
> - smu_pcie_arg = (i << 16) |
> - ((table_member1[i] <= pcie_gen_cap) ?
> - (table_member1[i] << 8) :
> - (pcie_gen_cap << 8)) |
> - ((table_member2[i] <= pcie_width_cap) ?
> - table_member2[i] :
> - pcie_width_cap);
> + smu_pcie_arg = (i << 16 |
> + pcie_table->pcie_gen[i] << 8 |
> + pcie_table->pcie_lane[i]);
>
> ret = smu_cmn_send_smc_msg_with_param(smu,
> SMU_MSG_OverridePcieParameters,
> @@ -2101,11 +2162,6 @@ static int
> sienna_cichlid_update_pcie_parameters(struct smu_context *smu,
> NULL);
> if (ret)
> return ret;
> -
> - if (table_member1[i] > pcie_gen_cap)
> - dpm_context->dpm_tables.pcie_table.pcie_gen[i] =
> pcie_gen_cap;
> - if (table_member2[i] > pcie_width_cap)
> - dpm_context->dpm_tables.pcie_table.pcie_lane[i] =
> pcie_width_cap;
> }
>
> return 0;
> --
> 2.34.1
More information about the amd-gfx
mailing list