[PATCH] drm/amdgpu: only export available rings to mesa for enabling kq|uq

Alex Deucher alexdeucher at gmail.com
Wed May 28 13:10:45 UTC 2025


On Wed, May 28, 2025 at 4:38 AM Prike Liang <Prike.Liang at amd.com> wrote:
>
> The kernel driver only requires exporting available rings to the mesa
> when the userq is disabled; otherwise, the userq IP mask will be cleaned
> up in the mesa.

The logic should work correctly as is.  There are three possible states:
1. KQ only
2. Both KQ and UQ
3. UQ only

Alex

>
> Signed-off-by: Prike Liang <Prike.Liang at amd.com>
> ---
>  drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c | 20 ++++++++++----------
>  1 file changed, 10 insertions(+), 10 deletions(-)
>
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
> index d2ce7d86dbc8..43d86c09d8bb 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
> @@ -409,7 +409,7 @@ static int amdgpu_hw_ip_info(struct amdgpu_device *adev,
>                 type = AMD_IP_BLOCK_TYPE_GFX;
>                 for (i = 0; i < adev->gfx.num_gfx_rings; i++)
>                         if (adev->gfx.gfx_ring[i].sched.ready &&
> -                           !adev->gfx.gfx_ring[i].no_user_submission)
> +                           adev->gfx.disable_uq)
>                                 ++num_rings;
>                 ib_start_alignment = 32;
>                 ib_size_alignment = 32;
> @@ -418,7 +418,7 @@ static int amdgpu_hw_ip_info(struct amdgpu_device *adev,
>                 type = AMD_IP_BLOCK_TYPE_GFX;
>                 for (i = 0; i < adev->gfx.num_compute_rings; i++)
>                         if (adev->gfx.compute_ring[i].sched.ready &&
> -                           !adev->gfx.compute_ring[i].no_user_submission)
> +                           adev->gfx.disable_uq)
>                                 ++num_rings;
>                 ib_start_alignment = 32;
>                 ib_size_alignment = 32;
> @@ -427,7 +427,7 @@ static int amdgpu_hw_ip_info(struct amdgpu_device *adev,
>                 type = AMD_IP_BLOCK_TYPE_SDMA;
>                 for (i = 0; i < adev->sdma.num_instances; i++)
>                         if (adev->sdma.instance[i].ring.sched.ready &&
> -                           !adev->sdma.instance[i].ring.no_user_submission)
> +                               adev->gfx.disable_uq)
>                                 ++num_rings;
>                 ib_start_alignment = 256;
>                 ib_size_alignment = 4;
> @@ -439,7 +439,7 @@ static int amdgpu_hw_ip_info(struct amdgpu_device *adev,
>                                 continue;
>
>                         if (adev->uvd.inst[i].ring.sched.ready &&
> -                           !adev->uvd.inst[i].ring.no_user_submission)
> +                           adev->gfx.disable_uq)
>                                 ++num_rings;
>                 }
>                 ib_start_alignment = 256;
> @@ -449,7 +449,7 @@ static int amdgpu_hw_ip_info(struct amdgpu_device *adev,
>                 type = AMD_IP_BLOCK_TYPE_VCE;
>                 for (i = 0; i < adev->vce.num_rings; i++)
>                         if (adev->vce.ring[i].sched.ready &&
> -                           !adev->vce.ring[i].no_user_submission)
> +                           adev->gfx.disable_uq)
>                                 ++num_rings;
>                 ib_start_alignment = 256;
>                 ib_size_alignment = 4;
> @@ -462,7 +462,7 @@ static int amdgpu_hw_ip_info(struct amdgpu_device *adev,
>
>                         for (j = 0; j < adev->uvd.num_enc_rings; j++)
>                                 if (adev->uvd.inst[i].ring_enc[j].sched.ready &&
> -                                   !adev->uvd.inst[i].ring_enc[j].no_user_submission)
> +                                   adev->gfx.disable_uq)
>                                         ++num_rings;
>                 }
>                 ib_start_alignment = 256;
> @@ -475,7 +475,7 @@ static int amdgpu_hw_ip_info(struct amdgpu_device *adev,
>                                 continue;
>
>                         if (adev->vcn.inst[i].ring_dec.sched.ready &&
> -                           !adev->vcn.inst[i].ring_dec.no_user_submission)
> +                           adev->gfx.disable_uq)
>                                 ++num_rings;
>                 }
>                 ib_start_alignment = 256;
> @@ -489,7 +489,7 @@ static int amdgpu_hw_ip_info(struct amdgpu_device *adev,
>
>                         for (j = 0; j < adev->vcn.inst[i].num_enc_rings; j++)
>                                 if (adev->vcn.inst[i].ring_enc[j].sched.ready &&
> -                                   !adev->vcn.inst[i].ring_enc[j].no_user_submission)
> +                                   adev->gfx.disable_uq)
>                                         ++num_rings;
>                 }
>                 ib_start_alignment = 256;
> @@ -505,7 +505,7 @@ static int amdgpu_hw_ip_info(struct amdgpu_device *adev,
>
>                         for (j = 0; j < adev->jpeg.num_jpeg_rings; j++)
>                                 if (adev->jpeg.inst[i].ring_dec[j].sched.ready &&
> -                                   !adev->jpeg.inst[i].ring_dec[j].no_user_submission)
> +                                   adev->gfx.disable_uq)
>                                         ++num_rings;
>                 }
>                 ib_start_alignment = 256;
> @@ -514,7 +514,7 @@ static int amdgpu_hw_ip_info(struct amdgpu_device *adev,
>         case AMDGPU_HW_IP_VPE:
>                 type = AMD_IP_BLOCK_TYPE_VPE;
>                 if (adev->vpe.ring.sched.ready &&
> -                   !adev->vpe.ring.no_user_submission)
> +                   adev->gfx.disable_uq)
>                         ++num_rings;
>                 ib_start_alignment = 256;
>                 ib_size_alignment = 4;
> --
> 2.34.1
>


More information about the amd-gfx mailing list