[PATCH 2/7] drm/amdgpu: use scheduler load balancing for SDMA CS

Zhang, Jerry (Junwei) Jerry.Zhang at amd.com
Thu Aug 2 05:50:33 UTC 2018


On 08/01/2018 07:31 PM, Christian König wrote:
> Start to use the scheduler load balancing for userspace SDMA
> command submissions.
>

In this case, each SDMA could load all SDMA(instances) rqs, and UMD will not specify a ring id.
If so, we may abstract a set of rings for each type of IP, associated with such kind of IP instances' rq.
Accordingly libdrm needs to update as well, but it may be more user-friendly, regardless of ring id when submits command.

And will it interfere ctx->ring's settings, like sequence?
e.g. submit a command to SDMA0, but SDMA0 is busy and SDMA1 is idle,
So the command will be pushed to SDMA1 rq, but update sequence for ctx->ring[SDMA0] actually.


Regards,
Jerry

> Signed-off-by: Christian König <christian.koenig at amd.com>
> ---
>   drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c | 25 +++++++++++++++++++++----
>   1 file changed, 21 insertions(+), 4 deletions(-)
>
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c
> index df6965761046..59046f68975a 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c
> @@ -48,7 +48,8 @@ static int amdgpu_ctx_init(struct amdgpu_device *adev,
>   			   struct drm_file *filp,
>   			   struct amdgpu_ctx *ctx)
>   {
> -	unsigned i, j;
> +	struct drm_sched_rq *sdma_rqs[AMDGPU_MAX_RINGS];
> +	unsigned i, j, num_sdma_rqs;
>   	int r;
>
>   	if (priority < 0 || priority >= DRM_SCHED_PRIORITY_MAX)
> @@ -80,18 +81,34 @@ static int amdgpu_ctx_init(struct amdgpu_device *adev,
>   	ctx->init_priority = priority;
>   	ctx->override_priority = DRM_SCHED_PRIORITY_UNSET;
>
> -	/* create context entity for each ring */
> +	num_sdma_rqs = 0;
>   	for (i = 0; i < adev->num_rings; i++) {
>   		struct amdgpu_ring *ring = adev->rings[i];
>   		struct drm_sched_rq *rq;
>
>   		rq = &ring->sched.sched_rq[priority];
> +		if (ring->funcs->type == AMDGPU_RING_TYPE_SDMA)
> +			sdma_rqs[num_sdma_rqs++] = rq;
> +	}
> +
> +	/* create context entity for each ring */
> +	for (i = 0; i < adev->num_rings; i++) {
> +		struct amdgpu_ring *ring = adev->rings[i];
>
>   		if (ring == &adev->gfx.kiq.ring)
>   			continue;
>
> -		r = drm_sched_entity_init(&ctx->rings[i].entity,
> -					  &rq, 1, &ctx->guilty);
> +		if (ring->funcs->type == AMDGPU_RING_TYPE_SDMA) {
> +			r = drm_sched_entity_init(&ctx->rings[i].entity,
> +						  sdma_rqs, num_sdma_rqs,
> +						  &ctx->guilty);
> +		} else {
> +			struct drm_sched_rq *rq;
> +
> +			rq = &ring->sched.sched_rq[priority];
> +			r = drm_sched_entity_init(&ctx->rings[i].entity,
> +						  &rq, 1, &ctx->guilty);
> +		}
>   		if (r)
>   			goto failed;
>   	}
>


More information about the amd-gfx mailing list