[PATCH 4/4] drm/scheduler: do not keep a copy of sched list

Nirmoy nirmodas at amd.com
Mon Dec 9 13:56:23 UTC 2019


Hi Christian,

I got a different idea, a bit more simple let me know what do you think 
about it:

diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu.h 
b/drivers/gpu/drm/amd/amdgpu/amdgpu.h
index 50bab33cba39..8de4de4f7a43 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu.h
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu.h
@@ -870,6 +870,7 @@ struct amdgpu_device {
         u64                             fence_context;
         unsigned                        num_rings;
         struct amdgpu_ring              *rings[AMDGPU_MAX_RINGS];
+      struct drm_gpu_scheduler *rings_sched_list[AMDGPU_MAX_RINGS];
         bool                            ib_pool_ready;
         struct amdgpu_sa_manager        ring_tmp_bo;

diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c 
b/drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c
index 1d6850af9908..52b3a5d85a1d 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c
@@ -122,9 +122,8 @@ static int amdgpu_ctx_init(struct amdgpu_device *adev,

         for (i = 0; i < AMDGPU_HW_IP_NUM; ++i) {
                 struct amdgpu_ring *rings[AMDGPU_MAX_RINGS];
-               struct drm_gpu_scheduler *sched_list[AMDGPU_MAX_RINGS]
+              struct drm_gpu_scheduler **sched_list;
                 unsigned num_rings = 0;
-               unsigned num_rqs = 0;

                 switch (i) {
                 case AMDGPU_HW_IP_GFX:
@@ -177,17 +176,11 @@ static int amdgpu_ctx_init(struct amdgpu_device *adev,
                         break;
                 }

-               for (j = 0; j < num_rings; ++j) {
-                       if (!rings[j]->adev)
-                               continue;
-
-                       sched_list[num_rqs++] = &rings[j]->sched;
-               }
-
+              sched_list= adev->rings_sched_list+rings[0]->idx;
                 for (j = 0; j < amdgpu_ctx_num_entities[i]; ++j)
                         r = 
drm_sched_entity_init(&ctx->entities[i][j].entity,
                                                   priority, sched_list,
-                                                 num_rqs, &ctx->guilty);
+                                                num_rings, &ctx->guilty);
                 if (r)
                         goto error_cleanup_entities;
         }
diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 
b/drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c
index 377fe20bce23..e8cfa357e445 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c
@@ -480,6 +480,8 @@ int amdgpu_fence_driver_init_ring(struct amdgpu_ring 
*ring,
                                   ring->name);
                         return r;
                 }
+
+               adev->rings_sched_list[ring->idx] = &ring->sched;
         }

         return 0;
diff --git a/drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 
b/drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c
index bd9ed33bab43..bfe36199ffed 100644
--- a/drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c
+++ b/drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c
@@ -1744,8 +1744,11 @@ static int sdma_v4_0_sw_init(void *handle)
                                      AMDGPU_SDMA_IRQ_INSTANCE0 + i);
                 if (r)
                         return r;
+       }
+
+       if (adev->sdma.has_page_queue) {
+               for (i = 0; i < adev->sdma.num_instances; i++) {

-               if (adev->sdma.has_page_queue) {
                         ring = &adev->sdma.instance[i].page;
                         ring->ring_obj = NULL;
                         ring->use_doorbell = true;

It relies on contiguous ring initialization that's why I had to change  
sdma_v4_0.c so that we do ring_init(sdma0, sdma1, page0, page1}

instead of ring_init{sdma0, page0, sdma1, page1}


Regards,

Nirmoy

On 12/9/19 1:20 PM, Christian König wrote:
> Yes, you need to do this for the SDMA as well but in general that 
> looks like the idea I had in mind as well.
>
> I would do it like this:
>
> 1. Change the special case when you only get one scheduler for an 
> entity to drop the pointer to the scheduler list.
>     This way we always use the same scheduler for the entity and can 
> pass in the array on the stack.
>
> 2. Change all callers which use more than one scheduler in the list to 
> pass in pointers which are not allocated on the stack.
>     This obviously also means that we build the list of schedulers for 
> each type only once during device init and not for each context init.
>
> 3. Make the scheduler list const and drop the kcalloc()/kfree() from 
> the entity code.
>
> Regards,
> Christian.
>
> Am 08.12.19 um 20:57 schrieb Nirmoy:
>>
>> On 12/6/19 8:41 PM, Christian König wrote:
>>> Am 06.12.19 um 18:33 schrieb Nirmoy Das:
>>>> entity should not keep copy and maintain sched list for
>>>> itself.
>>>
>>> That is a good step, but we need to take this further.
>>
>> How about  something like ?
>>
>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.h 
>> b/drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.h
>> index 0ae0a2715b0d..a71ee084b47a 100644
>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.h
>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.h
>> @@ -269,8 +269,10 @@ struct amdgpu_gfx {
>>         bool                            me_fw_write_wait;
>>         bool                            cp_fw_write_wait;
>>         struct amdgpu_ring gfx_ring[AMDGPU_MAX_GFX_RINGS];
>> +       struct drm_gpu_scheduler *gfx_sched_list[AMDGPU_MAX_GFX_RINGS];
>>         unsigned                        num_gfx_rings;
>>         struct amdgpu_ring compute_ring[AMDGPU_MAX_COMPUTE_RINGS];
>> +       struct drm_gpu_scheduler 
>> *compute_sched_list[AMDGPU_MAX_COMPUTE_RINGS];
>>         unsigned                        num_compute_rings;
>>         struct amdgpu_irq_src           eop_irq;
>>         struct amdgpu_irq_src           priv_reg_irq;
>>
>>
>> Regards,
>>
>> Nirmoy
>>
>


More information about the amd-gfx mailing list