[PATCH 1/4] drm/amdgpu: add sched sync for amdgpu job
Christian König
deathsimple at vodafone.de
Wed May 10 09:21:17 UTC 2017
Am 10.05.2017 um 11:00 schrieb zhoucm1:
>
>
> On 2017年05月10日 16:50, Christian König wrote:
>> Am 10.05.2017 um 10:38 schrieb zhoucm1:
>>>
>>>
>>> On 2017年05月10日 16:26, Christian König wrote:
>>>> Am 10.05.2017 um 09:31 schrieb Chunming Zhou:
>>>>> this is an improvement for previous patch, the sched_sync is to
>>>>> store fence
>>>>> that could be skipped as scheduled, when job is executed, we
>>>>> didn't need
>>>>> pipeline_sync if all fences in sched_sync are signalled, otherwise
>>>>> insert
>>>>> pipeline_sync still.
>>>>>
>>>>> Change-Id: I26d3a2794272ba94b25753d4bf367326d12f6939
>>>>> Signed-off-by: Chunming Zhou <David1.Zhou at amd.com>
>>>>> ---
>>>>> drivers/gpu/drm/amd/amdgpu/amdgpu.h | 1 +
>>>>> drivers/gpu/drm/amd/amdgpu/amdgpu_ib.c | 7 ++++++-
>>>>> drivers/gpu/drm/amd/amdgpu/amdgpu_job.c | 5 ++++-
>>>>> 3 files changed, 11 insertions(+), 2 deletions(-)
>>>>>
>>>>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu.h
>>>>> b/drivers/gpu/drm/amd/amdgpu/amdgpu.h
>>>>> index 787acd7..ef018bf 100644
>>>>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu.h
>>>>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu.h
>>>>> @@ -1162,6 +1162,7 @@ struct amdgpu_job {
>>>>> struct amdgpu_vm *vm;
>>>>> struct amdgpu_ring *ring;
>>>>> struct amdgpu_sync sync;
>>>>> + struct amdgpu_sync sched_sync;
>>>>> struct amdgpu_ib *ibs;
>>>>> struct fence *fence; /* the hw fence */
>>>>> uint32_t preamble_status;
>>>>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_ib.c
>>>>> b/drivers/gpu/drm/amd/amdgpu/amdgpu_ib.c
>>>>> index 2c6624d..86ad507 100644
>>>>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_ib.c
>>>>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_ib.c
>>>>> @@ -121,6 +121,7 @@ int amdgpu_ib_schedule(struct amdgpu_ring
>>>>> *ring, unsigned num_ibs,
>>>>> {
>>>>> struct amdgpu_device *adev = ring->adev;
>>>>> struct amdgpu_ib *ib = &ibs[0];
>>>>> + struct fence *tmp;
>>>>> bool skip_preamble, need_ctx_switch;
>>>>> unsigned patch_offset = ~0;
>>>>> struct amdgpu_vm *vm;
>>>>> @@ -167,8 +168,12 @@ int amdgpu_ib_schedule(struct amdgpu_ring
>>>>> *ring, unsigned num_ibs,
>>>>> return r;
>>>>> }
>>>>> - if (ring->funcs->emit_pipeline_sync && job &&
>>>>> job->need_pipeline_sync)
>>>>> + if (ring->funcs->emit_pipeline_sync && job &&
>>>>> + (tmp = amdgpu_sync_get_fence(&job->sched_sync))) {
>>>>> + job->need_pipeline_sync = true;
>>>>> amdgpu_ring_emit_pipeline_sync(ring);
>>>>> + fence_put(tmp);
>>>>> + }
>>>>> if (vm) {
>>>>> amdgpu_ring_insert_nop(ring, extra_nop); /* prevent CE
>>>>> go too fast than DE */
>>>>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c
>>>>> b/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c
>>>>> index cfa97ab..fa0c8b1 100644
>>>>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c
>>>>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c
>>>>> @@ -60,6 +60,7 @@ int amdgpu_job_alloc(struct amdgpu_device *adev,
>>>>> unsigned num_ibs,
>>>>> (*job)->need_pipeline_sync = false;
>>>>> amdgpu_sync_create(&(*job)->sync);
>>>>> + amdgpu_sync_create(&(*job)->sched_sync);
>>>>> return 0;
>>>>> }
>>>>> @@ -98,6 +99,7 @@ static void amdgpu_job_free_cb(struct
>>>>> amd_sched_job *s_job)
>>>>> fence_put(job->fence);
>>>>> amdgpu_sync_free(&job->sync);
>>>>> + amdgpu_sync_free(&job->sched_sync);
>>>>> kfree(job);
>>>>> }
>>>>> @@ -107,6 +109,7 @@ void amdgpu_job_free(struct amdgpu_job *job)
>>>>> fence_put(job->fence);
>>>>> amdgpu_sync_free(&job->sync);
>>>>> + amdgpu_sync_free(&job->sched_sync);
>>>>> kfree(job);
>>>>> }
>>>>> @@ -154,7 +157,7 @@ static struct fence
>>>>> *amdgpu_job_dependency(struct amd_sched_job *sched_job)
>>>>> }
>>>>> if (amd_sched_dependency_optimized(fence,
>>>>> sched_job->s_entity))
>>>>> - job->need_pipeline_sync = true;
>>>>> + amdgpu_sync_fence(job->adev, &job->sched_sync, fence);
>>>>
>>>> This can result in an -ENOMEM
>>> will handle it.
>>>> and additional to that we only need to remember the last fence
>>>> optimized like this, not all of them.
>>>>
>>>> So just keep the last one found here in job->sched_fence instead.
>>> I guess this isn't enough.
>>> The dependency is not in order when calling, so the last one is not
>>> always the last scheduled fence.
>>> And they could be sched fence not hw fence, although they are
>>> handled by same hw ring, but the sched fence context isn't same.
>>> so we still need sched_sync here, right?
>>
>> No, amdgpu_job_dependency is only called again when the returned
>> fence is signaled (or scheduled on the same ring).
> Let use give an example for it:
> Assume job->sync has two fences(fenceA and fenceB) which could be
> scheduled. fenceA is from entity1, fenceB is from entity2, but both
> for gfx engine, but fenceA could be submitted to hw ring behind fenceB.
> the order in job->sync list is: others---->fenceA---->fenceB--->others.
> when calling amdgpu_job_dependency, fenceA will be checked first, and
> then fenceB.
>
> If following your proposal, we only store fenceB, but fenceA is the
> later. Which isn't expected.
Ah! Indeed, I didn't realized that the dependent fence could have
already been scheduled.
Mhm, how are we going to handle the out of memory situation then? Sine
we are inside a kernel thread we are not supposed to fail at this point.
Regards,
Christian.
>
>
> Regards,
> David Zhou
>>
>> So when this is called and you find that you need to wait for another
>> fence the order is guaranteed.
>>
>> Regards,
>> Christian.
>>
>>>
>>> Regards,
>>> David zhou
>>>>
>>>> Regards,
>>>> Christian.
>>>>
>>>>> return fence;
>>>>> }
>>>>
>>>>
>>>
>>
>
More information about the amd-gfx
mailing list