[PATCH 2/3] drm/amdgpu: sync to KFD fences before clearing PTEs
Christian König
ckoenig.leichtzumerken at gmail.com
Thu Aug 29 09:48:38 UTC 2024
Am 29.08.24 um 00:40 schrieb Felix Kuehling:
>
> On 2024-08-22 05:07, Christian König wrote:
>> Am 21.08.24 um 22:01 schrieb Felix Kuehling:
>>> On 2024-08-21 08:03, Christian König wrote:
>>>> This patch tries to solve the basic problem we also need to sync to
>>>> the KFD fences of the BO because otherwise it can be that we clear
>>>> PTEs while the KFD queues are still running.
>>>
>>> This is going to trigger a lot of phantom KFD evictions and will
>>> tank performance. It's probably not what you intended.
>>
>> I tried to avoid that by only waiting for the KFD fence only in the
>> particular situation that we can't lock the cleared BO because of
>> contention.
>
> OK. It's hard to make out where you're adding that call with the small
> context in the patch. As far as I can tell it's in the "if (clear ||
> !bo)" branch. The "clear" case is as you mention, only used when the
> BO cannot be locked. The !bo case is PRT?
Yes, exactly that.
>
> Contention would happen, if this runs concurrently with a
> restore-from-eviction, in which case we're already on a slow path and
> another eviction doesn't matter (as long as we're not getting into a
> live-lock situation). Or if a KFD BO is in the middle of being mapped
> or unmapped by another thread, which should be unlikely. So maybe this
> won't have a huge impact in practice. It's worth a try.
>
> The patch is
>
> Acked-by: Felix Kuehling <felix.kuehling at amd.com>
Let's see if this works or not in practice.
Thanks,
Christian.
>
>
>>
>> The only short term alternative I can see is to lock all BOs during
>> CS and that is a) a really large rework and b) will most likely hurt
>> performance.
>>
>> Then there is the alternative to lock the VM during BO eviction, but
>> that means we need to wait on using the drm_exec object inside TTM as
>> well. So that won't get this fixed in the next halve year or so.
>>
>> Regards,
>> Christian.
>>
>>>
>>> Regards,
>>> Felix
>>>
>>>
>>>>
>>>> Signed-off-by: Christian König <christian.koenig at amd.com>
>>>> ---
>>>> drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c | 30
>>>> ++++++++++++++++++++++++
>>>> drivers/gpu/drm/amd/amdgpu/amdgpu_sync.h | 1 +
>>>> drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c | 6 +++++
>>>> 3 files changed, 37 insertions(+)
>>>>
>>>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
>>>> b/drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
>>>> index bdf1ef825d89..c586ab4c911b 100644
>>>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
>>>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
>>>> @@ -260,6 +260,36 @@ int amdgpu_sync_resv(struct amdgpu_device
>>>> *adev, struct amdgpu_sync *sync,
>>>> return 0;
>>>> }
>>>> +/**
>>>> + * amdgpu_sync_kfd - sync to KFD fences
>>>> + *
>>>> + * @sync: sync object to add KFD fences to
>>>> + * @resv: reservation object with KFD fences
>>>> + *
>>>> + * Extract all KFD fences and add them to the sync object.
>>>> + */
>>>> +int amdgpu_sync_kfd(struct amdgpu_sync *sync, struct dma_resv *resv)
>>>> +{
>>>> + struct dma_resv_iter cursor;
>>>> + struct dma_fence *f;
>>>> + int r = 0;
>>>> +
>>>> + dma_resv_iter_begin(&cursor, resv, DMA_RESV_USAGE_BOOKKEEP);
>>>> + dma_resv_for_each_fence_unlocked(&cursor, f) {
>>>> + void *fence_owner = amdgpu_sync_get_owner(f);
>>>> +
>>>> + if (fence_owner != AMDGPU_FENCE_OWNER_KFD)
>>>> + continue;
>>>> +
>>>> + r = amdgpu_sync_fence(sync, f);
>>>> + if (r)
>>>> + break;
>>>> + }
>>>> + dma_resv_iter_end(&cursor);
>>>> +
>>>> + return r;
>>>> +}
>>>> +
>>>> /* Free the entry back to the slab */
>>>> static void amdgpu_sync_entry_free(struct amdgpu_sync_entry *e)
>>>> {
>>>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_sync.h
>>>> b/drivers/gpu/drm/amd/amdgpu/amdgpu_sync.h
>>>> index cf1e9e858efd..e3272dce798d 100644
>>>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_sync.h
>>>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_sync.h
>>>> @@ -51,6 +51,7 @@ int amdgpu_sync_fence(struct amdgpu_sync *sync,
>>>> struct dma_fence *f);
>>>> int amdgpu_sync_resv(struct amdgpu_device *adev, struct
>>>> amdgpu_sync *sync,
>>>> struct dma_resv *resv, enum amdgpu_sync_mode mode,
>>>> void *owner);
>>>> +int amdgpu_sync_kfd(struct amdgpu_sync *sync, struct dma_resv *resv);
>>>> struct dma_fence *amdgpu_sync_peek_fence(struct amdgpu_sync *sync,
>>>> struct amdgpu_ring *ring);
>>>> struct dma_fence *amdgpu_sync_get_fence(struct amdgpu_sync *sync);
>>>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
>>>> b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
>>>> index ba99d428610a..13d429b91327 100644
>>>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
>>>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
>>>> @@ -1168,6 +1168,12 @@ int amdgpu_vm_bo_update(struct amdgpu_device
>>>> *adev, struct amdgpu_bo_va *bo_va,
>>>> AMDGPU_SYNC_EQ_OWNER, vm);
>>>> if (r)
>>>> goto error_free;
>>>> + if (bo) {
>>>> + r = amdgpu_sync_kfd(&sync, bo->tbo.base.resv);
>>>> + if (r)
>>>> + goto error_free;
>>>> + }
>>>> +
>>>> } else {
>>>> struct drm_gem_object *obj = &bo->tbo.base;
>>
More information about the amd-gfx
mailing list