[PATCH 4/4] drm/amdgpu: rework synchronization of VM updates v3
Christian König
ckoenig.leichtzumerken at gmail.com
Thu Jan 23 16:20:17 UTC 2020
Am 23.01.20 um 16:39 schrieb William Lewis:
> Was the change to true from the only use of intr in
> amdgpu_bo_sync_wait_resv intentional? If so, would it not make sense to
> remove the argument from the function signature while the API is changing?
That's indeed a copy & paste typo. Thanks for pointing it out.
Christian.
>
> On 1/23/20 8:21 AM, Christian König wrote:
>> If provided we only sync to the BOs reservation
>> object and no longer to the root PD.
>>
>> v2: update comment, cleanup amdgpu_bo_sync_wait_resv
>> v3: use correct reservation object while clearing
>>
>> Signed-off-by: Christian König <christian.koenig at amd.com>
>> ---
>> drivers/gpu/drm/amd/amdgpu/amdgpu_object.c | 37 ++++++++++++++++++++------
>> drivers/gpu/drm/amd/amdgpu/amdgpu_object.h | 3 +++
>> drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c | 7 -----
>> drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c | 41 +++++++++++++++++------------
>> drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h | 4 +--
>> drivers/gpu/drm/amd/amdgpu/amdgpu_vm_cpu.c | 22 ++++------------
>> drivers/gpu/drm/amd/amdgpu/amdgpu_vm_sdma.c | 15 +++--------
>> 7 files changed, 67 insertions(+), 62 deletions(-)
>>
>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_object.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
>> index 46c76e2e1281..c70bbdda078c 100644
>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
>> @@ -1403,30 +1403,51 @@ void amdgpu_bo_fence(struct amdgpu_bo *bo, struct dma_fence *fence,
>> }
>>
>> /**
>> - * amdgpu_sync_wait_resv - Wait for BO reservation fences
>> + * amdgpu_bo_sync_wait_resv - Wait for BO reservation fences
>> *
>> - * @bo: buffer object
>> + * @adev: amdgpu device pointer
>> + * @resv: reservation object to sync to
>> + * @sync_mode: synchronization mode
>> * @owner: fence owner
>> * @intr: Whether the wait is interruptible
>> *
>> + * Extract the fences from the reservation object and waits for them to finish.
>> + *
>> * Returns:
>> * 0 on success, errno otherwise.
>> */
>> -int amdgpu_bo_sync_wait(struct amdgpu_bo *bo, void *owner, bool intr)
>> +int amdgpu_bo_sync_wait_resv(struct amdgpu_device *adev, struct dma_resv *resv,
>> + enum amdgpu_sync_mode sync_mode, void *owner,
>> + bool intr)
>> {
>> - struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
>> struct amdgpu_sync sync;
>> int r;
>>
>> amdgpu_sync_create(&sync);
>> - amdgpu_sync_resv(adev, &sync, bo->tbo.base.resv,
>> - AMDGPU_SYNC_NE_OWNER, owner);
>> - r = amdgpu_sync_wait(&sync, intr);
>> + amdgpu_sync_resv(adev, &sync, resv, sync_mode, owner);
>> + r = amdgpu_sync_wait(&sync, true);
>> amdgpu_sync_free(&sync);
>> -
>> return r;
>> }
>>
>> +/**
>> + * amdgpu_bo_sync_wait - Wrapper for amdgpu_bo_sync_wait_resv
>> + * @bo: buffer object to wait for
>> + * @owner: fence owner
>> + * @intr: Whether the wait is interruptible
>> + *
>> + * Wrapper to wait for fences in a BO.
>> + * Returns:
>> + * 0 on success, errno otherwise.
>> + */
>> +int amdgpu_bo_sync_wait(struct amdgpu_bo *bo, void *owner, bool intr)
>> +{
>> + struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
>> +
>> + return amdgpu_bo_sync_wait_resv(adev, bo->tbo.base.resv,
>> + AMDGPU_SYNC_NE_OWNER, owner, intr);
>> +}
>> +
>> /**
>> * amdgpu_bo_gpu_offset - return GPU offset of bo
>> * @bo: amdgpu object for which we query the offset
>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_object.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_object.h
>> index 2eeafc77c9c1..96d805889e8d 100644
>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_object.h
>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_object.h
>> @@ -283,6 +283,9 @@ void amdgpu_bo_release_notify(struct ttm_buffer_object *bo);
>> int amdgpu_bo_fault_reserve_notify(struct ttm_buffer_object *bo);
>> void amdgpu_bo_fence(struct amdgpu_bo *bo, struct dma_fence *fence,
>> bool shared);
>> +int amdgpu_bo_sync_wait_resv(struct amdgpu_device *adev, struct dma_resv *resv,
>> + enum amdgpu_sync_mode sync_mode, void *owner,
>> + bool intr);
>> int amdgpu_bo_sync_wait(struct amdgpu_bo *bo, void *owner, bool intr);
>> u64 amdgpu_bo_gpu_offset(struct amdgpu_bo *bo);
>> int amdgpu_bo_validate(struct amdgpu_bo *bo);
>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
>> index 9f42032676da..b86392253696 100644
>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
>> @@ -249,13 +249,6 @@ int amdgpu_sync_resv(struct amdgpu_device *adev, struct amdgpu_sync *sync,
>> owner != AMDGPU_FENCE_OWNER_UNDEFINED)
>> continue;
>>
>> - /* VM updates only sync with moves but not with user
>> - * command submissions or KFD evictions fences
>> - */
>> - if (fence_owner != AMDGPU_FENCE_OWNER_UNDEFINED &&
>> - owner == AMDGPU_FENCE_OWNER_VM)
>> - continue;
>> -
>> /* Ignore fences depending on the sync mode */
>> switch (mode) {
>> case AMDGPU_SYNC_ALWAYS:
>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
>> index 0f79c17118bf..c268aa14381e 100644
>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
>> @@ -797,7 +797,7 @@ static int amdgpu_vm_clear_bo(struct amdgpu_device *adev,
>> params.vm = vm;
>> params.direct = direct;
>>
>> - r = vm->update_funcs->prepare(¶ms, AMDGPU_FENCE_OWNER_KFD, NULL);
>> + r = vm->update_funcs->prepare(¶ms, NULL, AMDGPU_SYNC_EXPLICIT);
>> if (r)
>> return r;
>>
>> @@ -1293,7 +1293,7 @@ int amdgpu_vm_update_pdes(struct amdgpu_device *adev,
>> params.vm = vm;
>> params.direct = direct;
>>
>> - r = vm->update_funcs->prepare(¶ms, AMDGPU_FENCE_OWNER_VM, NULL);
>> + r = vm->update_funcs->prepare(¶ms, NULL, AMDGPU_SYNC_EXPLICIT);
>> if (r)
>> return r;
>>
>> @@ -1552,7 +1552,7 @@ static int amdgpu_vm_update_ptes(struct amdgpu_vm_update_params *params,
>> * @adev: amdgpu_device pointer
>> * @vm: requested vm
>> * @direct: direct submission in a page fault
>> - * @exclusive: fence we need to sync to
>> + * @resv: fences we need to sync to
>> * @start: start of mapped range
>> * @last: last mapped entry
>> * @flags: flags for the entries
>> @@ -1567,14 +1567,14 @@ static int amdgpu_vm_update_ptes(struct amdgpu_vm_update_params *params,
>> */
>> static int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
>> struct amdgpu_vm *vm, bool direct,
>> - struct dma_fence *exclusive,
>> + struct dma_resv *resv,
>> uint64_t start, uint64_t last,
>> uint64_t flags, uint64_t addr,
>> dma_addr_t *pages_addr,
>> struct dma_fence **fence)
>> {
>> struct amdgpu_vm_update_params params;
>> - void *owner = AMDGPU_FENCE_OWNER_VM;
>> + enum amdgpu_sync_mode sync_mode;
>> int r;
>>
>> memset(¶ms, 0, sizeof(params));
>> @@ -1583,9 +1583,13 @@ static int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
>> params.direct = direct;
>> params.pages_addr = pages_addr;
>>
>> - /* sync to everything except eviction fences on unmapping */
>> + /* Implicitly sync to command submissions in the same VM before
>> + * unmapping. Sync to moving fences before mapping.
>> + */
>> if (!(flags & AMDGPU_PTE_VALID))
>> - owner = AMDGPU_FENCE_OWNER_KFD;
>> + sync_mode = AMDGPU_SYNC_EQ_OWNER;
>> + else
>> + sync_mode = AMDGPU_SYNC_EXPLICIT;
>>
>> amdgpu_vm_eviction_lock(vm);
>> if (vm->evicting) {
>> @@ -1593,7 +1597,7 @@ static int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
>> goto error_unlock;
>> }
>>
>> - r = vm->update_funcs->prepare(¶ms, owner, exclusive);
>> + r = vm->update_funcs->prepare(¶ms, resv, sync_mode);
>> if (r)
>> goto error_unlock;
>>
>> @@ -1612,7 +1616,7 @@ static int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
>> * amdgpu_vm_bo_split_mapping - split a mapping into smaller chunks
>> *
>> * @adev: amdgpu_device pointer
>> - * @exclusive: fence we need to sync to
>> + * @resv: fences we need to sync to
>> * @pages_addr: DMA addresses to use for mapping
>> * @vm: requested vm
>> * @mapping: mapped range and flags to use for the update
>> @@ -1628,7 +1632,7 @@ static int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
>> * 0 for success, -EINVAL for failure.
>> */
>> static int amdgpu_vm_bo_split_mapping(struct amdgpu_device *adev,
>> - struct dma_fence *exclusive,
>> + struct dma_resv *resv,
>> dma_addr_t *pages_addr,
>> struct amdgpu_vm *vm,
>> struct amdgpu_bo_va_mapping *mapping,
>> @@ -1704,7 +1708,7 @@ static int amdgpu_vm_bo_split_mapping(struct amdgpu_device *adev,
>> }
>>
>> last = min((uint64_t)mapping->last, start + max_entries - 1);
>> - r = amdgpu_vm_bo_update_mapping(adev, vm, false, exclusive,
>> + r = amdgpu_vm_bo_update_mapping(adev, vm, false, resv,
>> start, last, flags, addr,
>> dma_addr, fence);
>> if (r)
>> @@ -1743,7 +1747,8 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev, struct amdgpu_bo_va *bo_va,
>> dma_addr_t *pages_addr = NULL;
>> struct ttm_mem_reg *mem;
>> struct drm_mm_node *nodes;
>> - struct dma_fence *exclusive, **last_update;
>> + struct dma_fence **last_update;
>> + struct dma_resv *resv;
>> uint64_t flags;
>> struct amdgpu_device *bo_adev = adev;
>> int r;
>> @@ -1751,7 +1756,7 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev, struct amdgpu_bo_va *bo_va,
>> if (clear || !bo) {
>> mem = NULL;
>> nodes = NULL;
>> - exclusive = NULL;
>> + resv = vm->root.base.bo->tbo.base.resv;
>> } else {
>> struct ttm_dma_tt *ttm;
>>
>> @@ -1761,7 +1766,7 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev, struct amdgpu_bo_va *bo_va,
>> ttm = container_of(bo->tbo.ttm, struct ttm_dma_tt, ttm);
>> pages_addr = ttm->dma_address;
>> }
>> - exclusive = bo->tbo.moving;
>> + resv = bo->tbo.base.resv;
>> }
>>
>> if (bo) {
>> @@ -1775,7 +1780,8 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev, struct amdgpu_bo_va *bo_va,
>> flags = 0x0;
>> }
>>
>> - if (clear || (bo && bo->tbo.base.resv == vm->root.base.bo->tbo.base.resv))
>> + if (clear || (bo && bo->tbo.base.resv ==
>> + vm->root.base.bo->tbo.base.resv))
>> last_update = &vm->last_update;
>> else
>> last_update = &bo_va->last_pt_update;
>> @@ -1789,7 +1795,7 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev, struct amdgpu_bo_va *bo_va,
>> }
>>
>> list_for_each_entry(mapping, &bo_va->invalids, list) {
>> - r = amdgpu_vm_bo_split_mapping(adev, exclusive, pages_addr, vm,
>> + r = amdgpu_vm_bo_split_mapping(adev, resv, pages_addr, vm,
>> mapping, flags, bo_adev, nodes,
>> last_update);
>> if (r)
>> @@ -1984,6 +1990,7 @@ int amdgpu_vm_clear_freed(struct amdgpu_device *adev,
>> struct amdgpu_vm *vm,
>> struct dma_fence **fence)
>> {
>> + struct dma_resv *resv = vm->root.base.bo->tbo.base.resv;
>> struct amdgpu_bo_va_mapping *mapping;
>> uint64_t init_pte_value = 0;
>> struct dma_fence *f = NULL;
>> @@ -1998,7 +2005,7 @@ int amdgpu_vm_clear_freed(struct amdgpu_device *adev,
>> mapping->start < AMDGPU_GMC_HOLE_START)
>> init_pte_value = AMDGPU_PTE_DEFAULT_ATC;
>>
>> - r = amdgpu_vm_bo_update_mapping(adev, vm, false, NULL,
>> + r = amdgpu_vm_bo_update_mapping(adev, vm, false, resv,
>> mapping->start, mapping->last,
>> init_pte_value, 0, NULL, &f);
>> amdgpu_vm_free_mapping(adev, vm, mapping, f);
>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
>> index c21a36bebc0c..b5705fcfc935 100644
>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
>> @@ -230,8 +230,8 @@ struct amdgpu_vm_update_params {
>>
>> struct amdgpu_vm_update_funcs {
>> int (*map_table)(struct amdgpu_bo *bo);
>> - int (*prepare)(struct amdgpu_vm_update_params *p, void * owner,
>> - struct dma_fence *exclusive);
>> + int (*prepare)(struct amdgpu_vm_update_params *p, struct dma_resv *resv,
>> + enum amdgpu_sync_mode sync_mode);
>> int (*update)(struct amdgpu_vm_update_params *p,
>> struct amdgpu_bo *bo, uint64_t pe, uint64_t addr,
>> unsigned count, uint32_t incr, uint64_t flags);
>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm_cpu.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm_cpu.c
>> index 68b013be3837..e38516304070 100644
>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm_cpu.c
>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm_cpu.c
>> @@ -44,26 +44,14 @@ static int amdgpu_vm_cpu_map_table(struct amdgpu_bo *table)
>> * Returns:
>> * Negativ errno, 0 for success.
>> */
>> -static int amdgpu_vm_cpu_prepare(struct amdgpu_vm_update_params *p, void *owner,
>> - struct dma_fence *exclusive)
>> +static int amdgpu_vm_cpu_prepare(struct amdgpu_vm_update_params *p,
>> + struct dma_resv *resv,
>> + enum amdgpu_sync_mode sync_mode)
>> {
>> - int r;
>> -
>> - /* Wait for any BO move to be completed */
>> - if (exclusive) {
>> - r = dma_fence_wait(exclusive, true);
>> - if (unlikely(r))
>> - return r;
>> - }
>> -
>> - /* Don't wait for submissions during page fault */
>> - if (p->direct)
>> + if (!resv)
>> return 0;
>>
>> - /* Wait for PT BOs to be idle. PTs share the same resv. object
>> - * as the root PD BO
>> - */
>> - return amdgpu_bo_sync_wait(p->vm->root.base.bo, owner, true);
>> + return amdgpu_bo_sync_wait_resv(p->adev, resv, sync_mode, p->vm, true);
>> }
>>
>> /**
>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm_sdma.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm_sdma.c
>> index ab6481751763..4cc7881f438c 100644
>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm_sdma.c
>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm_sdma.c
>> @@ -58,9 +58,9 @@ static int amdgpu_vm_sdma_map_table(struct amdgpu_bo *table)
>> * Negativ errno, 0 for success.
>> */
>> static int amdgpu_vm_sdma_prepare(struct amdgpu_vm_update_params *p,
>> - void *owner, struct dma_fence *exclusive)
>> + struct dma_resv *resv,
>> + enum amdgpu_sync_mode sync_mode)
>> {
>> - struct amdgpu_bo *root = p->vm->root.base.bo;
>> unsigned int ndw = AMDGPU_VM_SDMA_MIN_NUM_DW;
>> int r;
>>
>> @@ -70,17 +70,10 @@ static int amdgpu_vm_sdma_prepare(struct amdgpu_vm_update_params *p,
>>
>> p->num_dw_left = ndw;
>>
>> - /* Wait for moves to be completed */
>> - r = amdgpu_sync_fence(&p->job->sync, exclusive, false);
>> - if (r)
>> - return r;
>> -
>> - /* Don't wait for any submissions during page fault handling */
>> - if (p->direct)
>> + if (!resv)
>> return 0;
>>
>> - return amdgpu_sync_resv(p->adev, &p->job->sync, root->tbo.base.resv,
>> - AMDGPU_SYNC_NE_OWNER, owner);
>> + return amdgpu_sync_resv(p->adev, &p->job->sync, resv, sync_mode, p->vm);
>> }
>>
>> /**
> _______________________________________________
> amd-gfx mailing list
> amd-gfx at lists.freedesktop.org
> https://lists.freedesktop.org/mailman/listinfo/amd-gfx
More information about the amd-gfx
mailing list