[PATCH 1/6] drm/amdgpu: rework VM state machine lock handling v2
Zhang, Jerry (Junwei)
Jerry.Zhang at amd.com
Fri May 18 09:55:31 UTC 2018
2, 3, 4, 5 are
Reviewed-by: Junwei Zhang <Jerry.Zhang at amd.com>
Patch 1:
could you show the reserving VM?
Patch 6:
I could read that code, but not sure the purpose.
Jerry
On 05/17/2018 05:49 PM, Christian König wrote:
> Only the moved state needs a separate spin lock protection. All other
> states are protected by reserving the VM anyway.
>
> v2: fix some more incorrect cases
>
> Signed-off-by: Christian König <christian.koenig at amd.com>
> ---
> drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c | 66 +++++++++++-----------------------
> drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h | 4 +--
> 2 files changed, 21 insertions(+), 49 deletions(-)
>
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
> index 1a8f4e0dd023..f0deedcaf1c9 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
> @@ -119,9 +119,7 @@ static void amdgpu_vm_bo_base_init(struct amdgpu_vm_bo_base *base,
> * is currently evicted. add the bo to the evicted list to make sure it
> * is validated on next vm use to avoid fault.
> * */
> - spin_lock(&vm->status_lock);
> list_move_tail(&base->vm_status, &vm->evicted);
> - spin_unlock(&vm->status_lock);
> }
>
> /**
> @@ -228,7 +226,6 @@ int amdgpu_vm_validate_pt_bos(struct amdgpu_device *adev, struct amdgpu_vm *vm,
> struct ttm_bo_global *glob = adev->mman.bdev.glob;
> int r;
>
> - spin_lock(&vm->status_lock);
> while (!list_empty(&vm->evicted)) {
> struct amdgpu_vm_bo_base *bo_base;
> struct amdgpu_bo *bo;
> @@ -236,10 +233,8 @@ int amdgpu_vm_validate_pt_bos(struct amdgpu_device *adev, struct amdgpu_vm *vm,
> bo_base = list_first_entry(&vm->evicted,
> struct amdgpu_vm_bo_base,
> vm_status);
> - spin_unlock(&vm->status_lock);
>
> bo = bo_base->bo;
> - BUG_ON(!bo);
> if (bo->parent) {
> r = validate(param, bo);
> if (r)
> @@ -259,13 +254,14 @@ int amdgpu_vm_validate_pt_bos(struct amdgpu_device *adev, struct amdgpu_vm *vm,
> return r;
> }
>
> - spin_lock(&vm->status_lock);
> - if (bo->tbo.type != ttm_bo_type_kernel)
> + if (bo->tbo.type != ttm_bo_type_kernel) {
> + spin_lock(&vm->moved_lock);
> list_move(&bo_base->vm_status, &vm->moved);
> - else
> + spin_unlock(&vm->moved_lock);
> + } else {
> list_move(&bo_base->vm_status, &vm->relocated);
> + }
> }
> - spin_unlock(&vm->status_lock);
>
> return 0;
> }
> @@ -279,13 +275,7 @@ int amdgpu_vm_validate_pt_bos(struct amdgpu_device *adev, struct amdgpu_vm *vm,
> */
> bool amdgpu_vm_ready(struct amdgpu_vm *vm)
> {
> - bool ready;
> -
> - spin_lock(&vm->status_lock);
> - ready = list_empty(&vm->evicted);
> - spin_unlock(&vm->status_lock);
> -
> - return ready;
> + return list_empty(&vm->evicted);
> }
>
> /**
> @@ -477,9 +467,7 @@ static int amdgpu_vm_alloc_levels(struct amdgpu_device *adev,
> pt->parent = amdgpu_bo_ref(parent->base.bo);
>
> amdgpu_vm_bo_base_init(&entry->base, vm, pt);
> - spin_lock(&vm->status_lock);
> list_move(&entry->base.vm_status, &vm->relocated);
> - spin_unlock(&vm->status_lock);
> }
>
> if (level < AMDGPU_VM_PTB) {
> @@ -926,10 +914,8 @@ static void amdgpu_vm_invalidate_level(struct amdgpu_device *adev,
> if (!entry->base.bo)
> continue;
>
> - spin_lock(&vm->status_lock);
> if (list_empty(&entry->base.vm_status))
> list_add(&entry->base.vm_status, &vm->relocated);
> - spin_unlock(&vm->status_lock);
> amdgpu_vm_invalidate_level(adev, vm, entry, level + 1);
> }
> }
> @@ -974,7 +960,6 @@ int amdgpu_vm_update_directories(struct amdgpu_device *adev,
> params.func = amdgpu_vm_do_set_ptes;
> }
>
> - spin_lock(&vm->status_lock);
> while (!list_empty(&vm->relocated)) {
> struct amdgpu_vm_bo_base *bo_base, *parent;
> struct amdgpu_vm_pt *pt, *entry;
> @@ -984,13 +969,10 @@ int amdgpu_vm_update_directories(struct amdgpu_device *adev,
> struct amdgpu_vm_bo_base,
> vm_status);
> list_del_init(&bo_base->vm_status);
> - spin_unlock(&vm->status_lock);
>
> bo = bo_base->bo->parent;
> - if (!bo) {
> - spin_lock(&vm->status_lock);
> + if (!bo)
> continue;
> - }
>
> parent = list_first_entry(&bo->va, struct amdgpu_vm_bo_base,
> bo_list);
> @@ -999,12 +981,10 @@ int amdgpu_vm_update_directories(struct amdgpu_device *adev,
>
> amdgpu_vm_update_pde(¶ms, vm, pt, entry);
>
> - spin_lock(&vm->status_lock);
> if (!vm->use_cpu_for_update &&
> (ndw - params.ib->length_dw) < 32)
> break;
> }
> - spin_unlock(&vm->status_lock);
>
> if (vm->use_cpu_for_update) {
> /* Flush HDP */
> @@ -1107,9 +1087,7 @@ static void amdgpu_vm_handle_huge_pages(struct amdgpu_pte_update_params *p,
> if (entry->huge) {
> /* Add the entry to the relocated list to update it. */
> entry->huge = false;
> - spin_lock(&p->vm->status_lock);
> list_move(&entry->base.vm_status, &p->vm->relocated);
> - spin_unlock(&p->vm->status_lock);
> }
> return;
> }
> @@ -1588,8 +1566,9 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev,
> amdgpu_asic_flush_hdp(adev, NULL);
> }
>
> - spin_lock(&vm->status_lock);
> + spin_lock(&vm->moved_lock);
> list_del_init(&bo_va->base.vm_status);
> + spin_unlock(&vm->moved_lock);
>
> /* If the BO is not in its preferred location add it back to
> * the evicted list so that it gets validated again on the
> @@ -1599,7 +1578,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev,
> !(bo->preferred_domains &
> amdgpu_mem_type_to_domain(bo->tbo.mem.mem_type)))
> list_add_tail(&bo_va->base.vm_status, &vm->evicted);
> - spin_unlock(&vm->status_lock);
>
> list_splice_init(&bo_va->invalids, &bo_va->valids);
> bo_va->cleared = clear;
> @@ -1811,14 +1789,14 @@ int amdgpu_vm_handle_moved(struct amdgpu_device *adev,
> bool clear;
> int r = 0;
>
> - spin_lock(&vm->status_lock);
> + spin_lock(&vm->moved_lock);
> while (!list_empty(&vm->moved)) {
> struct amdgpu_bo_va *bo_va;
> struct reservation_object *resv;
>
> bo_va = list_first_entry(&vm->moved,
> struct amdgpu_bo_va, base.vm_status);
> - spin_unlock(&vm->status_lock);
> + spin_unlock(&vm->moved_lock);
>
> resv = bo_va->base.bo->tbo.resv;
>
> @@ -1839,9 +1817,9 @@ int amdgpu_vm_handle_moved(struct amdgpu_device *adev,
> if (!clear && resv != vm->root.base.bo->tbo.resv)
> reservation_object_unlock(resv);
>
> - spin_lock(&vm->status_lock);
> + spin_lock(&vm->moved_lock);
> }
> - spin_unlock(&vm->status_lock);
> + spin_unlock(&vm->moved_lock);
>
> return r;
> }
> @@ -1903,10 +1881,10 @@ static void amdgpu_vm_bo_insert_map(struct amdgpu_device *adev,
> amdgpu_vm_prt_get(adev);
>
> if (bo && bo->tbo.resv == vm->root.base.bo->tbo.resv) {
> - spin_lock(&vm->status_lock);
> + spin_lock(&vm->moved_lock);
> if (list_empty(&bo_va->base.vm_status))
> list_add(&bo_va->base.vm_status, &vm->moved);
> - spin_unlock(&vm->status_lock);
> + spin_unlock(&vm->moved_lock);
> }
> trace_amdgpu_vm_bo_map(bo_va, mapping);
> }
> @@ -2216,9 +2194,9 @@ void amdgpu_vm_bo_rmv(struct amdgpu_device *adev,
>
> list_del(&bo_va->base.bo_list);
>
> - spin_lock(&vm->status_lock);
> + spin_lock(&vm->moved_lock);
> list_del(&bo_va->base.vm_status);
> - spin_unlock(&vm->status_lock);
> + spin_unlock(&vm->moved_lock);
>
> list_for_each_entry_safe(mapping, next, &bo_va->valids, list) {
> list_del(&mapping->list);
> @@ -2261,28 +2239,24 @@ void amdgpu_vm_bo_invalidate(struct amdgpu_device *adev,
>
> bo_base->moved = true;
> if (evicted && bo->tbo.resv == vm->root.base.bo->tbo.resv) {
> - spin_lock(&bo_base->vm->status_lock);
> if (bo->tbo.type == ttm_bo_type_kernel)
> list_move(&bo_base->vm_status, &vm->evicted);
> else
> list_move_tail(&bo_base->vm_status,
> &vm->evicted);
> - spin_unlock(&bo_base->vm->status_lock);
> continue;
> }
>
> if (bo->tbo.type == ttm_bo_type_kernel) {
> - spin_lock(&bo_base->vm->status_lock);
> if (list_empty(&bo_base->vm_status))
> list_add(&bo_base->vm_status, &vm->relocated);
> - spin_unlock(&bo_base->vm->status_lock);
> continue;
> }
>
> - spin_lock(&bo_base->vm->status_lock);
> + spin_lock(&bo_base->vm->moved_lock);
> if (list_empty(&bo_base->vm_status))
> list_add(&bo_base->vm_status, &vm->moved);
> - spin_unlock(&bo_base->vm->status_lock);
> + spin_unlock(&bo_base->vm->moved_lock);
> }
> }
>
> @@ -2391,9 +2365,9 @@ int amdgpu_vm_init(struct amdgpu_device *adev, struct amdgpu_vm *vm,
> vm->va = RB_ROOT_CACHED;
> for (i = 0; i < AMDGPU_MAX_VMHUBS; i++)
> vm->reserved_vmid[i] = NULL;
> - spin_lock_init(&vm->status_lock);
> INIT_LIST_HEAD(&vm->evicted);
> INIT_LIST_HEAD(&vm->relocated);
> + spin_lock_init(&vm->moved_lock);
> INIT_LIST_HEAD(&vm->moved);
> INIT_LIST_HEAD(&vm->freed);
>
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
> index d6827083572a..0196b9a782f2 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
> @@ -168,9 +168,6 @@ struct amdgpu_vm {
> /* tree of virtual addresses mapped */
> struct rb_root_cached va;
>
> - /* protecting invalidated */
> - spinlock_t status_lock;
> -
> /* BOs who needs a validation */
> struct list_head evicted;
>
> @@ -179,6 +176,7 @@ struct amdgpu_vm {
>
> /* BOs moved, but not yet updated in the PT */
> struct list_head moved;
> + spinlock_t moved_lock;
>
> /* BO mappings freed, but not yet updated in the PT */
> struct list_head freed;
>
More information about the amd-gfx
mailing list