[PATCH] drm/radeon: stop using addr to check for BO move
Alex Deucher
alexdeucher at gmail.com
Tue May 26 07:35:01 PDT 2015
On Tue, May 26, 2015 at 6:24 AM, Christian König
<deathsimple at vodafone.de> wrote:
> From: Christian König <christian.koenig at amd.com>
>
> It is theoretically possible that a swapped out BO gets the
> same GTT address, but different backing pages while being swapped in.
>
> Instead just use another VA state to note updated areas.
>
> Signed-off-by: Christian König <christian.koenig at amd.com>
Applied to my -next tree.
Thanks!
Alex
> ---
> drivers/gpu/drm/radeon/radeon.h | 4 ++-
> drivers/gpu/drm/radeon/radeon_vm.c | 53 +++++++++++++++++++++-----------------
> 2 files changed, 32 insertions(+), 25 deletions(-)
>
> diff --git a/drivers/gpu/drm/radeon/radeon.h b/drivers/gpu/drm/radeon/radeon.h
> index 46eb0fa..ef7df51 100644
> --- a/drivers/gpu/drm/radeon/radeon.h
> +++ b/drivers/gpu/drm/radeon/radeon.h
> @@ -467,7 +467,6 @@ struct radeon_bo_va {
> /* protected by bo being reserved */
> struct list_head bo_list;
> uint32_t flags;
> - uint64_t addr;
> struct radeon_fence *last_pt_update;
> unsigned ref_count;
>
> @@ -941,6 +940,9 @@ struct radeon_vm {
> /* BOs freed, but not yet updated in the PT */
> struct list_head freed;
>
> + /* BOs cleared in the PT */
> + struct list_head cleared;
> +
> /* contains the page directory */
> struct radeon_bo *page_directory;
> unsigned max_pde_used;
> diff --git a/drivers/gpu/drm/radeon/radeon_vm.c b/drivers/gpu/drm/radeon/radeon_vm.c
> index de42fc4..9739ded 100644
> --- a/drivers/gpu/drm/radeon/radeon_vm.c
> +++ b/drivers/gpu/drm/radeon/radeon_vm.c
> @@ -331,7 +331,6 @@ struct radeon_bo_va *radeon_vm_bo_add(struct radeon_device *rdev,
> bo_va->it.start = 0;
> bo_va->it.last = 0;
> bo_va->flags = 0;
> - bo_va->addr = 0;
> bo_va->ref_count = 1;
> INIT_LIST_HEAD(&bo_va->bo_list);
> INIT_LIST_HEAD(&bo_va->vm_status);
> @@ -491,9 +490,11 @@ int radeon_vm_bo_set_addr(struct radeon_device *rdev,
> }
>
> if (bo_va->it.start || bo_va->it.last) {
> - if (bo_va->addr) {
> + spin_lock(&vm->status_lock);
> + if (list_empty(&bo_va->vm_status)) {
> /* add a clone of the bo_va to clear the old address */
> struct radeon_bo_va *tmp;
> + spin_unlock(&vm->status_lock);
> tmp = kzalloc(sizeof(struct radeon_bo_va), GFP_KERNEL);
> if (!tmp) {
> mutex_unlock(&vm->mutex);
> @@ -502,14 +503,11 @@ int radeon_vm_bo_set_addr(struct radeon_device *rdev,
> tmp->it.start = bo_va->it.start;
> tmp->it.last = bo_va->it.last;
> tmp->vm = vm;
> - tmp->addr = bo_va->addr;
> tmp->bo = radeon_bo_ref(bo_va->bo);
> spin_lock(&vm->status_lock);
> list_add(&tmp->vm_status, &vm->freed);
> - spin_unlock(&vm->status_lock);
> -
> - bo_va->addr = 0;
> }
> + spin_unlock(&vm->status_lock);
>
> interval_tree_remove(&bo_va->it, &vm->va);
> bo_va->it.start = 0;
> @@ -520,10 +518,12 @@ int radeon_vm_bo_set_addr(struct radeon_device *rdev,
> bo_va->it.start = soffset;
> bo_va->it.last = eoffset - 1;
> interval_tree_insert(&bo_va->it, &vm->va);
> + spin_lock(&vm->status_lock);
> + list_add(&bo_va->vm_status, &vm->cleared);
> + spin_unlock(&vm->status_lock);
> }
>
> bo_va->flags = flags;
> - bo_va->addr = 0;
>
> soffset >>= radeon_vm_block_size;
> eoffset >>= radeon_vm_block_size;
> @@ -921,7 +921,16 @@ int radeon_vm_bo_update(struct radeon_device *rdev,
> }
>
> spin_lock(&vm->status_lock);
> - list_del_init(&bo_va->vm_status);
> + if (mem) {
> + if (list_empty(&bo_va->vm_status)) {
> + spin_unlock(&vm->status_lock);
> + return 0;
> + }
> + list_del_init(&bo_va->vm_status);
> + } else {
> + list_del(&bo_va->vm_status);
> + list_add(&bo_va->vm_status, &vm->cleared);
> + }
> spin_unlock(&vm->status_lock);
>
> bo_va->flags &= ~RADEON_VM_PAGE_VALID;
> @@ -947,10 +956,6 @@ int radeon_vm_bo_update(struct radeon_device *rdev,
> addr = 0;
> }
>
> - if (addr == bo_va->addr)
> - return 0;
> - bo_va->addr = addr;
> -
> trace_radeon_vm_bo_update(bo_va);
>
> nptes = bo_va->it.last - bo_va->it.start + 1;
> @@ -1038,7 +1043,7 @@ int radeon_vm_clear_freed(struct radeon_device *rdev,
> struct radeon_vm *vm)
> {
> struct radeon_bo_va *bo_va;
> - int r;
> + int r = 0;
>
> spin_lock(&vm->status_lock);
> while (!list_empty(&vm->freed)) {
> @@ -1049,14 +1054,15 @@ int radeon_vm_clear_freed(struct radeon_device *rdev,
> r = radeon_vm_bo_update(rdev, bo_va, NULL);
> radeon_bo_unref(&bo_va->bo);
> radeon_fence_unref(&bo_va->last_pt_update);
> + spin_lock(&vm->status_lock);
> + list_del(&bo_va->vm_status);
> kfree(bo_va);
> if (r)
> - return r;
> + break;
>
> - spin_lock(&vm->status_lock);
> }
> spin_unlock(&vm->status_lock);
> - return 0;
> + return r;
>
> }
>
> @@ -1114,14 +1120,14 @@ void radeon_vm_bo_rmv(struct radeon_device *rdev,
> mutex_lock(&vm->mutex);
> if (bo_va->it.start || bo_va->it.last)
> interval_tree_remove(&bo_va->it, &vm->va);
> - spin_lock(&vm->status_lock);
> - list_del(&bo_va->vm_status);
>
> - if (bo_va->addr) {
> + spin_lock(&vm->status_lock);
> + if (list_empty(&bo_va->vm_status)) {
> bo_va->bo = radeon_bo_ref(bo_va->bo);
> list_add(&bo_va->vm_status, &vm->freed);
> } else {
> radeon_fence_unref(&bo_va->last_pt_update);
> + list_del(&bo_va->vm_status);
> kfree(bo_va);
> }
> spin_unlock(&vm->status_lock);
> @@ -1144,12 +1150,10 @@ void radeon_vm_bo_invalidate(struct radeon_device *rdev,
> struct radeon_bo_va *bo_va;
>
> list_for_each_entry(bo_va, &bo->va, bo_list) {
> - if (bo_va->addr) {
> - spin_lock(&bo_va->vm->status_lock);
> - list_del(&bo_va->vm_status);
> + spin_lock(&bo_va->vm->status_lock);
> + if (list_empty(&bo_va->vm_status))
> list_add(&bo_va->vm_status, &bo_va->vm->invalidated);
> - spin_unlock(&bo_va->vm->status_lock);
> - }
> + spin_unlock(&bo_va->vm->status_lock);
> }
> }
>
> @@ -1179,6 +1183,7 @@ int radeon_vm_init(struct radeon_device *rdev, struct radeon_vm *vm)
> spin_lock_init(&vm->status_lock);
> INIT_LIST_HEAD(&vm->invalidated);
> INIT_LIST_HEAD(&vm->freed);
> + INIT_LIST_HEAD(&vm->cleared);
>
> pd_size = radeon_vm_directory_size(rdev);
> pd_entries = radeon_vm_num_pdes(rdev);
> --
> 1.9.1
>
More information about the dri-devel
mailing list