[PATCH] drm/radeon: stop using addr to check for BO move

Alex Deucher alexdeucher at gmail.com
Tue May 26 07:35:01 PDT 2015


On Tue, May 26, 2015 at 6:24 AM, Christian König
<deathsimple at vodafone.de> wrote:
> From: Christian König <christian.koenig at amd.com>
>
> It is theoretically possible that a swapped out BO gets the
> same GTT address, but different backing pages while being swapped in.
>
> Instead just use another VA state to note updated areas.
>
> Signed-off-by: Christian König <christian.koenig at amd.com>

Applied to my -next tree.

Thanks!

Alex

> ---
>  drivers/gpu/drm/radeon/radeon.h    |  4 ++-
>  drivers/gpu/drm/radeon/radeon_vm.c | 53 +++++++++++++++++++++-----------------
>  2 files changed, 32 insertions(+), 25 deletions(-)
>
> diff --git a/drivers/gpu/drm/radeon/radeon.h b/drivers/gpu/drm/radeon/radeon.h
> index 46eb0fa..ef7df51 100644
> --- a/drivers/gpu/drm/radeon/radeon.h
> +++ b/drivers/gpu/drm/radeon/radeon.h
> @@ -467,7 +467,6 @@ struct radeon_bo_va {
>         /* protected by bo being reserved */
>         struct list_head                bo_list;
>         uint32_t                        flags;
> -       uint64_t                        addr;
>         struct radeon_fence             *last_pt_update;
>         unsigned                        ref_count;
>
> @@ -941,6 +940,9 @@ struct radeon_vm {
>         /* BOs freed, but not yet updated in the PT */
>         struct list_head        freed;
>
> +       /* BOs cleared in the PT */
> +       struct list_head        cleared;
> +
>         /* contains the page directory */
>         struct radeon_bo        *page_directory;
>         unsigned                max_pde_used;
> diff --git a/drivers/gpu/drm/radeon/radeon_vm.c b/drivers/gpu/drm/radeon/radeon_vm.c
> index de42fc4..9739ded 100644
> --- a/drivers/gpu/drm/radeon/radeon_vm.c
> +++ b/drivers/gpu/drm/radeon/radeon_vm.c
> @@ -331,7 +331,6 @@ struct radeon_bo_va *radeon_vm_bo_add(struct radeon_device *rdev,
>         bo_va->it.start = 0;
>         bo_va->it.last = 0;
>         bo_va->flags = 0;
> -       bo_va->addr = 0;
>         bo_va->ref_count = 1;
>         INIT_LIST_HEAD(&bo_va->bo_list);
>         INIT_LIST_HEAD(&bo_va->vm_status);
> @@ -491,9 +490,11 @@ int radeon_vm_bo_set_addr(struct radeon_device *rdev,
>         }
>
>         if (bo_va->it.start || bo_va->it.last) {
> -               if (bo_va->addr) {
> +               spin_lock(&vm->status_lock);
> +               if (list_empty(&bo_va->vm_status)) {
>                         /* add a clone of the bo_va to clear the old address */
>                         struct radeon_bo_va *tmp;
> +                       spin_unlock(&vm->status_lock);
>                         tmp = kzalloc(sizeof(struct radeon_bo_va), GFP_KERNEL);
>                         if (!tmp) {
>                                 mutex_unlock(&vm->mutex);
> @@ -502,14 +503,11 @@ int radeon_vm_bo_set_addr(struct radeon_device *rdev,
>                         tmp->it.start = bo_va->it.start;
>                         tmp->it.last = bo_va->it.last;
>                         tmp->vm = vm;
> -                       tmp->addr = bo_va->addr;
>                         tmp->bo = radeon_bo_ref(bo_va->bo);
>                         spin_lock(&vm->status_lock);
>                         list_add(&tmp->vm_status, &vm->freed);
> -                       spin_unlock(&vm->status_lock);
> -
> -                       bo_va->addr = 0;
>                 }
> +               spin_unlock(&vm->status_lock);
>
>                 interval_tree_remove(&bo_va->it, &vm->va);
>                 bo_va->it.start = 0;
> @@ -520,10 +518,12 @@ int radeon_vm_bo_set_addr(struct radeon_device *rdev,
>                 bo_va->it.start = soffset;
>                 bo_va->it.last = eoffset - 1;
>                 interval_tree_insert(&bo_va->it, &vm->va);
> +               spin_lock(&vm->status_lock);
> +               list_add(&bo_va->vm_status, &vm->cleared);
> +               spin_unlock(&vm->status_lock);
>         }
>
>         bo_va->flags = flags;
> -       bo_va->addr = 0;
>
>         soffset >>= radeon_vm_block_size;
>         eoffset >>= radeon_vm_block_size;
> @@ -921,7 +921,16 @@ int radeon_vm_bo_update(struct radeon_device *rdev,
>         }
>
>         spin_lock(&vm->status_lock);
> -       list_del_init(&bo_va->vm_status);
> +       if (mem) {
> +               if (list_empty(&bo_va->vm_status)) {
> +                       spin_unlock(&vm->status_lock);
> +                       return 0;
> +               }
> +               list_del_init(&bo_va->vm_status);
> +       } else {
> +               list_del(&bo_va->vm_status);
> +               list_add(&bo_va->vm_status, &vm->cleared);
> +       }
>         spin_unlock(&vm->status_lock);
>
>         bo_va->flags &= ~RADEON_VM_PAGE_VALID;
> @@ -947,10 +956,6 @@ int radeon_vm_bo_update(struct radeon_device *rdev,
>                 addr = 0;
>         }
>
> -       if (addr == bo_va->addr)
> -               return 0;
> -       bo_va->addr = addr;
> -
>         trace_radeon_vm_bo_update(bo_va);
>
>         nptes = bo_va->it.last - bo_va->it.start + 1;
> @@ -1038,7 +1043,7 @@ int radeon_vm_clear_freed(struct radeon_device *rdev,
>                           struct radeon_vm *vm)
>  {
>         struct radeon_bo_va *bo_va;
> -       int r;
> +       int r = 0;
>
>         spin_lock(&vm->status_lock);
>         while (!list_empty(&vm->freed)) {
> @@ -1049,14 +1054,15 @@ int radeon_vm_clear_freed(struct radeon_device *rdev,
>                 r = radeon_vm_bo_update(rdev, bo_va, NULL);
>                 radeon_bo_unref(&bo_va->bo);
>                 radeon_fence_unref(&bo_va->last_pt_update);
> +               spin_lock(&vm->status_lock);
> +               list_del(&bo_va->vm_status);
>                 kfree(bo_va);
>                 if (r)
> -                       return r;
> +                       break;
>
> -               spin_lock(&vm->status_lock);
>         }
>         spin_unlock(&vm->status_lock);
> -       return 0;
> +       return r;
>
>  }
>
> @@ -1114,14 +1120,14 @@ void radeon_vm_bo_rmv(struct radeon_device *rdev,
>         mutex_lock(&vm->mutex);
>         if (bo_va->it.start || bo_va->it.last)
>                 interval_tree_remove(&bo_va->it, &vm->va);
> -       spin_lock(&vm->status_lock);
> -       list_del(&bo_va->vm_status);
>
> -       if (bo_va->addr) {
> +       spin_lock(&vm->status_lock);
> +       if (list_empty(&bo_va->vm_status)) {
>                 bo_va->bo = radeon_bo_ref(bo_va->bo);
>                 list_add(&bo_va->vm_status, &vm->freed);
>         } else {
>                 radeon_fence_unref(&bo_va->last_pt_update);
> +               list_del(&bo_va->vm_status);
>                 kfree(bo_va);
>         }
>         spin_unlock(&vm->status_lock);
> @@ -1144,12 +1150,10 @@ void radeon_vm_bo_invalidate(struct radeon_device *rdev,
>         struct radeon_bo_va *bo_va;
>
>         list_for_each_entry(bo_va, &bo->va, bo_list) {
> -               if (bo_va->addr) {
> -                       spin_lock(&bo_va->vm->status_lock);
> -                       list_del(&bo_va->vm_status);
> +               spin_lock(&bo_va->vm->status_lock);
> +               if (list_empty(&bo_va->vm_status))
>                         list_add(&bo_va->vm_status, &bo_va->vm->invalidated);
> -                       spin_unlock(&bo_va->vm->status_lock);
> -               }
> +               spin_unlock(&bo_va->vm->status_lock);
>         }
>  }
>
> @@ -1179,6 +1183,7 @@ int radeon_vm_init(struct radeon_device *rdev, struct radeon_vm *vm)
>         spin_lock_init(&vm->status_lock);
>         INIT_LIST_HEAD(&vm->invalidated);
>         INIT_LIST_HEAD(&vm->freed);
> +       INIT_LIST_HEAD(&vm->cleared);
>
>         pd_size = radeon_vm_directory_size(rdev);
>         pd_entries = radeon_vm_num_pdes(rdev);
> --
> 1.9.1
>


More information about the dri-devel mailing list