[PATCH 15/16] drm/xe: Adjust to commit "drm/gpuvm: add an abstraction for a VM/BO combination"
Matthew Brost
matthew.brost at intel.com
Thu Dec 7 12:38:50 UTC 2023
On Thu, Dec 07, 2023 at 03:11:55PM +0100, Thomas Hellström wrote:
> Signed-off-by: Thomas Hellström <thomas.hellstrom at linux.intel.com>
> ---
> drivers/gpu/drm/xe/xe_bo.c | 11 +++++++----
> drivers/gpu/drm/xe/xe_vm.c | 33 ++++++++++++++++++++++++++-------
> 2 files changed, 33 insertions(+), 11 deletions(-)
>
> diff --git a/drivers/gpu/drm/xe/xe_bo.c b/drivers/gpu/drm/xe/xe_bo.c
> index ad9d8793db3e..7e25c8b7a01a 100644
> --- a/drivers/gpu/drm/xe/xe_bo.c
> +++ b/drivers/gpu/drm/xe/xe_bo.c
> @@ -470,6 +470,7 @@ static int xe_bo_trigger_rebind(struct xe_device *xe, struct xe_bo *bo,
> struct dma_fence *fence;
> struct drm_gpuva *gpuva;
> struct drm_gem_object *obj = &bo->ttm.base;
> + struct drm_gpuvm_bo *vm_bo;
> int ret = 0;
>
> dma_resv_assert_held(bo->ttm.base.resv);
> @@ -482,11 +483,12 @@ static int xe_bo_trigger_rebind(struct xe_device *xe, struct xe_bo *bo,
> dma_resv_iter_end(&cursor);
> }
>
> - drm_gem_for_each_gpuva(gpuva, obj) {
> - struct xe_vma *vma = gpuva_to_vma(gpuva);
> - struct xe_vm *vm = xe_vma_vm(vma);
> + drm_gem_for_each_gpuvm_bo(vm_bo, obj) {
> + drm_gpuvm_bo_for_each_va(gpuva, vm_bo) {
> + struct xe_vma *vma = gpuva_to_vma(gpuva);
> + struct xe_vm *vm = xe_vma_vm(vma);
>
> - trace_xe_vma_evict(vma);
> + trace_xe_vma_evict(vma);
>
> if (xe_vm_in_fault_mode(vm)) {
> /* Wait for pending binds / unbinds. */
> @@ -540,6 +542,7 @@ static int xe_bo_trigger_rebind(struct xe_device *xe, struct xe_bo *bo,
> if (vm_resv_locked)
> dma_resv_unlock(xe_vm_resv(vm));
> }
> + }
The indentation looks incorrect here.
Everything else looks correct.
Matt
> }
>
> return ret;
> diff --git a/drivers/gpu/drm/xe/xe_vm.c b/drivers/gpu/drm/xe/xe_vm.c
> index 6c8a84026af4..f71285e8ef10 100644
> --- a/drivers/gpu/drm/xe/xe_vm.c
> +++ b/drivers/gpu/drm/xe/xe_vm.c
> @@ -911,12 +911,21 @@ static struct xe_vma *xe_vma_create(struct xe_vm *vm,
> vma->pat_index = pat_index;
>
> if (bo) {
> + struct drm_gpuvm_bo *vm_bo;
> +
> xe_bo_assert_held(bo);
>
> + vm_bo = drm_gpuvm_bo_obtain(vma->gpuva.vm, &bo->ttm.base);
> + if (IS_ERR(vm_bo)) {
> + kfree(vma);
> + return ERR_CAST(vm_bo);
> + }
> +
> drm_gem_object_get(&bo->ttm.base);
> vma->gpuva.gem.obj = &bo->ttm.base;
> vma->gpuva.gem.offset = bo_offset_or_userptr;
> - drm_gpuva_link(&vma->gpuva);
> + drm_gpuva_link(&vma->gpuva, vm_bo);
> + drm_gpuvm_bo_put(vm_bo);
> } else /* userptr or null */ {
> if (!is_null) {
> u64 size = end - start + 1;
> @@ -998,16 +1007,19 @@ static struct xe_vma *
> bo_has_vm_references_locked(struct xe_bo *bo, struct xe_vm *vm,
> struct xe_vma *ignore)
> {
> - struct drm_gpuva *gpuva;
> + struct drm_gpuvm_bo *vm_bo;
> + struct drm_gpuva *va;
> struct drm_gem_object *obj = &bo->ttm.base;
>
> xe_bo_assert_held(bo);
>
> - drm_gem_for_each_gpuva(gpuva, obj) {
> - struct xe_vma *vma = gpuva_to_vma(gpuva);
> + drm_gem_for_each_gpuvm_bo(vm_bo, obj) {
> + drm_gpuvm_bo_for_each_va(va, vm_bo) {
> + struct xe_vma *vma = gpuva_to_vma(va);
>
> - if (vma != ignore && xe_vma_vm(vma) == vm)
> - return vma;
> + if (vma != ignore && xe_vma_vm(vma) == vm)
> + return vma;
> + }
> }
>
> return NULL;
> @@ -2156,6 +2168,7 @@ vm_bind_ioctl_ops_create(struct xe_vm *vm, struct xe_bo *bo,
> struct drm_gpuva_ops *ops;
> struct drm_gpuva_op *__op;
> struct xe_vma_op *op;
> + struct drm_gpuvm_bo *vm_bo;
> int err;
>
> lockdep_assert_held_write(&vm->lock);
> @@ -2183,7 +2196,13 @@ vm_bind_ioctl_ops_create(struct xe_vm *vm, struct xe_bo *bo,
> err = xe_bo_lock(bo, true);
> if (err)
> return ERR_PTR(err);
> - ops = drm_gpuvm_gem_unmap_ops_create(&vm->gpuvm, obj);
> +
> + vm_bo = drm_gpuvm_bo_find(&vm->gpuvm, obj);
> + if (!vm_bo)
> + break;
> +
> + ops = drm_gpuvm_bo_unmap_ops_create(vm_bo);
> + drm_gpuvm_bo_put(vm_bo);
> xe_bo_unlock(bo);
> break;
> default:
> --
> 2.42.0
>
More information about the Intel-xe
mailing list