[PATCH 2/2] drm/amdgpu: Use drm_exec for seq64 bo lock
Arunpravin Paneer Selvam
Arunpravin.PaneerSelvam at amd.com
Wed Nov 1 16:26:02 UTC 2023
Replace seq64 bo lock sequences with drm_exec.
Signed-off-by: Alex Deucher <alexander.deucher at amd.com>
---
drivers/gpu/drm/amd/amdgpu/amdgpu_seq64.c | 73 ++++++++++-------------
1 file changed, 33 insertions(+), 40 deletions(-)
diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_seq64.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_seq64.c
index 63d8b68023be..810f7637096e 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_seq64.c
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_seq64.c
@@ -25,6 +25,8 @@
#include "amdgpu.h"
#include "amdgpu_seq64.h"
+#include <drm/drm_exec.h>
+
/**
* DOC: amdgpu_seq64
*
@@ -68,11 +70,8 @@ static inline u64 amdgpu_seq64_get_va_base(struct amdgpu_device *adev)
int amdgpu_seq64_map(struct amdgpu_device *adev, struct amdgpu_vm *vm,
struct amdgpu_bo_va **bo_va)
{
- struct ttm_validate_buffer seq64_tv;
- struct amdgpu_bo_list_entry pd;
- struct ww_acquire_ctx ticket;
- struct list_head list;
struct amdgpu_bo *bo;
+ struct drm_exec exec;
u64 seq64_addr;
int r;
@@ -80,23 +79,20 @@ int amdgpu_seq64_map(struct amdgpu_device *adev, struct amdgpu_vm *vm,
if (!bo)
return -EINVAL;
- INIT_LIST_HEAD(&list);
- INIT_LIST_HEAD(&seq64_tv.head);
-
- seq64_tv.bo = &bo->tbo;
- seq64_tv.num_shared = 1;
-
- list_add(&seq64_tv.head, &list);
- amdgpu_vm_get_pd_bo(vm, &list, &pd);
-
- r = ttm_eu_reserve_buffers(&ticket, &list, true, NULL);
- if (r)
- return r;
+ drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT);
+ drm_exec_until_all_locked(&exec) {
+ r = amdgpu_vm_lock_pd(vm, &exec, 0);
+ if (likely(!r))
+ r = drm_exec_lock_obj(&exec, &bo->tbo.base);
+ drm_exec_retry_on_contention(&exec);
+ if (unlikely(r))
+ goto error;
+ }
*bo_va = amdgpu_vm_bo_add(adev, vm, bo);
if (!*bo_va) {
r = -ENOMEM;
- goto error_vm;
+ goto error;
}
seq64_addr = amdgpu_seq64_get_va_base(adev);
@@ -104,23 +100,19 @@ int amdgpu_seq64_map(struct amdgpu_device *adev, struct amdgpu_vm *vm,
AMDGPU_PTE_READABLE);
if (r) {
DRM_ERROR("failed to do bo_map on userq sem, err=%d\n", r);
- goto error_map;
+ amdgpu_vm_bo_del(adev, *bo_va);
+ goto error;
}
r = amdgpu_vm_bo_update(adev, *bo_va, false);
if (r) {
DRM_ERROR("failed to do vm_bo_update on userq sem\n");
- goto error_map;
+ amdgpu_vm_bo_del(adev, *bo_va);
+ goto error;
}
- ttm_eu_backoff_reservation(&ticket, &list);
-
- return 0;
-
-error_map:
- amdgpu_vm_bo_del(adev, *bo_va);
-error_vm:
- ttm_eu_backoff_reservation(&ticket, &list);
+error:
+ drm_exec_fini(&exec);
return r;
}
@@ -134,12 +126,10 @@ int amdgpu_seq64_map(struct amdgpu_device *adev, struct amdgpu_vm *vm,
*/
void amdgpu_seq64_unmap(struct amdgpu_device *adev, struct amdgpu_fpriv *fpriv)
{
- struct ttm_validate_buffer seq64_tv;
- struct amdgpu_bo_list_entry pd;
- struct ww_acquire_ctx ticket;
- struct list_head list;
struct amdgpu_vm *vm;
struct amdgpu_bo *bo;
+ struct drm_exec exec;
+ int r;
if (!fpriv->seq64_va)
return;
@@ -149,20 +139,23 @@ void amdgpu_seq64_unmap(struct amdgpu_device *adev, struct amdgpu_fpriv *fpriv)
return;
vm = &fpriv->vm;
- INIT_LIST_HEAD(&list);
- INIT_LIST_HEAD(&seq64_tv.head);
- seq64_tv.bo = &bo->tbo;
- seq64_tv.num_shared = 1;
-
- list_add(&seq64_tv.head, &list);
- amdgpu_vm_get_pd_bo(vm, &list, &pd);
+ drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT);
+ drm_exec_until_all_locked(&exec) {
+ r = amdgpu_vm_lock_pd(vm, &exec, 0);
+ if (likely(!r))
+ r = drm_exec_lock_obj(&exec, &bo->tbo.base);
+ drm_exec_retry_on_contention(&exec);
+ if (unlikely(r))
+ goto error;
+ }
- ttm_eu_reserve_buffers(&ticket, &list, false, NULL);
amdgpu_vm_bo_del(adev, fpriv->seq64_va);
- ttm_eu_backoff_reservation(&ticket, &list);
fpriv->seq64_va = NULL;
+
+error:
+ drm_exec_fini(&exec);
}
/**
--
2.25.1
More information about the amd-gfx
mailing list