[Intel-xe] [PATCH v2] drm/xe: use variable instead of multiple function calls

Ruhl, Michael J michael.j.ruhl at intel.com
Tue Oct 31 20:09:24 UTC 2023


>-----Original Message-----
>From: Intel-xe <intel-xe-bounces at lists.freedesktop.org> On Behalf Of Dani
>Liberman
>Sent: Tuesday, October 31, 2023 1:12 PM
>To: intel-xe at lists.freedesktop.org
>Subject: [Intel-xe] [PATCH v2] drm/xe: use variable instead of multiple function
>calls
>
>Using function calls with negative logic was a bit confusing,
>positive logic is more readable.
>
>v2:
> - Update commit message
>
>Cc: Matthew Brost <matthew.brost at intel.com>
>Cc: Ville Syrjala <ville.syrjala at linux.intel.com>
>Signed-off-by: Dani Liberman <dliberman at habana.ai>
>Reviewed-by: Matthew Brost <matthew.brost at intel.com>
>---
> drivers/gpu/drm/xe/xe_exec.c | 15 ++++++++-------
> 1 file changed, 8 insertions(+), 7 deletions(-)
>
>diff --git a/drivers/gpu/drm/xe/xe_exec.c b/drivers/gpu/drm/xe/xe_exec.c
>index 28e84a0bbeb0..2de6c2c05078 100644
>--- a/drivers/gpu/drm/xe/xe_exec.c
>+++ b/drivers/gpu/drm/xe/xe_exec.c
>@@ -145,7 +145,7 @@ int xe_exec_ioctl(struct drm_device *dev, void *data,
>struct drm_file *file)
> 	struct xe_sched_job *job;
> 	struct dma_fence *rebind_fence;
> 	struct xe_vm *vm;
>-	bool write_locked;
>+	bool write_locked, vm_with_dma_fences;
> 	ktime_t end = 0;
> 	int err = 0;
>
>@@ -196,8 +196,9 @@ int xe_exec_ioctl(struct drm_device *dev, void *data,
>struct drm_file *file)
> 		}
> 	}
>
>+	vm_with_dma_fences = !xe_vm_no_dma_fences(vm);
> retry:

Are you 100% certain that this value will never change on retry?

M

>-	if (!xe_vm_no_dma_fences(vm) && xe_vm_userptr_check_repin(vm)) {
>+	if (vm_with_dma_fences && xe_vm_userptr_check_repin(vm)) {
> 		err = down_write_killable(&vm->lock);
> 		write_locked = true;
> 	} else {
>@@ -279,7 +280,7 @@ int xe_exec_ioctl(struct drm_device *dev, void *data,
>struct drm_file *file)
> 	}
>
> 	/* Wait behind munmap style rebinds */
>-	if (!xe_vm_no_dma_fences(vm)) {
>+	if (vm_with_dma_fences) {
> 		err = drm_sched_job_add_resv_dependencies(&job->drm,
> 							  &vm->resv,
>
>DMA_RESV_USAGE_KERNEL);
>@@ -292,7 +293,7 @@ int xe_exec_ioctl(struct drm_device *dev, void *data,
>struct drm_file *file)
> 	if (err)
> 		goto err_put_job;
>
>-	if (!xe_vm_no_dma_fences(vm)) {
>+	if (vm_with_dma_fences) {
> 		err = down_read_interruptible(&vm->userptr.notifier_lock);
> 		if (err)
> 			goto err_put_job;
>@@ -307,7 +308,7 @@ int xe_exec_ioctl(struct drm_device *dev, void *data,
>struct drm_file *file)
> 	 * the job and let the DRM scheduler / backend clean up the job.
> 	 */
> 	xe_sched_job_arm(job);
>-	if (!xe_vm_no_dma_fences(vm)) {
>+	if (vm_with_dma_fences) {
> 		/* Block userptr invalidations / BO eviction */
> 		dma_resv_add_fence(&vm->resv,
> 				   &job->drm.s_fence->finished,
>@@ -330,14 +331,14 @@ int xe_exec_ioctl(struct drm_device *dev, void
>*data, struct drm_file *file)
> 	xe_sched_job_push(job);
> 	xe_vm_reactivate_rebind(vm);
>
>-	if (!err && !xe_vm_no_dma_fences(vm)) {
>+	if (!err && vm_with_dma_fences) {
> 		spin_lock(&xe->ttm.lru_lock);
> 		ttm_lru_bulk_move_tail(&vm->lru_bulk_move);
> 		spin_unlock(&xe->ttm.lru_lock);
> 	}
>
> err_repin:
>-	if (!xe_vm_no_dma_fences(vm))
>+	if (vm_with_dma_fences)
> 		up_read(&vm->userptr.notifier_lock);
> err_put_job:
> 	if (err)
>--
>2.34.1



More information about the Intel-xe mailing list