[Intel-xe] [PATCH] drm/xe/vm: uncombine the combined_links.destroy

Matthew Brost matthew.brost at intel.com
Mon Aug 7 13:28:20 UTC 2023


On Mon, Aug 07, 2023 at 02:02:27PM +0100, Matthew Auld wrote:
> It looks like it's plausible for eviction to be happening as we are
> closing the vm, however the rebind_link and destroy links are in this
> case not mutually exclusive, since eviction only needs the vm dma-resv
> lock which the vm close drops when accessing the contested list. If we
> race with eviction here it can run rampant and corrupt the list since
> both the destroy and rebind links are the same list entry underneath.
> Simplest is to just split these entries into separate links.
> 

I think there is another problem here too, I think
vm->notifier.rebind_list could be corrupted too.

I think a better solution is to check the XE_VMA_DESTROYED bit in
xe_bo_trigger_rebind before adding to either the
vm->notifier.rebind_list or vm->rebind_list. Both checks should done
under the list locks.

Matt

> References: https://gitlab.freedesktop.org/drm/xe/kernel/-/issues/514
> Signed-off-by: Matthew Auld <matthew.auld at intel.com>
> Cc: Matthew Brost <matthew.brost at intel.com>
> ---
>  drivers/gpu/drm/xe/xe_vm.c       | 10 +++++-----
>  drivers/gpu/drm/xe/xe_vm_types.h | 11 ++++++-----
>  2 files changed, 11 insertions(+), 10 deletions(-)
> 
> diff --git a/drivers/gpu/drm/xe/xe_vm.c b/drivers/gpu/drm/xe/xe_vm.c
> index cb28dbc2bdbb..aa13b2bcda86 100644
> --- a/drivers/gpu/drm/xe/xe_vm.c
> +++ b/drivers/gpu/drm/xe/xe_vm.c
> @@ -891,6 +891,7 @@ static struct xe_vma *xe_vma_create(struct xe_vm *vm,
>  	}
>  
>  	INIT_LIST_HEAD(&vma->combined_links.rebind);
> +	INIT_LIST_HEAD(&vma->destroy_link);
>  	INIT_LIST_HEAD(&vma->notifier.rebind_link);
>  	INIT_LIST_HEAD(&vma->extobj.link);
>  
> @@ -1063,7 +1064,7 @@ static void xe_vma_destroy(struct xe_vma *vma, struct dma_fence *fence)
>  	struct xe_vm *vm = xe_vma_vm(vma);
>  
>  	lockdep_assert_held_write(&vm->lock);
> -	XE_WARN_ON(!list_empty(&vma->combined_links.destroy));
> +	XE_WARN_ON(!list_empty(&vma->destroy_link));
>  
>  	if (xe_vma_is_userptr(vma)) {
>  		XE_WARN_ON(!(vma->gpuva.flags & XE_VMA_DESTROYED));
> @@ -1449,7 +1450,7 @@ void xe_vm_close_and_put(struct xe_vm *vm)
>  			continue;
>  		}
>  
> -		list_move_tail(&vma->combined_links.destroy, &contested);
> +		list_add_tail(&vma->destroy_link, &contested);
>  	}
>  
>  	/*
> @@ -1477,9 +1478,8 @@ void xe_vm_close_and_put(struct xe_vm *vm)
>  	 * Since we hold a refcount to the bo, we can remove and free
>  	 * the members safely without locking.
>  	 */
> -	list_for_each_entry_safe(vma, next_vma, &contested,
> -				 combined_links.destroy) {
> -		list_del_init(&vma->combined_links.destroy);
> +	list_for_each_entry_safe(vma, next_vma, &contested, destroy_link) {
> +		list_del_init(&vma->destroy_link);
>  		xe_vma_destroy_unlocked(vma);
>  	}
>  
> diff --git a/drivers/gpu/drm/xe/xe_vm_types.h b/drivers/gpu/drm/xe/xe_vm_types.h
> index 3681a5ff588b..a01dde418803 100644
> --- a/drivers/gpu/drm/xe/xe_vm_types.h
> +++ b/drivers/gpu/drm/xe/xe_vm_types.h
> @@ -79,13 +79,14 @@ struct xe_vma {
>  		 * mutually exclusive execution with userptr.
>  		 */
>  		struct list_head rebind;
> -		/**
> -		 * @destroy: link to contested list when VM is being closed.
> -		 * Protected by vm->lock in write mode and vm's resv lock.
> -		 */
> -		struct list_head destroy;
>  	} combined_links;
>  
> +	/**
> +	 * @destroy_link: link to contested list when VM is being closed.
> +	 * Protected by vm->lock in write mode and vm's resv lock.
> +	 */
> +	struct list_head destroy_link;
> +
>  	union {
>  		/** @destroy_cb: callback to destroy VMA when unbind job is done */
>  		struct dma_fence_cb destroy_cb;
> -- 
> 2.41.0
> 


More information about the Intel-xe mailing list