[PATCH] drm/amdgpu: Enable P2P dmabuf over XGMI

Christian König ckoenig.leichtzumerken at gmail.com
Thu Aug 6 09:09:43 UTC 2020


Am 06.08.20 um 11:04 schrieb Arunpravin:
> Access the exported P2P dmabuf over XGMI, if available.
> Otherwise, fall back to the existing PCIe method.
>
> Signed-off-by: Arunpravin <apaneers at amd.com>

Reviewed-by: Christian König <christian.koenig at amd.com>

> ---
>   drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c | 34 +++++++++++++++++++++
>   drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.h |  2 ++
>   drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c      | 19 ++++++++++--
>   3 files changed, 52 insertions(+), 3 deletions(-)
>
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c
> index ffeb20f11c07..589d008f91df 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c
> @@ -35,6 +35,7 @@
>   #include "amdgpu_display.h"
>   #include "amdgpu_gem.h"
>   #include "amdgpu_dma_buf.h"
> +#include "amdgpu_xgmi.h"
>   #include <drm/amdgpu_drm.h>
>   #include <linux/dma-buf.h>
>   #include <linux/dma-fence-array.h>
> @@ -560,3 +561,36 @@ struct drm_gem_object *amdgpu_gem_prime_import(struct drm_device *dev,
>   	obj->import_attach = attach;
>   	return obj;
>   }
> +
> +/**
> + * amdgpu_dmabuf_is_xgmi_accessible - Check if xgmi available for P2P transfer
> + *
> + * @adev: amdgpu_device pointer of the importer
> + * @bo: amdgpu buffer object
> + *
> + * Returns:
> + * True if dmabuf accessible over xgmi, false otherwise.
> + */
> +bool amdgpu_dmabuf_is_xgmi_accessible(struct amdgpu_device *adev,
> +				      struct amdgpu_bo *bo)
> +{
> +	struct drm_gem_object *obj = &bo->tbo.base;
> +	struct drm_gem_object *gobj;
> +
> +	if (obj->import_attach) {
> +		struct dma_buf *dma_buf = obj->import_attach->dmabuf;
> +
> +		if (dma_buf->ops != &amdgpu_dmabuf_ops)
> +			/* No XGMI with non AMD GPUs */
> +			return false;
> +
> +		gobj = dma_buf->priv;
> +		bo = gem_to_amdgpu_bo(gobj);
> +	}
> +
> +	if (amdgpu_xgmi_same_hive(adev, amdgpu_ttm_adev(bo->tbo.bdev)) &&
> +			(bo->preferred_domains & AMDGPU_GEM_DOMAIN_VRAM))
> +		return true;
> +
> +	return false;
> +}
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.h
> index ec447a7b6b28..2c5c84a06bb9 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.h
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.h
> @@ -29,6 +29,8 @@ struct dma_buf *amdgpu_gem_prime_export(struct drm_gem_object *gobj,
>   					int flags);
>   struct drm_gem_object *amdgpu_gem_prime_import(struct drm_device *dev,
>   					    struct dma_buf *dma_buf);
> +bool amdgpu_dmabuf_is_xgmi_accessible(struct amdgpu_device *adev,
> +				      struct amdgpu_bo *bo);
>   void *amdgpu_gem_prime_vmap(struct drm_gem_object *obj);
>   void amdgpu_gem_prime_vunmap(struct drm_gem_object *obj, void *vaddr);
>   int amdgpu_gem_prime_mmap(struct drm_gem_object *obj,
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
> index 71e005cf2952..771c27478bb1 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
> @@ -28,6 +28,7 @@
>   #include <linux/dma-fence-array.h>
>   #include <linux/interval_tree_generic.h>
>   #include <linux/idr.h>
> +#include <linux/dma-buf.h>
>   
>   #include <drm/amdgpu_drm.h>
>   #include "amdgpu.h"
> @@ -35,6 +36,7 @@
>   #include "amdgpu_amdkfd.h"
>   #include "amdgpu_gmc.h"
>   #include "amdgpu_xgmi.h"
> +#include "amdgpu_dma_buf.h"
>   
>   /**
>    * DOC: GPUVM
> @@ -1778,15 +1780,24 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev, struct amdgpu_bo_va *bo_va,
>   		nodes = NULL;
>   		resv = vm->root.base.bo->tbo.base.resv;
>   	} else {
> +		struct drm_gem_object *obj = &bo->tbo.base;
>   		struct ttm_dma_tt *ttm;
>   
> +		resv = bo->tbo.base.resv;
> +		if (obj->import_attach && bo_va->is_xgmi) {
> +			struct dma_buf *dma_buf = obj->import_attach->dmabuf;
> +			struct drm_gem_object *gobj = dma_buf->priv;
> +			struct amdgpu_bo *abo = gem_to_amdgpu_bo(gobj);
> +
> +			if (abo->tbo.mem.mem_type == TTM_PL_VRAM)
> +				bo = gem_to_amdgpu_bo(gobj);
> +		}
>   		mem = &bo->tbo.mem;
>   		nodes = mem->mm_node;
>   		if (mem->mem_type == TTM_PL_TT) {
>   			ttm = container_of(bo->tbo.ttm, struct ttm_dma_tt, ttm);
>   			pages_addr = ttm->dma_address;
>   		}
> -		resv = bo->tbo.base.resv;
>   	}
>   
>   	if (bo) {
> @@ -2132,8 +2143,10 @@ struct amdgpu_bo_va *amdgpu_vm_bo_add(struct amdgpu_device *adev,
>   	INIT_LIST_HEAD(&bo_va->valids);
>   	INIT_LIST_HEAD(&bo_va->invalids);
>   
> -	if (bo && amdgpu_xgmi_same_hive(adev, amdgpu_ttm_adev(bo->tbo.bdev)) &&
> -	    (bo->preferred_domains & AMDGPU_GEM_DOMAIN_VRAM)) {
> +	if (!bo)
> +		return bo_va;
> +
> +	if (amdgpu_dmabuf_is_xgmi_accessible(adev, bo)) {
>   		bo_va->is_xgmi = true;
>   		/* Power up XGMI if it can be potentially used */
>   		amdgpu_xgmi_set_pstate(adev, AMDGPU_XGMI_PSTATE_MAX_VEGA20);



More information about the amd-gfx mailing list