[PATCH 3/5] drm/i915/dmabuf: Add LMEM knowledge to dmabuf map handler

Daniel Vetter daniel at ffwll.ch
Tue Apr 28 15:01:40 UTC 2020


On Wed, Apr 22, 2020 at 05:25:17PM -0400, Michael J. Ruhl wrote:
> LMEM backed buffer objects do not have struct page information.
> Instead the dma_address of the struct sg is used to store the
> LMEM address information (relative to the device, this is not
> the CPU physical address).
> 
> The dmabuf map handler requires pages to do a dma_map_xx.
> 
> Add new mapping/unmapping functions, based on the LMEM usage
> of the dma_address to allow LMEM backed buffer objects to be
> mapped.
> 
> Before mapping check the peer2peer distance to verify that P2P
> dma can occur.

So this is supposed to check the importer's allow_peer2peer flag, and that
one is supposed to require the implementation of ->move_notify. Which
requires a pile of locking changes to align with dma_resv.

By not doing all that you avoid the lockdep splats, but you're also
breaking the peer2peer dma-buf contract big time :-)

I think this needs more work, or I need better glasses in case I'm not
spotting where this is all done.
-Daniel

> 
> Signed-off-by: Michael J. Ruhl <michael.j.ruhl at intel.com>
> ---
>  drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c | 82 ++++++++++++++++++++--
>  1 file changed, 76 insertions(+), 6 deletions(-)
> 
> diff --git a/drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c b/drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c
> index 7ea4abb6a896..402c989cc23d 100644
> --- a/drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c
> +++ b/drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c
> @@ -7,6 +7,7 @@
>  #include <linux/dma-buf.h>
>  #include <linux/highmem.h>
>  #include <linux/dma-resv.h>
> +#include <linux/pci-p2pdma.h>
>  #include <linux/scatterlist.h>
>  
>  #include "i915_drv.h"
> @@ -18,6 +19,67 @@ static struct drm_i915_gem_object *dma_buf_to_obj(struct dma_buf *buf)
>  	return to_intel_bo(buf->priv);
>  }
>  
> +static void dmabuf_unmap_addr(struct device *dev, struct scatterlist *sgl,
> +			      int nents, enum dma_data_direction dir)
> +{
> +	struct scatterlist *sg;
> +	int i;
> +
> +	for_each_sg(sgl, sg, nents, i)
> +		dma_unmap_resource(dev, sg_dma_address(sg), sg_dma_len(sg),
> +				   dir, 0);
> +}
> +
> +/**
> + * dmabuf_map_addr - Update LMEM address to a physical address and map the
> + * resource.
> + * @dev: valid device
> + * @obj: valid i915 GEM object
> + * @sg: scatter list to appy mapping to
> + * @nents: number of entries in the scatter list
> + * @dir: DMA direction
> + *
> + * The dma_address of the scatter list is the LMEM "address".  From this the
> + * actual physical address can be determined.
> + *
> + * Return of 0 means error.
> + *
> + */
> +static int dmabuf_map_addr(struct device *dev, struct drm_i915_gem_object *obj,
> +			   struct scatterlist *sgl, int nents,
> +			   enum dma_data_direction dir)
> +{
> +	struct scatterlist *sg;
> +	phys_addr_t addr;
> +	int distance;
> +	int i;
> +
> +	distance = pci_p2pdma_distance_many(obj->base.dev->pdev, &dev, 1,
> +					    true);
> +	if (distance < 0) {
> +		pr_info("%s: from: %s  to: %s  distance: %d\n", __func__,
> +			pci_name(obj->base.dev->pdev), dev_name(dev),
> +			distance);
> +		return 0;
> +	}
> +
> +	for_each_sg(sgl, sg, nents, i) {
> +		addr = sg_dma_address(sg) + obj->mm.region->io_start;
> +
> +		sg->dma_address = dma_map_resource(dev, addr, sg->length, dir,
> +						   0);
> +		if (dma_mapping_error(dev, sg->dma_address))
> +			goto unmap;
> +		sg->dma_length = sg->length;
> +	}
> +
> +	return nents;
> +
> +unmap:
> +	dmabuf_unmap_addr(dev, sgl, i, dir);
> +	return 0;
> +}
> +
>  static struct sg_table *i915_gem_map_dma_buf(struct dma_buf_attachment *attach,
>  					     enum dma_data_direction dir)
>  {
> @@ -44,12 +106,17 @@ static struct sg_table *i915_gem_map_dma_buf(struct dma_buf_attachment *attach,
>  	dst = sgt->sgl;
>  	for_each_sg(obj->mm.pages->sgl, src, obj->mm.pages->nents, i) {
>  		sg_set_page(dst, sg_page(src), src->length, 0);
> +		sg_dma_address(dst) = sg_dma_address(src);
>  		dst = sg_next(dst);
>  	}
>  
> -	if (!dma_map_sg_attrs(attach->dev,
> -			      sgt->sgl, sgt->nents, dir,
> -			      DMA_ATTR_SKIP_CPU_SYNC)) {
> +	if (i915_gem_object_has_struct_page(obj))
> +		ret = dma_map_sg_attrs(attach->dev, sgt->sgl, sgt->nents, dir,
> +				       DMA_ATTR_SKIP_CPU_SYNC);
> +	else
> +		ret = dmabuf_map_addr(attach->dev, obj, sgt->sgl, sgt->nents,
> +				      dir);
> +	if (!ret) {
>  		ret = -ENOMEM;
>  		goto err_free_sg;
>  	}
> @@ -72,9 +139,12 @@ static void i915_gem_unmap_dma_buf(struct dma_buf_attachment *attach,
>  {
>  	struct drm_i915_gem_object *obj = dma_buf_to_obj(attach->dmabuf);
>  
> -	dma_unmap_sg_attrs(attach->dev,
> -			   sgt->sgl, sgt->nents, dir,
> -			   DMA_ATTR_SKIP_CPU_SYNC);
> +	if (i915_gem_object_has_struct_page(obj))
> +		dma_unmap_sg_attrs(attach->dev, sgt->sgl, sgt->nents, dir,
> +				   DMA_ATTR_SKIP_CPU_SYNC);
> +	else
> +		dmabuf_unmap_addr(attach->dev, sgt->sgl, sgt->nents, dir);
> +
>  	sg_free_table(sgt);
>  	kfree(sgt);
>  
> -- 
> 2.21.0
> 

-- 
Daniel Vetter
Software Engineer, Intel Corporation
http://blog.ffwll.ch


More information about the dri-devel mailing list