[PATCH 6/9] drm/exynos: switch to new buffer allocation

Joonyoung Shim jy0922.shim at samsung.com
Tue Oct 13 00:00:51 PDT 2015


The buffer allocation using DMA mapping API can't support non-continuous
buffer on non-iommu and cachable buffer, so switch to new buffer
allocation using drm_gem_get/put_pages() and doesn't use DMA mapping API
for mmap except allocation of physically continuous buffer on non-iommu.

Signed-off-by: Joonyoung Shim <jy0922.shim at samsung.com>
---
 drivers/gpu/drm/exynos/exynos_drm_gem.c | 90 +++++++++++----------------------
 1 file changed, 29 insertions(+), 61 deletions(-)

diff --git a/drivers/gpu/drm/exynos/exynos_drm_gem.c b/drivers/gpu/drm/exynos/exynos_drm_gem.c
index d982d46b04da..163d113df1ab 100644
--- a/drivers/gpu/drm/exynos/exynos_drm_gem.c
+++ b/drivers/gpu/drm/exynos/exynos_drm_gem.c
@@ -77,10 +77,7 @@ static int exynos_drm_alloc_dma(struct exynos_drm_gem *exynos_gem)
 
 	init_dma_attrs(&exynos_gem->dma_attrs);
 
-	if (exynos_gem->flags & EXYNOS_BO_WC ||
-			!(exynos_gem->flags & EXYNOS_BO_CACHABLE))
-		dma_set_attr(DMA_ATTR_WRITE_COMBINE, &exynos_gem->dma_attrs);
-
+	dma_set_attr(DMA_ATTR_WRITE_COMBINE, &exynos_gem->dma_attrs);
 	dma_set_attr(DMA_ATTR_NO_KERNEL_MAPPING, &exynos_gem->dma_attrs);
 
 	nr_pages = exynos_gem->size >> PAGE_SHIFT;
@@ -128,51 +125,21 @@ static void exynos_drm_free_dma(struct exynos_drm_gem *exynos_gem)
 static int exynos_drm_alloc_buf(struct exynos_drm_gem *exynos_gem)
 {
 	struct drm_device *dev = exynos_gem->base.dev;
-	enum dma_attr attr;
-	unsigned int nr_pages;
+	int ret;
 
 	if (exynos_gem->dma_addr) {
 		DRM_DEBUG_KMS("already allocated.\n");
 		return 0;
 	}
 
-	if (!is_drm_iommu_supported(dev))
-		return exynos_drm_alloc_dma(exynos_gem);
-
-	init_dma_attrs(&exynos_gem->dma_attrs);
-
-	/*
-	 * if EXYNOS_BO_CONTIG, fully physically contiguous memory
-	 * region will be allocated else physically contiguous
-	 * as possible.
-	 */
-	if (!(exynos_gem->flags & EXYNOS_BO_NONCONTIG))
-		dma_set_attr(DMA_ATTR_FORCE_CONTIGUOUS, &exynos_gem->dma_attrs);
-
-	/* if EXYNOS_BO_WC or EXYNOS_BO_NONCACHABLE, writecombine mapping */
-	if (exynos_gem->flags & EXYNOS_BO_WC ||
-			!(exynos_gem->flags & EXYNOS_BO_CACHABLE))
-		attr = DMA_ATTR_WRITE_COMBINE;
-
-	dma_set_attr(attr, &exynos_gem->dma_attrs);
-	dma_set_attr(DMA_ATTR_NO_KERNEL_MAPPING, &exynos_gem->dma_attrs);
-
-	nr_pages = exynos_gem->size >> PAGE_SHIFT;
-
-	exynos_gem->cookie = dma_alloc_attrs(dev->dev, exynos_gem->size,
-					     &exynos_gem->dma_addr, GFP_KERNEL,
-					     &exynos_gem->dma_attrs);
-	if (!exynos_gem->cookie) {
-		DRM_ERROR("failed to allocate buffer.\n");
-		if (exynos_gem->pages)
-			drm_free_large(exynos_gem->pages);
-		return -ENOMEM;
+	if (!is_drm_iommu_supported(dev)) {
+		if (!(exynos_gem->flags & EXYNOS_BO_NONCONTIG))
+			return exynos_drm_alloc_dma(exynos_gem);
 	}
 
-	exynos_gem->pages = exynos_gem->cookie;
-
-	DRM_DEBUG_KMS("dma_addr(0x%lx), size(0x%lx)\n",
-			(unsigned long)exynos_gem->dma_addr, exynos_gem->size);
+	ret = exynos_drm_get_pages(exynos_gem);
+	if (ret < 0)
+		return ret;
 
 	return 0;
 }
@@ -186,15 +153,12 @@ static void exynos_drm_free_buf(struct exynos_drm_gem *exynos_gem)
 		return;
 	}
 
-	if (!is_drm_iommu_supported(dev))
-		return exynos_drm_free_dma(exynos_gem);
-
-	DRM_DEBUG_KMS("dma_addr(0x%lx), size(0x%lx)\n",
-			(unsigned long)exynos_gem->dma_addr, exynos_gem->size);
+	if (!is_drm_iommu_supported(dev)) {
+		if (!(exynos_gem->flags & EXYNOS_BO_NONCONTIG))
+			return exynos_drm_free_dma(exynos_gem);
+	}
 
-	dma_free_attrs(dev->dev, exynos_gem->size, exynos_gem->cookie,
-			(dma_addr_t)exynos_gem->dma_addr,
-			&exynos_gem->dma_attrs);
+	exynos_drm_put_pages(exynos_gem);
 }
 
 static int exynos_drm_gem_handle_create(struct drm_gem_object *obj,
@@ -400,8 +364,8 @@ void exynos_drm_gem_put_dma_addr(struct drm_device *dev,
 	drm_gem_object_unreference_unlocked(obj);
 }
 
-static int exynos_drm_gem_mmap_buffer(struct exynos_drm_gem *exynos_gem,
-				      struct vm_area_struct *vma)
+static int exynos_drm_gem_mmap_dma(struct exynos_drm_gem *exynos_gem,
+				   struct vm_area_struct *vma)
 {
 	struct drm_device *drm_dev = exynos_gem->base.dev;
 	unsigned long vm_size;
@@ -579,6 +543,19 @@ int exynos_drm_gem_mmap(struct file *filp, struct vm_area_struct *vma)
 
 	DRM_DEBUG_KMS("flags = 0x%x\n", exynos_gem->flags);
 
+	if (!is_drm_iommu_supported(obj->dev)) {
+		if (!(exynos_gem->flags & EXYNOS_BO_NONCONTIG)) {
+			ret = exynos_drm_gem_mmap_dma(exynos_gem, vma);
+			if (ret < 0)
+				drm_gem_vm_close(vma);
+
+			return ret;
+		}
+	}
+
+	vma->vm_flags &= ~VM_PFNMAP;
+	vma->vm_flags |= VM_MIXEDMAP;
+
 	/* non-cachable as default. */
 	if (exynos_gem->flags & EXYNOS_BO_CACHABLE)
 		vma->vm_page_prot = vm_get_page_prot(vma->vm_flags);
@@ -589,16 +566,7 @@ int exynos_drm_gem_mmap(struct file *filp, struct vm_area_struct *vma)
 		vma->vm_page_prot =
 			pgprot_noncached(vm_get_page_prot(vma->vm_flags));
 
-	ret = exynos_drm_gem_mmap_buffer(exynos_gem, vma);
-	if (ret)
-		goto err_close_vm;
-
-	return ret;
-
-err_close_vm:
-	drm_gem_vm_close(vma);
-
-	return ret;
+	return 0;
 }
 
 /* low-level interface prime helpers */
-- 
1.9.1



More information about the dri-devel mailing list