[PATCH 4/6] drm/ttm: Support kmap for single-page mappings in ttm_bo_vmap()

Thomas Zimmermann tzimmermann at suse.de
Fri Jun 14 13:21:58 UTC 2024


In ttm_bo_vmap(), set up single-page mappings with kmap() in certain
cases. The feature is already present in ttm_bo_kmap().

This functionality is require by DRM's xe driver, which claims that
using kmap() is an optimization over vmap(). [1] Reading the commit
at [2] indicates otherwise. It is not possible to use kmap_local_page()
and kunmap_local_page(), as TTM cannot guarantee the requirements for
ordering these calls. [3]

Signed-off-by: Thomas Zimmermann <tzimmermann at suse.de>
Link: https://elixir.bootlin.com/linux/v6.9/source/drivers/gpu/drm/xe/xe_bo.c#L1870 # 1
Link: https://lore.kernel.org/all/20220813220034.806698-1-ira.weiny@intel.com/T/#u # 2
Link: https://elixir.bootlin.com/linux/v6.9/source/include/linux/highmem.h#L70 # 3
---
 drivers/gpu/drm/ttm/ttm_bo_util.c | 33 ++++++++++++++++++++++---------
 1 file changed, 24 insertions(+), 9 deletions(-)

diff --git a/drivers/gpu/drm/ttm/ttm_bo_util.c b/drivers/gpu/drm/ttm/ttm_bo_util.c
index 31f9772f05dac..c06cfccace39d 100644
--- a/drivers/gpu/drm/ttm/ttm_bo_util.c
+++ b/drivers/gpu/drm/ttm/ttm_bo_util.c
@@ -516,6 +516,8 @@ int ttm_bo_vmap(struct ttm_buffer_object *bo,
 			.no_wait_gpu = false
 		};
 		struct ttm_tt *ttm = bo->ttm;
+		struct ttm_resource_manager *man =
+			ttm_manager_type(bo->bdev, bo->resource->mem_type);
 		unsigned long start_page = offset >> PAGE_SHIFT;
 		unsigned long aligned_size = size + (offset - (start_page << PAGE_SHIFT));
 		unsigned long num_pages = DIV_ROUND_UP(aligned_size, PAGE_SIZE);
@@ -527,15 +529,25 @@ int ttm_bo_vmap(struct ttm_buffer_object *bo,
 		if (ret)
 			return ret;
 
-		/*
-		 * We need to use vmap to get the desired page protection
-		 * or to make the buffer object look contiguous.
-		 */
-		prot = ttm_io_prot(bo, mem, PAGE_KERNEL);
-		vaddr = vmap(ttm->pages + start_page, num_pages, 0, prot);
-		if (!vaddr)
-			return -ENOMEM;
-		alloc_flags = ttm_bo_map_vmap;
+		if (num_pages == 1 && ttm->caching == ttm_cached &&
+		    !(man->use_tt && (ttm->page_flags & TTM_TT_FLAG_DECRYPTED))) {
+			/*
+			 * We're mapping a single page, and the desired
+			 * page protection is consistent with the bo.
+			 */
+			vaddr = kmap(ttm->pages[start_page]);
+			alloc_flags = ttm_bo_map_kmap;
+		} else {
+			/*
+			 * We need to use vmap to get the desired page protection
+			 * or to make the buffer object look contiguous.
+			 */
+			prot = ttm_io_prot(bo, mem, PAGE_KERNEL);
+			vaddr = vmap(ttm->pages + start_page, num_pages, 0, prot);
+			if (!vaddr)
+				return -ENOMEM;
+			alloc_flags = ttm_bo_map_vmap;
+		}
 
 		iosys_map_set_vaddr(map, vaddr);
 		map->alloc_flags = alloc_flags;
@@ -567,6 +579,9 @@ void ttm_bo_vunmap(struct ttm_buffer_object *bo, struct iosys_map *map)
 	case ttm_bo_map_vmap:
 		vunmap(map->vaddr);
 		break;
+	case ttm_bo_map_kmap:
+		kunmap(kmap_to_page(map->vaddr));
+		break;
 	case ttm_bo_map_premapped:
 		break;
 	default:
-- 
2.45.2



More information about the Intel-xe mailing list