[Intel-gfx] [PATCH 2/2] drm/i915: Disable page-faults around the fast pwrite/pread paths

Chris Wilson chris at chris-wilson.co.uk
Sat Jul 9 10:38:51 CEST 2011


These paths hold onto the struct mutex whilst accessing pages. In
order, to prevent a recursive dead-lock should we fault-in a GTT mapped
page we need to return -EFAULT and fallback to the slow path.

Lockdep has complained before about the potential dead-lock, but rvis is
the first application found to sufficiently abuse the API to trigger it.

Cursory performance regression testing on a 1GiB PineView system using
x11perf, cairo-perf-trace, glxgears and a few game benchmarks suggested
no large regressions with just a 2% slowdown for firefox. The caveat is
that this was an otherwise idle system and that for 32-bit systems
io_mapping_map_atomic_wc() already disabled page-faults.

Bugzilla: https://bugs.freedesktop.org/show_bug.cgi?id=38115
Signed-off-by: Chris Wilson <chris at chris-wilson.co.uk>
---
 drivers/gpu/drm/i915/i915_gem.c |   22 ++++++++++++++++++++--
 1 files changed, 20 insertions(+), 2 deletions(-)

diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c
index 2fce620..ecb27fd 100644
--- a/drivers/gpu/drm/i915/i915_gem.c
+++ b/drivers/gpu/drm/i915/i915_gem.c
@@ -365,9 +365,15 @@ i915_gem_shmem_pread_fast(struct drm_device *dev,
 			return PTR_ERR(page);
 
 		vaddr = kmap_atomic(page);
+		/* We have to disable faulting here in case the user address
+		 * is really a GTT mapping and so we can not enter
+		 * i915_gem_fault() whilst already holding struct_mutex.
+		 */
+		pagefault_disable();
 		ret = __copy_to_user_inatomic(user_data,
 					      vaddr + page_offset,
 					      page_length);
+		pagefault_enable();
 		kunmap_atomic(vaddr);
 
 		mark_page_accessed(page);
@@ -593,8 +599,14 @@ fast_user_write(struct io_mapping *mapping,
 	unsigned long unwritten;
 
 	vaddr_atomic = io_mapping_map_atomic_wc(mapping, page_base);
+	/* We have to disable faulting here in case the user address
+	 * is really a GTT mapping and so we can not enter
+	 * i915_gem_fault() whilst already holding struct_mutex.
+	 */
+	pagefault_disable();
 	unwritten = __copy_from_user_inatomic_nocache(vaddr_atomic + page_offset,
 						      user_data, length);
+	pagefault_enable();
 	io_mapping_unmap_atomic(vaddr_atomic);
 	return unwritten;
 }
@@ -812,11 +824,17 @@ i915_gem_shmem_pwrite_fast(struct drm_device *dev,
 		if (IS_ERR(page))
 			return PTR_ERR(page);
 
-		vaddr = kmap_atomic(page, KM_USER0);
+		vaddr = kmap_atomic(page);
+		/* We have to disable faulting here in case the user address
+		 * is really a GTT mapping and so we can not enter
+		 * i915_gem_fault() whilst already holding struct_mutex.
+		 */
+		pagefault_disable();
 		ret = __copy_from_user_inatomic(vaddr + page_offset,
 						user_data,
 						page_length);
-		kunmap_atomic(vaddr, KM_USER0);
+		pagefault_enable();
+		kunmap_atomic(vaddr);
 
 		set_page_dirty(page);
 		mark_page_accessed(page);
-- 
1.7.5.4




More information about the Intel-gfx mailing list