[Intel-gfx] [PATCH] drm/i915: Redirect GTT mappings to the CPU page if cache-coherent
Chris Wilson
chris at chris-wilson.co.uk
Wed Apr 13 20:35:52 CEST 2011
... or if we will need to perform a cache-flush on the object anyway.
Unless, of course, we need to use a fence register to perform tiling
operations during the transfer (in which case we are no longer on a
chipset for which we need to be extra careful not to write through the
GTT to a snooped page).
Signed-off-by: Chris Wilson <chris at chris-wilson.co.uk>
---
drivers/gpu/drm/i915/i915_gem.c | 41 ++++++++++++++++++++++++++++++++++++++-
1 files changed, 40 insertions(+), 1 deletions(-)
diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c
index 8b3007c..3c7443d 100644
--- a/drivers/gpu/drm/i915/i915_gem.c
+++ b/drivers/gpu/drm/i915/i915_gem.c
@@ -1211,12 +1211,43 @@ int i915_gem_fault(struct vm_area_struct *vma, struct vm_fault *vmf)
trace_i915_gem_object_fault(obj, page_offset, true, write);
- /* Now bind it into the GTT if needed */
if (!obj->map_and_fenceable) {
ret = i915_gem_object_unbind(obj);
if (ret)
goto unlock;
}
+
+ /* If it is unbound or we are currently writing through the CPU
+ * domain, continue to do so. On older chipsets it is
+ * particularly important to avoid writing through the GTT to
+ * snooped pages or face dire consequences. At least that's what
+ * the docs say...
+ */
+ if (obj->tiling_mode == I915_TILING_NONE &&
+ (obj->cache_level != I915_CACHE_NONE ||
+ obj->base.write_domain == I915_GEM_DOMAIN_CPU)) {
+ struct page *page;
+
+ ret = i915_gem_object_set_to_cpu_domain(obj, write);
+ if (ret)
+ goto unlock;
+
+ obj->dirty = 1;
+ obj->fault_mappable = true;
+ mutex_unlock(&dev->struct_mutex);
+
+ page = read_cache_page_gfp(obj->base.filp->f_path.dentry->d_inode->i_mapping,
+ page_offset,
+ GFP_HIGHUSER | __GFP_RECLAIMABLE);
+ if (IS_ERR(page)) {
+ ret = PTR_ERR(page);
+ goto out;
+ }
+
+ vmf->page = page;
+ return VM_FAULT_LOCKED;
+ }
+
if (!obj->gtt_space) {
ret = i915_gem_object_bind_to_gtt(obj, 0, true);
if (ret)
@@ -1699,6 +1730,11 @@ i915_gem_object_truncate(struct drm_i915_gem_object *obj)
{
struct inode *inode;
+ /* We may have inserted the backing pages into our vma
+ * when fulfilling a pagefault whilst in the CPU domain.
+ */
+ i915_gem_release_mmap(obj);
+
/* Our goal here is to return as much of the memory as
* is possible back to the system as we are called from OOM.
* To do this we must instruct the shmfs to drop all of its
@@ -3691,6 +3727,9 @@ void i915_gem_free_object(struct drm_gem_object *gem_obj)
if (obj->phys_obj)
i915_gem_detach_phys_object(dev, obj);
+ /* Discard all references to the backing storage for this object */
+ i915_gem_object_truncate(obj);
+
i915_gem_free_object_tail(obj);
}
--
1.7.4.1
More information about the Intel-gfx
mailing list