[Intel-gfx] [PATCH] drm/i915: Recover batch pool caches from shrinker

Chris Wilson chris at chris-wilson.co.uk
Wed Sep 19 19:49:11 UTC 2018


Discard all of our batch pools under mempressure to make their pages
available to the shrinker. We will quickly reacquire them when necessary
for more GPU relocations or for the command parser.

v2: Init the lists for mock_engine
v3: Return a strong ref from i915_gem_batch_pool_get() and convert it
into an active reference to protect ourselves against all allocations
while the object is in play.

Bugzilla: https://bugs.freedesktop.org/show_bug.cgi?id=107936
Signed-off-by: Chris Wilson <chris at chris-wilson.co.uk>
Cc: Matthew Auld <matthew.william.auld at gmail.com>
---
 drivers/gpu/drm/i915/i915_gem_batch_pool.c   |  9 ++++++---
 drivers/gpu/drm/i915/i915_gem_execbuffer.c   | 15 +++++++++++----
 drivers/gpu/drm/i915/i915_gem_object.h       |  1 +
 drivers/gpu/drm/i915/i915_gem_shrinker.c     | 11 +++++++++++
 drivers/gpu/drm/i915/selftests/mock_engine.c |  2 ++
 5 files changed, 31 insertions(+), 7 deletions(-)

diff --git a/drivers/gpu/drm/i915/i915_gem_batch_pool.c b/drivers/gpu/drm/i915/i915_gem_batch_pool.c
index f3890b664e3f..b560350cc138 100644
--- a/drivers/gpu/drm/i915/i915_gem_batch_pool.c
+++ b/drivers/gpu/drm/i915/i915_gem_batch_pool.c
@@ -131,10 +131,13 @@ i915_gem_batch_pool_get(struct i915_gem_batch_pool *pool,
 		return obj;
 
 found:
+	list_del(&obj->batch_pool_link);
 	ret = i915_gem_object_pin_pages(obj);
-	if (ret)
+	if (ret) {
+		i915_gem_object_put(obj);
 		return ERR_PTR(ret);
+	}
 
-	list_move_tail(&obj->batch_pool_link, list);
-	return obj;
+	list_add_tail(&obj->batch_pool_link, list);
+	return i915_gem_object_get(obj);
 }
diff --git a/drivers/gpu/drm/i915/i915_gem_execbuffer.c b/drivers/gpu/drm/i915/i915_gem_execbuffer.c
index 7f70188809b0..9a565d6f2864 100644
--- a/drivers/gpu/drm/i915/i915_gem_execbuffer.c
+++ b/drivers/gpu/drm/i915/i915_gem_execbuffer.c
@@ -1150,8 +1150,10 @@ static int __reloc_gpu_alloc(struct i915_execbuffer *eb,
 				      I915_MAP_FORCE_WB :
 				      I915_MAP_FORCE_WC);
 	i915_gem_object_unpin_pages(obj);
-	if (IS_ERR(cmd))
-		return PTR_ERR(cmd);
+	if (IS_ERR(cmd)) {
+		err = PTR_ERR(cmd);
+		goto err_put;
+	}
 
 	err = i915_gem_object_set_to_wc_domain(obj, false);
 	if (err)
@@ -1194,6 +1196,7 @@ static int __reloc_gpu_alloc(struct i915_execbuffer *eb,
 
 	rq->batch = batch;
 	i915_vma_unpin(batch);
+	i915_gem_object_set_active_reference(obj);
 
 	cache->rq = rq;
 	cache->rq_cmd = cmd;
@@ -1210,6 +1213,8 @@ static int __reloc_gpu_alloc(struct i915_execbuffer *eb,
 	i915_vma_unpin(batch);
 err_unmap:
 	i915_gem_object_unpin_map(obj);
+err_put:
+	i915_gem_object_put(obj);
 	return err;
 }
 
@@ -1943,13 +1948,15 @@ static struct i915_vma *eb_parse(struct i915_execbuffer *eb, bool is_master)
 		goto out;
 
 	eb->vma[eb->buffer_count] = i915_vma_get(vma);
-	eb->flags[eb->buffer_count] =
-		__EXEC_OBJECT_HAS_PIN | __EXEC_OBJECT_HAS_REF;
+	eb->flags[eb->buffer_count] = __EXEC_OBJECT_HAS_PIN;
 	vma->exec_flags = &eb->flags[eb->buffer_count];
 	eb->buffer_count++;
 
+	i915_gem_object_set_active_reference(shadow_batch_obj);
+
 out:
 	i915_gem_object_unpin_pages(shadow_batch_obj);
+	i915_gem_object_put(shadow_batch_obj);
 	return vma;
 }
 
diff --git a/drivers/gpu/drm/i915/i915_gem_object.h b/drivers/gpu/drm/i915/i915_gem_object.h
index a6dd7c46de0d..e756b55190b7 100644
--- a/drivers/gpu/drm/i915/i915_gem_object.h
+++ b/drivers/gpu/drm/i915/i915_gem_object.h
@@ -402,6 +402,7 @@ static inline void
 i915_gem_object_set_active_reference(struct drm_i915_gem_object *obj)
 {
 	lockdep_assert_held(&obj->base.dev->struct_mutex);
+	GEM_BUG_ON(test_bit(I915_BO_ACTIVE_REF, &obj->flags));
 	__set_bit(I915_BO_ACTIVE_REF, &obj->flags);
 }
 
diff --git a/drivers/gpu/drm/i915/i915_gem_shrinker.c b/drivers/gpu/drm/i915/i915_gem_shrinker.c
index 681226992705..6036fe8495d4 100644
--- a/drivers/gpu/drm/i915/i915_gem_shrinker.c
+++ b/drivers/gpu/drm/i915/i915_gem_shrinker.c
@@ -117,6 +117,15 @@ static bool unsafe_drop_pages(struct drm_i915_gem_object *obj)
 	return !i915_gem_object_has_pages(obj);
 }
 
+static void shrink_caches(struct drm_i915_private *i915)
+{
+	struct intel_engine_cs *engine;
+	enum intel_engine_id id;
+
+	for_each_engine(engine, i915, id)
+		i915_gem_batch_pool_fini(&engine->batch_pool);
+}
+
 /**
  * i915_gem_shrink - Shrink buffer object caches
  * @i915: i915 device
@@ -181,6 +190,8 @@ i915_gem_shrink(struct drm_i915_private *i915,
 	trace_i915_gem_shrink(i915, target, flags);
 	i915_retire_requests(i915);
 
+	shrink_caches(i915);
+
 	/*
 	 * Unbinding of objects will require HW access; Let us not wake the
 	 * device just to recover a little memory. If absolutely necessary,
diff --git a/drivers/gpu/drm/i915/selftests/mock_engine.c b/drivers/gpu/drm/i915/selftests/mock_engine.c
index 22a73da45ad5..8fc511eb3f13 100644
--- a/drivers/gpu/drm/i915/selftests/mock_engine.c
+++ b/drivers/gpu/drm/i915/selftests/mock_engine.c
@@ -205,6 +205,8 @@ struct intel_engine_cs *mock_engine(struct drm_i915_private *i915,
 	intel_engine_init_breadcrumbs(&engine->base);
 	engine->base.breadcrumbs.mock = true; /* prevent touching HW for irqs */
 
+	intel_engine_init_batch_pool(&engine->base);
+
 	/* fake hw queue */
 	spin_lock_init(&engine->hw_lock);
 	timer_setup(&engine->hw_delay, hw_delay_complete, 0);
-- 
2.19.0



More information about the Intel-gfx mailing list