diff mbox series

drm/i915: Recover batch pool caches from shrinker

Message ID 20180919194911.1084-1-chris@chris-wilson.co.uk (mailing list archive)
State New, archived
Headers show
Series drm/i915: Recover batch pool caches from shrinker | expand

Commit Message

Chris Wilson Sept. 19, 2018, 7:49 p.m. UTC
Discard all of our batch pools under mempressure to make their pages
available to the shrinker. We will quickly reacquire them when necessary
for more GPU relocations or for the command parser.

v2: Init the lists for mock_engine
v3: Return a strong ref from i915_gem_batch_pool_get() and convert it
into an active reference to protect ourselves against all allocations
while the object is in play.

Bugzilla: https://bugs.freedesktop.org/show_bug.cgi?id=107936
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Matthew Auld <matthew.william.auld@gmail.com>
---
 drivers/gpu/drm/i915/i915_gem_batch_pool.c   |  9 ++++++---
 drivers/gpu/drm/i915/i915_gem_execbuffer.c   | 15 +++++++++++----
 drivers/gpu/drm/i915/i915_gem_object.h       |  1 +
 drivers/gpu/drm/i915/i915_gem_shrinker.c     | 11 +++++++++++
 drivers/gpu/drm/i915/selftests/mock_engine.c |  2 ++
 5 files changed, 31 insertions(+), 7 deletions(-)

Comments

Matthew Auld Sept. 20, 2018, 9:57 a.m. UTC | #1
On Wed, 19 Sep 2018 at 20:49, Chris Wilson <chris@chris-wilson.co.uk> wrote:
>
> Discard all of our batch pools under mempressure to make their pages
> available to the shrinker. We will quickly reacquire them when necessary
> for more GPU relocations or for the command parser.
>
> v2: Init the lists for mock_engine
> v3: Return a strong ref from i915_gem_batch_pool_get() and convert it
> into an active reference to protect ourselves against all allocations
> while the object is in play.
>
> Bugzilla: https://bugs.freedesktop.org/show_bug.cgi?id=107936
> Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
> Cc: Matthew Auld <matthew.william.auld@gmail.com>
Reviewed-by: Matthew Auld <matthew.william.auld@gmail.com>
Chris Wilson Sept. 20, 2018, 10:02 a.m. UTC | #2
Quoting Matthew Auld (2018-09-20 10:57:34)
> On Wed, 19 Sep 2018 at 20:49, Chris Wilson <chris@chris-wilson.co.uk> wrote:
> >
> > Discard all of our batch pools under mempressure to make their pages
> > available to the shrinker. We will quickly reacquire them when necessary
> > for more GPU relocations or for the command parser.
> >
> > v2: Init the lists for mock_engine
> > v3: Return a strong ref from i915_gem_batch_pool_get() and convert it
> > into an active reference to protect ourselves against all allocations
> > while the object is in play.
> >
> > Bugzilla: https://bugs.freedesktop.org/show_bug.cgi?id=107936
> > Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
> > Cc: Matthew Auld <matthew.william.auld@gmail.com>
> Reviewed-by: Matthew Auld <matthew.william.auld@gmail.com>

Sorry, still not quite right yet. Closer, but the conversion into an
active reference is not guaranteed (as the execbuf may be interrupted).
-Chris
diff mbox series

Patch

diff --git a/drivers/gpu/drm/i915/i915_gem_batch_pool.c b/drivers/gpu/drm/i915/i915_gem_batch_pool.c
index f3890b664e3f..b560350cc138 100644
--- a/drivers/gpu/drm/i915/i915_gem_batch_pool.c
+++ b/drivers/gpu/drm/i915/i915_gem_batch_pool.c
@@ -131,10 +131,13 @@  i915_gem_batch_pool_get(struct i915_gem_batch_pool *pool,
 		return obj;
 
 found:
+	list_del(&obj->batch_pool_link);
 	ret = i915_gem_object_pin_pages(obj);
-	if (ret)
+	if (ret) {
+		i915_gem_object_put(obj);
 		return ERR_PTR(ret);
+	}
 
-	list_move_tail(&obj->batch_pool_link, list);
-	return obj;
+	list_add_tail(&obj->batch_pool_link, list);
+	return i915_gem_object_get(obj);
 }
diff --git a/drivers/gpu/drm/i915/i915_gem_execbuffer.c b/drivers/gpu/drm/i915/i915_gem_execbuffer.c
index 7f70188809b0..9a565d6f2864 100644
--- a/drivers/gpu/drm/i915/i915_gem_execbuffer.c
+++ b/drivers/gpu/drm/i915/i915_gem_execbuffer.c
@@ -1150,8 +1150,10 @@  static int __reloc_gpu_alloc(struct i915_execbuffer *eb,
 				      I915_MAP_FORCE_WB :
 				      I915_MAP_FORCE_WC);
 	i915_gem_object_unpin_pages(obj);
-	if (IS_ERR(cmd))
-		return PTR_ERR(cmd);
+	if (IS_ERR(cmd)) {
+		err = PTR_ERR(cmd);
+		goto err_put;
+	}
 
 	err = i915_gem_object_set_to_wc_domain(obj, false);
 	if (err)
@@ -1194,6 +1196,7 @@  static int __reloc_gpu_alloc(struct i915_execbuffer *eb,
 
 	rq->batch = batch;
 	i915_vma_unpin(batch);
+	i915_gem_object_set_active_reference(obj);
 
 	cache->rq = rq;
 	cache->rq_cmd = cmd;
@@ -1210,6 +1213,8 @@  static int __reloc_gpu_alloc(struct i915_execbuffer *eb,
 	i915_vma_unpin(batch);
 err_unmap:
 	i915_gem_object_unpin_map(obj);
+err_put:
+	i915_gem_object_put(obj);
 	return err;
 }
 
@@ -1943,13 +1948,15 @@  static struct i915_vma *eb_parse(struct i915_execbuffer *eb, bool is_master)
 		goto out;
 
 	eb->vma[eb->buffer_count] = i915_vma_get(vma);
-	eb->flags[eb->buffer_count] =
-		__EXEC_OBJECT_HAS_PIN | __EXEC_OBJECT_HAS_REF;
+	eb->flags[eb->buffer_count] = __EXEC_OBJECT_HAS_PIN;
 	vma->exec_flags = &eb->flags[eb->buffer_count];
 	eb->buffer_count++;
 
+	i915_gem_object_set_active_reference(shadow_batch_obj);
+
 out:
 	i915_gem_object_unpin_pages(shadow_batch_obj);
+	i915_gem_object_put(shadow_batch_obj);
 	return vma;
 }
 
diff --git a/drivers/gpu/drm/i915/i915_gem_object.h b/drivers/gpu/drm/i915/i915_gem_object.h
index a6dd7c46de0d..e756b55190b7 100644
--- a/drivers/gpu/drm/i915/i915_gem_object.h
+++ b/drivers/gpu/drm/i915/i915_gem_object.h
@@ -402,6 +402,7 @@  static inline void
 i915_gem_object_set_active_reference(struct drm_i915_gem_object *obj)
 {
 	lockdep_assert_held(&obj->base.dev->struct_mutex);
+	GEM_BUG_ON(test_bit(I915_BO_ACTIVE_REF, &obj->flags));
 	__set_bit(I915_BO_ACTIVE_REF, &obj->flags);
 }
 
diff --git a/drivers/gpu/drm/i915/i915_gem_shrinker.c b/drivers/gpu/drm/i915/i915_gem_shrinker.c
index 681226992705..6036fe8495d4 100644
--- a/drivers/gpu/drm/i915/i915_gem_shrinker.c
+++ b/drivers/gpu/drm/i915/i915_gem_shrinker.c
@@ -117,6 +117,15 @@  static bool unsafe_drop_pages(struct drm_i915_gem_object *obj)
 	return !i915_gem_object_has_pages(obj);
 }
 
+static void shrink_caches(struct drm_i915_private *i915)
+{
+	struct intel_engine_cs *engine;
+	enum intel_engine_id id;
+
+	for_each_engine(engine, i915, id)
+		i915_gem_batch_pool_fini(&engine->batch_pool);
+}
+
 /**
  * i915_gem_shrink - Shrink buffer object caches
  * @i915: i915 device
@@ -181,6 +190,8 @@  i915_gem_shrink(struct drm_i915_private *i915,
 	trace_i915_gem_shrink(i915, target, flags);
 	i915_retire_requests(i915);
 
+	shrink_caches(i915);
+
 	/*
 	 * Unbinding of objects will require HW access; Let us not wake the
 	 * device just to recover a little memory. If absolutely necessary,
diff --git a/drivers/gpu/drm/i915/selftests/mock_engine.c b/drivers/gpu/drm/i915/selftests/mock_engine.c
index 22a73da45ad5..8fc511eb3f13 100644
--- a/drivers/gpu/drm/i915/selftests/mock_engine.c
+++ b/drivers/gpu/drm/i915/selftests/mock_engine.c
@@ -205,6 +205,8 @@  struct intel_engine_cs *mock_engine(struct drm_i915_private *i915,
 	intel_engine_init_breadcrumbs(&engine->base);
 	engine->base.breadcrumbs.mock = true; /* prevent touching HW for irqs */
 
+	intel_engine_init_batch_pool(&engine->base);
+
 	/* fake hw queue */
 	spin_lock_init(&engine->hw_lock);
 	timer_setup(&engine->hw_delay, hw_delay_complete, 0);