Message ID | 1353408318-4104-3-git-send-email-chris@chris-wilson.co.uk (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
On Tue, Nov 20, 2012 at 10:45:18AM +0000, Chris Wilson wrote: > If we have hit oom whilst holding our struct_mutex, then currently we > cannot reap our own GPU buffers which likely pin most of memory, making > an outright OOM more likely. So if we are running in direct reclaim and > already hold the mutex, attempt to free buffers knowing that the > original function can not continue until we return. > > Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk> Ok, I've merged the two prep patches, but I think this one here will blow up, see below. > --- > drivers/gpu/drm/i915/i915_gem.c | 24 +++++++++++++++++++++--- > 1 file changed, 21 insertions(+), 3 deletions(-) > > diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c > index a7067e0..38d8fa3 100644 > --- a/drivers/gpu/drm/i915/i915_gem.c > +++ b/drivers/gpu/drm/i915/i915_gem.c > @@ -4340,6 +4340,18 @@ void i915_gem_release(struct drm_device *dev, struct drm_file *file) > spin_unlock(&file_priv->mm.lock); > } > > +static bool mutex_is_locked_by(struct mutex *mutex, struct task_struct *task) > +{ > + if (!mutex_is_locked(mutex)) > + return false; > + > +#if defined(CONFIG_SMP) || defined(CONFIG_DEBUG_MUTEXES) > + return mutex->owner == task; > +#else > + return true; I guess the idea is that on UP we can steal the lock, but that will introduce tons of new places where the backing storage can disappear (everywhere we block, instead of just allocations). And with preempt it won't work at all, since we could changed the lru lists in another task while the first one is walking them, leading ot all sorts of corruptions. So I think it's better to just ignore UP (for now) and return a conservative false here. Plus a comment explaining why. -Daniel > +#endif > +} > + > static int > i915_gem_inactive_shrink(struct shrinker *shrinker, struct shrink_control *sc) > { > @@ -4350,10 +4362,15 @@ i915_gem_inactive_shrink(struct shrinker *shrinker, struct shrink_control *sc) > struct drm_device *dev = dev_priv->dev; > struct drm_i915_gem_object *obj; > int nr_to_scan = sc->nr_to_scan; > + bool unlock = true; > int cnt; > > - if (!mutex_trylock(&dev->struct_mutex)) > - return 0; > + if (!mutex_trylock(&dev->struct_mutex)) { > + if (!mutex_is_locked_by(&dev->struct_mutex, current)) > + return 0; > + > + unlock = false; > + } > > if (nr_to_scan) { > nr_to_scan -= i915_gem_purge(dev_priv, nr_to_scan); > @@ -4369,6 +4386,7 @@ i915_gem_inactive_shrink(struct shrinker *shrinker, struct shrink_control *sc) > if (obj->pin_count == 0 && obj->pages_pin_count == 0) > cnt += obj->base.size >> PAGE_SHIFT; > > - mutex_unlock(&dev->struct_mutex); > + if (unlock) > + mutex_unlock(&dev->struct_mutex); > return cnt; > } > -- > 1.7.10.4 > > _______________________________________________ > Intel-gfx mailing list > Intel-gfx@lists.freedesktop.org > http://lists.freedesktop.org/mailman/listinfo/intel-gfx
diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c index a7067e0..38d8fa3 100644 --- a/drivers/gpu/drm/i915/i915_gem.c +++ b/drivers/gpu/drm/i915/i915_gem.c @@ -4340,6 +4340,18 @@ void i915_gem_release(struct drm_device *dev, struct drm_file *file) spin_unlock(&file_priv->mm.lock); } +static bool mutex_is_locked_by(struct mutex *mutex, struct task_struct *task) +{ + if (!mutex_is_locked(mutex)) + return false; + +#if defined(CONFIG_SMP) || defined(CONFIG_DEBUG_MUTEXES) + return mutex->owner == task; +#else + return true; +#endif +} + static int i915_gem_inactive_shrink(struct shrinker *shrinker, struct shrink_control *sc) { @@ -4350,10 +4362,15 @@ i915_gem_inactive_shrink(struct shrinker *shrinker, struct shrink_control *sc) struct drm_device *dev = dev_priv->dev; struct drm_i915_gem_object *obj; int nr_to_scan = sc->nr_to_scan; + bool unlock = true; int cnt; - if (!mutex_trylock(&dev->struct_mutex)) - return 0; + if (!mutex_trylock(&dev->struct_mutex)) { + if (!mutex_is_locked_by(&dev->struct_mutex, current)) + return 0; + + unlock = false; + } if (nr_to_scan) { nr_to_scan -= i915_gem_purge(dev_priv, nr_to_scan); @@ -4369,6 +4386,7 @@ i915_gem_inactive_shrink(struct shrinker *shrinker, struct shrink_control *sc) if (obj->pin_count == 0 && obj->pages_pin_count == 0) cnt += obj->base.size >> PAGE_SHIFT; - mutex_unlock(&dev->struct_mutex); + if (unlock) + mutex_unlock(&dev->struct_mutex); return cnt; }
If we have hit oom whilst holding our struct_mutex, then currently we cannot reap our own GPU buffers which likely pin most of memory, making an outright OOM more likely. So if we are running in direct reclaim and already hold the mutex, attempt to free buffers knowing that the original function can not continue until we return. Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk> --- drivers/gpu/drm/i915/i915_gem.c | 24 +++++++++++++++++++++--- 1 file changed, 21 insertions(+), 3 deletions(-)