Message ID | 20190408091728.20207-9-chris@chris-wilson.co.uk (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | [01/29] drm/i915: Mark up ips for RCU protection | expand |
On 08/04/2019 10:17, Chris Wilson wrote: > Split out the powermanagement portion (GT wakeref, suspend/resume) of > GEM from i915_gem.c into its own file. This is the same patch I gave r-b on before? Judging from the title and commit message I'd say yes. But I know it is not, don't worry. Regards, Tvrtko > Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk> > --- > drivers/gpu/drm/i915/Makefile | 1 + > drivers/gpu/drm/i915/Makefile.header-test | 1 + > drivers/gpu/drm/i915/i915_debugfs.c | 4 +- > drivers/gpu/drm/i915/i915_drv.h | 12 +- > drivers/gpu/drm/i915/i915_gem.c | 363 +---------------- > drivers/gpu/drm/i915/i915_gem_pm.c | 365 ++++++++++++++++++ > drivers/gpu/drm/i915/i915_gem_pm.h | 28 ++ > .../gpu/drm/i915/selftests/i915_gem_context.c | 2 +- > .../gpu/drm/i915/selftests/i915_gem_object.c | 8 +- > .../gpu/drm/i915/selftests/mock_gem_device.c | 10 +- > 10 files changed, 418 insertions(+), 376 deletions(-) > create mode 100644 drivers/gpu/drm/i915/i915_gem_pm.c > create mode 100644 drivers/gpu/drm/i915/i915_gem_pm.h > > diff --git a/drivers/gpu/drm/i915/Makefile b/drivers/gpu/drm/i915/Makefile > index 233bad5e361f..858642c7bc40 100644 > --- a/drivers/gpu/drm/i915/Makefile > +++ b/drivers/gpu/drm/i915/Makefile > @@ -97,6 +97,7 @@ i915-y += \ > i915_gem_internal.o \ > i915_gem.o \ > i915_gem_object.o \ > + i915_gem_pm.o \ > i915_gem_render_state.o \ > i915_gem_shrinker.o \ > i915_gem_stolen.o \ > diff --git a/drivers/gpu/drm/i915/Makefile.header-test b/drivers/gpu/drm/i915/Makefile.header-test > index e6b3e7588860..702e3a7ade4c 100644 > --- a/drivers/gpu/drm/i915/Makefile.header-test > +++ b/drivers/gpu/drm/i915/Makefile.header-test > @@ -5,6 +5,7 @@ > header_test := \ > i915_active_types.h \ > i915_gem_context_types.h \ > + i915_gem_pm.h \ > i915_priolist_types.h \ > i915_scheduler_types.h \ > i915_timeline_types.h \ > diff --git a/drivers/gpu/drm/i915/i915_debugfs.c b/drivers/gpu/drm/i915/i915_debugfs.c > index eac4095278b2..e3eca88614b2 100644 > --- a/drivers/gpu/drm/i915/i915_debugfs.c > +++ b/drivers/gpu/drm/i915/i915_debugfs.c > @@ -3941,8 +3941,8 @@ i915_drop_caches_set(void *data, u64 val) > if (val & DROP_IDLE) { > do { > if (READ_ONCE(i915->gt.active_requests)) > - flush_delayed_work(&i915->gt.retire_work); > - drain_delayed_work(&i915->gt.idle_work); > + flush_delayed_work(&i915->gem.retire_work); > + drain_delayed_work(&i915->gem.idle_work); > } while (READ_ONCE(i915->gt.awake)); > } > > diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h > index 62a7e91acd7f..cbae9be052e0 100644 > --- a/drivers/gpu/drm/i915/i915_drv.h > +++ b/drivers/gpu/drm/i915/i915_drv.h > @@ -2020,6 +2020,12 @@ struct drm_i915_private { > */ > intel_wakeref_t awake; > > + ktime_t last_init_time; > + > + struct i915_vma *scratch; > + } gt; > + > + struct { > /** > * We leave the user IRQ off as much as possible, > * but this means that requests will finish and never > @@ -2037,11 +2043,7 @@ struct drm_i915_private { > * off the idle_work. > */ > struct delayed_work idle_work; > - > - ktime_t last_init_time; > - > - struct i915_vma *scratch; > - } gt; > + } gem; > > /* For i945gm vblank irq vs. C3 workaround */ > struct { > diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c > index a8b0aa05288b..c56633e87c32 100644 > --- a/drivers/gpu/drm/i915/i915_gem.c > +++ b/drivers/gpu/drm/i915/i915_gem.c > @@ -46,7 +46,7 @@ > #include "i915_drv.h" > #include "i915_gem_clflush.h" > #include "i915_gemfs.h" > -#include "i915_globals.h" > +#include "i915_gem_pm.h" > #include "i915_trace.h" > #include "i915_vgpu.h" > > @@ -103,105 +103,6 @@ static void i915_gem_info_remove_obj(struct drm_i915_private *dev_priv, > spin_unlock(&dev_priv->mm.object_stat_lock); > } > > -static void __i915_gem_park(struct drm_i915_private *i915) > -{ > - intel_wakeref_t wakeref; > - > - GEM_TRACE("\n"); > - > - lockdep_assert_held(&i915->drm.struct_mutex); > - GEM_BUG_ON(i915->gt.active_requests); > - GEM_BUG_ON(!list_empty(&i915->gt.active_rings)); > - > - if (!i915->gt.awake) > - return; > - > - /* > - * Be paranoid and flush a concurrent interrupt to make sure > - * we don't reactivate any irq tasklets after parking. > - * > - * FIXME: Note that even though we have waited for execlists to be idle, > - * there may still be an in-flight interrupt even though the CSB > - * is now empty. synchronize_irq() makes sure that a residual interrupt > - * is completed before we continue, but it doesn't prevent the HW from > - * raising a spurious interrupt later. To complete the shield we should > - * coordinate disabling the CS irq with flushing the interrupts. > - */ > - synchronize_irq(i915->drm.irq); > - > - intel_engines_park(i915); > - i915_timelines_park(i915); > - > - i915_pmu_gt_parked(i915); > - i915_vma_parked(i915); > - > - wakeref = fetch_and_zero(&i915->gt.awake); > - GEM_BUG_ON(!wakeref); > - > - if (INTEL_GEN(i915) >= 6) > - gen6_rps_idle(i915); > - > - intel_display_power_put(i915, POWER_DOMAIN_GT_IRQ, wakeref); > - > - i915_globals_park(); > -} > - > -void i915_gem_park(struct drm_i915_private *i915) > -{ > - GEM_TRACE("\n"); > - > - lockdep_assert_held(&i915->drm.struct_mutex); > - GEM_BUG_ON(i915->gt.active_requests); > - > - if (!i915->gt.awake) > - return; > - > - /* Defer the actual call to __i915_gem_park() to prevent ping-pongs */ > - mod_delayed_work(i915->wq, &i915->gt.idle_work, msecs_to_jiffies(100)); > -} > - > -void i915_gem_unpark(struct drm_i915_private *i915) > -{ > - GEM_TRACE("\n"); > - > - lockdep_assert_held(&i915->drm.struct_mutex); > - GEM_BUG_ON(!i915->gt.active_requests); > - assert_rpm_wakelock_held(i915); > - > - if (i915->gt.awake) > - return; > - > - /* > - * It seems that the DMC likes to transition between the DC states a lot > - * when there are no connected displays (no active power domains) during > - * command submission. > - * > - * This activity has negative impact on the performance of the chip with > - * huge latencies observed in the interrupt handler and elsewhere. > - * > - * Work around it by grabbing a GT IRQ power domain whilst there is any > - * GT activity, preventing any DC state transitions. > - */ > - i915->gt.awake = intel_display_power_get(i915, POWER_DOMAIN_GT_IRQ); > - GEM_BUG_ON(!i915->gt.awake); > - > - i915_globals_unpark(); > - > - intel_enable_gt_powersave(i915); > - i915_update_gfx_val(i915); > - if (INTEL_GEN(i915) >= 6) > - gen6_rps_busy(i915); > - i915_pmu_gt_unparked(i915); > - > - intel_engines_unpark(i915); > - > - i915_queue_hangcheck(i915); > - > - queue_delayed_work(i915->wq, > - &i915->gt.retire_work, > - round_jiffies_up_relative(HZ)); > -} > - > int > i915_gem_get_aperture_ioctl(struct drm_device *dev, void *data, > struct drm_file *file) > @@ -2088,7 +1989,7 @@ static int i915_gem_object_create_mmap_offset(struct drm_i915_gem_object *obj) > if (!err) > break; > > - } while (flush_delayed_work(&dev_priv->gt.retire_work)); > + } while (flush_delayed_work(&dev_priv->gem.retire_work)); > > return err; > } > @@ -2871,132 +2772,6 @@ i915_gem_object_pwrite_gtt(struct drm_i915_gem_object *obj, > return 0; > } > > -static void > -i915_gem_retire_work_handler(struct work_struct *work) > -{ > - struct drm_i915_private *dev_priv = > - container_of(work, typeof(*dev_priv), gt.retire_work.work); > - struct drm_device *dev = &dev_priv->drm; > - > - /* Come back later if the device is busy... */ > - if (mutex_trylock(&dev->struct_mutex)) { > - i915_retire_requests(dev_priv); > - mutex_unlock(&dev->struct_mutex); > - } > - > - /* > - * Keep the retire handler running until we are finally idle. > - * We do not need to do this test under locking as in the worst-case > - * we queue the retire worker once too often. > - */ > - if (READ_ONCE(dev_priv->gt.awake)) > - queue_delayed_work(dev_priv->wq, > - &dev_priv->gt.retire_work, > - round_jiffies_up_relative(HZ)); > -} > - > -static bool switch_to_kernel_context_sync(struct drm_i915_private *i915, > - unsigned long mask) > -{ > - bool result = true; > - > - /* > - * Even if we fail to switch, give whatever is running a small chance > - * to save itself before we report the failure. Yes, this may be a > - * false positive due to e.g. ENOMEM, caveat emptor! > - */ > - if (i915_gem_switch_to_kernel_context(i915, mask)) > - result = false; > - > - if (i915_gem_wait_for_idle(i915, > - I915_WAIT_LOCKED | > - I915_WAIT_FOR_IDLE_BOOST, > - I915_GEM_IDLE_TIMEOUT)) > - result = false; > - > - if (!result) { > - if (i915_modparams.reset) { /* XXX hide warning from gem_eio */ > - dev_err(i915->drm.dev, > - "Failed to idle engines, declaring wedged!\n"); > - GEM_TRACE_DUMP(); > - } > - > - /* Forcibly cancel outstanding work and leave the gpu quiet. */ > - i915_gem_set_wedged(i915); > - } > - > - i915_retire_requests(i915); /* ensure we flush after wedging */ > - return result; > -} > - > -static bool load_power_context(struct drm_i915_private *i915) > -{ > - /* Force loading the kernel context on all engines */ > - if (!switch_to_kernel_context_sync(i915, ALL_ENGINES)) > - return false; > - > - /* > - * Immediately park the GPU so that we enable powersaving and > - * treat it as idle. The next time we issue a request, we will > - * unpark and start using the engine->pinned_default_state, otherwise > - * it is in limbo and an early reset may fail. > - */ > - __i915_gem_park(i915); > - > - return true; > -} > - > -static void > -i915_gem_idle_work_handler(struct work_struct *work) > -{ > - struct drm_i915_private *i915 = > - container_of(work, typeof(*i915), gt.idle_work.work); > - bool rearm_hangcheck; > - > - if (!READ_ONCE(i915->gt.awake)) > - return; > - > - if (READ_ONCE(i915->gt.active_requests)) > - return; > - > - rearm_hangcheck = > - cancel_delayed_work_sync(&i915->gpu_error.hangcheck_work); > - > - if (!mutex_trylock(&i915->drm.struct_mutex)) { > - /* Currently busy, come back later */ > - mod_delayed_work(i915->wq, > - &i915->gt.idle_work, > - msecs_to_jiffies(50)); > - goto out_rearm; > - } > - > - /* > - * Flush out the last user context, leaving only the pinned > - * kernel context resident. Should anything unfortunate happen > - * while we are idle (such as the GPU being power cycled), no users > - * will be harmed. > - */ > - if (!work_pending(&i915->gt.idle_work.work) && > - !i915->gt.active_requests) { > - ++i915->gt.active_requests; /* don't requeue idle */ > - > - switch_to_kernel_context_sync(i915, i915->gt.active_engines); > - > - if (!--i915->gt.active_requests) { > - __i915_gem_park(i915); > - rearm_hangcheck = false; > - } > - } > - > - mutex_unlock(&i915->drm.struct_mutex); > - > -out_rearm: > - if (rearm_hangcheck) { > - GEM_BUG_ON(!i915->gt.awake); > - i915_queue_hangcheck(i915); > - } > -} > - > void i915_gem_close_object(struct drm_gem_object *gem, struct drm_file *file) > { > struct drm_i915_private *i915 = to_i915(gem->dev); > @@ -4412,133 +4187,6 @@ void i915_gem_sanitize(struct drm_i915_private *i915) > mutex_unlock(&i915->drm.struct_mutex); > } > > -void i915_gem_suspend(struct drm_i915_private *i915) > -{ > - intel_wakeref_t wakeref; > - > - GEM_TRACE("\n"); > - > - wakeref = intel_runtime_pm_get(i915); > - > - flush_workqueue(i915->wq); > - > - mutex_lock(&i915->drm.struct_mutex); > - > - /* > - * We have to flush all the executing contexts to main memory so > - * that they can saved in the hibernation image. To ensure the last > - * context image is coherent, we have to switch away from it. That > - * leaves the i915->kernel_context still active when > - * we actually suspend, and its image in memory may not match the GPU > - * state. Fortunately, the kernel_context is disposable and we do > - * not rely on its state. > - */ > - switch_to_kernel_context_sync(i915, i915->gt.active_engines); > - > - mutex_unlock(&i915->drm.struct_mutex); > - i915_reset_flush(i915); > - > - drain_delayed_work(&i915->gt.retire_work); > - > - /* > - * As the idle_work is rearming if it detects a race, play safe and > - * repeat the flush until it is definitely idle. > - */ > - drain_delayed_work(&i915->gt.idle_work); > - > - /* > - * Assert that we successfully flushed all the work and > - * reset the GPU back to its idle, low power state. > - */ > - GEM_BUG_ON(i915->gt.awake); > - > - intel_uc_suspend(i915); > - > - intel_runtime_pm_put(i915, wakeref); > -} > - > -void i915_gem_suspend_late(struct drm_i915_private *i915) > -{ > - struct drm_i915_gem_object *obj; > - struct list_head *phases[] = { > - &i915->mm.unbound_list, > - &i915->mm.bound_list, > - NULL > - }, **phase; > - > - /* > - * Neither the BIOS, ourselves or any other kernel > - * expects the system to be in execlists mode on startup, > - * so we need to reset the GPU back to legacy mode. And the only > - * known way to disable logical contexts is through a GPU reset. > - * > - * So in order to leave the system in a known default configuration, > - * always reset the GPU upon unload and suspend. Afterwards we then > - * clean up the GEM state tracking, flushing off the requests and > - * leaving the system in a known idle state. > - * > - * Note that is of the upmost importance that the GPU is idle and > - * all stray writes are flushed *before* we dismantle the backing > - * storage for the pinned objects. > - * > - * However, since we are uncertain that resetting the GPU on older > - * machines is a good idea, we don't - just in case it leaves the > - * machine in an unusable condition. > - */ > - > - mutex_lock(&i915->drm.struct_mutex); > - for (phase = phases; *phase; phase++) { > - list_for_each_entry(obj, *phase, mm.link) > - WARN_ON(i915_gem_object_set_to_gtt_domain(obj, false)); > - } > - mutex_unlock(&i915->drm.struct_mutex); > - > - intel_uc_sanitize(i915); > - i915_gem_sanitize(i915); > -} > - > -void i915_gem_resume(struct drm_i915_private *i915) > -{ > - GEM_TRACE("\n"); > - > - WARN_ON(i915->gt.awake); > - > - mutex_lock(&i915->drm.struct_mutex); > - intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL); > - > - i915_gem_restore_gtt_mappings(i915); > - i915_gem_restore_fences(i915); > - > - /* > - * As we didn't flush the kernel context before suspend, we cannot > - * guarantee that the context image is complete. So let's just reset > - * it and start again. > - */ > - intel_gt_resume(i915); > - > - if (i915_gem_init_hw(i915)) > - goto err_wedged; > - > - intel_uc_resume(i915); > - > - /* Always reload a context for powersaving. */ > - if (!load_power_context(i915)) > - goto err_wedged; > - > -out_unlock: > - intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); > - mutex_unlock(&i915->drm.struct_mutex); > - return; > - > -err_wedged: > - if (!i915_reset_failed(i915)) { > - dev_err(i915->drm.dev, > - "Failed to re-initialize GPU, declaring it wedged!\n"); > - i915_gem_set_wedged(i915); > - } > - goto out_unlock; > -} > - > void i915_gem_init_swizzling(struct drm_i915_private *dev_priv) > { > if (INTEL_GEN(dev_priv) < 5 || > @@ -4721,7 +4369,7 @@ static int __intel_engines_record_defaults(struct drm_i915_private *i915) > } > > /* Flush the default context image to memory, and enable powersaving. */ > - if (!load_power_context(i915)) { > + if (!i915_gem_load_power_context(i915)) { > err = -EIO; > goto err_active; > } > @@ -5115,11 +4763,8 @@ int i915_gem_init_early(struct drm_i915_private *dev_priv) > INIT_LIST_HEAD(&dev_priv->gt.closed_vma); > > i915_gem_init__mm(dev_priv); > + i915_gem_init__pm(dev_priv); > > - INIT_DELAYED_WORK(&dev_priv->gt.retire_work, > - i915_gem_retire_work_handler); > - INIT_DELAYED_WORK(&dev_priv->gt.idle_work, > - i915_gem_idle_work_handler); > init_waitqueue_head(&dev_priv->gpu_error.wait_queue); > init_waitqueue_head(&dev_priv->gpu_error.reset_queue); > mutex_init(&dev_priv->gpu_error.wedge_mutex); > diff --git a/drivers/gpu/drm/i915/i915_gem_pm.c b/drivers/gpu/drm/i915/i915_gem_pm.c > new file mode 100644 > index 000000000000..bc0663cac7b6 > --- /dev/null > +++ b/drivers/gpu/drm/i915/i915_gem_pm.c > @@ -0,0 +1,365 @@ > +/* > + * SPDX-License-Identifier: MIT > + * > + * Copyright © 2019 Intel Corporation > + */ > + > +#include "i915_drv.h" > +#include "i915_gem_pm.h" > +#include "i915_globals.h" > +#include "intel_pm.h" > + > +static void __i915_gem_park(struct drm_i915_private *i915) > +{ > + intel_wakeref_t wakeref; > + > + GEM_TRACE("\n"); > + > + lockdep_assert_held(&i915->drm.struct_mutex); > + GEM_BUG_ON(i915->gt.active_requests); > + GEM_BUG_ON(!list_empty(&i915->gt.active_rings)); > + > + if (!i915->gt.awake) > + return; > + > + /* > + * Be paranoid and flush a concurrent interrupt to make sure > + * we don't reactivate any irq tasklets after parking. > + * > + * FIXME: Note that even though we have waited for execlists to be idle, > + * there may still be an in-flight interrupt even though the CSB > + * is now empty. synchronize_irq() makes sure that a residual interrupt > + * is completed before we continue, but it doesn't prevent the HW from > + * raising a spurious interrupt later. To complete the shield we should > + * coordinate disabling the CS irq with flushing the interrupts. > + */ > + synchronize_irq(i915->drm.irq); > + > + intel_engines_park(i915); > + i915_timelines_park(i915); > + > + i915_pmu_gt_parked(i915); > + i915_vma_parked(i915); > + > + wakeref = fetch_and_zero(&i915->gt.awake); > + GEM_BUG_ON(!wakeref); > + > + if (INTEL_GEN(i915) >= 6) > + gen6_rps_idle(i915); > + > + intel_display_power_put(i915, POWER_DOMAIN_GT_IRQ, wakeref); > + > + i915_globals_park(); > +} > + > +static bool switch_to_kernel_context_sync(struct drm_i915_private *i915, > + unsigned long mask) > +{ > + bool result = true; > + > + /* > + * Even if we fail to switch, give whatever is running a small chance > + * to save itself before we report the failure. Yes, this may be a > + * false positive due to e.g. ENOMEM, caveat emptor! > + */ > + if (i915_gem_switch_to_kernel_context(i915, mask)) > + result = false; > + > + if (i915_gem_wait_for_idle(i915, > + I915_WAIT_LOCKED | > + I915_WAIT_FOR_IDLE_BOOST, > + I915_GEM_IDLE_TIMEOUT)) > + result = false; > + > + if (!result) { > + if (i915_modparams.reset) { /* XXX hide warning from gem_eio */ > + dev_err(i915->drm.dev, > + "Failed to idle engines, declaring wedged!\n"); > + GEM_TRACE_DUMP(); > + } > + > + /* Forcibly cancel outstanding work and leave the gpu quiet. */ > + i915_gem_set_wedged(i915); > + } > + > + i915_retire_requests(i915); /* ensure we flush after wedging */ > + return result; > +} > + > +static void idle_work_handler(struct work_struct *work) > +{ > + struct drm_i915_private *i915 = > + container_of(work, typeof(*i915), gem.idle_work.work); > + bool rearm_hangcheck; > + > + if (!READ_ONCE(i915->gt.awake)) > + return; > + > + if (READ_ONCE(i915->gt.active_requests)) > + return; > + > + rearm_hangcheck = > + cancel_delayed_work_sync(&i915->gpu_error.hangcheck_work); > + > + if (!mutex_trylock(&i915->drm.struct_mutex)) { > + /* Currently busy, come back later */ > + mod_delayed_work(i915->wq, > + &i915->gem.idle_work, > + msecs_to_jiffies(50)); > + goto out_rearm; > + } > + > + /* > + * Flush out the last user context, leaving only the pinned > + * kernel context resident. Should anything unfortunate happen > + * while we are idle (such as the GPU being power cycled), no users > + * will be harmed. > + */ > + if (!work_pending(&i915->gem.idle_work.work) && > + !i915->gt.active_requests) { > + ++i915->gt.active_requests; /* don't requeue idle */ > + > + switch_to_kernel_context_sync(i915, i915->gt.active_engines); > + > + if (!--i915->gt.active_requests) { > + __i915_gem_park(i915); > + rearm_hangcheck = false; > + } > + } > + > + mutex_unlock(&i915->drm.struct_mutex); > + > +out_rearm: > + if (rearm_hangcheck) { > + GEM_BUG_ON(!i915->gt.awake); > + i915_queue_hangcheck(i915); > + } > +} > + > +static void retire_work_handler(struct work_struct *work) > +{ > + struct drm_i915_private *i915 = > + container_of(work, typeof(*i915), gem.retire_work.work); > + > + /* Come back later if the device is busy... */ > + if (mutex_trylock(&i915->drm.struct_mutex)) { > + i915_retire_requests(i915); > + mutex_unlock(&i915->drm.struct_mutex); > + } > + > + /* > + * Keep the retire handler running until we are finally idle. > + * We do not need to do this test under locking as in the worst-case > + * we queue the retire worker once too often. > + */ > + if (READ_ONCE(i915->gt.awake)) > + queue_delayed_work(i915->wq, > + &i915->gem.retire_work, > + round_jiffies_up_relative(HZ)); > +} > + > +void i915_gem_park(struct drm_i915_private *i915) > +{ > + GEM_TRACE("\n"); > + > + lockdep_assert_held(&i915->drm.struct_mutex); > + GEM_BUG_ON(i915->gt.active_requests); > + > + if (!i915->gt.awake) > + return; > + > + /* Defer the actual call to __i915_gem_park() to prevent ping-pongs */ > + mod_delayed_work(i915->wq, &i915->gem.idle_work, msecs_to_jiffies(100)); > +} > + > +void i915_gem_unpark(struct drm_i915_private *i915) > +{ > + GEM_TRACE("\n"); > + > + lockdep_assert_held(&i915->drm.struct_mutex); > + GEM_BUG_ON(!i915->gt.active_requests); > + assert_rpm_wakelock_held(i915); > + > + if (i915->gt.awake) > + return; > + > + /* > + * It seems that the DMC likes to transition between the DC states a lot > + * when there are no connected displays (no active power domains) during > + * command submission. > + * > + * This activity has negative impact on the performance of the chip with > + * huge latencies observed in the interrupt handler and elsewhere. > + * > + * Work around it by grabbing a GT IRQ power domain whilst there is any > + * GT activity, preventing any DC state transitions. > + */ > + i915->gt.awake = intel_display_power_get(i915, POWER_DOMAIN_GT_IRQ); > + GEM_BUG_ON(!i915->gt.awake); > + > + i915_globals_unpark(); > + > + intel_enable_gt_powersave(i915); > + i915_update_gfx_val(i915); > + if (INTEL_GEN(i915) >= 6) > + gen6_rps_busy(i915); > + i915_pmu_gt_unparked(i915); > + > + intel_engines_unpark(i915); > + > + i915_queue_hangcheck(i915); > + > + queue_delayed_work(i915->wq, > + &i915->gem.retire_work, > + round_jiffies_up_relative(HZ)); > +} > + > +bool i915_gem_load_power_context(struct drm_i915_private *i915) > +{ > + /* Force loading the kernel context on all engines */ > + if (!switch_to_kernel_context_sync(i915, ALL_ENGINES)) > + return false; > + > + /* > + * Immediately park the GPU so that we enable powersaving and > + * treat it as idle. The next time we issue a request, we will > + * unpark and start using the engine->pinned_default_state, otherwise > + * it is in limbo and an early reset may fail. > + */ > + __i915_gem_park(i915); > + > + return true; > +} > + > +void i915_gem_suspend(struct drm_i915_private *i915) > +{ > + intel_wakeref_t wakeref; > + > + GEM_TRACE("\n"); > + > + wakeref = intel_runtime_pm_get(i915); > + > + flush_workqueue(i915->wq); > + > + mutex_lock(&i915->drm.struct_mutex); > + > + /* > + * We have to flush all the executing contexts to main memory so > + * that they can saved in the hibernation image. To ensure the last > + * context image is coherent, we have to switch away from it. That > + * leaves the i915->kernel_context still active when > + * we actually suspend, and its image in memory may not match the GPU > + * state. Fortunately, the kernel_context is disposable and we do > + * not rely on its state. > + */ > + switch_to_kernel_context_sync(i915, i915->gt.active_engines); > + > + mutex_unlock(&i915->drm.struct_mutex); > + i915_reset_flush(i915); > + > + drain_delayed_work(&i915->gem.retire_work); > + > + /* > + * As the idle_work is rearming if it detects a race, play safe and > + * repeat the flush until it is definitely idle. > + */ > + drain_delayed_work(&i915->gem.idle_work); > + > + /* > + * Assert that we successfully flushed all the work and > + * reset the GPU back to its idle, low power state. > + */ > + GEM_BUG_ON(i915->gt.awake); > + > + intel_uc_suspend(i915); > + > + intel_runtime_pm_put(i915, wakeref); > +} > + > +void i915_gem_suspend_late(struct drm_i915_private *i915) > +{ > + struct drm_i915_gem_object *obj; > + struct list_head *phases[] = { > + &i915->mm.unbound_list, > + &i915->mm.bound_list, > + NULL > + }, **phase; > + > + /* > + * Neither the BIOS, ourselves or any other kernel > + * expects the system to be in execlists mode on startup, > + * so we need to reset the GPU back to legacy mode. And the only > + * known way to disable logical contexts is through a GPU reset. > + * > + * So in order to leave the system in a known default configuration, > + * always reset the GPU upon unload and suspend. Afterwards we then > + * clean up the GEM state tracking, flushing off the requests and > + * leaving the system in a known idle state. > + * > + * Note that is of the upmost importance that the GPU is idle and > + * all stray writes are flushed *before* we dismantle the backing > + * storage for the pinned objects. > + * > + * However, since we are uncertain that resetting the GPU on older > + * machines is a good idea, we don't - just in case it leaves the > + * machine in an unusable condition. > + */ > + > + mutex_lock(&i915->drm.struct_mutex); > + for (phase = phases; *phase; phase++) { > + list_for_each_entry(obj, *phase, mm.link) > + WARN_ON(i915_gem_object_set_to_gtt_domain(obj, false)); > + } > + mutex_unlock(&i915->drm.struct_mutex); > + > + intel_uc_sanitize(i915); > + i915_gem_sanitize(i915); > +} > + > +void i915_gem_resume(struct drm_i915_private *i915) > +{ > + GEM_TRACE("\n"); > + > + WARN_ON(i915->gt.awake); > + > + mutex_lock(&i915->drm.struct_mutex); > + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL); > + > + i915_gem_restore_gtt_mappings(i915); > + i915_gem_restore_fences(i915); > + > + /* > + * As we didn't flush the kernel context before suspend, we cannot > + * guarantee that the context image is complete. So let's just reset > + * it and start again. > + */ > + intel_gt_resume(i915); > + > + if (i915_gem_init_hw(i915)) > + goto err_wedged; > + > + intel_uc_resume(i915); > + > + /* Always reload a context for powersaving. */ > + if (!i915_gem_load_power_context(i915)) > + goto err_wedged; > + > +out_unlock: > + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); > + mutex_unlock(&i915->drm.struct_mutex); > + return; > + > +err_wedged: > + if (!i915_reset_failed(i915)) { > + dev_err(i915->drm.dev, > + "Failed to re-initialize GPU, declaring it wedged!\n"); > + i915_gem_set_wedged(i915); > + } > + goto out_unlock; > +} > + > +void i915_gem_init__pm(struct drm_i915_private *i915) > +{ > + INIT_DELAYED_WORK(&i915->gem.idle_work, idle_work_handler); > + INIT_DELAYED_WORK(&i915->gem.retire_work, retire_work_handler); > +} > diff --git a/drivers/gpu/drm/i915/i915_gem_pm.h b/drivers/gpu/drm/i915/i915_gem_pm.h > new file mode 100644 > index 000000000000..52f65e3f06b5 > --- /dev/null > +++ b/drivers/gpu/drm/i915/i915_gem_pm.h > @@ -0,0 +1,28 @@ > +/* > + * SPDX-License-Identifier: MIT > + * > + * Copyright © 2019 Intel Corporation > + */ > + > +#ifndef __I915_GEM_PM_H__ > +#define __I915_GEM_PM_H__ > + > +#include <linux/types.h> > + > +struct drm_i915_private; > +struct work_struct; > + > +void i915_gem_init__pm(struct drm_i915_private *i915); > + > +bool i915_gem_load_power_context(struct drm_i915_private *i915); > +void i915_gem_resume(struct drm_i915_private *i915); > + > +void i915_gem_unpark(struct drm_i915_private *i915); > +void i915_gem_park(struct drm_i915_private *i915); > + > +void i915_gem_idle_work_handler(struct work_struct *work); > + > +void i915_gem_suspend(struct drm_i915_private *i915); > +void i915_gem_suspend_late(struct drm_i915_private *i915); > + > +#endif /* __I915_GEM_PM_H__ */ > diff --git a/drivers/gpu/drm/i915/selftests/i915_gem_context.c b/drivers/gpu/drm/i915/selftests/i915_gem_context.c > index 6f52ca881173..9d646fa1b74e 100644 > --- a/drivers/gpu/drm/i915/selftests/i915_gem_context.c > +++ b/drivers/gpu/drm/i915/selftests/i915_gem_context.c > @@ -1658,7 +1658,7 @@ static int __igt_switch_to_kernel_context(struct drm_i915_private *i915, > /* XXX Bonus points for proving we are the kernel context! */ > > mutex_unlock(&i915->drm.struct_mutex); > - drain_delayed_work(&i915->gt.idle_work); > + drain_delayed_work(&i915->gem.idle_work); > mutex_lock(&i915->drm.struct_mutex); > } > > diff --git a/drivers/gpu/drm/i915/selftests/i915_gem_object.c b/drivers/gpu/drm/i915/selftests/i915_gem_object.c > index 971148fbe6f5..12fc53c694a6 100644 > --- a/drivers/gpu/drm/i915/selftests/i915_gem_object.c > +++ b/drivers/gpu/drm/i915/selftests/i915_gem_object.c > @@ -514,8 +514,8 @@ static void disable_retire_worker(struct drm_i915_private *i915) > } > mutex_unlock(&i915->drm.struct_mutex); > > - cancel_delayed_work_sync(&i915->gt.retire_work); > - cancel_delayed_work_sync(&i915->gt.idle_work); > + cancel_delayed_work_sync(&i915->gem.retire_work); > + cancel_delayed_work_sync(&i915->gem.idle_work); > } > > static int igt_mmap_offset_exhaustion(void *arg) > @@ -617,9 +617,9 @@ static int igt_mmap_offset_exhaustion(void *arg) > out_park: > mutex_lock(&i915->drm.struct_mutex); > if (--i915->gt.active_requests) > - queue_delayed_work(i915->wq, &i915->gt.retire_work, 0); > + queue_delayed_work(i915->wq, &i915->gem.retire_work, 0); > else > - queue_delayed_work(i915->wq, &i915->gt.idle_work, 0); > + queue_delayed_work(i915->wq, &i915->gem.idle_work, 0); > mutex_unlock(&i915->drm.struct_mutex); > i915_gem_shrinker_register(i915); > return err; > diff --git a/drivers/gpu/drm/i915/selftests/mock_gem_device.c b/drivers/gpu/drm/i915/selftests/mock_gem_device.c > index f444ee5add27..fb677b4019a0 100644 > --- a/drivers/gpu/drm/i915/selftests/mock_gem_device.c > +++ b/drivers/gpu/drm/i915/selftests/mock_gem_device.c > @@ -59,8 +59,8 @@ static void mock_device_release(struct drm_device *dev) > i915_gem_contexts_lost(i915); > mutex_unlock(&i915->drm.struct_mutex); > > - drain_delayed_work(&i915->gt.retire_work); > - drain_delayed_work(&i915->gt.idle_work); > + drain_delayed_work(&i915->gem.retire_work); > + drain_delayed_work(&i915->gem.idle_work); > i915_gem_drain_workqueue(i915); > > mutex_lock(&i915->drm.struct_mutex); > @@ -111,7 +111,7 @@ static void mock_retire_work_handler(struct work_struct *work) > static void mock_idle_work_handler(struct work_struct *work) > { > struct drm_i915_private *i915 = > - container_of(work, typeof(*i915), gt.idle_work.work); > + container_of(work, typeof(*i915), gem.idle_work.work); > > i915->gt.active_engines = 0; > } > @@ -197,8 +197,8 @@ struct drm_i915_private *mock_gem_device(void) > > mock_init_contexts(i915); > > - INIT_DELAYED_WORK(&i915->gt.retire_work, mock_retire_work_handler); > - INIT_DELAYED_WORK(&i915->gt.idle_work, mock_idle_work_handler); > + INIT_DELAYED_WORK(&i915->gem.retire_work, mock_retire_work_handler); > + INIT_DELAYED_WORK(&i915->gem.idle_work, mock_idle_work_handler); > > i915->gt.awake = true; > >
Quoting Tvrtko Ursulin (2019-04-08 15:56:39) > > On 08/04/2019 10:17, Chris Wilson wrote: > > Split out the powermanagement portion (GT wakeref, suspend/resume) of > > GEM from i915_gem.c into its own file. > > This is the same patch I gave r-b on before? Judging from the title and > commit message I'd say yes. But I know it is not, don't worry. We are in a different series where this step may or may not make sense. -Chris
On 08/04/2019 10:17, Chris Wilson wrote: > Split out the powermanagement portion (GT wakeref, suspend/resume) of > GEM from i915_gem.c into its own file. > > Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk> > --- > drivers/gpu/drm/i915/Makefile | 1 + > drivers/gpu/drm/i915/Makefile.header-test | 1 + > drivers/gpu/drm/i915/i915_debugfs.c | 4 +- > drivers/gpu/drm/i915/i915_drv.h | 12 +- > drivers/gpu/drm/i915/i915_gem.c | 363 +---------------- > drivers/gpu/drm/i915/i915_gem_pm.c | 365 ++++++++++++++++++ > drivers/gpu/drm/i915/i915_gem_pm.h | 28 ++ > .../gpu/drm/i915/selftests/i915_gem_context.c | 2 +- > .../gpu/drm/i915/selftests/i915_gem_object.c | 8 +- > .../gpu/drm/i915/selftests/mock_gem_device.c | 10 +- > 10 files changed, 418 insertions(+), 376 deletions(-) > create mode 100644 drivers/gpu/drm/i915/i915_gem_pm.c > create mode 100644 drivers/gpu/drm/i915/i915_gem_pm.h > > diff --git a/drivers/gpu/drm/i915/Makefile b/drivers/gpu/drm/i915/Makefile > index 233bad5e361f..858642c7bc40 100644 > --- a/drivers/gpu/drm/i915/Makefile > +++ b/drivers/gpu/drm/i915/Makefile > @@ -97,6 +97,7 @@ i915-y += \ > i915_gem_internal.o \ > i915_gem.o \ > i915_gem_object.o \ > + i915_gem_pm.o \ > i915_gem_render_state.o \ > i915_gem_shrinker.o \ > i915_gem_stolen.o \ > diff --git a/drivers/gpu/drm/i915/Makefile.header-test b/drivers/gpu/drm/i915/Makefile.header-test > index e6b3e7588860..702e3a7ade4c 100644 > --- a/drivers/gpu/drm/i915/Makefile.header-test > +++ b/drivers/gpu/drm/i915/Makefile.header-test > @@ -5,6 +5,7 @@ > header_test := \ > i915_active_types.h \ > i915_gem_context_types.h \ > + i915_gem_pm.h \ > i915_priolist_types.h \ > i915_scheduler_types.h \ > i915_timeline_types.h \ > diff --git a/drivers/gpu/drm/i915/i915_debugfs.c b/drivers/gpu/drm/i915/i915_debugfs.c > index eac4095278b2..e3eca88614b2 100644 > --- a/drivers/gpu/drm/i915/i915_debugfs.c > +++ b/drivers/gpu/drm/i915/i915_debugfs.c > @@ -3941,8 +3941,8 @@ i915_drop_caches_set(void *data, u64 val) > if (val & DROP_IDLE) { > do { > if (READ_ONCE(i915->gt.active_requests)) > - flush_delayed_work(&i915->gt.retire_work); > - drain_delayed_work(&i915->gt.idle_work); > + flush_delayed_work(&i915->gem.retire_work); > + drain_delayed_work(&i915->gem.idle_work); > } while (READ_ONCE(i915->gt.awake)); > } > > diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h > index 62a7e91acd7f..cbae9be052e0 100644 > --- a/drivers/gpu/drm/i915/i915_drv.h > +++ b/drivers/gpu/drm/i915/i915_drv.h > @@ -2020,6 +2020,12 @@ struct drm_i915_private { > */ > intel_wakeref_t awake; > > + ktime_t last_init_time; > + > + struct i915_vma *scratch; > + } gt; > + > + struct { > /** > * We leave the user IRQ off as much as possible, > * but this means that requests will finish and never > @@ -2037,11 +2043,7 @@ struct drm_i915_private { > * off the idle_work. > */ > struct delayed_work idle_work; > - > - ktime_t last_init_time; > - > - struct i915_vma *scratch; > - } gt; > + } gem; > > /* For i945gm vblank irq vs. C3 workaround */ > struct { > diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c > index a8b0aa05288b..c56633e87c32 100644 > --- a/drivers/gpu/drm/i915/i915_gem.c > +++ b/drivers/gpu/drm/i915/i915_gem.c > @@ -46,7 +46,7 @@ > #include "i915_drv.h" > #include "i915_gem_clflush.h" > #include "i915_gemfs.h" > -#include "i915_globals.h" > +#include "i915_gem_pm.h" > #include "i915_trace.h" > #include "i915_vgpu.h" > > @@ -103,105 +103,6 @@ static void i915_gem_info_remove_obj(struct drm_i915_private *dev_priv, > spin_unlock(&dev_priv->mm.object_stat_lock); > } > > -static void __i915_gem_park(struct drm_i915_private *i915) > -{ > - intel_wakeref_t wakeref; > - > - GEM_TRACE("\n"); > - > - lockdep_assert_held(&i915->drm.struct_mutex); > - GEM_BUG_ON(i915->gt.active_requests); > - GEM_BUG_ON(!list_empty(&i915->gt.active_rings)); > - > - if (!i915->gt.awake) > - return; > - > - /* > - * Be paranoid and flush a concurrent interrupt to make sure > - * we don't reactivate any irq tasklets after parking. > - * > - * FIXME: Note that even though we have waited for execlists to be idle, > - * there may still be an in-flight interrupt even though the CSB > - * is now empty. synchronize_irq() makes sure that a residual interrupt > - * is completed before we continue, but it doesn't prevent the HW from > - * raising a spurious interrupt later. To complete the shield we should > - * coordinate disabling the CS irq with flushing the interrupts. > - */ > - synchronize_irq(i915->drm.irq); > - > - intel_engines_park(i915); > - i915_timelines_park(i915); > - > - i915_pmu_gt_parked(i915); > - i915_vma_parked(i915); > - > - wakeref = fetch_and_zero(&i915->gt.awake); > - GEM_BUG_ON(!wakeref); > - > - if (INTEL_GEN(i915) >= 6) > - gen6_rps_idle(i915); > - > - intel_display_power_put(i915, POWER_DOMAIN_GT_IRQ, wakeref); > - > - i915_globals_park(); > -} > - > -void i915_gem_park(struct drm_i915_private *i915) > -{ > - GEM_TRACE("\n"); > - > - lockdep_assert_held(&i915->drm.struct_mutex); > - GEM_BUG_ON(i915->gt.active_requests); > - > - if (!i915->gt.awake) > - return; > - > - /* Defer the actual call to __i915_gem_park() to prevent ping-pongs */ > - mod_delayed_work(i915->wq, &i915->gt.idle_work, msecs_to_jiffies(100)); > -} > - > -void i915_gem_unpark(struct drm_i915_private *i915) > -{ > - GEM_TRACE("\n"); > - > - lockdep_assert_held(&i915->drm.struct_mutex); > - GEM_BUG_ON(!i915->gt.active_requests); > - assert_rpm_wakelock_held(i915); > - > - if (i915->gt.awake) > - return; > - > - /* > - * It seems that the DMC likes to transition between the DC states a lot > - * when there are no connected displays (no active power domains) during > - * command submission. > - * > - * This activity has negative impact on the performance of the chip with > - * huge latencies observed in the interrupt handler and elsewhere. > - * > - * Work around it by grabbing a GT IRQ power domain whilst there is any > - * GT activity, preventing any DC state transitions. > - */ > - i915->gt.awake = intel_display_power_get(i915, POWER_DOMAIN_GT_IRQ); > - GEM_BUG_ON(!i915->gt.awake); > - > - i915_globals_unpark(); > - > - intel_enable_gt_powersave(i915); > - i915_update_gfx_val(i915); > - if (INTEL_GEN(i915) >= 6) > - gen6_rps_busy(i915); > - i915_pmu_gt_unparked(i915); > - > - intel_engines_unpark(i915); > - > - i915_queue_hangcheck(i915); > - > - queue_delayed_work(i915->wq, > - &i915->gt.retire_work, > - round_jiffies_up_relative(HZ)); > -} > - > int > i915_gem_get_aperture_ioctl(struct drm_device *dev, void *data, > struct drm_file *file) > @@ -2088,7 +1989,7 @@ static int i915_gem_object_create_mmap_offset(struct drm_i915_gem_object *obj) > if (!err) > break; > > - } while (flush_delayed_work(&dev_priv->gt.retire_work)); > + } while (flush_delayed_work(&dev_priv->gem.retire_work)); > > return err; > } > @@ -2871,132 +2772,6 @@ i915_gem_object_pwrite_gtt(struct drm_i915_gem_object *obj, > return 0; > } > > -static void > -i915_gem_retire_work_handler(struct work_struct *work) > -{ > - struct drm_i915_private *dev_priv = > - container_of(work, typeof(*dev_priv), gt.retire_work.work); > - struct drm_device *dev = &dev_priv->drm; > - > - /* Come back later if the device is busy... */ > - if (mutex_trylock(&dev->struct_mutex)) { > - i915_retire_requests(dev_priv); > - mutex_unlock(&dev->struct_mutex); > - } > - > - /* > - * Keep the retire handler running until we are finally idle. > - * We do not need to do this test under locking as in the worst-case > - * we queue the retire worker once too often. > - */ > - if (READ_ONCE(dev_priv->gt.awake)) > - queue_delayed_work(dev_priv->wq, > - &dev_priv->gt.retire_work, > - round_jiffies_up_relative(HZ)); > -} > - > -static bool switch_to_kernel_context_sync(struct drm_i915_private *i915, > - unsigned long mask) > -{ > - bool result = true; > - > - /* > - * Even if we fail to switch, give whatever is running a small chance > - * to save itself before we report the failure. Yes, this may be a > - * false positive due to e.g. ENOMEM, caveat emptor! > - */ > - if (i915_gem_switch_to_kernel_context(i915, mask)) > - result = false; > - > - if (i915_gem_wait_for_idle(i915, > - I915_WAIT_LOCKED | > - I915_WAIT_FOR_IDLE_BOOST, > - I915_GEM_IDLE_TIMEOUT)) > - result = false; > - > - if (!result) { > - if (i915_modparams.reset) { /* XXX hide warning from gem_eio */ > - dev_err(i915->drm.dev, > - "Failed to idle engines, declaring wedged!\n"); > - GEM_TRACE_DUMP(); > - } > - > - /* Forcibly cancel outstanding work and leave the gpu quiet. */ > - i915_gem_set_wedged(i915); > - } > - > - i915_retire_requests(i915); /* ensure we flush after wedging */ > - return result; > -} > - > -static bool load_power_context(struct drm_i915_private *i915) > -{ > - /* Force loading the kernel context on all engines */ > - if (!switch_to_kernel_context_sync(i915, ALL_ENGINES)) > - return false; > - > - /* > - * Immediately park the GPU so that we enable powersaving and > - * treat it as idle. The next time we issue a request, we will > - * unpark and start using the engine->pinned_default_state, otherwise > - * it is in limbo and an early reset may fail. > - */ > - __i915_gem_park(i915); > - > - return true; > -} > - > -static void > -i915_gem_idle_work_handler(struct work_struct *work) > -{ > - struct drm_i915_private *i915 = > - container_of(work, typeof(*i915), gt.idle_work.work); > - bool rearm_hangcheck; > - > - if (!READ_ONCE(i915->gt.awake)) > - return; > - > - if (READ_ONCE(i915->gt.active_requests)) > - return; > - > - rearm_hangcheck = > - cancel_delayed_work_sync(&i915->gpu_error.hangcheck_work); > - > - if (!mutex_trylock(&i915->drm.struct_mutex)) { > - /* Currently busy, come back later */ > - mod_delayed_work(i915->wq, > - &i915->gt.idle_work, > - msecs_to_jiffies(50)); > - goto out_rearm; > - } > - > - /* > - * Flush out the last user context, leaving only the pinned > - * kernel context resident. Should anything unfortunate happen > - * while we are idle (such as the GPU being power cycled), no users > - * will be harmed. > - */ > - if (!work_pending(&i915->gt.idle_work.work) && > - !i915->gt.active_requests) { > - ++i915->gt.active_requests; /* don't requeue idle */ > - > - switch_to_kernel_context_sync(i915, i915->gt.active_engines); > - > - if (!--i915->gt.active_requests) { > - __i915_gem_park(i915); > - rearm_hangcheck = false; > - } > - } > - > - mutex_unlock(&i915->drm.struct_mutex); > - > -out_rearm: > - if (rearm_hangcheck) { > - GEM_BUG_ON(!i915->gt.awake); > - i915_queue_hangcheck(i915); > - } > -} > - > void i915_gem_close_object(struct drm_gem_object *gem, struct drm_file *file) > { > struct drm_i915_private *i915 = to_i915(gem->dev); > @@ -4412,133 +4187,6 @@ void i915_gem_sanitize(struct drm_i915_private *i915) > mutex_unlock(&i915->drm.struct_mutex); > } > > -void i915_gem_suspend(struct drm_i915_private *i915) > -{ > - intel_wakeref_t wakeref; > - > - GEM_TRACE("\n"); > - > - wakeref = intel_runtime_pm_get(i915); > - > - flush_workqueue(i915->wq); > - > - mutex_lock(&i915->drm.struct_mutex); > - > - /* > - * We have to flush all the executing contexts to main memory so > - * that they can saved in the hibernation image. To ensure the last > - * context image is coherent, we have to switch away from it. That > - * leaves the i915->kernel_context still active when > - * we actually suspend, and its image in memory may not match the GPU > - * state. Fortunately, the kernel_context is disposable and we do > - * not rely on its state. > - */ > - switch_to_kernel_context_sync(i915, i915->gt.active_engines); > - > - mutex_unlock(&i915->drm.struct_mutex); > - i915_reset_flush(i915); > - > - drain_delayed_work(&i915->gt.retire_work); > - > - /* > - * As the idle_work is rearming if it detects a race, play safe and > - * repeat the flush until it is definitely idle. > - */ > - drain_delayed_work(&i915->gt.idle_work); > - > - /* > - * Assert that we successfully flushed all the work and > - * reset the GPU back to its idle, low power state. > - */ > - GEM_BUG_ON(i915->gt.awake); > - > - intel_uc_suspend(i915); > - > - intel_runtime_pm_put(i915, wakeref); > -} > - > -void i915_gem_suspend_late(struct drm_i915_private *i915) > -{ > - struct drm_i915_gem_object *obj; > - struct list_head *phases[] = { > - &i915->mm.unbound_list, > - &i915->mm.bound_list, > - NULL > - }, **phase; > - > - /* > - * Neither the BIOS, ourselves or any other kernel > - * expects the system to be in execlists mode on startup, > - * so we need to reset the GPU back to legacy mode. And the only > - * known way to disable logical contexts is through a GPU reset. > - * > - * So in order to leave the system in a known default configuration, > - * always reset the GPU upon unload and suspend. Afterwards we then > - * clean up the GEM state tracking, flushing off the requests and > - * leaving the system in a known idle state. > - * > - * Note that is of the upmost importance that the GPU is idle and > - * all stray writes are flushed *before* we dismantle the backing > - * storage for the pinned objects. > - * > - * However, since we are uncertain that resetting the GPU on older > - * machines is a good idea, we don't - just in case it leaves the > - * machine in an unusable condition. > - */ > - > - mutex_lock(&i915->drm.struct_mutex); > - for (phase = phases; *phase; phase++) { > - list_for_each_entry(obj, *phase, mm.link) > - WARN_ON(i915_gem_object_set_to_gtt_domain(obj, false)); > - } > - mutex_unlock(&i915->drm.struct_mutex); > - > - intel_uc_sanitize(i915); > - i915_gem_sanitize(i915); > -} > - > -void i915_gem_resume(struct drm_i915_private *i915) > -{ > - GEM_TRACE("\n"); > - > - WARN_ON(i915->gt.awake); > - > - mutex_lock(&i915->drm.struct_mutex); > - intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL); > - > - i915_gem_restore_gtt_mappings(i915); > - i915_gem_restore_fences(i915); > - > - /* > - * As we didn't flush the kernel context before suspend, we cannot > - * guarantee that the context image is complete. So let's just reset > - * it and start again. > - */ > - intel_gt_resume(i915); > - > - if (i915_gem_init_hw(i915)) > - goto err_wedged; > - > - intel_uc_resume(i915); > - > - /* Always reload a context for powersaving. */ > - if (!load_power_context(i915)) > - goto err_wedged; > - > -out_unlock: > - intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); > - mutex_unlock(&i915->drm.struct_mutex); > - return; > - > -err_wedged: > - if (!i915_reset_failed(i915)) { > - dev_err(i915->drm.dev, > - "Failed to re-initialize GPU, declaring it wedged!\n"); > - i915_gem_set_wedged(i915); > - } > - goto out_unlock; > -} > - > void i915_gem_init_swizzling(struct drm_i915_private *dev_priv) > { > if (INTEL_GEN(dev_priv) < 5 || > @@ -4721,7 +4369,7 @@ static int __intel_engines_record_defaults(struct drm_i915_private *i915) > } > > /* Flush the default context image to memory, and enable powersaving. */ > - if (!load_power_context(i915)) { > + if (!i915_gem_load_power_context(i915)) { > err = -EIO; > goto err_active; > } > @@ -5115,11 +4763,8 @@ int i915_gem_init_early(struct drm_i915_private *dev_priv) > INIT_LIST_HEAD(&dev_priv->gt.closed_vma); > > i915_gem_init__mm(dev_priv); > + i915_gem_init__pm(dev_priv); > > - INIT_DELAYED_WORK(&dev_priv->gt.retire_work, > - i915_gem_retire_work_handler); > - INIT_DELAYED_WORK(&dev_priv->gt.idle_work, > - i915_gem_idle_work_handler); > init_waitqueue_head(&dev_priv->gpu_error.wait_queue); > init_waitqueue_head(&dev_priv->gpu_error.reset_queue); > mutex_init(&dev_priv->gpu_error.wedge_mutex); > diff --git a/drivers/gpu/drm/i915/i915_gem_pm.c b/drivers/gpu/drm/i915/i915_gem_pm.c > new file mode 100644 > index 000000000000..bc0663cac7b6 > --- /dev/null > +++ b/drivers/gpu/drm/i915/i915_gem_pm.c > @@ -0,0 +1,365 @@ > +/* > + * SPDX-License-Identifier: MIT > + * > + * Copyright © 2019 Intel Corporation > + */ > + > +#include "i915_drv.h" > +#include "i915_gem_pm.h" > +#include "i915_globals.h" > +#include "intel_pm.h" > + > +static void __i915_gem_park(struct drm_i915_private *i915) > +{ > + intel_wakeref_t wakeref; > + > + GEM_TRACE("\n"); > + > + lockdep_assert_held(&i915->drm.struct_mutex); > + GEM_BUG_ON(i915->gt.active_requests); > + GEM_BUG_ON(!list_empty(&i915->gt.active_rings)); > + > + if (!i915->gt.awake) > + return; > + > + /* > + * Be paranoid and flush a concurrent interrupt to make sure > + * we don't reactivate any irq tasklets after parking. > + * > + * FIXME: Note that even though we have waited for execlists to be idle, > + * there may still be an in-flight interrupt even though the CSB > + * is now empty. synchronize_irq() makes sure that a residual interrupt > + * is completed before we continue, but it doesn't prevent the HW from > + * raising a spurious interrupt later. To complete the shield we should > + * coordinate disabling the CS irq with flushing the interrupts. > + */ > + synchronize_irq(i915->drm.irq); > + > + intel_engines_park(i915); > + i915_timelines_park(i915); > + > + i915_pmu_gt_parked(i915); > + i915_vma_parked(i915); > + > + wakeref = fetch_and_zero(&i915->gt.awake); > + GEM_BUG_ON(!wakeref); > + > + if (INTEL_GEN(i915) >= 6) > + gen6_rps_idle(i915); > + > + intel_display_power_put(i915, POWER_DOMAIN_GT_IRQ, wakeref); > + > + i915_globals_park(); > +} > + > +static bool switch_to_kernel_context_sync(struct drm_i915_private *i915, > + unsigned long mask) > +{ > + bool result = true; > + > + /* > + * Even if we fail to switch, give whatever is running a small chance > + * to save itself before we report the failure. Yes, this may be a > + * false positive due to e.g. ENOMEM, caveat emptor! > + */ > + if (i915_gem_switch_to_kernel_context(i915, mask)) > + result = false; > + > + if (i915_gem_wait_for_idle(i915, > + I915_WAIT_LOCKED | > + I915_WAIT_FOR_IDLE_BOOST, > + I915_GEM_IDLE_TIMEOUT)) > + result = false; > + > + if (!result) { > + if (i915_modparams.reset) { /* XXX hide warning from gem_eio */ > + dev_err(i915->drm.dev, > + "Failed to idle engines, declaring wedged!\n"); > + GEM_TRACE_DUMP(); > + } > + > + /* Forcibly cancel outstanding work and leave the gpu quiet. */ > + i915_gem_set_wedged(i915); > + } > + > + i915_retire_requests(i915); /* ensure we flush after wedging */ > + return result; > +} > + > +static void idle_work_handler(struct work_struct *work) > +{ > + struct drm_i915_private *i915 = > + container_of(work, typeof(*i915), gem.idle_work.work); > + bool rearm_hangcheck; > + > + if (!READ_ONCE(i915->gt.awake)) > + return; > + > + if (READ_ONCE(i915->gt.active_requests)) > + return; > + > + rearm_hangcheck = > + cancel_delayed_work_sync(&i915->gpu_error.hangcheck_work); > + > + if (!mutex_trylock(&i915->drm.struct_mutex)) { > + /* Currently busy, come back later */ > + mod_delayed_work(i915->wq, > + &i915->gem.idle_work, > + msecs_to_jiffies(50)); > + goto out_rearm; > + } > + > + /* > + * Flush out the last user context, leaving only the pinned > + * kernel context resident. Should anything unfortunate happen > + * while we are idle (such as the GPU being power cycled), no users > + * will be harmed. > + */ > + if (!work_pending(&i915->gem.idle_work.work) && > + !i915->gt.active_requests) { > + ++i915->gt.active_requests; /* don't requeue idle */ > + > + switch_to_kernel_context_sync(i915, i915->gt.active_engines); > + > + if (!--i915->gt.active_requests) { > + __i915_gem_park(i915); > + rearm_hangcheck = false; > + } > + } > + > + mutex_unlock(&i915->drm.struct_mutex); > + > +out_rearm: > + if (rearm_hangcheck) { > + GEM_BUG_ON(!i915->gt.awake); > + i915_queue_hangcheck(i915); > + } > +} > + > +static void retire_work_handler(struct work_struct *work) > +{ > + struct drm_i915_private *i915 = > + container_of(work, typeof(*i915), gem.retire_work.work); > + > + /* Come back later if the device is busy... */ > + if (mutex_trylock(&i915->drm.struct_mutex)) { > + i915_retire_requests(i915); > + mutex_unlock(&i915->drm.struct_mutex); > + } > + > + /* > + * Keep the retire handler running until we are finally idle. > + * We do not need to do this test under locking as in the worst-case > + * we queue the retire worker once too often. > + */ > + if (READ_ONCE(i915->gt.awake)) > + queue_delayed_work(i915->wq, > + &i915->gem.retire_work, > + round_jiffies_up_relative(HZ)); > +} > + > +void i915_gem_park(struct drm_i915_private *i915) > +{ > + GEM_TRACE("\n"); > + > + lockdep_assert_held(&i915->drm.struct_mutex); > + GEM_BUG_ON(i915->gt.active_requests); > + > + if (!i915->gt.awake) > + return; > + > + /* Defer the actual call to __i915_gem_park() to prevent ping-pongs */ > + mod_delayed_work(i915->wq, &i915->gem.idle_work, msecs_to_jiffies(100)); > +} > + > +void i915_gem_unpark(struct drm_i915_private *i915) > +{ > + GEM_TRACE("\n"); > + > + lockdep_assert_held(&i915->drm.struct_mutex); > + GEM_BUG_ON(!i915->gt.active_requests); > + assert_rpm_wakelock_held(i915); > + > + if (i915->gt.awake) > + return; > + > + /* > + * It seems that the DMC likes to transition between the DC states a lot > + * when there are no connected displays (no active power domains) during > + * command submission. > + * > + * This activity has negative impact on the performance of the chip with > + * huge latencies observed in the interrupt handler and elsewhere. > + * > + * Work around it by grabbing a GT IRQ power domain whilst there is any > + * GT activity, preventing any DC state transitions. > + */ > + i915->gt.awake = intel_display_power_get(i915, POWER_DOMAIN_GT_IRQ); > + GEM_BUG_ON(!i915->gt.awake); > + > + i915_globals_unpark(); > + > + intel_enable_gt_powersave(i915); > + i915_update_gfx_val(i915); > + if (INTEL_GEN(i915) >= 6) > + gen6_rps_busy(i915); > + i915_pmu_gt_unparked(i915); > + > + intel_engines_unpark(i915); > + > + i915_queue_hangcheck(i915); > + > + queue_delayed_work(i915->wq, > + &i915->gem.retire_work, > + round_jiffies_up_relative(HZ)); > +} > + > +bool i915_gem_load_power_context(struct drm_i915_private *i915) > +{ > + /* Force loading the kernel context on all engines */ > + if (!switch_to_kernel_context_sync(i915, ALL_ENGINES)) > + return false; > + > + /* > + * Immediately park the GPU so that we enable powersaving and > + * treat it as idle. The next time we issue a request, we will > + * unpark and start using the engine->pinned_default_state, otherwise > + * it is in limbo and an early reset may fail. > + */ > + __i915_gem_park(i915); > + > + return true; > +} > + > +void i915_gem_suspend(struct drm_i915_private *i915) > +{ > + intel_wakeref_t wakeref; > + > + GEM_TRACE("\n"); > + > + wakeref = intel_runtime_pm_get(i915); > + > + flush_workqueue(i915->wq); > + > + mutex_lock(&i915->drm.struct_mutex); > + > + /* > + * We have to flush all the executing contexts to main memory so > + * that they can saved in the hibernation image. To ensure the last > + * context image is coherent, we have to switch away from it. That > + * leaves the i915->kernel_context still active when > + * we actually suspend, and its image in memory may not match the GPU > + * state. Fortunately, the kernel_context is disposable and we do > + * not rely on its state. > + */ > + switch_to_kernel_context_sync(i915, i915->gt.active_engines); > + > + mutex_unlock(&i915->drm.struct_mutex); > + i915_reset_flush(i915); > + > + drain_delayed_work(&i915->gem.retire_work); > + > + /* > + * As the idle_work is rearming if it detects a race, play safe and > + * repeat the flush until it is definitely idle. > + */ > + drain_delayed_work(&i915->gem.idle_work); > + > + /* > + * Assert that we successfully flushed all the work and > + * reset the GPU back to its idle, low power state. > + */ > + GEM_BUG_ON(i915->gt.awake); > + > + intel_uc_suspend(i915); > + > + intel_runtime_pm_put(i915, wakeref); > +} > + > +void i915_gem_suspend_late(struct drm_i915_private *i915) > +{ > + struct drm_i915_gem_object *obj; > + struct list_head *phases[] = { > + &i915->mm.unbound_list, > + &i915->mm.bound_list, > + NULL > + }, **phase; > + > + /* > + * Neither the BIOS, ourselves or any other kernel > + * expects the system to be in execlists mode on startup, > + * so we need to reset the GPU back to legacy mode. And the only > + * known way to disable logical contexts is through a GPU reset. > + * > + * So in order to leave the system in a known default configuration, > + * always reset the GPU upon unload and suspend. Afterwards we then > + * clean up the GEM state tracking, flushing off the requests and > + * leaving the system in a known idle state. > + * > + * Note that is of the upmost importance that the GPU is idle and > + * all stray writes are flushed *before* we dismantle the backing > + * storage for the pinned objects. > + * > + * However, since we are uncertain that resetting the GPU on older > + * machines is a good idea, we don't - just in case it leaves the > + * machine in an unusable condition. > + */ > + > + mutex_lock(&i915->drm.struct_mutex); > + for (phase = phases; *phase; phase++) { > + list_for_each_entry(obj, *phase, mm.link) > + WARN_ON(i915_gem_object_set_to_gtt_domain(obj, false)); > + } > + mutex_unlock(&i915->drm.struct_mutex); > + > + intel_uc_sanitize(i915); > + i915_gem_sanitize(i915); > +} > + > +void i915_gem_resume(struct drm_i915_private *i915) > +{ > + GEM_TRACE("\n"); > + > + WARN_ON(i915->gt.awake); > + > + mutex_lock(&i915->drm.struct_mutex); > + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL); > + > + i915_gem_restore_gtt_mappings(i915); > + i915_gem_restore_fences(i915); > + > + /* > + * As we didn't flush the kernel context before suspend, we cannot > + * guarantee that the context image is complete. So let's just reset > + * it and start again. > + */ > + intel_gt_resume(i915); > + > + if (i915_gem_init_hw(i915)) > + goto err_wedged; > + > + intel_uc_resume(i915); > + > + /* Always reload a context for powersaving. */ > + if (!i915_gem_load_power_context(i915)) > + goto err_wedged; > + > +out_unlock: > + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); > + mutex_unlock(&i915->drm.struct_mutex); > + return; > + > +err_wedged: > + if (!i915_reset_failed(i915)) { > + dev_err(i915->drm.dev, > + "Failed to re-initialize GPU, declaring it wedged!\n"); > + i915_gem_set_wedged(i915); > + } > + goto out_unlock; > +} > + > +void i915_gem_init__pm(struct drm_i915_private *i915) > +{ > + INIT_DELAYED_WORK(&i915->gem.idle_work, idle_work_handler); > + INIT_DELAYED_WORK(&i915->gem.retire_work, retire_work_handler); > +} > diff --git a/drivers/gpu/drm/i915/i915_gem_pm.h b/drivers/gpu/drm/i915/i915_gem_pm.h > new file mode 100644 > index 000000000000..52f65e3f06b5 > --- /dev/null > +++ b/drivers/gpu/drm/i915/i915_gem_pm.h > @@ -0,0 +1,28 @@ > +/* > + * SPDX-License-Identifier: MIT > + * > + * Copyright © 2019 Intel Corporation > + */ > + > +#ifndef __I915_GEM_PM_H__ > +#define __I915_GEM_PM_H__ > + > +#include <linux/types.h> > + > +struct drm_i915_private; > +struct work_struct; > + > +void i915_gem_init__pm(struct drm_i915_private *i915); > + > +bool i915_gem_load_power_context(struct drm_i915_private *i915); > +void i915_gem_resume(struct drm_i915_private *i915); > + > +void i915_gem_unpark(struct drm_i915_private *i915); > +void i915_gem_park(struct drm_i915_private *i915); > + > +void i915_gem_idle_work_handler(struct work_struct *work); > + > +void i915_gem_suspend(struct drm_i915_private *i915); > +void i915_gem_suspend_late(struct drm_i915_private *i915); > + > +#endif /* __I915_GEM_PM_H__ */ > diff --git a/drivers/gpu/drm/i915/selftests/i915_gem_context.c b/drivers/gpu/drm/i915/selftests/i915_gem_context.c > index 6f52ca881173..9d646fa1b74e 100644 > --- a/drivers/gpu/drm/i915/selftests/i915_gem_context.c > +++ b/drivers/gpu/drm/i915/selftests/i915_gem_context.c > @@ -1658,7 +1658,7 @@ static int __igt_switch_to_kernel_context(struct drm_i915_private *i915, > /* XXX Bonus points for proving we are the kernel context! */ > > mutex_unlock(&i915->drm.struct_mutex); > - drain_delayed_work(&i915->gt.idle_work); > + drain_delayed_work(&i915->gem.idle_work); > mutex_lock(&i915->drm.struct_mutex); > } > > diff --git a/drivers/gpu/drm/i915/selftests/i915_gem_object.c b/drivers/gpu/drm/i915/selftests/i915_gem_object.c > index 971148fbe6f5..12fc53c694a6 100644 > --- a/drivers/gpu/drm/i915/selftests/i915_gem_object.c > +++ b/drivers/gpu/drm/i915/selftests/i915_gem_object.c > @@ -514,8 +514,8 @@ static void disable_retire_worker(struct drm_i915_private *i915) > } > mutex_unlock(&i915->drm.struct_mutex); > > - cancel_delayed_work_sync(&i915->gt.retire_work); > - cancel_delayed_work_sync(&i915->gt.idle_work); > + cancel_delayed_work_sync(&i915->gem.retire_work); > + cancel_delayed_work_sync(&i915->gem.idle_work); > } > > static int igt_mmap_offset_exhaustion(void *arg) > @@ -617,9 +617,9 @@ static int igt_mmap_offset_exhaustion(void *arg) > out_park: > mutex_lock(&i915->drm.struct_mutex); > if (--i915->gt.active_requests) > - queue_delayed_work(i915->wq, &i915->gt.retire_work, 0); > + queue_delayed_work(i915->wq, &i915->gem.retire_work, 0); > else > - queue_delayed_work(i915->wq, &i915->gt.idle_work, 0); > + queue_delayed_work(i915->wq, &i915->gem.idle_work, 0); > mutex_unlock(&i915->drm.struct_mutex); > i915_gem_shrinker_register(i915); > return err; > diff --git a/drivers/gpu/drm/i915/selftests/mock_gem_device.c b/drivers/gpu/drm/i915/selftests/mock_gem_device.c > index f444ee5add27..fb677b4019a0 100644 > --- a/drivers/gpu/drm/i915/selftests/mock_gem_device.c > +++ b/drivers/gpu/drm/i915/selftests/mock_gem_device.c > @@ -59,8 +59,8 @@ static void mock_device_release(struct drm_device *dev) > i915_gem_contexts_lost(i915); > mutex_unlock(&i915->drm.struct_mutex); > > - drain_delayed_work(&i915->gt.retire_work); > - drain_delayed_work(&i915->gt.idle_work); > + drain_delayed_work(&i915->gem.retire_work); > + drain_delayed_work(&i915->gem.idle_work); > i915_gem_drain_workqueue(i915); > > mutex_lock(&i915->drm.struct_mutex); > @@ -111,7 +111,7 @@ static void mock_retire_work_handler(struct work_struct *work) > static void mock_idle_work_handler(struct work_struct *work) > { > struct drm_i915_private *i915 = > - container_of(work, typeof(*i915), gt.idle_work.work); > + container_of(work, typeof(*i915), gem.idle_work.work); > > i915->gt.active_engines = 0; > } > @@ -197,8 +197,8 @@ struct drm_i915_private *mock_gem_device(void) > > mock_init_contexts(i915); > > - INIT_DELAYED_WORK(&i915->gt.retire_work, mock_retire_work_handler); > - INIT_DELAYED_WORK(&i915->gt.idle_work, mock_idle_work_handler); > + INIT_DELAYED_WORK(&i915->gem.retire_work, mock_retire_work_handler); > + INIT_DELAYED_WORK(&i915->gem.idle_work, mock_idle_work_handler); > > i915->gt.awake = true; > > Reviewed-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com> Regards, Tvrtko
diff --git a/drivers/gpu/drm/i915/Makefile b/drivers/gpu/drm/i915/Makefile index 233bad5e361f..858642c7bc40 100644 --- a/drivers/gpu/drm/i915/Makefile +++ b/drivers/gpu/drm/i915/Makefile @@ -97,6 +97,7 @@ i915-y += \ i915_gem_internal.o \ i915_gem.o \ i915_gem_object.o \ + i915_gem_pm.o \ i915_gem_render_state.o \ i915_gem_shrinker.o \ i915_gem_stolen.o \ diff --git a/drivers/gpu/drm/i915/Makefile.header-test b/drivers/gpu/drm/i915/Makefile.header-test index e6b3e7588860..702e3a7ade4c 100644 --- a/drivers/gpu/drm/i915/Makefile.header-test +++ b/drivers/gpu/drm/i915/Makefile.header-test @@ -5,6 +5,7 @@ header_test := \ i915_active_types.h \ i915_gem_context_types.h \ + i915_gem_pm.h \ i915_priolist_types.h \ i915_scheduler_types.h \ i915_timeline_types.h \ diff --git a/drivers/gpu/drm/i915/i915_debugfs.c b/drivers/gpu/drm/i915/i915_debugfs.c index eac4095278b2..e3eca88614b2 100644 --- a/drivers/gpu/drm/i915/i915_debugfs.c +++ b/drivers/gpu/drm/i915/i915_debugfs.c @@ -3941,8 +3941,8 @@ i915_drop_caches_set(void *data, u64 val) if (val & DROP_IDLE) { do { if (READ_ONCE(i915->gt.active_requests)) - flush_delayed_work(&i915->gt.retire_work); - drain_delayed_work(&i915->gt.idle_work); + flush_delayed_work(&i915->gem.retire_work); + drain_delayed_work(&i915->gem.idle_work); } while (READ_ONCE(i915->gt.awake)); } diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h index 62a7e91acd7f..cbae9be052e0 100644 --- a/drivers/gpu/drm/i915/i915_drv.h +++ b/drivers/gpu/drm/i915/i915_drv.h @@ -2020,6 +2020,12 @@ struct drm_i915_private { */ intel_wakeref_t awake; + ktime_t last_init_time; + + struct i915_vma *scratch; + } gt; + + struct { /** * We leave the user IRQ off as much as possible, * but this means that requests will finish and never @@ -2037,11 +2043,7 @@ struct drm_i915_private { * off the idle_work. */ struct delayed_work idle_work; - - ktime_t last_init_time; - - struct i915_vma *scratch; - } gt; + } gem; /* For i945gm vblank irq vs. C3 workaround */ struct { diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c index a8b0aa05288b..c56633e87c32 100644 --- a/drivers/gpu/drm/i915/i915_gem.c +++ b/drivers/gpu/drm/i915/i915_gem.c @@ -46,7 +46,7 @@ #include "i915_drv.h" #include "i915_gem_clflush.h" #include "i915_gemfs.h" -#include "i915_globals.h" +#include "i915_gem_pm.h" #include "i915_trace.h" #include "i915_vgpu.h" @@ -103,105 +103,6 @@ static void i915_gem_info_remove_obj(struct drm_i915_private *dev_priv, spin_unlock(&dev_priv->mm.object_stat_lock); } -static void __i915_gem_park(struct drm_i915_private *i915) -{ - intel_wakeref_t wakeref; - - GEM_TRACE("\n"); - - lockdep_assert_held(&i915->drm.struct_mutex); - GEM_BUG_ON(i915->gt.active_requests); - GEM_BUG_ON(!list_empty(&i915->gt.active_rings)); - - if (!i915->gt.awake) - return; - - /* - * Be paranoid and flush a concurrent interrupt to make sure - * we don't reactivate any irq tasklets after parking. - * - * FIXME: Note that even though we have waited for execlists to be idle, - * there may still be an in-flight interrupt even though the CSB - * is now empty. synchronize_irq() makes sure that a residual interrupt - * is completed before we continue, but it doesn't prevent the HW from - * raising a spurious interrupt later. To complete the shield we should - * coordinate disabling the CS irq with flushing the interrupts. - */ - synchronize_irq(i915->drm.irq); - - intel_engines_park(i915); - i915_timelines_park(i915); - - i915_pmu_gt_parked(i915); - i915_vma_parked(i915); - - wakeref = fetch_and_zero(&i915->gt.awake); - GEM_BUG_ON(!wakeref); - - if (INTEL_GEN(i915) >= 6) - gen6_rps_idle(i915); - - intel_display_power_put(i915, POWER_DOMAIN_GT_IRQ, wakeref); - - i915_globals_park(); -} - -void i915_gem_park(struct drm_i915_private *i915) -{ - GEM_TRACE("\n"); - - lockdep_assert_held(&i915->drm.struct_mutex); - GEM_BUG_ON(i915->gt.active_requests); - - if (!i915->gt.awake) - return; - - /* Defer the actual call to __i915_gem_park() to prevent ping-pongs */ - mod_delayed_work(i915->wq, &i915->gt.idle_work, msecs_to_jiffies(100)); -} - -void i915_gem_unpark(struct drm_i915_private *i915) -{ - GEM_TRACE("\n"); - - lockdep_assert_held(&i915->drm.struct_mutex); - GEM_BUG_ON(!i915->gt.active_requests); - assert_rpm_wakelock_held(i915); - - if (i915->gt.awake) - return; - - /* - * It seems that the DMC likes to transition between the DC states a lot - * when there are no connected displays (no active power domains) during - * command submission. - * - * This activity has negative impact on the performance of the chip with - * huge latencies observed in the interrupt handler and elsewhere. - * - * Work around it by grabbing a GT IRQ power domain whilst there is any - * GT activity, preventing any DC state transitions. - */ - i915->gt.awake = intel_display_power_get(i915, POWER_DOMAIN_GT_IRQ); - GEM_BUG_ON(!i915->gt.awake); - - i915_globals_unpark(); - - intel_enable_gt_powersave(i915); - i915_update_gfx_val(i915); - if (INTEL_GEN(i915) >= 6) - gen6_rps_busy(i915); - i915_pmu_gt_unparked(i915); - - intel_engines_unpark(i915); - - i915_queue_hangcheck(i915); - - queue_delayed_work(i915->wq, - &i915->gt.retire_work, - round_jiffies_up_relative(HZ)); -} - int i915_gem_get_aperture_ioctl(struct drm_device *dev, void *data, struct drm_file *file) @@ -2088,7 +1989,7 @@ static int i915_gem_object_create_mmap_offset(struct drm_i915_gem_object *obj) if (!err) break; - } while (flush_delayed_work(&dev_priv->gt.retire_work)); + } while (flush_delayed_work(&dev_priv->gem.retire_work)); return err; } @@ -2871,132 +2772,6 @@ i915_gem_object_pwrite_gtt(struct drm_i915_gem_object *obj, return 0; } -static void -i915_gem_retire_work_handler(struct work_struct *work) -{ - struct drm_i915_private *dev_priv = - container_of(work, typeof(*dev_priv), gt.retire_work.work); - struct drm_device *dev = &dev_priv->drm; - - /* Come back later if the device is busy... */ - if (mutex_trylock(&dev->struct_mutex)) { - i915_retire_requests(dev_priv); - mutex_unlock(&dev->struct_mutex); - } - - /* - * Keep the retire handler running until we are finally idle. - * We do not need to do this test under locking as in the worst-case - * we queue the retire worker once too often. - */ - if (READ_ONCE(dev_priv->gt.awake)) - queue_delayed_work(dev_priv->wq, - &dev_priv->gt.retire_work, - round_jiffies_up_relative(HZ)); -} - -static bool switch_to_kernel_context_sync(struct drm_i915_private *i915, - unsigned long mask) -{ - bool result = true; - - /* - * Even if we fail to switch, give whatever is running a small chance - * to save itself before we report the failure. Yes, this may be a - * false positive due to e.g. ENOMEM, caveat emptor! - */ - if (i915_gem_switch_to_kernel_context(i915, mask)) - result = false; - - if (i915_gem_wait_for_idle(i915, - I915_WAIT_LOCKED | - I915_WAIT_FOR_IDLE_BOOST, - I915_GEM_IDLE_TIMEOUT)) - result = false; - - if (!result) { - if (i915_modparams.reset) { /* XXX hide warning from gem_eio */ - dev_err(i915->drm.dev, - "Failed to idle engines, declaring wedged!\n"); - GEM_TRACE_DUMP(); - } - - /* Forcibly cancel outstanding work and leave the gpu quiet. */ - i915_gem_set_wedged(i915); - } - - i915_retire_requests(i915); /* ensure we flush after wedging */ - return result; -} - -static bool load_power_context(struct drm_i915_private *i915) -{ - /* Force loading the kernel context on all engines */ - if (!switch_to_kernel_context_sync(i915, ALL_ENGINES)) - return false; - - /* - * Immediately park the GPU so that we enable powersaving and - * treat it as idle. The next time we issue a request, we will - * unpark and start using the engine->pinned_default_state, otherwise - * it is in limbo and an early reset may fail. - */ - __i915_gem_park(i915); - - return true; -} - -static void -i915_gem_idle_work_handler(struct work_struct *work) -{ - struct drm_i915_private *i915 = - container_of(work, typeof(*i915), gt.idle_work.work); - bool rearm_hangcheck; - - if (!READ_ONCE(i915->gt.awake)) - return; - - if (READ_ONCE(i915->gt.active_requests)) - return; - - rearm_hangcheck = - cancel_delayed_work_sync(&i915->gpu_error.hangcheck_work); - - if (!mutex_trylock(&i915->drm.struct_mutex)) { - /* Currently busy, come back later */ - mod_delayed_work(i915->wq, - &i915->gt.idle_work, - msecs_to_jiffies(50)); - goto out_rearm; - } - - /* - * Flush out the last user context, leaving only the pinned - * kernel context resident. Should anything unfortunate happen - * while we are idle (such as the GPU being power cycled), no users - * will be harmed. - */ - if (!work_pending(&i915->gt.idle_work.work) && - !i915->gt.active_requests) { - ++i915->gt.active_requests; /* don't requeue idle */ - - switch_to_kernel_context_sync(i915, i915->gt.active_engines); - - if (!--i915->gt.active_requests) { - __i915_gem_park(i915); - rearm_hangcheck = false; - } - } - - mutex_unlock(&i915->drm.struct_mutex); - -out_rearm: - if (rearm_hangcheck) { - GEM_BUG_ON(!i915->gt.awake); - i915_queue_hangcheck(i915); - } -} - void i915_gem_close_object(struct drm_gem_object *gem, struct drm_file *file) { struct drm_i915_private *i915 = to_i915(gem->dev); @@ -4412,133 +4187,6 @@ void i915_gem_sanitize(struct drm_i915_private *i915) mutex_unlock(&i915->drm.struct_mutex); } -void i915_gem_suspend(struct drm_i915_private *i915) -{ - intel_wakeref_t wakeref; - - GEM_TRACE("\n"); - - wakeref = intel_runtime_pm_get(i915); - - flush_workqueue(i915->wq); - - mutex_lock(&i915->drm.struct_mutex); - - /* - * We have to flush all the executing contexts to main memory so - * that they can saved in the hibernation image. To ensure the last - * context image is coherent, we have to switch away from it. That - * leaves the i915->kernel_context still active when - * we actually suspend, and its image in memory may not match the GPU - * state. Fortunately, the kernel_context is disposable and we do - * not rely on its state. - */ - switch_to_kernel_context_sync(i915, i915->gt.active_engines); - - mutex_unlock(&i915->drm.struct_mutex); - i915_reset_flush(i915); - - drain_delayed_work(&i915->gt.retire_work); - - /* - * As the idle_work is rearming if it detects a race, play safe and - * repeat the flush until it is definitely idle. - */ - drain_delayed_work(&i915->gt.idle_work); - - /* - * Assert that we successfully flushed all the work and - * reset the GPU back to its idle, low power state. - */ - GEM_BUG_ON(i915->gt.awake); - - intel_uc_suspend(i915); - - intel_runtime_pm_put(i915, wakeref); -} - -void i915_gem_suspend_late(struct drm_i915_private *i915) -{ - struct drm_i915_gem_object *obj; - struct list_head *phases[] = { - &i915->mm.unbound_list, - &i915->mm.bound_list, - NULL - }, **phase; - - /* - * Neither the BIOS, ourselves or any other kernel - * expects the system to be in execlists mode on startup, - * so we need to reset the GPU back to legacy mode. And the only - * known way to disable logical contexts is through a GPU reset. - * - * So in order to leave the system in a known default configuration, - * always reset the GPU upon unload and suspend. Afterwards we then - * clean up the GEM state tracking, flushing off the requests and - * leaving the system in a known idle state. - * - * Note that is of the upmost importance that the GPU is idle and - * all stray writes are flushed *before* we dismantle the backing - * storage for the pinned objects. - * - * However, since we are uncertain that resetting the GPU on older - * machines is a good idea, we don't - just in case it leaves the - * machine in an unusable condition. - */ - - mutex_lock(&i915->drm.struct_mutex); - for (phase = phases; *phase; phase++) { - list_for_each_entry(obj, *phase, mm.link) - WARN_ON(i915_gem_object_set_to_gtt_domain(obj, false)); - } - mutex_unlock(&i915->drm.struct_mutex); - - intel_uc_sanitize(i915); - i915_gem_sanitize(i915); -} - -void i915_gem_resume(struct drm_i915_private *i915) -{ - GEM_TRACE("\n"); - - WARN_ON(i915->gt.awake); - - mutex_lock(&i915->drm.struct_mutex); - intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL); - - i915_gem_restore_gtt_mappings(i915); - i915_gem_restore_fences(i915); - - /* - * As we didn't flush the kernel context before suspend, we cannot - * guarantee that the context image is complete. So let's just reset - * it and start again. - */ - intel_gt_resume(i915); - - if (i915_gem_init_hw(i915)) - goto err_wedged; - - intel_uc_resume(i915); - - /* Always reload a context for powersaving. */ - if (!load_power_context(i915)) - goto err_wedged; - -out_unlock: - intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); - mutex_unlock(&i915->drm.struct_mutex); - return; - -err_wedged: - if (!i915_reset_failed(i915)) { - dev_err(i915->drm.dev, - "Failed to re-initialize GPU, declaring it wedged!\n"); - i915_gem_set_wedged(i915); - } - goto out_unlock; -} - void i915_gem_init_swizzling(struct drm_i915_private *dev_priv) { if (INTEL_GEN(dev_priv) < 5 || @@ -4721,7 +4369,7 @@ static int __intel_engines_record_defaults(struct drm_i915_private *i915) } /* Flush the default context image to memory, and enable powersaving. */ - if (!load_power_context(i915)) { + if (!i915_gem_load_power_context(i915)) { err = -EIO; goto err_active; } @@ -5115,11 +4763,8 @@ int i915_gem_init_early(struct drm_i915_private *dev_priv) INIT_LIST_HEAD(&dev_priv->gt.closed_vma); i915_gem_init__mm(dev_priv); + i915_gem_init__pm(dev_priv); - INIT_DELAYED_WORK(&dev_priv->gt.retire_work, - i915_gem_retire_work_handler); - INIT_DELAYED_WORK(&dev_priv->gt.idle_work, - i915_gem_idle_work_handler); init_waitqueue_head(&dev_priv->gpu_error.wait_queue); init_waitqueue_head(&dev_priv->gpu_error.reset_queue); mutex_init(&dev_priv->gpu_error.wedge_mutex); diff --git a/drivers/gpu/drm/i915/i915_gem_pm.c b/drivers/gpu/drm/i915/i915_gem_pm.c new file mode 100644 index 000000000000..bc0663cac7b6 --- /dev/null +++ b/drivers/gpu/drm/i915/i915_gem_pm.c @@ -0,0 +1,365 @@ +/* + * SPDX-License-Identifier: MIT + * + * Copyright © 2019 Intel Corporation + */ + +#include "i915_drv.h" +#include "i915_gem_pm.h" +#include "i915_globals.h" +#include "intel_pm.h" + +static void __i915_gem_park(struct drm_i915_private *i915) +{ + intel_wakeref_t wakeref; + + GEM_TRACE("\n"); + + lockdep_assert_held(&i915->drm.struct_mutex); + GEM_BUG_ON(i915->gt.active_requests); + GEM_BUG_ON(!list_empty(&i915->gt.active_rings)); + + if (!i915->gt.awake) + return; + + /* + * Be paranoid and flush a concurrent interrupt to make sure + * we don't reactivate any irq tasklets after parking. + * + * FIXME: Note that even though we have waited for execlists to be idle, + * there may still be an in-flight interrupt even though the CSB + * is now empty. synchronize_irq() makes sure that a residual interrupt + * is completed before we continue, but it doesn't prevent the HW from + * raising a spurious interrupt later. To complete the shield we should + * coordinate disabling the CS irq with flushing the interrupts. + */ + synchronize_irq(i915->drm.irq); + + intel_engines_park(i915); + i915_timelines_park(i915); + + i915_pmu_gt_parked(i915); + i915_vma_parked(i915); + + wakeref = fetch_and_zero(&i915->gt.awake); + GEM_BUG_ON(!wakeref); + + if (INTEL_GEN(i915) >= 6) + gen6_rps_idle(i915); + + intel_display_power_put(i915, POWER_DOMAIN_GT_IRQ, wakeref); + + i915_globals_park(); +} + +static bool switch_to_kernel_context_sync(struct drm_i915_private *i915, + unsigned long mask) +{ + bool result = true; + + /* + * Even if we fail to switch, give whatever is running a small chance + * to save itself before we report the failure. Yes, this may be a + * false positive due to e.g. ENOMEM, caveat emptor! + */ + if (i915_gem_switch_to_kernel_context(i915, mask)) + result = false; + + if (i915_gem_wait_for_idle(i915, + I915_WAIT_LOCKED | + I915_WAIT_FOR_IDLE_BOOST, + I915_GEM_IDLE_TIMEOUT)) + result = false; + + if (!result) { + if (i915_modparams.reset) { /* XXX hide warning from gem_eio */ + dev_err(i915->drm.dev, + "Failed to idle engines, declaring wedged!\n"); + GEM_TRACE_DUMP(); + } + + /* Forcibly cancel outstanding work and leave the gpu quiet. */ + i915_gem_set_wedged(i915); + } + + i915_retire_requests(i915); /* ensure we flush after wedging */ + return result; +} + +static void idle_work_handler(struct work_struct *work) +{ + struct drm_i915_private *i915 = + container_of(work, typeof(*i915), gem.idle_work.work); + bool rearm_hangcheck; + + if (!READ_ONCE(i915->gt.awake)) + return; + + if (READ_ONCE(i915->gt.active_requests)) + return; + + rearm_hangcheck = + cancel_delayed_work_sync(&i915->gpu_error.hangcheck_work); + + if (!mutex_trylock(&i915->drm.struct_mutex)) { + /* Currently busy, come back later */ + mod_delayed_work(i915->wq, + &i915->gem.idle_work, + msecs_to_jiffies(50)); + goto out_rearm; + } + + /* + * Flush out the last user context, leaving only the pinned + * kernel context resident. Should anything unfortunate happen + * while we are idle (such as the GPU being power cycled), no users + * will be harmed. + */ + if (!work_pending(&i915->gem.idle_work.work) && + !i915->gt.active_requests) { + ++i915->gt.active_requests; /* don't requeue idle */ + + switch_to_kernel_context_sync(i915, i915->gt.active_engines); + + if (!--i915->gt.active_requests) { + __i915_gem_park(i915); + rearm_hangcheck = false; + } + } + + mutex_unlock(&i915->drm.struct_mutex); + +out_rearm: + if (rearm_hangcheck) { + GEM_BUG_ON(!i915->gt.awake); + i915_queue_hangcheck(i915); + } +} + +static void retire_work_handler(struct work_struct *work) +{ + struct drm_i915_private *i915 = + container_of(work, typeof(*i915), gem.retire_work.work); + + /* Come back later if the device is busy... */ + if (mutex_trylock(&i915->drm.struct_mutex)) { + i915_retire_requests(i915); + mutex_unlock(&i915->drm.struct_mutex); + } + + /* + * Keep the retire handler running until we are finally idle. + * We do not need to do this test under locking as in the worst-case + * we queue the retire worker once too often. + */ + if (READ_ONCE(i915->gt.awake)) + queue_delayed_work(i915->wq, + &i915->gem.retire_work, + round_jiffies_up_relative(HZ)); +} + +void i915_gem_park(struct drm_i915_private *i915) +{ + GEM_TRACE("\n"); + + lockdep_assert_held(&i915->drm.struct_mutex); + GEM_BUG_ON(i915->gt.active_requests); + + if (!i915->gt.awake) + return; + + /* Defer the actual call to __i915_gem_park() to prevent ping-pongs */ + mod_delayed_work(i915->wq, &i915->gem.idle_work, msecs_to_jiffies(100)); +} + +void i915_gem_unpark(struct drm_i915_private *i915) +{ + GEM_TRACE("\n"); + + lockdep_assert_held(&i915->drm.struct_mutex); + GEM_BUG_ON(!i915->gt.active_requests); + assert_rpm_wakelock_held(i915); + + if (i915->gt.awake) + return; + + /* + * It seems that the DMC likes to transition between the DC states a lot + * when there are no connected displays (no active power domains) during + * command submission. + * + * This activity has negative impact on the performance of the chip with + * huge latencies observed in the interrupt handler and elsewhere. + * + * Work around it by grabbing a GT IRQ power domain whilst there is any + * GT activity, preventing any DC state transitions. + */ + i915->gt.awake = intel_display_power_get(i915, POWER_DOMAIN_GT_IRQ); + GEM_BUG_ON(!i915->gt.awake); + + i915_globals_unpark(); + + intel_enable_gt_powersave(i915); + i915_update_gfx_val(i915); + if (INTEL_GEN(i915) >= 6) + gen6_rps_busy(i915); + i915_pmu_gt_unparked(i915); + + intel_engines_unpark(i915); + + i915_queue_hangcheck(i915); + + queue_delayed_work(i915->wq, + &i915->gem.retire_work, + round_jiffies_up_relative(HZ)); +} + +bool i915_gem_load_power_context(struct drm_i915_private *i915) +{ + /* Force loading the kernel context on all engines */ + if (!switch_to_kernel_context_sync(i915, ALL_ENGINES)) + return false; + + /* + * Immediately park the GPU so that we enable powersaving and + * treat it as idle. The next time we issue a request, we will + * unpark and start using the engine->pinned_default_state, otherwise + * it is in limbo and an early reset may fail. + */ + __i915_gem_park(i915); + + return true; +} + +void i915_gem_suspend(struct drm_i915_private *i915) +{ + intel_wakeref_t wakeref; + + GEM_TRACE("\n"); + + wakeref = intel_runtime_pm_get(i915); + + flush_workqueue(i915->wq); + + mutex_lock(&i915->drm.struct_mutex); + + /* + * We have to flush all the executing contexts to main memory so + * that they can saved in the hibernation image. To ensure the last + * context image is coherent, we have to switch away from it. That + * leaves the i915->kernel_context still active when + * we actually suspend, and its image in memory may not match the GPU + * state. Fortunately, the kernel_context is disposable and we do + * not rely on its state. + */ + switch_to_kernel_context_sync(i915, i915->gt.active_engines); + + mutex_unlock(&i915->drm.struct_mutex); + i915_reset_flush(i915); + + drain_delayed_work(&i915->gem.retire_work); + + /* + * As the idle_work is rearming if it detects a race, play safe and + * repeat the flush until it is definitely idle. + */ + drain_delayed_work(&i915->gem.idle_work); + + /* + * Assert that we successfully flushed all the work and + * reset the GPU back to its idle, low power state. + */ + GEM_BUG_ON(i915->gt.awake); + + intel_uc_suspend(i915); + + intel_runtime_pm_put(i915, wakeref); +} + +void i915_gem_suspend_late(struct drm_i915_private *i915) +{ + struct drm_i915_gem_object *obj; + struct list_head *phases[] = { + &i915->mm.unbound_list, + &i915->mm.bound_list, + NULL + }, **phase; + + /* + * Neither the BIOS, ourselves or any other kernel + * expects the system to be in execlists mode on startup, + * so we need to reset the GPU back to legacy mode. And the only + * known way to disable logical contexts is through a GPU reset. + * + * So in order to leave the system in a known default configuration, + * always reset the GPU upon unload and suspend. Afterwards we then + * clean up the GEM state tracking, flushing off the requests and + * leaving the system in a known idle state. + * + * Note that is of the upmost importance that the GPU is idle and + * all stray writes are flushed *before* we dismantle the backing + * storage for the pinned objects. + * + * However, since we are uncertain that resetting the GPU on older + * machines is a good idea, we don't - just in case it leaves the + * machine in an unusable condition. + */ + + mutex_lock(&i915->drm.struct_mutex); + for (phase = phases; *phase; phase++) { + list_for_each_entry(obj, *phase, mm.link) + WARN_ON(i915_gem_object_set_to_gtt_domain(obj, false)); + } + mutex_unlock(&i915->drm.struct_mutex); + + intel_uc_sanitize(i915); + i915_gem_sanitize(i915); +} + +void i915_gem_resume(struct drm_i915_private *i915) +{ + GEM_TRACE("\n"); + + WARN_ON(i915->gt.awake); + + mutex_lock(&i915->drm.struct_mutex); + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL); + + i915_gem_restore_gtt_mappings(i915); + i915_gem_restore_fences(i915); + + /* + * As we didn't flush the kernel context before suspend, we cannot + * guarantee that the context image is complete. So let's just reset + * it and start again. + */ + intel_gt_resume(i915); + + if (i915_gem_init_hw(i915)) + goto err_wedged; + + intel_uc_resume(i915); + + /* Always reload a context for powersaving. */ + if (!i915_gem_load_power_context(i915)) + goto err_wedged; + +out_unlock: + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); + mutex_unlock(&i915->drm.struct_mutex); + return; + +err_wedged: + if (!i915_reset_failed(i915)) { + dev_err(i915->drm.dev, + "Failed to re-initialize GPU, declaring it wedged!\n"); + i915_gem_set_wedged(i915); + } + goto out_unlock; +} + +void i915_gem_init__pm(struct drm_i915_private *i915) +{ + INIT_DELAYED_WORK(&i915->gem.idle_work, idle_work_handler); + INIT_DELAYED_WORK(&i915->gem.retire_work, retire_work_handler); +} diff --git a/drivers/gpu/drm/i915/i915_gem_pm.h b/drivers/gpu/drm/i915/i915_gem_pm.h new file mode 100644 index 000000000000..52f65e3f06b5 --- /dev/null +++ b/drivers/gpu/drm/i915/i915_gem_pm.h @@ -0,0 +1,28 @@ +/* + * SPDX-License-Identifier: MIT + * + * Copyright © 2019 Intel Corporation + */ + +#ifndef __I915_GEM_PM_H__ +#define __I915_GEM_PM_H__ + +#include <linux/types.h> + +struct drm_i915_private; +struct work_struct; + +void i915_gem_init__pm(struct drm_i915_private *i915); + +bool i915_gem_load_power_context(struct drm_i915_private *i915); +void i915_gem_resume(struct drm_i915_private *i915); + +void i915_gem_unpark(struct drm_i915_private *i915); +void i915_gem_park(struct drm_i915_private *i915); + +void i915_gem_idle_work_handler(struct work_struct *work); + +void i915_gem_suspend(struct drm_i915_private *i915); +void i915_gem_suspend_late(struct drm_i915_private *i915); + +#endif /* __I915_GEM_PM_H__ */ diff --git a/drivers/gpu/drm/i915/selftests/i915_gem_context.c b/drivers/gpu/drm/i915/selftests/i915_gem_context.c index 6f52ca881173..9d646fa1b74e 100644 --- a/drivers/gpu/drm/i915/selftests/i915_gem_context.c +++ b/drivers/gpu/drm/i915/selftests/i915_gem_context.c @@ -1658,7 +1658,7 @@ static int __igt_switch_to_kernel_context(struct drm_i915_private *i915, /* XXX Bonus points for proving we are the kernel context! */ mutex_unlock(&i915->drm.struct_mutex); - drain_delayed_work(&i915->gt.idle_work); + drain_delayed_work(&i915->gem.idle_work); mutex_lock(&i915->drm.struct_mutex); } diff --git a/drivers/gpu/drm/i915/selftests/i915_gem_object.c b/drivers/gpu/drm/i915/selftests/i915_gem_object.c index 971148fbe6f5..12fc53c694a6 100644 --- a/drivers/gpu/drm/i915/selftests/i915_gem_object.c +++ b/drivers/gpu/drm/i915/selftests/i915_gem_object.c @@ -514,8 +514,8 @@ static void disable_retire_worker(struct drm_i915_private *i915) } mutex_unlock(&i915->drm.struct_mutex); - cancel_delayed_work_sync(&i915->gt.retire_work); - cancel_delayed_work_sync(&i915->gt.idle_work); + cancel_delayed_work_sync(&i915->gem.retire_work); + cancel_delayed_work_sync(&i915->gem.idle_work); } static int igt_mmap_offset_exhaustion(void *arg) @@ -617,9 +617,9 @@ static int igt_mmap_offset_exhaustion(void *arg) out_park: mutex_lock(&i915->drm.struct_mutex); if (--i915->gt.active_requests) - queue_delayed_work(i915->wq, &i915->gt.retire_work, 0); + queue_delayed_work(i915->wq, &i915->gem.retire_work, 0); else - queue_delayed_work(i915->wq, &i915->gt.idle_work, 0); + queue_delayed_work(i915->wq, &i915->gem.idle_work, 0); mutex_unlock(&i915->drm.struct_mutex); i915_gem_shrinker_register(i915); return err; diff --git a/drivers/gpu/drm/i915/selftests/mock_gem_device.c b/drivers/gpu/drm/i915/selftests/mock_gem_device.c index f444ee5add27..fb677b4019a0 100644 --- a/drivers/gpu/drm/i915/selftests/mock_gem_device.c +++ b/drivers/gpu/drm/i915/selftests/mock_gem_device.c @@ -59,8 +59,8 @@ static void mock_device_release(struct drm_device *dev) i915_gem_contexts_lost(i915); mutex_unlock(&i915->drm.struct_mutex); - drain_delayed_work(&i915->gt.retire_work); - drain_delayed_work(&i915->gt.idle_work); + drain_delayed_work(&i915->gem.retire_work); + drain_delayed_work(&i915->gem.idle_work); i915_gem_drain_workqueue(i915); mutex_lock(&i915->drm.struct_mutex); @@ -111,7 +111,7 @@ static void mock_retire_work_handler(struct work_struct *work) static void mock_idle_work_handler(struct work_struct *work) { struct drm_i915_private *i915 = - container_of(work, typeof(*i915), gt.idle_work.work); + container_of(work, typeof(*i915), gem.idle_work.work); i915->gt.active_engines = 0; } @@ -197,8 +197,8 @@ struct drm_i915_private *mock_gem_device(void) mock_init_contexts(i915); - INIT_DELAYED_WORK(&i915->gt.retire_work, mock_retire_work_handler); - INIT_DELAYED_WORK(&i915->gt.idle_work, mock_idle_work_handler); + INIT_DELAYED_WORK(&i915->gem.retire_work, mock_retire_work_handler); + INIT_DELAYED_WORK(&i915->gem.idle_work, mock_idle_work_handler); i915->gt.awake = true;
Split out the powermanagement portion (GT wakeref, suspend/resume) of GEM from i915_gem.c into its own file. Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk> --- drivers/gpu/drm/i915/Makefile | 1 + drivers/gpu/drm/i915/Makefile.header-test | 1 + drivers/gpu/drm/i915/i915_debugfs.c | 4 +- drivers/gpu/drm/i915/i915_drv.h | 12 +- drivers/gpu/drm/i915/i915_gem.c | 363 +---------------- drivers/gpu/drm/i915/i915_gem_pm.c | 365 ++++++++++++++++++ drivers/gpu/drm/i915/i915_gem_pm.h | 28 ++ .../gpu/drm/i915/selftests/i915_gem_context.c | 2 +- .../gpu/drm/i915/selftests/i915_gem_object.c | 8 +- .../gpu/drm/i915/selftests/mock_gem_device.c | 10 +- 10 files changed, 418 insertions(+), 376 deletions(-) create mode 100644 drivers/gpu/drm/i915/i915_gem_pm.c create mode 100644 drivers/gpu/drm/i915/i915_gem_pm.h