[Intel-gfx] [PATCH 09/29] drm/i915: Pull the GEM powermangement coupling into its own file
Tvrtko Ursulin
tvrtko.ursulin at linux.intel.com
Mon Apr 8 14:56:39 UTC 2019
On 08/04/2019 10:17, Chris Wilson wrote:
> Split out the powermanagement portion (GT wakeref, suspend/resume) of
> GEM from i915_gem.c into its own file.
This is the same patch I gave r-b on before? Judging from the title and
commit message I'd say yes. But I know it is not, don't worry.
Regards,
Tvrtko
> Signed-off-by: Chris Wilson <chris at chris-wilson.co.uk>
> ---
> drivers/gpu/drm/i915/Makefile | 1 +
> drivers/gpu/drm/i915/Makefile.header-test | 1 +
> drivers/gpu/drm/i915/i915_debugfs.c | 4 +-
> drivers/gpu/drm/i915/i915_drv.h | 12 +-
> drivers/gpu/drm/i915/i915_gem.c | 363 +----------------
> drivers/gpu/drm/i915/i915_gem_pm.c | 365 ++++++++++++++++++
> drivers/gpu/drm/i915/i915_gem_pm.h | 28 ++
> .../gpu/drm/i915/selftests/i915_gem_context.c | 2 +-
> .../gpu/drm/i915/selftests/i915_gem_object.c | 8 +-
> .../gpu/drm/i915/selftests/mock_gem_device.c | 10 +-
> 10 files changed, 418 insertions(+), 376 deletions(-)
> create mode 100644 drivers/gpu/drm/i915/i915_gem_pm.c
> create mode 100644 drivers/gpu/drm/i915/i915_gem_pm.h
>
> diff --git a/drivers/gpu/drm/i915/Makefile b/drivers/gpu/drm/i915/Makefile
> index 233bad5e361f..858642c7bc40 100644
> --- a/drivers/gpu/drm/i915/Makefile
> +++ b/drivers/gpu/drm/i915/Makefile
> @@ -97,6 +97,7 @@ i915-y += \
> i915_gem_internal.o \
> i915_gem.o \
> i915_gem_object.o \
> + i915_gem_pm.o \
> i915_gem_render_state.o \
> i915_gem_shrinker.o \
> i915_gem_stolen.o \
> diff --git a/drivers/gpu/drm/i915/Makefile.header-test b/drivers/gpu/drm/i915/Makefile.header-test
> index e6b3e7588860..702e3a7ade4c 100644
> --- a/drivers/gpu/drm/i915/Makefile.header-test
> +++ b/drivers/gpu/drm/i915/Makefile.header-test
> @@ -5,6 +5,7 @@
> header_test := \
> i915_active_types.h \
> i915_gem_context_types.h \
> + i915_gem_pm.h \
> i915_priolist_types.h \
> i915_scheduler_types.h \
> i915_timeline_types.h \
> diff --git a/drivers/gpu/drm/i915/i915_debugfs.c b/drivers/gpu/drm/i915/i915_debugfs.c
> index eac4095278b2..e3eca88614b2 100644
> --- a/drivers/gpu/drm/i915/i915_debugfs.c
> +++ b/drivers/gpu/drm/i915/i915_debugfs.c
> @@ -3941,8 +3941,8 @@ i915_drop_caches_set(void *data, u64 val)
> if (val & DROP_IDLE) {
> do {
> if (READ_ONCE(i915->gt.active_requests))
> - flush_delayed_work(&i915->gt.retire_work);
> - drain_delayed_work(&i915->gt.idle_work);
> + flush_delayed_work(&i915->gem.retire_work);
> + drain_delayed_work(&i915->gem.idle_work);
> } while (READ_ONCE(i915->gt.awake));
> }
>
> diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h
> index 62a7e91acd7f..cbae9be052e0 100644
> --- a/drivers/gpu/drm/i915/i915_drv.h
> +++ b/drivers/gpu/drm/i915/i915_drv.h
> @@ -2020,6 +2020,12 @@ struct drm_i915_private {
> */
> intel_wakeref_t awake;
>
> + ktime_t last_init_time;
> +
> + struct i915_vma *scratch;
> + } gt;
> +
> + struct {
> /**
> * We leave the user IRQ off as much as possible,
> * but this means that requests will finish and never
> @@ -2037,11 +2043,7 @@ struct drm_i915_private {
> * off the idle_work.
> */
> struct delayed_work idle_work;
> -
> - ktime_t last_init_time;
> -
> - struct i915_vma *scratch;
> - } gt;
> + } gem;
>
> /* For i945gm vblank irq vs. C3 workaround */
> struct {
> diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c
> index a8b0aa05288b..c56633e87c32 100644
> --- a/drivers/gpu/drm/i915/i915_gem.c
> +++ b/drivers/gpu/drm/i915/i915_gem.c
> @@ -46,7 +46,7 @@
> #include "i915_drv.h"
> #include "i915_gem_clflush.h"
> #include "i915_gemfs.h"
> -#include "i915_globals.h"
> +#include "i915_gem_pm.h"
> #include "i915_trace.h"
> #include "i915_vgpu.h"
>
> @@ -103,105 +103,6 @@ static void i915_gem_info_remove_obj(struct drm_i915_private *dev_priv,
> spin_unlock(&dev_priv->mm.object_stat_lock);
> }
>
> -static void __i915_gem_park(struct drm_i915_private *i915)
> -{
> - intel_wakeref_t wakeref;
> -
> - GEM_TRACE("\n");
> -
> - lockdep_assert_held(&i915->drm.struct_mutex);
> - GEM_BUG_ON(i915->gt.active_requests);
> - GEM_BUG_ON(!list_empty(&i915->gt.active_rings));
> -
> - if (!i915->gt.awake)
> - return;
> -
> - /*
> - * Be paranoid and flush a concurrent interrupt to make sure
> - * we don't reactivate any irq tasklets after parking.
> - *
> - * FIXME: Note that even though we have waited for execlists to be idle,
> - * there may still be an in-flight interrupt even though the CSB
> - * is now empty. synchronize_irq() makes sure that a residual interrupt
> - * is completed before we continue, but it doesn't prevent the HW from
> - * raising a spurious interrupt later. To complete the shield we should
> - * coordinate disabling the CS irq with flushing the interrupts.
> - */
> - synchronize_irq(i915->drm.irq);
> -
> - intel_engines_park(i915);
> - i915_timelines_park(i915);
> -
> - i915_pmu_gt_parked(i915);
> - i915_vma_parked(i915);
> -
> - wakeref = fetch_and_zero(&i915->gt.awake);
> - GEM_BUG_ON(!wakeref);
> -
> - if (INTEL_GEN(i915) >= 6)
> - gen6_rps_idle(i915);
> -
> - intel_display_power_put(i915, POWER_DOMAIN_GT_IRQ, wakeref);
> -
> - i915_globals_park();
> -}
> -
> -void i915_gem_park(struct drm_i915_private *i915)
> -{
> - GEM_TRACE("\n");
> -
> - lockdep_assert_held(&i915->drm.struct_mutex);
> - GEM_BUG_ON(i915->gt.active_requests);
> -
> - if (!i915->gt.awake)
> - return;
> -
> - /* Defer the actual call to __i915_gem_park() to prevent ping-pongs */
> - mod_delayed_work(i915->wq, &i915->gt.idle_work, msecs_to_jiffies(100));
> -}
> -
> -void i915_gem_unpark(struct drm_i915_private *i915)
> -{
> - GEM_TRACE("\n");
> -
> - lockdep_assert_held(&i915->drm.struct_mutex);
> - GEM_BUG_ON(!i915->gt.active_requests);
> - assert_rpm_wakelock_held(i915);
> -
> - if (i915->gt.awake)
> - return;
> -
> - /*
> - * It seems that the DMC likes to transition between the DC states a lot
> - * when there are no connected displays (no active power domains) during
> - * command submission.
> - *
> - * This activity has negative impact on the performance of the chip with
> - * huge latencies observed in the interrupt handler and elsewhere.
> - *
> - * Work around it by grabbing a GT IRQ power domain whilst there is any
> - * GT activity, preventing any DC state transitions.
> - */
> - i915->gt.awake = intel_display_power_get(i915, POWER_DOMAIN_GT_IRQ);
> - GEM_BUG_ON(!i915->gt.awake);
> -
> - i915_globals_unpark();
> -
> - intel_enable_gt_powersave(i915);
> - i915_update_gfx_val(i915);
> - if (INTEL_GEN(i915) >= 6)
> - gen6_rps_busy(i915);
> - i915_pmu_gt_unparked(i915);
> -
> - intel_engines_unpark(i915);
> -
> - i915_queue_hangcheck(i915);
> -
> - queue_delayed_work(i915->wq,
> - &i915->gt.retire_work,
> - round_jiffies_up_relative(HZ));
> -}
> -
> int
> i915_gem_get_aperture_ioctl(struct drm_device *dev, void *data,
> struct drm_file *file)
> @@ -2088,7 +1989,7 @@ static int i915_gem_object_create_mmap_offset(struct drm_i915_gem_object *obj)
> if (!err)
> break;
>
> - } while (flush_delayed_work(&dev_priv->gt.retire_work));
> + } while (flush_delayed_work(&dev_priv->gem.retire_work));
>
> return err;
> }
> @@ -2871,132 +2772,6 @@ i915_gem_object_pwrite_gtt(struct drm_i915_gem_object *obj,
> return 0;
> }
>
> -static void
> -i915_gem_retire_work_handler(struct work_struct *work)
> -{
> - struct drm_i915_private *dev_priv =
> - container_of(work, typeof(*dev_priv), gt.retire_work.work);
> - struct drm_device *dev = &dev_priv->drm;
> -
> - /* Come back later if the device is busy... */
> - if (mutex_trylock(&dev->struct_mutex)) {
> - i915_retire_requests(dev_priv);
> - mutex_unlock(&dev->struct_mutex);
> - }
> -
> - /*
> - * Keep the retire handler running until we are finally idle.
> - * We do not need to do this test under locking as in the worst-case
> - * we queue the retire worker once too often.
> - */
> - if (READ_ONCE(dev_priv->gt.awake))
> - queue_delayed_work(dev_priv->wq,
> - &dev_priv->gt.retire_work,
> - round_jiffies_up_relative(HZ));
> -}
> -
> -static bool switch_to_kernel_context_sync(struct drm_i915_private *i915,
> - unsigned long mask)
> -{
> - bool result = true;
> -
> - /*
> - * Even if we fail to switch, give whatever is running a small chance
> - * to save itself before we report the failure. Yes, this may be a
> - * false positive due to e.g. ENOMEM, caveat emptor!
> - */
> - if (i915_gem_switch_to_kernel_context(i915, mask))
> - result = false;
> -
> - if (i915_gem_wait_for_idle(i915,
> - I915_WAIT_LOCKED |
> - I915_WAIT_FOR_IDLE_BOOST,
> - I915_GEM_IDLE_TIMEOUT))
> - result = false;
> -
> - if (!result) {
> - if (i915_modparams.reset) { /* XXX hide warning from gem_eio */
> - dev_err(i915->drm.dev,
> - "Failed to idle engines, declaring wedged!\n");
> - GEM_TRACE_DUMP();
> - }
> -
> - /* Forcibly cancel outstanding work and leave the gpu quiet. */
> - i915_gem_set_wedged(i915);
> - }
> -
> - i915_retire_requests(i915); /* ensure we flush after wedging */
> - return result;
> -}
> -
> -static bool load_power_context(struct drm_i915_private *i915)
> -{
> - /* Force loading the kernel context on all engines */
> - if (!switch_to_kernel_context_sync(i915, ALL_ENGINES))
> - return false;
> -
> - /*
> - * Immediately park the GPU so that we enable powersaving and
> - * treat it as idle. The next time we issue a request, we will
> - * unpark and start using the engine->pinned_default_state, otherwise
> - * it is in limbo and an early reset may fail.
> - */
> - __i915_gem_park(i915);
> -
> - return true;
> -}
> -
> -static void
> -i915_gem_idle_work_handler(struct work_struct *work)
> -{
> - struct drm_i915_private *i915 =
> - container_of(work, typeof(*i915), gt.idle_work.work);
> - bool rearm_hangcheck;
> -
> - if (!READ_ONCE(i915->gt.awake))
> - return;
> -
> - if (READ_ONCE(i915->gt.active_requests))
> - return;
> -
> - rearm_hangcheck =
> - cancel_delayed_work_sync(&i915->gpu_error.hangcheck_work);
> -
> - if (!mutex_trylock(&i915->drm.struct_mutex)) {
> - /* Currently busy, come back later */
> - mod_delayed_work(i915->wq,
> - &i915->gt.idle_work,
> - msecs_to_jiffies(50));
> - goto out_rearm;
> - }
> -
> - /*
> - * Flush out the last user context, leaving only the pinned
> - * kernel context resident. Should anything unfortunate happen
> - * while we are idle (such as the GPU being power cycled), no users
> - * will be harmed.
> - */
> - if (!work_pending(&i915->gt.idle_work.work) &&
> - !i915->gt.active_requests) {
> - ++i915->gt.active_requests; /* don't requeue idle */
> -
> - switch_to_kernel_context_sync(i915, i915->gt.active_engines);
> -
> - if (!--i915->gt.active_requests) {
> - __i915_gem_park(i915);
> - rearm_hangcheck = false;
> - }
> - }
> -
> - mutex_unlock(&i915->drm.struct_mutex);
> -
> -out_rearm:
> - if (rearm_hangcheck) {
> - GEM_BUG_ON(!i915->gt.awake);
> - i915_queue_hangcheck(i915);
> - }
> -}
> -
> void i915_gem_close_object(struct drm_gem_object *gem, struct drm_file *file)
> {
> struct drm_i915_private *i915 = to_i915(gem->dev);
> @@ -4412,133 +4187,6 @@ void i915_gem_sanitize(struct drm_i915_private *i915)
> mutex_unlock(&i915->drm.struct_mutex);
> }
>
> -void i915_gem_suspend(struct drm_i915_private *i915)
> -{
> - intel_wakeref_t wakeref;
> -
> - GEM_TRACE("\n");
> -
> - wakeref = intel_runtime_pm_get(i915);
> -
> - flush_workqueue(i915->wq);
> -
> - mutex_lock(&i915->drm.struct_mutex);
> -
> - /*
> - * We have to flush all the executing contexts to main memory so
> - * that they can saved in the hibernation image. To ensure the last
> - * context image is coherent, we have to switch away from it. That
> - * leaves the i915->kernel_context still active when
> - * we actually suspend, and its image in memory may not match the GPU
> - * state. Fortunately, the kernel_context is disposable and we do
> - * not rely on its state.
> - */
> - switch_to_kernel_context_sync(i915, i915->gt.active_engines);
> -
> - mutex_unlock(&i915->drm.struct_mutex);
> - i915_reset_flush(i915);
> -
> - drain_delayed_work(&i915->gt.retire_work);
> -
> - /*
> - * As the idle_work is rearming if it detects a race, play safe and
> - * repeat the flush until it is definitely idle.
> - */
> - drain_delayed_work(&i915->gt.idle_work);
> -
> - /*
> - * Assert that we successfully flushed all the work and
> - * reset the GPU back to its idle, low power state.
> - */
> - GEM_BUG_ON(i915->gt.awake);
> -
> - intel_uc_suspend(i915);
> -
> - intel_runtime_pm_put(i915, wakeref);
> -}
> -
> -void i915_gem_suspend_late(struct drm_i915_private *i915)
> -{
> - struct drm_i915_gem_object *obj;
> - struct list_head *phases[] = {
> - &i915->mm.unbound_list,
> - &i915->mm.bound_list,
> - NULL
> - }, **phase;
> -
> - /*
> - * Neither the BIOS, ourselves or any other kernel
> - * expects the system to be in execlists mode on startup,
> - * so we need to reset the GPU back to legacy mode. And the only
> - * known way to disable logical contexts is through a GPU reset.
> - *
> - * So in order to leave the system in a known default configuration,
> - * always reset the GPU upon unload and suspend. Afterwards we then
> - * clean up the GEM state tracking, flushing off the requests and
> - * leaving the system in a known idle state.
> - *
> - * Note that is of the upmost importance that the GPU is idle and
> - * all stray writes are flushed *before* we dismantle the backing
> - * storage for the pinned objects.
> - *
> - * However, since we are uncertain that resetting the GPU on older
> - * machines is a good idea, we don't - just in case it leaves the
> - * machine in an unusable condition.
> - */
> -
> - mutex_lock(&i915->drm.struct_mutex);
> - for (phase = phases; *phase; phase++) {
> - list_for_each_entry(obj, *phase, mm.link)
> - WARN_ON(i915_gem_object_set_to_gtt_domain(obj, false));
> - }
> - mutex_unlock(&i915->drm.struct_mutex);
> -
> - intel_uc_sanitize(i915);
> - i915_gem_sanitize(i915);
> -}
> -
> -void i915_gem_resume(struct drm_i915_private *i915)
> -{
> - GEM_TRACE("\n");
> -
> - WARN_ON(i915->gt.awake);
> -
> - mutex_lock(&i915->drm.struct_mutex);
> - intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
> -
> - i915_gem_restore_gtt_mappings(i915);
> - i915_gem_restore_fences(i915);
> -
> - /*
> - * As we didn't flush the kernel context before suspend, we cannot
> - * guarantee that the context image is complete. So let's just reset
> - * it and start again.
> - */
> - intel_gt_resume(i915);
> -
> - if (i915_gem_init_hw(i915))
> - goto err_wedged;
> -
> - intel_uc_resume(i915);
> -
> - /* Always reload a context for powersaving. */
> - if (!load_power_context(i915))
> - goto err_wedged;
> -
> -out_unlock:
> - intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> - mutex_unlock(&i915->drm.struct_mutex);
> - return;
> -
> -err_wedged:
> - if (!i915_reset_failed(i915)) {
> - dev_err(i915->drm.dev,
> - "Failed to re-initialize GPU, declaring it wedged!\n");
> - i915_gem_set_wedged(i915);
> - }
> - goto out_unlock;
> -}
> -
> void i915_gem_init_swizzling(struct drm_i915_private *dev_priv)
> {
> if (INTEL_GEN(dev_priv) < 5 ||
> @@ -4721,7 +4369,7 @@ static int __intel_engines_record_defaults(struct drm_i915_private *i915)
> }
>
> /* Flush the default context image to memory, and enable powersaving. */
> - if (!load_power_context(i915)) {
> + if (!i915_gem_load_power_context(i915)) {
> err = -EIO;
> goto err_active;
> }
> @@ -5115,11 +4763,8 @@ int i915_gem_init_early(struct drm_i915_private *dev_priv)
> INIT_LIST_HEAD(&dev_priv->gt.closed_vma);
>
> i915_gem_init__mm(dev_priv);
> + i915_gem_init__pm(dev_priv);
>
> - INIT_DELAYED_WORK(&dev_priv->gt.retire_work,
> - i915_gem_retire_work_handler);
> - INIT_DELAYED_WORK(&dev_priv->gt.idle_work,
> - i915_gem_idle_work_handler);
> init_waitqueue_head(&dev_priv->gpu_error.wait_queue);
> init_waitqueue_head(&dev_priv->gpu_error.reset_queue);
> mutex_init(&dev_priv->gpu_error.wedge_mutex);
> diff --git a/drivers/gpu/drm/i915/i915_gem_pm.c b/drivers/gpu/drm/i915/i915_gem_pm.c
> new file mode 100644
> index 000000000000..bc0663cac7b6
> --- /dev/null
> +++ b/drivers/gpu/drm/i915/i915_gem_pm.c
> @@ -0,0 +1,365 @@
> +/*
> + * SPDX-License-Identifier: MIT
> + *
> + * Copyright © 2019 Intel Corporation
> + */
> +
> +#include "i915_drv.h"
> +#include "i915_gem_pm.h"
> +#include "i915_globals.h"
> +#include "intel_pm.h"
> +
> +static void __i915_gem_park(struct drm_i915_private *i915)
> +{
> + intel_wakeref_t wakeref;
> +
> + GEM_TRACE("\n");
> +
> + lockdep_assert_held(&i915->drm.struct_mutex);
> + GEM_BUG_ON(i915->gt.active_requests);
> + GEM_BUG_ON(!list_empty(&i915->gt.active_rings));
> +
> + if (!i915->gt.awake)
> + return;
> +
> + /*
> + * Be paranoid and flush a concurrent interrupt to make sure
> + * we don't reactivate any irq tasklets after parking.
> + *
> + * FIXME: Note that even though we have waited for execlists to be idle,
> + * there may still be an in-flight interrupt even though the CSB
> + * is now empty. synchronize_irq() makes sure that a residual interrupt
> + * is completed before we continue, but it doesn't prevent the HW from
> + * raising a spurious interrupt later. To complete the shield we should
> + * coordinate disabling the CS irq with flushing the interrupts.
> + */
> + synchronize_irq(i915->drm.irq);
> +
> + intel_engines_park(i915);
> + i915_timelines_park(i915);
> +
> + i915_pmu_gt_parked(i915);
> + i915_vma_parked(i915);
> +
> + wakeref = fetch_and_zero(&i915->gt.awake);
> + GEM_BUG_ON(!wakeref);
> +
> + if (INTEL_GEN(i915) >= 6)
> + gen6_rps_idle(i915);
> +
> + intel_display_power_put(i915, POWER_DOMAIN_GT_IRQ, wakeref);
> +
> + i915_globals_park();
> +}
> +
> +static bool switch_to_kernel_context_sync(struct drm_i915_private *i915,
> + unsigned long mask)
> +{
> + bool result = true;
> +
> + /*
> + * Even if we fail to switch, give whatever is running a small chance
> + * to save itself before we report the failure. Yes, this may be a
> + * false positive due to e.g. ENOMEM, caveat emptor!
> + */
> + if (i915_gem_switch_to_kernel_context(i915, mask))
> + result = false;
> +
> + if (i915_gem_wait_for_idle(i915,
> + I915_WAIT_LOCKED |
> + I915_WAIT_FOR_IDLE_BOOST,
> + I915_GEM_IDLE_TIMEOUT))
> + result = false;
> +
> + if (!result) {
> + if (i915_modparams.reset) { /* XXX hide warning from gem_eio */
> + dev_err(i915->drm.dev,
> + "Failed to idle engines, declaring wedged!\n");
> + GEM_TRACE_DUMP();
> + }
> +
> + /* Forcibly cancel outstanding work and leave the gpu quiet. */
> + i915_gem_set_wedged(i915);
> + }
> +
> + i915_retire_requests(i915); /* ensure we flush after wedging */
> + return result;
> +}
> +
> +static void idle_work_handler(struct work_struct *work)
> +{
> + struct drm_i915_private *i915 =
> + container_of(work, typeof(*i915), gem.idle_work.work);
> + bool rearm_hangcheck;
> +
> + if (!READ_ONCE(i915->gt.awake))
> + return;
> +
> + if (READ_ONCE(i915->gt.active_requests))
> + return;
> +
> + rearm_hangcheck =
> + cancel_delayed_work_sync(&i915->gpu_error.hangcheck_work);
> +
> + if (!mutex_trylock(&i915->drm.struct_mutex)) {
> + /* Currently busy, come back later */
> + mod_delayed_work(i915->wq,
> + &i915->gem.idle_work,
> + msecs_to_jiffies(50));
> + goto out_rearm;
> + }
> +
> + /*
> + * Flush out the last user context, leaving only the pinned
> + * kernel context resident. Should anything unfortunate happen
> + * while we are idle (such as the GPU being power cycled), no users
> + * will be harmed.
> + */
> + if (!work_pending(&i915->gem.idle_work.work) &&
> + !i915->gt.active_requests) {
> + ++i915->gt.active_requests; /* don't requeue idle */
> +
> + switch_to_kernel_context_sync(i915, i915->gt.active_engines);
> +
> + if (!--i915->gt.active_requests) {
> + __i915_gem_park(i915);
> + rearm_hangcheck = false;
> + }
> + }
> +
> + mutex_unlock(&i915->drm.struct_mutex);
> +
> +out_rearm:
> + if (rearm_hangcheck) {
> + GEM_BUG_ON(!i915->gt.awake);
> + i915_queue_hangcheck(i915);
> + }
> +}
> +
> +static void retire_work_handler(struct work_struct *work)
> +{
> + struct drm_i915_private *i915 =
> + container_of(work, typeof(*i915), gem.retire_work.work);
> +
> + /* Come back later if the device is busy... */
> + if (mutex_trylock(&i915->drm.struct_mutex)) {
> + i915_retire_requests(i915);
> + mutex_unlock(&i915->drm.struct_mutex);
> + }
> +
> + /*
> + * Keep the retire handler running until we are finally idle.
> + * We do not need to do this test under locking as in the worst-case
> + * we queue the retire worker once too often.
> + */
> + if (READ_ONCE(i915->gt.awake))
> + queue_delayed_work(i915->wq,
> + &i915->gem.retire_work,
> + round_jiffies_up_relative(HZ));
> +}
> +
> +void i915_gem_park(struct drm_i915_private *i915)
> +{
> + GEM_TRACE("\n");
> +
> + lockdep_assert_held(&i915->drm.struct_mutex);
> + GEM_BUG_ON(i915->gt.active_requests);
> +
> + if (!i915->gt.awake)
> + return;
> +
> + /* Defer the actual call to __i915_gem_park() to prevent ping-pongs */
> + mod_delayed_work(i915->wq, &i915->gem.idle_work, msecs_to_jiffies(100));
> +}
> +
> +void i915_gem_unpark(struct drm_i915_private *i915)
> +{
> + GEM_TRACE("\n");
> +
> + lockdep_assert_held(&i915->drm.struct_mutex);
> + GEM_BUG_ON(!i915->gt.active_requests);
> + assert_rpm_wakelock_held(i915);
> +
> + if (i915->gt.awake)
> + return;
> +
> + /*
> + * It seems that the DMC likes to transition between the DC states a lot
> + * when there are no connected displays (no active power domains) during
> + * command submission.
> + *
> + * This activity has negative impact on the performance of the chip with
> + * huge latencies observed in the interrupt handler and elsewhere.
> + *
> + * Work around it by grabbing a GT IRQ power domain whilst there is any
> + * GT activity, preventing any DC state transitions.
> + */
> + i915->gt.awake = intel_display_power_get(i915, POWER_DOMAIN_GT_IRQ);
> + GEM_BUG_ON(!i915->gt.awake);
> +
> + i915_globals_unpark();
> +
> + intel_enable_gt_powersave(i915);
> + i915_update_gfx_val(i915);
> + if (INTEL_GEN(i915) >= 6)
> + gen6_rps_busy(i915);
> + i915_pmu_gt_unparked(i915);
> +
> + intel_engines_unpark(i915);
> +
> + i915_queue_hangcheck(i915);
> +
> + queue_delayed_work(i915->wq,
> + &i915->gem.retire_work,
> + round_jiffies_up_relative(HZ));
> +}
> +
> +bool i915_gem_load_power_context(struct drm_i915_private *i915)
> +{
> + /* Force loading the kernel context on all engines */
> + if (!switch_to_kernel_context_sync(i915, ALL_ENGINES))
> + return false;
> +
> + /*
> + * Immediately park the GPU so that we enable powersaving and
> + * treat it as idle. The next time we issue a request, we will
> + * unpark and start using the engine->pinned_default_state, otherwise
> + * it is in limbo and an early reset may fail.
> + */
> + __i915_gem_park(i915);
> +
> + return true;
> +}
> +
> +void i915_gem_suspend(struct drm_i915_private *i915)
> +{
> + intel_wakeref_t wakeref;
> +
> + GEM_TRACE("\n");
> +
> + wakeref = intel_runtime_pm_get(i915);
> +
> + flush_workqueue(i915->wq);
> +
> + mutex_lock(&i915->drm.struct_mutex);
> +
> + /*
> + * We have to flush all the executing contexts to main memory so
> + * that they can saved in the hibernation image. To ensure the last
> + * context image is coherent, we have to switch away from it. That
> + * leaves the i915->kernel_context still active when
> + * we actually suspend, and its image in memory may not match the GPU
> + * state. Fortunately, the kernel_context is disposable and we do
> + * not rely on its state.
> + */
> + switch_to_kernel_context_sync(i915, i915->gt.active_engines);
> +
> + mutex_unlock(&i915->drm.struct_mutex);
> + i915_reset_flush(i915);
> +
> + drain_delayed_work(&i915->gem.retire_work);
> +
> + /*
> + * As the idle_work is rearming if it detects a race, play safe and
> + * repeat the flush until it is definitely idle.
> + */
> + drain_delayed_work(&i915->gem.idle_work);
> +
> + /*
> + * Assert that we successfully flushed all the work and
> + * reset the GPU back to its idle, low power state.
> + */
> + GEM_BUG_ON(i915->gt.awake);
> +
> + intel_uc_suspend(i915);
> +
> + intel_runtime_pm_put(i915, wakeref);
> +}
> +
> +void i915_gem_suspend_late(struct drm_i915_private *i915)
> +{
> + struct drm_i915_gem_object *obj;
> + struct list_head *phases[] = {
> + &i915->mm.unbound_list,
> + &i915->mm.bound_list,
> + NULL
> + }, **phase;
> +
> + /*
> + * Neither the BIOS, ourselves or any other kernel
> + * expects the system to be in execlists mode on startup,
> + * so we need to reset the GPU back to legacy mode. And the only
> + * known way to disable logical contexts is through a GPU reset.
> + *
> + * So in order to leave the system in a known default configuration,
> + * always reset the GPU upon unload and suspend. Afterwards we then
> + * clean up the GEM state tracking, flushing off the requests and
> + * leaving the system in a known idle state.
> + *
> + * Note that is of the upmost importance that the GPU is idle and
> + * all stray writes are flushed *before* we dismantle the backing
> + * storage for the pinned objects.
> + *
> + * However, since we are uncertain that resetting the GPU on older
> + * machines is a good idea, we don't - just in case it leaves the
> + * machine in an unusable condition.
> + */
> +
> + mutex_lock(&i915->drm.struct_mutex);
> + for (phase = phases; *phase; phase++) {
> + list_for_each_entry(obj, *phase, mm.link)
> + WARN_ON(i915_gem_object_set_to_gtt_domain(obj, false));
> + }
> + mutex_unlock(&i915->drm.struct_mutex);
> +
> + intel_uc_sanitize(i915);
> + i915_gem_sanitize(i915);
> +}
> +
> +void i915_gem_resume(struct drm_i915_private *i915)
> +{
> + GEM_TRACE("\n");
> +
> + WARN_ON(i915->gt.awake);
> +
> + mutex_lock(&i915->drm.struct_mutex);
> + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
> +
> + i915_gem_restore_gtt_mappings(i915);
> + i915_gem_restore_fences(i915);
> +
> + /*
> + * As we didn't flush the kernel context before suspend, we cannot
> + * guarantee that the context image is complete. So let's just reset
> + * it and start again.
> + */
> + intel_gt_resume(i915);
> +
> + if (i915_gem_init_hw(i915))
> + goto err_wedged;
> +
> + intel_uc_resume(i915);
> +
> + /* Always reload a context for powersaving. */
> + if (!i915_gem_load_power_context(i915))
> + goto err_wedged;
> +
> +out_unlock:
> + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> + mutex_unlock(&i915->drm.struct_mutex);
> + return;
> +
> +err_wedged:
> + if (!i915_reset_failed(i915)) {
> + dev_err(i915->drm.dev,
> + "Failed to re-initialize GPU, declaring it wedged!\n");
> + i915_gem_set_wedged(i915);
> + }
> + goto out_unlock;
> +}
> +
> +void i915_gem_init__pm(struct drm_i915_private *i915)
> +{
> + INIT_DELAYED_WORK(&i915->gem.idle_work, idle_work_handler);
> + INIT_DELAYED_WORK(&i915->gem.retire_work, retire_work_handler);
> +}
> diff --git a/drivers/gpu/drm/i915/i915_gem_pm.h b/drivers/gpu/drm/i915/i915_gem_pm.h
> new file mode 100644
> index 000000000000..52f65e3f06b5
> --- /dev/null
> +++ b/drivers/gpu/drm/i915/i915_gem_pm.h
> @@ -0,0 +1,28 @@
> +/*
> + * SPDX-License-Identifier: MIT
> + *
> + * Copyright © 2019 Intel Corporation
> + */
> +
> +#ifndef __I915_GEM_PM_H__
> +#define __I915_GEM_PM_H__
> +
> +#include <linux/types.h>
> +
> +struct drm_i915_private;
> +struct work_struct;
> +
> +void i915_gem_init__pm(struct drm_i915_private *i915);
> +
> +bool i915_gem_load_power_context(struct drm_i915_private *i915);
> +void i915_gem_resume(struct drm_i915_private *i915);
> +
> +void i915_gem_unpark(struct drm_i915_private *i915);
> +void i915_gem_park(struct drm_i915_private *i915);
> +
> +void i915_gem_idle_work_handler(struct work_struct *work);
> +
> +void i915_gem_suspend(struct drm_i915_private *i915);
> +void i915_gem_suspend_late(struct drm_i915_private *i915);
> +
> +#endif /* __I915_GEM_PM_H__ */
> diff --git a/drivers/gpu/drm/i915/selftests/i915_gem_context.c b/drivers/gpu/drm/i915/selftests/i915_gem_context.c
> index 6f52ca881173..9d646fa1b74e 100644
> --- a/drivers/gpu/drm/i915/selftests/i915_gem_context.c
> +++ b/drivers/gpu/drm/i915/selftests/i915_gem_context.c
> @@ -1658,7 +1658,7 @@ static int __igt_switch_to_kernel_context(struct drm_i915_private *i915,
> /* XXX Bonus points for proving we are the kernel context! */
>
> mutex_unlock(&i915->drm.struct_mutex);
> - drain_delayed_work(&i915->gt.idle_work);
> + drain_delayed_work(&i915->gem.idle_work);
> mutex_lock(&i915->drm.struct_mutex);
> }
>
> diff --git a/drivers/gpu/drm/i915/selftests/i915_gem_object.c b/drivers/gpu/drm/i915/selftests/i915_gem_object.c
> index 971148fbe6f5..12fc53c694a6 100644
> --- a/drivers/gpu/drm/i915/selftests/i915_gem_object.c
> +++ b/drivers/gpu/drm/i915/selftests/i915_gem_object.c
> @@ -514,8 +514,8 @@ static void disable_retire_worker(struct drm_i915_private *i915)
> }
> mutex_unlock(&i915->drm.struct_mutex);
>
> - cancel_delayed_work_sync(&i915->gt.retire_work);
> - cancel_delayed_work_sync(&i915->gt.idle_work);
> + cancel_delayed_work_sync(&i915->gem.retire_work);
> + cancel_delayed_work_sync(&i915->gem.idle_work);
> }
>
> static int igt_mmap_offset_exhaustion(void *arg)
> @@ -617,9 +617,9 @@ static int igt_mmap_offset_exhaustion(void *arg)
> out_park:
> mutex_lock(&i915->drm.struct_mutex);
> if (--i915->gt.active_requests)
> - queue_delayed_work(i915->wq, &i915->gt.retire_work, 0);
> + queue_delayed_work(i915->wq, &i915->gem.retire_work, 0);
> else
> - queue_delayed_work(i915->wq, &i915->gt.idle_work, 0);
> + queue_delayed_work(i915->wq, &i915->gem.idle_work, 0);
> mutex_unlock(&i915->drm.struct_mutex);
> i915_gem_shrinker_register(i915);
> return err;
> diff --git a/drivers/gpu/drm/i915/selftests/mock_gem_device.c b/drivers/gpu/drm/i915/selftests/mock_gem_device.c
> index f444ee5add27..fb677b4019a0 100644
> --- a/drivers/gpu/drm/i915/selftests/mock_gem_device.c
> +++ b/drivers/gpu/drm/i915/selftests/mock_gem_device.c
> @@ -59,8 +59,8 @@ static void mock_device_release(struct drm_device *dev)
> i915_gem_contexts_lost(i915);
> mutex_unlock(&i915->drm.struct_mutex);
>
> - drain_delayed_work(&i915->gt.retire_work);
> - drain_delayed_work(&i915->gt.idle_work);
> + drain_delayed_work(&i915->gem.retire_work);
> + drain_delayed_work(&i915->gem.idle_work);
> i915_gem_drain_workqueue(i915);
>
> mutex_lock(&i915->drm.struct_mutex);
> @@ -111,7 +111,7 @@ static void mock_retire_work_handler(struct work_struct *work)
> static void mock_idle_work_handler(struct work_struct *work)
> {
> struct drm_i915_private *i915 =
> - container_of(work, typeof(*i915), gt.idle_work.work);
> + container_of(work, typeof(*i915), gem.idle_work.work);
>
> i915->gt.active_engines = 0;
> }
> @@ -197,8 +197,8 @@ struct drm_i915_private *mock_gem_device(void)
>
> mock_init_contexts(i915);
>
> - INIT_DELAYED_WORK(&i915->gt.retire_work, mock_retire_work_handler);
> - INIT_DELAYED_WORK(&i915->gt.idle_work, mock_idle_work_handler);
> + INIT_DELAYED_WORK(&i915->gem.retire_work, mock_retire_work_handler);
> + INIT_DELAYED_WORK(&i915->gem.idle_work, mock_idle_work_handler);
>
> i915->gt.awake = true;
>
>
More information about the Intel-gfx
mailing list