Split out the powermanagement portion (GT wakeref, suspend/resume) of
GEM from i915_gem.c into its own file.

Signed-off-by: Chris Wilson <ch...@chris-wilson.co.uk>
Reviewed-by: Tvrtko Ursulin <tvrtko.ursu...@intel.com>
---
 drivers/gpu/drm/i915/Makefile                 |   1 +
 drivers/gpu/drm/i915/Makefile.header-test     |   1 +
 drivers/gpu/drm/i915/i915_debugfs.c           |   4 +-
 drivers/gpu/drm/i915/i915_drv.h               |  12 +-
 drivers/gpu/drm/i915/i915_gem.c               | 363 +----------------
 drivers/gpu/drm/i915/i915_gem_pm.c            | 365 ++++++++++++++++++
 drivers/gpu/drm/i915/i915_gem_pm.h            |  28 ++
 .../gpu/drm/i915/selftests/i915_gem_context.c |   2 +-
 .../gpu/drm/i915/selftests/i915_gem_object.c  |   8 +-
 .../gpu/drm/i915/selftests/mock_gem_device.c  |  10 +-
 10 files changed, 418 insertions(+), 376 deletions(-)
 create mode 100644 drivers/gpu/drm/i915/i915_gem_pm.c
 create mode 100644 drivers/gpu/drm/i915/i915_gem_pm.h

diff --git a/drivers/gpu/drm/i915/Makefile b/drivers/gpu/drm/i915/Makefile
index 233bad5e361f..858642c7bc40 100644
--- a/drivers/gpu/drm/i915/Makefile
+++ b/drivers/gpu/drm/i915/Makefile
@@ -97,6 +97,7 @@ i915-y += \
          i915_gem_internal.o \
          i915_gem.o \
          i915_gem_object.o \
+         i915_gem_pm.o \
          i915_gem_render_state.o \
          i915_gem_shrinker.o \
          i915_gem_stolen.o \
diff --git a/drivers/gpu/drm/i915/Makefile.header-test 
b/drivers/gpu/drm/i915/Makefile.header-test
index e6b3e7588860..702e3a7ade4c 100644
--- a/drivers/gpu/drm/i915/Makefile.header-test
+++ b/drivers/gpu/drm/i915/Makefile.header-test
@@ -5,6 +5,7 @@
 header_test := \
        i915_active_types.h \
        i915_gem_context_types.h \
+       i915_gem_pm.h \
        i915_priolist_types.h \
        i915_scheduler_types.h \
        i915_timeline_types.h \
diff --git a/drivers/gpu/drm/i915/i915_debugfs.c 
b/drivers/gpu/drm/i915/i915_debugfs.c
index 3f039758b152..8dcba78fb43b 100644
--- a/drivers/gpu/drm/i915/i915_debugfs.c
+++ b/drivers/gpu/drm/i915/i915_debugfs.c
@@ -3941,8 +3941,8 @@ i915_drop_caches_set(void *data, u64 val)
        if (val & DROP_IDLE) {
                do {
                        if (READ_ONCE(i915->gt.active_requests))
-                               flush_delayed_work(&i915->gt.retire_work);
-                       drain_delayed_work(&i915->gt.idle_work);
+                               flush_delayed_work(&i915->gem.retire_work);
+                       drain_delayed_work(&i915->gem.idle_work);
                } while (READ_ONCE(i915->gt.awake));
        }
 
diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h
index 437e394d9fa6..45e027f45e62 100644
--- a/drivers/gpu/drm/i915/i915_drv.h
+++ b/drivers/gpu/drm/i915/i915_drv.h
@@ -2020,6 +2020,12 @@ struct drm_i915_private {
                 */
                intel_wakeref_t awake;
 
+               ktime_t last_init_time;
+
+               struct i915_vma *scratch;
+       } gt;
+
+       struct {
                /**
                 * We leave the user IRQ off as much as possible,
                 * but this means that requests will finish and never
@@ -2037,11 +2043,7 @@ struct drm_i915_private {
                 * off the idle_work.
                 */
                struct delayed_work idle_work;
-
-               ktime_t last_init_time;
-
-               struct i915_vma *scratch;
-       } gt;
+       } gem;
 
        /* For i945gm vblank irq vs. C3 workaround */
        struct {
diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c
index 21adeb340357..7f833c97138e 100644
--- a/drivers/gpu/drm/i915/i915_gem.c
+++ b/drivers/gpu/drm/i915/i915_gem.c
@@ -46,7 +46,7 @@
 #include "i915_drv.h"
 #include "i915_gem_clflush.h"
 #include "i915_gemfs.h"
-#include "i915_globals.h"
+#include "i915_gem_pm.h"
 #include "i915_trace.h"
 #include "i915_vgpu.h"
 
@@ -103,105 +103,6 @@ static void i915_gem_info_remove_obj(struct 
drm_i915_private *dev_priv,
        spin_unlock(&dev_priv->mm.object_stat_lock);
 }
 
-static void __i915_gem_park(struct drm_i915_private *i915)
-{
-       intel_wakeref_t wakeref;
-
-       GEM_TRACE("\n");
-
-       lockdep_assert_held(&i915->drm.struct_mutex);
-       GEM_BUG_ON(i915->gt.active_requests);
-       GEM_BUG_ON(!list_empty(&i915->gt.active_rings));
-
-       if (!i915->gt.awake)
-               return;
-
-       /*
-        * Be paranoid and flush a concurrent interrupt to make sure
-        * we don't reactivate any irq tasklets after parking.
-        *
-        * FIXME: Note that even though we have waited for execlists to be idle,
-        * there may still be an in-flight interrupt even though the CSB
-        * is now empty. synchronize_irq() makes sure that a residual interrupt
-        * is completed before we continue, but it doesn't prevent the HW from
-        * raising a spurious interrupt later. To complete the shield we should
-        * coordinate disabling the CS irq with flushing the interrupts.
-        */
-       synchronize_irq(i915->drm.irq);
-
-       intel_engines_park(i915);
-       i915_timelines_park(i915);
-
-       i915_pmu_gt_parked(i915);
-       i915_vma_parked(i915);
-
-       wakeref = fetch_and_zero(&i915->gt.awake);
-       GEM_BUG_ON(!wakeref);
-
-       if (INTEL_GEN(i915) >= 6)
-               gen6_rps_idle(i915);
-
-       intel_display_power_put(i915, POWER_DOMAIN_GT_IRQ, wakeref);
-
-       i915_globals_park();
-}
-
-void i915_gem_park(struct drm_i915_private *i915)
-{
-       GEM_TRACE("\n");
-
-       lockdep_assert_held(&i915->drm.struct_mutex);
-       GEM_BUG_ON(i915->gt.active_requests);
-
-       if (!i915->gt.awake)
-               return;
-
-       /* Defer the actual call to __i915_gem_park() to prevent ping-pongs */
-       mod_delayed_work(i915->wq, &i915->gt.idle_work, msecs_to_jiffies(100));
-}
-
-void i915_gem_unpark(struct drm_i915_private *i915)
-{
-       GEM_TRACE("\n");
-
-       lockdep_assert_held(&i915->drm.struct_mutex);
-       GEM_BUG_ON(!i915->gt.active_requests);
-       assert_rpm_wakelock_held(i915);
-
-       if (i915->gt.awake)
-               return;
-
-       /*
-        * It seems that the DMC likes to transition between the DC states a lot
-        * when there are no connected displays (no active power domains) during
-        * command submission.
-        *
-        * This activity has negative impact on the performance of the chip with
-        * huge latencies observed in the interrupt handler and elsewhere.
-        *
-        * Work around it by grabbing a GT IRQ power domain whilst there is any
-        * GT activity, preventing any DC state transitions.
-        */
-       i915->gt.awake = intel_display_power_get(i915, POWER_DOMAIN_GT_IRQ);
-       GEM_BUG_ON(!i915->gt.awake);
-
-       i915_globals_unpark();
-
-       intel_enable_gt_powersave(i915);
-       i915_update_gfx_val(i915);
-       if (INTEL_GEN(i915) >= 6)
-               gen6_rps_busy(i915);
-       i915_pmu_gt_unparked(i915);
-
-       intel_engines_unpark(i915);
-
-       i915_queue_hangcheck(i915);
-
-       queue_delayed_work(i915->wq,
-                          &i915->gt.retire_work,
-                          round_jiffies_up_relative(HZ));
-}
-
 int
 i915_gem_get_aperture_ioctl(struct drm_device *dev, void *data,
                            struct drm_file *file)
@@ -2088,7 +1989,7 @@ static int i915_gem_object_create_mmap_offset(struct 
drm_i915_gem_object *obj)
                if (!err)
                        break;
 
-       } while (flush_delayed_work(&dev_priv->gt.retire_work));
+       } while (flush_delayed_work(&dev_priv->gem.retire_work));
 
        return err;
 }
@@ -2848,132 +2749,6 @@ i915_gem_object_pwrite_gtt(struct drm_i915_gem_object 
*obj,
        return 0;
 }
 
-static void
-i915_gem_retire_work_handler(struct work_struct *work)
-{
-       struct drm_i915_private *dev_priv =
-               container_of(work, typeof(*dev_priv), gt.retire_work.work);
-       struct drm_device *dev = &dev_priv->drm;
-
-       /* Come back later if the device is busy... */
-       if (mutex_trylock(&dev->struct_mutex)) {
-               i915_retire_requests(dev_priv);
-               mutex_unlock(&dev->struct_mutex);
-       }
-
-       /*
-        * Keep the retire handler running until we are finally idle.
-        * We do not need to do this test under locking as in the worst-case
-        * we queue the retire worker once too often.
-        */
-       if (READ_ONCE(dev_priv->gt.awake))
-               queue_delayed_work(dev_priv->wq,
-                                  &dev_priv->gt.retire_work,
-                                  round_jiffies_up_relative(HZ));
-}
-
-static bool switch_to_kernel_context_sync(struct drm_i915_private *i915,
-                                         unsigned long mask)
-{
-       bool result = true;
-
-       /*
-        * Even if we fail to switch, give whatever is running a small chance
-        * to save itself before we report the failure. Yes, this may be a
-        * false positive due to e.g. ENOMEM, caveat emptor!
-        */
-       if (i915_gem_switch_to_kernel_context(i915, mask))
-               result = false;
-
-       if (i915_gem_wait_for_idle(i915,
-                                  I915_WAIT_LOCKED |
-                                  I915_WAIT_FOR_IDLE_BOOST,
-                                  I915_GEM_IDLE_TIMEOUT))
-               result = false;
-
-       if (!result) {
-               if (i915_modparams.reset) { /* XXX hide warning from gem_eio */
-                       dev_err(i915->drm.dev,
-                               "Failed to idle engines, declaring wedged!\n");
-                       GEM_TRACE_DUMP();
-               }
-
-               /* Forcibly cancel outstanding work and leave the gpu quiet. */
-               i915_gem_set_wedged(i915);
-       }
-
-       i915_retire_requests(i915); /* ensure we flush after wedging */
-       return result;
-}
-
-static bool load_power_context(struct drm_i915_private *i915)
-{
-       /* Force loading the kernel context on all engines */
-       if (!switch_to_kernel_context_sync(i915, ALL_ENGINES))
-               return false;
-
-       /*
-        * Immediately park the GPU so that we enable powersaving and
-        * treat it as idle. The next time we issue a request, we will
-        * unpark and start using the engine->pinned_default_state, otherwise
-        * it is in limbo and an early reset may fail.
-        */
-       __i915_gem_park(i915);
-
-       return true;
-}
-
-static void
-i915_gem_idle_work_handler(struct work_struct *work)
-{
-       struct drm_i915_private *i915 =
-               container_of(work, typeof(*i915), gt.idle_work.work);
-       bool rearm_hangcheck;
-
-       if (!READ_ONCE(i915->gt.awake))
-               return;
-
-       if (READ_ONCE(i915->gt.active_requests))
-               return;
-
-       rearm_hangcheck =
-               cancel_delayed_work_sync(&i915->gpu_error.hangcheck_work);
-
-       if (!mutex_trylock(&i915->drm.struct_mutex)) {
-               /* Currently busy, come back later */
-               mod_delayed_work(i915->wq,
-                                &i915->gt.idle_work,
-                                msecs_to_jiffies(50));
-               goto out_rearm;
-       }
-
-       /*
-        * Flush out the last user context, leaving only the pinned
-        * kernel context resident. Should anything unfortunate happen
-        * while we are idle (such as the GPU being power cycled), no users
-        * will be harmed.
-        */
-       if (!work_pending(&i915->gt.idle_work.work) &&
-           !i915->gt.active_requests) {
-               ++i915->gt.active_requests; /* don't requeue idle */
-
-               switch_to_kernel_context_sync(i915, i915->gt.active_engines);
-
-               if (!--i915->gt.active_requests) {
-                       __i915_gem_park(i915);
-                       rearm_hangcheck = false;
-               }
-       }
-
-       mutex_unlock(&i915->drm.struct_mutex);
-
-out_rearm:
-       if (rearm_hangcheck) {
-               GEM_BUG_ON(!i915->gt.awake);
-               i915_queue_hangcheck(i915);
-       }
-}
-
 void i915_gem_close_object(struct drm_gem_object *gem, struct drm_file *file)
 {
        struct drm_i915_private *i915 = to_i915(gem->dev);
@@ -4389,133 +4164,6 @@ void i915_gem_sanitize(struct drm_i915_private *i915)
        mutex_unlock(&i915->drm.struct_mutex);
 }
 
-void i915_gem_suspend(struct drm_i915_private *i915)
-{
-       intel_wakeref_t wakeref;
-
-       GEM_TRACE("\n");
-
-       wakeref = intel_runtime_pm_get(i915);
-
-       flush_workqueue(i915->wq);
-
-       mutex_lock(&i915->drm.struct_mutex);
-
-       /*
-        * We have to flush all the executing contexts to main memory so
-        * that they can saved in the hibernation image. To ensure the last
-        * context image is coherent, we have to switch away from it. That
-        * leaves the i915->kernel_context still active when
-        * we actually suspend, and its image in memory may not match the GPU
-        * state. Fortunately, the kernel_context is disposable and we do
-        * not rely on its state.
-        */
-       switch_to_kernel_context_sync(i915, i915->gt.active_engines);
-
-       mutex_unlock(&i915->drm.struct_mutex);
-       i915_reset_flush(i915);
-
-       drain_delayed_work(&i915->gt.retire_work);
-
-       /*
-        * As the idle_work is rearming if it detects a race, play safe and
-        * repeat the flush until it is definitely idle.
-        */
-       drain_delayed_work(&i915->gt.idle_work);
-
-       /*
-        * Assert that we successfully flushed all the work and
-        * reset the GPU back to its idle, low power state.
-        */
-       GEM_BUG_ON(i915->gt.awake);
-
-       intel_uc_suspend(i915);
-
-       intel_runtime_pm_put(i915, wakeref);
-}
-
-void i915_gem_suspend_late(struct drm_i915_private *i915)
-{
-       struct drm_i915_gem_object *obj;
-       struct list_head *phases[] = {
-               &i915->mm.unbound_list,
-               &i915->mm.bound_list,
-               NULL
-       }, **phase;
-
-       /*
-        * Neither the BIOS, ourselves or any other kernel
-        * expects the system to be in execlists mode on startup,
-        * so we need to reset the GPU back to legacy mode. And the only
-        * known way to disable logical contexts is through a GPU reset.
-        *
-        * So in order to leave the system in a known default configuration,
-        * always reset the GPU upon unload and suspend. Afterwards we then
-        * clean up the GEM state tracking, flushing off the requests and
-        * leaving the system in a known idle state.
-        *
-        * Note that is of the upmost importance that the GPU is idle and
-        * all stray writes are flushed *before* we dismantle the backing
-        * storage for the pinned objects.
-        *
-        * However, since we are uncertain that resetting the GPU on older
-        * machines is a good idea, we don't - just in case it leaves the
-        * machine in an unusable condition.
-        */
-
-       mutex_lock(&i915->drm.struct_mutex);
-       for (phase = phases; *phase; phase++) {
-               list_for_each_entry(obj, *phase, mm.link)
-                       WARN_ON(i915_gem_object_set_to_gtt_domain(obj, false));
-       }
-       mutex_unlock(&i915->drm.struct_mutex);
-
-       intel_uc_sanitize(i915);
-       i915_gem_sanitize(i915);
-}
-
-void i915_gem_resume(struct drm_i915_private *i915)
-{
-       GEM_TRACE("\n");
-
-       WARN_ON(i915->gt.awake);
-
-       mutex_lock(&i915->drm.struct_mutex);
-       intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
-
-       i915_gem_restore_gtt_mappings(i915);
-       i915_gem_restore_fences(i915);
-
-       /*
-        * As we didn't flush the kernel context before suspend, we cannot
-        * guarantee that the context image is complete. So let's just reset
-        * it and start again.
-        */
-       intel_gt_resume(i915);
-
-       if (i915_gem_init_hw(i915))
-               goto err_wedged;
-
-       intel_uc_resume(i915);
-
-       /* Always reload a context for powersaving. */
-       if (!load_power_context(i915))
-               goto err_wedged;
-
-out_unlock:
-       intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
-       mutex_unlock(&i915->drm.struct_mutex);
-       return;
-
-err_wedged:
-       if (!i915_reset_failed(i915)) {
-               dev_err(i915->drm.dev,
-                       "Failed to re-initialize GPU, declaring it wedged!\n");
-               i915_gem_set_wedged(i915);
-       }
-       goto out_unlock;
-}
-
 void i915_gem_init_swizzling(struct drm_i915_private *dev_priv)
 {
        if (INTEL_GEN(dev_priv) < 5 ||
@@ -4698,7 +4346,7 @@ static int __intel_engines_record_defaults(struct 
drm_i915_private *i915)
        }
 
        /* Flush the default context image to memory, and enable powersaving. */
-       if (!load_power_context(i915)) {
+       if (!i915_gem_load_power_context(i915)) {
                err = -EIO;
                goto err_active;
        }
@@ -5113,11 +4761,8 @@ int i915_gem_init_early(struct drm_i915_private 
*dev_priv)
        INIT_LIST_HEAD(&dev_priv->gt.closed_vma);
 
        i915_gem_init__mm(dev_priv);
+       i915_gem_init__pm(dev_priv);
 
-       INIT_DELAYED_WORK(&dev_priv->gt.retire_work,
-                         i915_gem_retire_work_handler);
-       INIT_DELAYED_WORK(&dev_priv->gt.idle_work,
-                         i915_gem_idle_work_handler);
        init_waitqueue_head(&dev_priv->gpu_error.wait_queue);
        init_waitqueue_head(&dev_priv->gpu_error.reset_queue);
        mutex_init(&dev_priv->gpu_error.wedge_mutex);
diff --git a/drivers/gpu/drm/i915/i915_gem_pm.c 
b/drivers/gpu/drm/i915/i915_gem_pm.c
new file mode 100644
index 000000000000..9fb0e8d567a2
--- /dev/null
+++ b/drivers/gpu/drm/i915/i915_gem_pm.c
@@ -0,0 +1,365 @@
+/*
+ * SPDX-License-Identifier: MIT
+ *
+ * Copyright © 2019 Intel Corporation
+ */
+
+#include "i915_drv.h"
+#include "i915_gem_pm.h"
+#include "i915_globals.h"
+#include "intel_pm.h"
+
+static void __i915_gem_park(struct drm_i915_private *i915)
+{
+       intel_wakeref_t wakeref;
+
+       GEM_TRACE("\n");
+
+       lockdep_assert_held(&i915->drm.struct_mutex);
+       GEM_BUG_ON(i915->gt.active_requests);
+       GEM_BUG_ON(!list_empty(&i915->gt.active_rings));
+
+       if (!i915->gt.awake)
+               return;
+
+       /*
+        * Be paranoid and flush a concurrent interrupt to make sure
+        * we don't reactivate any irq tasklets after parking.
+        *
+        * FIXME: Note that even though we have waited for execlists to be idle,
+        * there may still be an in-flight interrupt even though the CSB
+        * is now empty. synchronize_irq() makes sure that a residual interrupt
+        * is completed before we continue, but it doesn't prevent the HW from
+        * raising a spurious interrupt later. To complete the shield we should
+        * coordinate disabling the CS irq with flushing the interrupts.
+        */
+       synchronize_irq(i915->drm.irq);
+
+       intel_engines_park(i915);
+       i915_timelines_park(i915);
+
+       i915_pmu_gt_parked(i915);
+       i915_vma_parked(i915);
+
+       wakeref = fetch_and_zero(&i915->gt.awake);
+       GEM_BUG_ON(!wakeref);
+
+       if (INTEL_GEN(i915) >= 6)
+               gen6_rps_idle(i915);
+
+       intel_display_power_put(i915, POWER_DOMAIN_GT_IRQ, wakeref);
+
+       i915_globals_park();
+}
+
+static bool switch_to_kernel_context_sync(struct drm_i915_private *i915,
+                                         unsigned long mask)
+{
+       bool result = true;
+
+       /*
+        * Even if we fail to switch, give whatever is running a small chance
+        * to save itself before we report the failure. Yes, this may be a
+        * false positive due to e.g. ENOMEM, caveat emptor!
+        */
+       if (i915_gem_switch_to_kernel_context(i915, mask))
+               result = false;
+
+       if (i915_gem_wait_for_idle(i915,
+                                  I915_WAIT_LOCKED |
+                                  I915_WAIT_FOR_IDLE_BOOST,
+                                  I915_GEM_IDLE_TIMEOUT))
+               result = false;
+
+       if (!result) {
+               if (i915_modparams.reset) { /* XXX hide warning from gem_eio */
+                       dev_err(i915->drm.dev,
+                               "Failed to idle engines, declaring wedged!\n");
+                       GEM_TRACE_DUMP();
+               }
+
+               /* Forcibly cancel outstanding work and leave the gpu quiet. */
+               i915_gem_set_wedged(i915);
+       }
+
+       i915_retire_requests(i915); /* ensure we flush after wedging */
+       return result;
+}
+
+static void idle_work_handler(struct work_struct *work)
+{
+       struct drm_i915_private *i915 =
+               container_of(work, typeof(*i915), gem.idle_work.work);
+       bool rearm_hangcheck;
+
+       if (!READ_ONCE(i915->gt.awake))
+               return;
+
+       if (READ_ONCE(i915->gt.active_requests))
+               return;
+
+       rearm_hangcheck =
+               cancel_delayed_work_sync(&i915->gpu_error.hangcheck_work);
+
+       if (!mutex_trylock(&i915->drm.struct_mutex)) {
+               /* Currently busy, come back later */
+               mod_delayed_work(i915->wq,
+                                &i915->gem.idle_work,
+                                msecs_to_jiffies(50));
+               goto out_rearm;
+       }
+
+       /*
+        * Flush out the last user context, leaving only the pinned
+        * kernel context resident. Should anything unfortunate happen
+        * while we are idle (such as the GPU being power cycled), no users
+        * will be harmed.
+        */
+       if (!work_pending(&i915->gem.idle_work.work) &&
+           !i915->gt.active_requests) {
+               ++i915->gt.active_requests; /* don't requeue idle */
+
+               switch_to_kernel_context_sync(i915, i915->gt.active_engines);
+
+               if (!--i915->gt.active_requests) {
+                       __i915_gem_park(i915);
+                       rearm_hangcheck = false;
+               }
+       }
+
+       mutex_unlock(&i915->drm.struct_mutex);
+
+out_rearm:
+       if (rearm_hangcheck) {
+               GEM_BUG_ON(!i915->gt.awake);
+               i915_queue_hangcheck(i915);
+       }
+}
+
+static void retire_work_handler(struct work_struct *work)
+{
+       struct drm_i915_private *i915 =
+               container_of(work, typeof(*i915), gem.retire_work.work);
+
+       /* Come back later if the device is busy... */
+       if (mutex_trylock(&i915->drm.struct_mutex)) {
+               i915_retire_requests(i915);
+               mutex_unlock(&i915->drm.struct_mutex);
+       }
+
+       /*
+        * Keep the retire handler running until we are finally idle.
+        * We do not need to do this test under locking as in the worst-case
+        * we queue the retire worker once too often.
+        */
+       if (READ_ONCE(i915->gt.awake))
+               queue_delayed_work(i915->wq,
+                                  &i915->gem.retire_work,
+                                  round_jiffies_up_relative(HZ));
+}
+
+void i915_gem_park(struct drm_i915_private *i915)
+{
+       GEM_TRACE("\n");
+
+       lockdep_assert_held(&i915->drm.struct_mutex);
+       GEM_BUG_ON(i915->gt.active_requests);
+
+       if (!i915->gt.awake)
+               return;
+
+       /* Defer the actual call to __i915_gem_park() to prevent ping-pongs */
+       mod_delayed_work(i915->wq, &i915->gem.idle_work, msecs_to_jiffies(100));
+}
+
+void i915_gem_unpark(struct drm_i915_private *i915)
+{
+       GEM_TRACE("\n");
+
+       lockdep_assert_held(&i915->drm.struct_mutex);
+       GEM_BUG_ON(!i915->gt.active_requests);
+       assert_rpm_wakelock_held(i915);
+
+       if (i915->gt.awake)
+               return;
+
+       /*
+        * It seems that the DMC likes to transition between the DC states a lot
+        * when there are no connected displays (no active power domains) during
+        * command submission.
+        *
+        * This activity has negative impact on the performance of the chip with
+        * huge latencies observed in the interrupt handler and elsewhere.
+        *
+        * Work around it by grabbing a GT IRQ power domain whilst there is any
+        * GT activity, preventing any DC state transitions.
+        */
+       i915->gt.awake = intel_display_power_get(i915, POWER_DOMAIN_GT_IRQ);
+       GEM_BUG_ON(!i915->gt.awake);
+
+       i915_globals_unpark();
+
+       intel_enable_gt_powersave(i915);
+       i915_update_gfx_val(i915);
+       if (INTEL_GEN(i915) >= 6)
+               gen6_rps_busy(i915);
+       i915_pmu_gt_unparked(i915);
+
+       intel_engines_unpark(i915);
+
+       i915_queue_hangcheck(i915);
+
+       queue_delayed_work(i915->wq,
+                          &i915->gem.retire_work,
+                          round_jiffies_up_relative(HZ));
+}
+
+bool i915_gem_load_power_context(struct drm_i915_private *i915)
+{
+       /* Force loading the kernel context on all engines */
+       if (!switch_to_kernel_context_sync(i915, ALL_ENGINES))
+               return false;
+
+       /*
+        * Immediately park the GPU so that we enable powersaving and
+        * treat it as idle. The next time we issue a request, we will
+        * unpark and start using the engine->pinned_default_state, otherwise
+        * it is in limbo and an early reset may fail.
+        */
+       __i915_gem_park(i915);
+
+       return true;
+}
+
+void i915_gem_suspend(struct drm_i915_private *i915)
+{
+       intel_wakeref_t wakeref;
+
+       GEM_TRACE("\n");
+
+       wakeref = intel_runtime_pm_get(i915);
+
+       mutex_lock(&i915->drm.struct_mutex);
+
+       /*
+        * We have to flush all the executing contexts to main memory so
+        * that they can saved in the hibernation image. To ensure the last
+        * context image is coherent, we have to switch away from it. That
+        * leaves the i915->kernel_context still active when
+        * we actually suspend, and its image in memory may not match the GPU
+        * state. Fortunately, the kernel_context is disposable and we do
+        * not rely on its state.
+        */
+       switch_to_kernel_context_sync(i915, i915->gt.active_engines);
+
+       mutex_unlock(&i915->drm.struct_mutex);
+       i915_reset_flush(i915);
+
+       drain_delayed_work(&i915->gem.retire_work);
+
+       /*
+        * As the idle_work is rearming if it detects a race, play safe and
+        * repeat the flush until it is definitely idle.
+        */
+       drain_delayed_work(&i915->gem.idle_work);
+
+       flush_workqueue(i915->wq);
+
+       /*
+        * Assert that we successfully flushed all the work and
+        * reset the GPU back to its idle, low power state.
+        */
+       GEM_BUG_ON(i915->gt.awake);
+
+       intel_uc_suspend(i915);
+
+       intel_runtime_pm_put(i915, wakeref);
+}
+
+void i915_gem_suspend_late(struct drm_i915_private *i915)
+{
+       struct drm_i915_gem_object *obj;
+       struct list_head *phases[] = {
+               &i915->mm.unbound_list,
+               &i915->mm.bound_list,
+               NULL
+       }, **phase;
+
+       /*
+        * Neither the BIOS, ourselves or any other kernel
+        * expects the system to be in execlists mode on startup,
+        * so we need to reset the GPU back to legacy mode. And the only
+        * known way to disable logical contexts is through a GPU reset.
+        *
+        * So in order to leave the system in a known default configuration,
+        * always reset the GPU upon unload and suspend. Afterwards we then
+        * clean up the GEM state tracking, flushing off the requests and
+        * leaving the system in a known idle state.
+        *
+        * Note that is of the upmost importance that the GPU is idle and
+        * all stray writes are flushed *before* we dismantle the backing
+        * storage for the pinned objects.
+        *
+        * However, since we are uncertain that resetting the GPU on older
+        * machines is a good idea, we don't - just in case it leaves the
+        * machine in an unusable condition.
+        */
+
+       mutex_lock(&i915->drm.struct_mutex);
+       for (phase = phases; *phase; phase++) {
+               list_for_each_entry(obj, *phase, mm.link)
+                       WARN_ON(i915_gem_object_set_to_gtt_domain(obj, false));
+       }
+       mutex_unlock(&i915->drm.struct_mutex);
+
+       intel_uc_sanitize(i915);
+       i915_gem_sanitize(i915);
+}
+
+void i915_gem_resume(struct drm_i915_private *i915)
+{
+       GEM_TRACE("\n");
+
+       WARN_ON(i915->gt.awake);
+
+       mutex_lock(&i915->drm.struct_mutex);
+       intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
+
+       i915_gem_restore_gtt_mappings(i915);
+       i915_gem_restore_fences(i915);
+
+       /*
+        * As we didn't flush the kernel context before suspend, we cannot
+        * guarantee that the context image is complete. So let's just reset
+        * it and start again.
+        */
+       intel_gt_resume(i915);
+
+       if (i915_gem_init_hw(i915))
+               goto err_wedged;
+
+       intel_uc_resume(i915);
+
+       /* Always reload a context for powersaving. */
+       if (!i915_gem_load_power_context(i915))
+               goto err_wedged;
+
+out_unlock:
+       intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
+       mutex_unlock(&i915->drm.struct_mutex);
+       return;
+
+err_wedged:
+       if (!i915_reset_failed(i915)) {
+               dev_err(i915->drm.dev,
+                       "Failed to re-initialize GPU, declaring it wedged!\n");
+               i915_gem_set_wedged(i915);
+       }
+       goto out_unlock;
+}
+
+void i915_gem_init__pm(struct drm_i915_private *i915)
+{
+       INIT_DELAYED_WORK(&i915->gem.idle_work, idle_work_handler);
+       INIT_DELAYED_WORK(&i915->gem.retire_work, retire_work_handler);
+}
diff --git a/drivers/gpu/drm/i915/i915_gem_pm.h 
b/drivers/gpu/drm/i915/i915_gem_pm.h
new file mode 100644
index 000000000000..52f65e3f06b5
--- /dev/null
+++ b/drivers/gpu/drm/i915/i915_gem_pm.h
@@ -0,0 +1,28 @@
+/*
+ * SPDX-License-Identifier: MIT
+ *
+ * Copyright © 2019 Intel Corporation
+ */
+
+#ifndef __I915_GEM_PM_H__
+#define __I915_GEM_PM_H__
+
+#include <linux/types.h>
+
+struct drm_i915_private;
+struct work_struct;
+
+void i915_gem_init__pm(struct drm_i915_private *i915);
+
+bool i915_gem_load_power_context(struct drm_i915_private *i915);
+void i915_gem_resume(struct drm_i915_private *i915);
+
+void i915_gem_unpark(struct drm_i915_private *i915);
+void i915_gem_park(struct drm_i915_private *i915);
+
+void i915_gem_idle_work_handler(struct work_struct *work);
+
+void i915_gem_suspend(struct drm_i915_private *i915);
+void i915_gem_suspend_late(struct drm_i915_private *i915);
+
+#endif /* __I915_GEM_PM_H__ */
diff --git a/drivers/gpu/drm/i915/selftests/i915_gem_context.c 
b/drivers/gpu/drm/i915/selftests/i915_gem_context.c
index 6f52ca881173..9d646fa1b74e 100644
--- a/drivers/gpu/drm/i915/selftests/i915_gem_context.c
+++ b/drivers/gpu/drm/i915/selftests/i915_gem_context.c
@@ -1658,7 +1658,7 @@ static int __igt_switch_to_kernel_context(struct 
drm_i915_private *i915,
                /* XXX Bonus points for proving we are the kernel context! */
 
                mutex_unlock(&i915->drm.struct_mutex);
-               drain_delayed_work(&i915->gt.idle_work);
+               drain_delayed_work(&i915->gem.idle_work);
                mutex_lock(&i915->drm.struct_mutex);
        }
 
diff --git a/drivers/gpu/drm/i915/selftests/i915_gem_object.c 
b/drivers/gpu/drm/i915/selftests/i915_gem_object.c
index 971148fbe6f5..12fc53c694a6 100644
--- a/drivers/gpu/drm/i915/selftests/i915_gem_object.c
+++ b/drivers/gpu/drm/i915/selftests/i915_gem_object.c
@@ -514,8 +514,8 @@ static void disable_retire_worker(struct drm_i915_private 
*i915)
        }
        mutex_unlock(&i915->drm.struct_mutex);
 
-       cancel_delayed_work_sync(&i915->gt.retire_work);
-       cancel_delayed_work_sync(&i915->gt.idle_work);
+       cancel_delayed_work_sync(&i915->gem.retire_work);
+       cancel_delayed_work_sync(&i915->gem.idle_work);
 }
 
 static int igt_mmap_offset_exhaustion(void *arg)
@@ -617,9 +617,9 @@ static int igt_mmap_offset_exhaustion(void *arg)
 out_park:
        mutex_lock(&i915->drm.struct_mutex);
        if (--i915->gt.active_requests)
-               queue_delayed_work(i915->wq, &i915->gt.retire_work, 0);
+               queue_delayed_work(i915->wq, &i915->gem.retire_work, 0);
        else
-               queue_delayed_work(i915->wq, &i915->gt.idle_work, 0);
+               queue_delayed_work(i915->wq, &i915->gem.idle_work, 0);
        mutex_unlock(&i915->drm.struct_mutex);
        i915_gem_shrinker_register(i915);
        return err;
diff --git a/drivers/gpu/drm/i915/selftests/mock_gem_device.c 
b/drivers/gpu/drm/i915/selftests/mock_gem_device.c
index f444ee5add27..fb677b4019a0 100644
--- a/drivers/gpu/drm/i915/selftests/mock_gem_device.c
+++ b/drivers/gpu/drm/i915/selftests/mock_gem_device.c
@@ -59,8 +59,8 @@ static void mock_device_release(struct drm_device *dev)
        i915_gem_contexts_lost(i915);
        mutex_unlock(&i915->drm.struct_mutex);
 
-       drain_delayed_work(&i915->gt.retire_work);
-       drain_delayed_work(&i915->gt.idle_work);
+       drain_delayed_work(&i915->gem.retire_work);
+       drain_delayed_work(&i915->gem.idle_work);
        i915_gem_drain_workqueue(i915);
 
        mutex_lock(&i915->drm.struct_mutex);
@@ -111,7 +111,7 @@ static void mock_retire_work_handler(struct work_struct 
*work)
 static void mock_idle_work_handler(struct work_struct *work)
 {
        struct drm_i915_private *i915 =
-               container_of(work, typeof(*i915), gt.idle_work.work);
+               container_of(work, typeof(*i915), gem.idle_work.work);
 
        i915->gt.active_engines = 0;
 }
@@ -197,8 +197,8 @@ struct drm_i915_private *mock_gem_device(void)
 
        mock_init_contexts(i915);
 
-       INIT_DELAYED_WORK(&i915->gt.retire_work, mock_retire_work_handler);
-       INIT_DELAYED_WORK(&i915->gt.idle_work, mock_idle_work_handler);
+       INIT_DELAYED_WORK(&i915->gem.retire_work, mock_retire_work_handler);
+       INIT_DELAYED_WORK(&i915->gem.idle_work, mock_idle_work_handler);
 
        i915->gt.awake = true;
 
-- 
2.20.1

_______________________________________________
Intel-gfx mailing list
Intel-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/intel-gfx

Reply via email to