drm/i915: Pull the GEM powermangement coupling into its own file
Split out the powermanagement portion (GT wakeref, suspend/resume) of GEM from i915_gem.c into its own file. Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk> Reviewed-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com> Link: https://patchwork.freedesktop.org/patch/msgid/20190424200717.1686-2-chris@chris-wilson.co.uk
This commit is contained in:
@@ -97,6 +97,7 @@ i915-y += \
|
|||||||
i915_gem_internal.o \
|
i915_gem_internal.o \
|
||||||
i915_gem.o \
|
i915_gem.o \
|
||||||
i915_gem_object.o \
|
i915_gem_object.o \
|
||||||
|
i915_gem_pm.o \
|
||||||
i915_gem_render_state.o \
|
i915_gem_render_state.o \
|
||||||
i915_gem_shrinker.o \
|
i915_gem_shrinker.o \
|
||||||
i915_gem_stolen.o \
|
i915_gem_stolen.o \
|
||||||
|
|||||||
@@ -5,6 +5,7 @@
|
|||||||
header_test := \
|
header_test := \
|
||||||
i915_active_types.h \
|
i915_active_types.h \
|
||||||
i915_gem_context_types.h \
|
i915_gem_context_types.h \
|
||||||
|
i915_gem_pm.h \
|
||||||
i915_priolist_types.h \
|
i915_priolist_types.h \
|
||||||
i915_scheduler_types.h \
|
i915_scheduler_types.h \
|
||||||
i915_timeline_types.h \
|
i915_timeline_types.h \
|
||||||
|
|||||||
@@ -3942,8 +3942,8 @@ i915_drop_caches_set(void *data, u64 val)
|
|||||||
if (val & DROP_IDLE) {
|
if (val & DROP_IDLE) {
|
||||||
do {
|
do {
|
||||||
if (READ_ONCE(i915->gt.active_requests))
|
if (READ_ONCE(i915->gt.active_requests))
|
||||||
flush_delayed_work(&i915->gt.retire_work);
|
flush_delayed_work(&i915->gem.retire_work);
|
||||||
drain_delayed_work(&i915->gt.idle_work);
|
drain_delayed_work(&i915->gem.idle_work);
|
||||||
} while (READ_ONCE(i915->gt.awake));
|
} while (READ_ONCE(i915->gt.awake));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -2020,6 +2020,12 @@ struct drm_i915_private {
|
|||||||
*/
|
*/
|
||||||
intel_wakeref_t awake;
|
intel_wakeref_t awake;
|
||||||
|
|
||||||
|
ktime_t last_init_time;
|
||||||
|
|
||||||
|
struct i915_vma *scratch;
|
||||||
|
} gt;
|
||||||
|
|
||||||
|
struct {
|
||||||
/**
|
/**
|
||||||
* We leave the user IRQ off as much as possible,
|
* We leave the user IRQ off as much as possible,
|
||||||
* but this means that requests will finish and never
|
* but this means that requests will finish and never
|
||||||
@@ -2037,11 +2043,7 @@ struct drm_i915_private {
|
|||||||
* off the idle_work.
|
* off the idle_work.
|
||||||
*/
|
*/
|
||||||
struct delayed_work idle_work;
|
struct delayed_work idle_work;
|
||||||
|
} gem;
|
||||||
ktime_t last_init_time;
|
|
||||||
|
|
||||||
struct i915_vma *scratch;
|
|
||||||
} gt;
|
|
||||||
|
|
||||||
/* For i945gm vblank irq vs. C3 workaround */
|
/* For i945gm vblank irq vs. C3 workaround */
|
||||||
struct {
|
struct {
|
||||||
|
|||||||
@@ -46,7 +46,7 @@
|
|||||||
#include "i915_drv.h"
|
#include "i915_drv.h"
|
||||||
#include "i915_gem_clflush.h"
|
#include "i915_gem_clflush.h"
|
||||||
#include "i915_gemfs.h"
|
#include "i915_gemfs.h"
|
||||||
#include "i915_globals.h"
|
#include "i915_gem_pm.h"
|
||||||
#include "i915_trace.h"
|
#include "i915_trace.h"
|
||||||
#include "i915_vgpu.h"
|
#include "i915_vgpu.h"
|
||||||
|
|
||||||
@@ -103,105 +103,6 @@ static void i915_gem_info_remove_obj(struct drm_i915_private *dev_priv,
|
|||||||
spin_unlock(&dev_priv->mm.object_stat_lock);
|
spin_unlock(&dev_priv->mm.object_stat_lock);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void __i915_gem_park(struct drm_i915_private *i915)
|
|
||||||
{
|
|
||||||
intel_wakeref_t wakeref;
|
|
||||||
|
|
||||||
GEM_TRACE("\n");
|
|
||||||
|
|
||||||
lockdep_assert_held(&i915->drm.struct_mutex);
|
|
||||||
GEM_BUG_ON(i915->gt.active_requests);
|
|
||||||
GEM_BUG_ON(!list_empty(&i915->gt.active_rings));
|
|
||||||
|
|
||||||
if (!i915->gt.awake)
|
|
||||||
return;
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Be paranoid and flush a concurrent interrupt to make sure
|
|
||||||
* we don't reactivate any irq tasklets after parking.
|
|
||||||
*
|
|
||||||
* FIXME: Note that even though we have waited for execlists to be idle,
|
|
||||||
* there may still be an in-flight interrupt even though the CSB
|
|
||||||
* is now empty. synchronize_irq() makes sure that a residual interrupt
|
|
||||||
* is completed before we continue, but it doesn't prevent the HW from
|
|
||||||
* raising a spurious interrupt later. To complete the shield we should
|
|
||||||
* coordinate disabling the CS irq with flushing the interrupts.
|
|
||||||
*/
|
|
||||||
synchronize_irq(i915->drm.irq);
|
|
||||||
|
|
||||||
intel_engines_park(i915);
|
|
||||||
i915_timelines_park(i915);
|
|
||||||
|
|
||||||
i915_pmu_gt_parked(i915);
|
|
||||||
i915_vma_parked(i915);
|
|
||||||
|
|
||||||
wakeref = fetch_and_zero(&i915->gt.awake);
|
|
||||||
GEM_BUG_ON(!wakeref);
|
|
||||||
|
|
||||||
if (INTEL_GEN(i915) >= 6)
|
|
||||||
gen6_rps_idle(i915);
|
|
||||||
|
|
||||||
intel_display_power_put(i915, POWER_DOMAIN_GT_IRQ, wakeref);
|
|
||||||
|
|
||||||
i915_globals_park();
|
|
||||||
}
|
|
||||||
|
|
||||||
void i915_gem_park(struct drm_i915_private *i915)
|
|
||||||
{
|
|
||||||
GEM_TRACE("\n");
|
|
||||||
|
|
||||||
lockdep_assert_held(&i915->drm.struct_mutex);
|
|
||||||
GEM_BUG_ON(i915->gt.active_requests);
|
|
||||||
|
|
||||||
if (!i915->gt.awake)
|
|
||||||
return;
|
|
||||||
|
|
||||||
/* Defer the actual call to __i915_gem_park() to prevent ping-pongs */
|
|
||||||
mod_delayed_work(i915->wq, &i915->gt.idle_work, msecs_to_jiffies(100));
|
|
||||||
}
|
|
||||||
|
|
||||||
void i915_gem_unpark(struct drm_i915_private *i915)
|
|
||||||
{
|
|
||||||
GEM_TRACE("\n");
|
|
||||||
|
|
||||||
lockdep_assert_held(&i915->drm.struct_mutex);
|
|
||||||
GEM_BUG_ON(!i915->gt.active_requests);
|
|
||||||
assert_rpm_wakelock_held(i915);
|
|
||||||
|
|
||||||
if (i915->gt.awake)
|
|
||||||
return;
|
|
||||||
|
|
||||||
/*
|
|
||||||
* It seems that the DMC likes to transition between the DC states a lot
|
|
||||||
* when there are no connected displays (no active power domains) during
|
|
||||||
* command submission.
|
|
||||||
*
|
|
||||||
* This activity has negative impact on the performance of the chip with
|
|
||||||
* huge latencies observed in the interrupt handler and elsewhere.
|
|
||||||
*
|
|
||||||
* Work around it by grabbing a GT IRQ power domain whilst there is any
|
|
||||||
* GT activity, preventing any DC state transitions.
|
|
||||||
*/
|
|
||||||
i915->gt.awake = intel_display_power_get(i915, POWER_DOMAIN_GT_IRQ);
|
|
||||||
GEM_BUG_ON(!i915->gt.awake);
|
|
||||||
|
|
||||||
i915_globals_unpark();
|
|
||||||
|
|
||||||
intel_enable_gt_powersave(i915);
|
|
||||||
i915_update_gfx_val(i915);
|
|
||||||
if (INTEL_GEN(i915) >= 6)
|
|
||||||
gen6_rps_busy(i915);
|
|
||||||
i915_pmu_gt_unparked(i915);
|
|
||||||
|
|
||||||
intel_engines_unpark(i915);
|
|
||||||
|
|
||||||
i915_queue_hangcheck(i915);
|
|
||||||
|
|
||||||
queue_delayed_work(i915->wq,
|
|
||||||
&i915->gt.retire_work,
|
|
||||||
round_jiffies_up_relative(HZ));
|
|
||||||
}
|
|
||||||
|
|
||||||
int
|
int
|
||||||
i915_gem_get_aperture_ioctl(struct drm_device *dev, void *data,
|
i915_gem_get_aperture_ioctl(struct drm_device *dev, void *data,
|
||||||
struct drm_file *file)
|
struct drm_file *file)
|
||||||
@@ -2088,7 +1989,7 @@ static int i915_gem_object_create_mmap_offset(struct drm_i915_gem_object *obj)
|
|||||||
if (!err)
|
if (!err)
|
||||||
break;
|
break;
|
||||||
|
|
||||||
} while (flush_delayed_work(&dev_priv->gt.retire_work));
|
} while (flush_delayed_work(&dev_priv->gem.retire_work));
|
||||||
|
|
||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
@@ -2848,132 +2749,6 @@ i915_gem_object_pwrite_gtt(struct drm_i915_gem_object *obj,
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void
|
|
||||||
i915_gem_retire_work_handler(struct work_struct *work)
|
|
||||||
{
|
|
||||||
struct drm_i915_private *dev_priv =
|
|
||||||
container_of(work, typeof(*dev_priv), gt.retire_work.work);
|
|
||||||
struct drm_device *dev = &dev_priv->drm;
|
|
||||||
|
|
||||||
/* Come back later if the device is busy... */
|
|
||||||
if (mutex_trylock(&dev->struct_mutex)) {
|
|
||||||
i915_retire_requests(dev_priv);
|
|
||||||
mutex_unlock(&dev->struct_mutex);
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Keep the retire handler running until we are finally idle.
|
|
||||||
* We do not need to do this test under locking as in the worst-case
|
|
||||||
* we queue the retire worker once too often.
|
|
||||||
*/
|
|
||||||
if (READ_ONCE(dev_priv->gt.awake))
|
|
||||||
queue_delayed_work(dev_priv->wq,
|
|
||||||
&dev_priv->gt.retire_work,
|
|
||||||
round_jiffies_up_relative(HZ));
|
|
||||||
}
|
|
||||||
|
|
||||||
static bool switch_to_kernel_context_sync(struct drm_i915_private *i915,
|
|
||||||
unsigned long mask)
|
|
||||||
{
|
|
||||||
bool result = true;
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Even if we fail to switch, give whatever is running a small chance
|
|
||||||
* to save itself before we report the failure. Yes, this may be a
|
|
||||||
* false positive due to e.g. ENOMEM, caveat emptor!
|
|
||||||
*/
|
|
||||||
if (i915_gem_switch_to_kernel_context(i915, mask))
|
|
||||||
result = false;
|
|
||||||
|
|
||||||
if (i915_gem_wait_for_idle(i915,
|
|
||||||
I915_WAIT_LOCKED |
|
|
||||||
I915_WAIT_FOR_IDLE_BOOST,
|
|
||||||
I915_GEM_IDLE_TIMEOUT))
|
|
||||||
result = false;
|
|
||||||
|
|
||||||
if (!result) {
|
|
||||||
if (i915_modparams.reset) { /* XXX hide warning from gem_eio */
|
|
||||||
dev_err(i915->drm.dev,
|
|
||||||
"Failed to idle engines, declaring wedged!\n");
|
|
||||||
GEM_TRACE_DUMP();
|
|
||||||
}
|
|
||||||
|
|
||||||
/* Forcibly cancel outstanding work and leave the gpu quiet. */
|
|
||||||
i915_gem_set_wedged(i915);
|
|
||||||
}
|
|
||||||
|
|
||||||
i915_retire_requests(i915); /* ensure we flush after wedging */
|
|
||||||
return result;
|
|
||||||
}
|
|
||||||
|
|
||||||
static bool load_power_context(struct drm_i915_private *i915)
|
|
||||||
{
|
|
||||||
/* Force loading the kernel context on all engines */
|
|
||||||
if (!switch_to_kernel_context_sync(i915, ALL_ENGINES))
|
|
||||||
return false;
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Immediately park the GPU so that we enable powersaving and
|
|
||||||
* treat it as idle. The next time we issue a request, we will
|
|
||||||
* unpark and start using the engine->pinned_default_state, otherwise
|
|
||||||
* it is in limbo and an early reset may fail.
|
|
||||||
*/
|
|
||||||
__i915_gem_park(i915);
|
|
||||||
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
static void
|
|
||||||
i915_gem_idle_work_handler(struct work_struct *work)
|
|
||||||
{
|
|
||||||
struct drm_i915_private *i915 =
|
|
||||||
container_of(work, typeof(*i915), gt.idle_work.work);
|
|
||||||
bool rearm_hangcheck;
|
|
||||||
|
|
||||||
if (!READ_ONCE(i915->gt.awake))
|
|
||||||
return;
|
|
||||||
|
|
||||||
if (READ_ONCE(i915->gt.active_requests))
|
|
||||||
return;
|
|
||||||
|
|
||||||
rearm_hangcheck =
|
|
||||||
cancel_delayed_work_sync(&i915->gpu_error.hangcheck_work);
|
|
||||||
|
|
||||||
if (!mutex_trylock(&i915->drm.struct_mutex)) {
|
|
||||||
/* Currently busy, come back later */
|
|
||||||
mod_delayed_work(i915->wq,
|
|
||||||
&i915->gt.idle_work,
|
|
||||||
msecs_to_jiffies(50));
|
|
||||||
goto out_rearm;
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Flush out the last user context, leaving only the pinned
|
|
||||||
* kernel context resident. Should anything unfortunate happen
|
|
||||||
* while we are idle (such as the GPU being power cycled), no users
|
|
||||||
* will be harmed.
|
|
||||||
*/
|
|
||||||
if (!work_pending(&i915->gt.idle_work.work) &&
|
|
||||||
!i915->gt.active_requests) {
|
|
||||||
++i915->gt.active_requests; /* don't requeue idle */
|
|
||||||
|
|
||||||
switch_to_kernel_context_sync(i915, i915->gt.active_engines);
|
|
||||||
|
|
||||||
if (!--i915->gt.active_requests) {
|
|
||||||
__i915_gem_park(i915);
|
|
||||||
rearm_hangcheck = false;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
mutex_unlock(&i915->drm.struct_mutex);
|
|
||||||
|
|
||||||
out_rearm:
|
|
||||||
if (rearm_hangcheck) {
|
|
||||||
GEM_BUG_ON(!i915->gt.awake);
|
|
||||||
i915_queue_hangcheck(i915);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
void i915_gem_close_object(struct drm_gem_object *gem, struct drm_file *file)
|
void i915_gem_close_object(struct drm_gem_object *gem, struct drm_file *file)
|
||||||
{
|
{
|
||||||
struct drm_i915_private *i915 = to_i915(gem->dev);
|
struct drm_i915_private *i915 = to_i915(gem->dev);
|
||||||
@@ -4389,133 +4164,6 @@ void i915_gem_sanitize(struct drm_i915_private *i915)
|
|||||||
mutex_unlock(&i915->drm.struct_mutex);
|
mutex_unlock(&i915->drm.struct_mutex);
|
||||||
}
|
}
|
||||||
|
|
||||||
void i915_gem_suspend(struct drm_i915_private *i915)
|
|
||||||
{
|
|
||||||
intel_wakeref_t wakeref;
|
|
||||||
|
|
||||||
GEM_TRACE("\n");
|
|
||||||
|
|
||||||
wakeref = intel_runtime_pm_get(i915);
|
|
||||||
|
|
||||||
flush_workqueue(i915->wq);
|
|
||||||
|
|
||||||
mutex_lock(&i915->drm.struct_mutex);
|
|
||||||
|
|
||||||
/*
|
|
||||||
* We have to flush all the executing contexts to main memory so
|
|
||||||
* that they can saved in the hibernation image. To ensure the last
|
|
||||||
* context image is coherent, we have to switch away from it. That
|
|
||||||
* leaves the i915->kernel_context still active when
|
|
||||||
* we actually suspend, and its image in memory may not match the GPU
|
|
||||||
* state. Fortunately, the kernel_context is disposable and we do
|
|
||||||
* not rely on its state.
|
|
||||||
*/
|
|
||||||
switch_to_kernel_context_sync(i915, i915->gt.active_engines);
|
|
||||||
|
|
||||||
mutex_unlock(&i915->drm.struct_mutex);
|
|
||||||
i915_reset_flush(i915);
|
|
||||||
|
|
||||||
drain_delayed_work(&i915->gt.retire_work);
|
|
||||||
|
|
||||||
/*
|
|
||||||
* As the idle_work is rearming if it detects a race, play safe and
|
|
||||||
* repeat the flush until it is definitely idle.
|
|
||||||
*/
|
|
||||||
drain_delayed_work(&i915->gt.idle_work);
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Assert that we successfully flushed all the work and
|
|
||||||
* reset the GPU back to its idle, low power state.
|
|
||||||
*/
|
|
||||||
GEM_BUG_ON(i915->gt.awake);
|
|
||||||
|
|
||||||
intel_uc_suspend(i915);
|
|
||||||
|
|
||||||
intel_runtime_pm_put(i915, wakeref);
|
|
||||||
}
|
|
||||||
|
|
||||||
void i915_gem_suspend_late(struct drm_i915_private *i915)
|
|
||||||
{
|
|
||||||
struct drm_i915_gem_object *obj;
|
|
||||||
struct list_head *phases[] = {
|
|
||||||
&i915->mm.unbound_list,
|
|
||||||
&i915->mm.bound_list,
|
|
||||||
NULL
|
|
||||||
}, **phase;
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Neither the BIOS, ourselves or any other kernel
|
|
||||||
* expects the system to be in execlists mode on startup,
|
|
||||||
* so we need to reset the GPU back to legacy mode. And the only
|
|
||||||
* known way to disable logical contexts is through a GPU reset.
|
|
||||||
*
|
|
||||||
* So in order to leave the system in a known default configuration,
|
|
||||||
* always reset the GPU upon unload and suspend. Afterwards we then
|
|
||||||
* clean up the GEM state tracking, flushing off the requests and
|
|
||||||
* leaving the system in a known idle state.
|
|
||||||
*
|
|
||||||
* Note that is of the upmost importance that the GPU is idle and
|
|
||||||
* all stray writes are flushed *before* we dismantle the backing
|
|
||||||
* storage for the pinned objects.
|
|
||||||
*
|
|
||||||
* However, since we are uncertain that resetting the GPU on older
|
|
||||||
* machines is a good idea, we don't - just in case it leaves the
|
|
||||||
* machine in an unusable condition.
|
|
||||||
*/
|
|
||||||
|
|
||||||
mutex_lock(&i915->drm.struct_mutex);
|
|
||||||
for (phase = phases; *phase; phase++) {
|
|
||||||
list_for_each_entry(obj, *phase, mm.link)
|
|
||||||
WARN_ON(i915_gem_object_set_to_gtt_domain(obj, false));
|
|
||||||
}
|
|
||||||
mutex_unlock(&i915->drm.struct_mutex);
|
|
||||||
|
|
||||||
intel_uc_sanitize(i915);
|
|
||||||
i915_gem_sanitize(i915);
|
|
||||||
}
|
|
||||||
|
|
||||||
void i915_gem_resume(struct drm_i915_private *i915)
|
|
||||||
{
|
|
||||||
GEM_TRACE("\n");
|
|
||||||
|
|
||||||
WARN_ON(i915->gt.awake);
|
|
||||||
|
|
||||||
mutex_lock(&i915->drm.struct_mutex);
|
|
||||||
intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
|
|
||||||
|
|
||||||
i915_gem_restore_gtt_mappings(i915);
|
|
||||||
i915_gem_restore_fences(i915);
|
|
||||||
|
|
||||||
/*
|
|
||||||
* As we didn't flush the kernel context before suspend, we cannot
|
|
||||||
* guarantee that the context image is complete. So let's just reset
|
|
||||||
* it and start again.
|
|
||||||
*/
|
|
||||||
intel_gt_resume(i915);
|
|
||||||
|
|
||||||
if (i915_gem_init_hw(i915))
|
|
||||||
goto err_wedged;
|
|
||||||
|
|
||||||
intel_uc_resume(i915);
|
|
||||||
|
|
||||||
/* Always reload a context for powersaving. */
|
|
||||||
if (!load_power_context(i915))
|
|
||||||
goto err_wedged;
|
|
||||||
|
|
||||||
out_unlock:
|
|
||||||
intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
|
|
||||||
mutex_unlock(&i915->drm.struct_mutex);
|
|
||||||
return;
|
|
||||||
|
|
||||||
err_wedged:
|
|
||||||
if (!i915_reset_failed(i915)) {
|
|
||||||
dev_err(i915->drm.dev,
|
|
||||||
"Failed to re-initialize GPU, declaring it wedged!\n");
|
|
||||||
i915_gem_set_wedged(i915);
|
|
||||||
}
|
|
||||||
goto out_unlock;
|
|
||||||
}
|
|
||||||
|
|
||||||
void i915_gem_init_swizzling(struct drm_i915_private *dev_priv)
|
void i915_gem_init_swizzling(struct drm_i915_private *dev_priv)
|
||||||
{
|
{
|
||||||
if (INTEL_GEN(dev_priv) < 5 ||
|
if (INTEL_GEN(dev_priv) < 5 ||
|
||||||
@@ -4698,7 +4346,7 @@ static int __intel_engines_record_defaults(struct drm_i915_private *i915)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/* Flush the default context image to memory, and enable powersaving. */
|
/* Flush the default context image to memory, and enable powersaving. */
|
||||||
if (!load_power_context(i915)) {
|
if (!i915_gem_load_power_context(i915)) {
|
||||||
err = -EIO;
|
err = -EIO;
|
||||||
goto err_active;
|
goto err_active;
|
||||||
}
|
}
|
||||||
@@ -5113,11 +4761,8 @@ int i915_gem_init_early(struct drm_i915_private *dev_priv)
|
|||||||
INIT_LIST_HEAD(&dev_priv->gt.closed_vma);
|
INIT_LIST_HEAD(&dev_priv->gt.closed_vma);
|
||||||
|
|
||||||
i915_gem_init__mm(dev_priv);
|
i915_gem_init__mm(dev_priv);
|
||||||
|
i915_gem_init__pm(dev_priv);
|
||||||
|
|
||||||
INIT_DELAYED_WORK(&dev_priv->gt.retire_work,
|
|
||||||
i915_gem_retire_work_handler);
|
|
||||||
INIT_DELAYED_WORK(&dev_priv->gt.idle_work,
|
|
||||||
i915_gem_idle_work_handler);
|
|
||||||
init_waitqueue_head(&dev_priv->gpu_error.wait_queue);
|
init_waitqueue_head(&dev_priv->gpu_error.wait_queue);
|
||||||
init_waitqueue_head(&dev_priv->gpu_error.reset_queue);
|
init_waitqueue_head(&dev_priv->gpu_error.reset_queue);
|
||||||
mutex_init(&dev_priv->gpu_error.wedge_mutex);
|
mutex_init(&dev_priv->gpu_error.wedge_mutex);
|
||||||
|
|||||||
365
drivers/gpu/drm/i915/i915_gem_pm.c
Normal file
365
drivers/gpu/drm/i915/i915_gem_pm.c
Normal file
@@ -0,0 +1,365 @@
|
|||||||
|
/*
|
||||||
|
* SPDX-License-Identifier: MIT
|
||||||
|
*
|
||||||
|
* Copyright © 2019 Intel Corporation
|
||||||
|
*/
|
||||||
|
|
||||||
|
#include "i915_drv.h"
|
||||||
|
#include "i915_gem_pm.h"
|
||||||
|
#include "i915_globals.h"
|
||||||
|
#include "intel_pm.h"
|
||||||
|
|
||||||
|
static void __i915_gem_park(struct drm_i915_private *i915)
|
||||||
|
{
|
||||||
|
intel_wakeref_t wakeref;
|
||||||
|
|
||||||
|
GEM_TRACE("\n");
|
||||||
|
|
||||||
|
lockdep_assert_held(&i915->drm.struct_mutex);
|
||||||
|
GEM_BUG_ON(i915->gt.active_requests);
|
||||||
|
GEM_BUG_ON(!list_empty(&i915->gt.active_rings));
|
||||||
|
|
||||||
|
if (!i915->gt.awake)
|
||||||
|
return;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Be paranoid and flush a concurrent interrupt to make sure
|
||||||
|
* we don't reactivate any irq tasklets after parking.
|
||||||
|
*
|
||||||
|
* FIXME: Note that even though we have waited for execlists to be idle,
|
||||||
|
* there may still be an in-flight interrupt even though the CSB
|
||||||
|
* is now empty. synchronize_irq() makes sure that a residual interrupt
|
||||||
|
* is completed before we continue, but it doesn't prevent the HW from
|
||||||
|
* raising a spurious interrupt later. To complete the shield we should
|
||||||
|
* coordinate disabling the CS irq with flushing the interrupts.
|
||||||
|
*/
|
||||||
|
synchronize_irq(i915->drm.irq);
|
||||||
|
|
||||||
|
intel_engines_park(i915);
|
||||||
|
i915_timelines_park(i915);
|
||||||
|
|
||||||
|
i915_pmu_gt_parked(i915);
|
||||||
|
i915_vma_parked(i915);
|
||||||
|
|
||||||
|
wakeref = fetch_and_zero(&i915->gt.awake);
|
||||||
|
GEM_BUG_ON(!wakeref);
|
||||||
|
|
||||||
|
if (INTEL_GEN(i915) >= 6)
|
||||||
|
gen6_rps_idle(i915);
|
||||||
|
|
||||||
|
intel_display_power_put(i915, POWER_DOMAIN_GT_IRQ, wakeref);
|
||||||
|
|
||||||
|
i915_globals_park();
|
||||||
|
}
|
||||||
|
|
||||||
|
static bool switch_to_kernel_context_sync(struct drm_i915_private *i915,
|
||||||
|
unsigned long mask)
|
||||||
|
{
|
||||||
|
bool result = true;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Even if we fail to switch, give whatever is running a small chance
|
||||||
|
* to save itself before we report the failure. Yes, this may be a
|
||||||
|
* false positive due to e.g. ENOMEM, caveat emptor!
|
||||||
|
*/
|
||||||
|
if (i915_gem_switch_to_kernel_context(i915, mask))
|
||||||
|
result = false;
|
||||||
|
|
||||||
|
if (i915_gem_wait_for_idle(i915,
|
||||||
|
I915_WAIT_LOCKED |
|
||||||
|
I915_WAIT_FOR_IDLE_BOOST,
|
||||||
|
I915_GEM_IDLE_TIMEOUT))
|
||||||
|
result = false;
|
||||||
|
|
||||||
|
if (!result) {
|
||||||
|
if (i915_modparams.reset) { /* XXX hide warning from gem_eio */
|
||||||
|
dev_err(i915->drm.dev,
|
||||||
|
"Failed to idle engines, declaring wedged!\n");
|
||||||
|
GEM_TRACE_DUMP();
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Forcibly cancel outstanding work and leave the gpu quiet. */
|
||||||
|
i915_gem_set_wedged(i915);
|
||||||
|
}
|
||||||
|
|
||||||
|
i915_retire_requests(i915); /* ensure we flush after wedging */
|
||||||
|
return result;
|
||||||
|
}
|
||||||
|
|
||||||
|
static void idle_work_handler(struct work_struct *work)
|
||||||
|
{
|
||||||
|
struct drm_i915_private *i915 =
|
||||||
|
container_of(work, typeof(*i915), gem.idle_work.work);
|
||||||
|
bool rearm_hangcheck;
|
||||||
|
|
||||||
|
if (!READ_ONCE(i915->gt.awake))
|
||||||
|
return;
|
||||||
|
|
||||||
|
if (READ_ONCE(i915->gt.active_requests))
|
||||||
|
return;
|
||||||
|
|
||||||
|
rearm_hangcheck =
|
||||||
|
cancel_delayed_work_sync(&i915->gpu_error.hangcheck_work);
|
||||||
|
|
||||||
|
if (!mutex_trylock(&i915->drm.struct_mutex)) {
|
||||||
|
/* Currently busy, come back later */
|
||||||
|
mod_delayed_work(i915->wq,
|
||||||
|
&i915->gem.idle_work,
|
||||||
|
msecs_to_jiffies(50));
|
||||||
|
goto out_rearm;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Flush out the last user context, leaving only the pinned
|
||||||
|
* kernel context resident. Should anything unfortunate happen
|
||||||
|
* while we are idle (such as the GPU being power cycled), no users
|
||||||
|
* will be harmed.
|
||||||
|
*/
|
||||||
|
if (!work_pending(&i915->gem.idle_work.work) &&
|
||||||
|
!i915->gt.active_requests) {
|
||||||
|
++i915->gt.active_requests; /* don't requeue idle */
|
||||||
|
|
||||||
|
switch_to_kernel_context_sync(i915, i915->gt.active_engines);
|
||||||
|
|
||||||
|
if (!--i915->gt.active_requests) {
|
||||||
|
__i915_gem_park(i915);
|
||||||
|
rearm_hangcheck = false;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
mutex_unlock(&i915->drm.struct_mutex);
|
||||||
|
|
||||||
|
out_rearm:
|
||||||
|
if (rearm_hangcheck) {
|
||||||
|
GEM_BUG_ON(!i915->gt.awake);
|
||||||
|
i915_queue_hangcheck(i915);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
static void retire_work_handler(struct work_struct *work)
|
||||||
|
{
|
||||||
|
struct drm_i915_private *i915 =
|
||||||
|
container_of(work, typeof(*i915), gem.retire_work.work);
|
||||||
|
|
||||||
|
/* Come back later if the device is busy... */
|
||||||
|
if (mutex_trylock(&i915->drm.struct_mutex)) {
|
||||||
|
i915_retire_requests(i915);
|
||||||
|
mutex_unlock(&i915->drm.struct_mutex);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Keep the retire handler running until we are finally idle.
|
||||||
|
* We do not need to do this test under locking as in the worst-case
|
||||||
|
* we queue the retire worker once too often.
|
||||||
|
*/
|
||||||
|
if (READ_ONCE(i915->gt.awake))
|
||||||
|
queue_delayed_work(i915->wq,
|
||||||
|
&i915->gem.retire_work,
|
||||||
|
round_jiffies_up_relative(HZ));
|
||||||
|
}
|
||||||
|
|
||||||
|
void i915_gem_park(struct drm_i915_private *i915)
|
||||||
|
{
|
||||||
|
GEM_TRACE("\n");
|
||||||
|
|
||||||
|
lockdep_assert_held(&i915->drm.struct_mutex);
|
||||||
|
GEM_BUG_ON(i915->gt.active_requests);
|
||||||
|
|
||||||
|
if (!i915->gt.awake)
|
||||||
|
return;
|
||||||
|
|
||||||
|
/* Defer the actual call to __i915_gem_park() to prevent ping-pongs */
|
||||||
|
mod_delayed_work(i915->wq, &i915->gem.idle_work, msecs_to_jiffies(100));
|
||||||
|
}
|
||||||
|
|
||||||
|
void i915_gem_unpark(struct drm_i915_private *i915)
|
||||||
|
{
|
||||||
|
GEM_TRACE("\n");
|
||||||
|
|
||||||
|
lockdep_assert_held(&i915->drm.struct_mutex);
|
||||||
|
GEM_BUG_ON(!i915->gt.active_requests);
|
||||||
|
assert_rpm_wakelock_held(i915);
|
||||||
|
|
||||||
|
if (i915->gt.awake)
|
||||||
|
return;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* It seems that the DMC likes to transition between the DC states a lot
|
||||||
|
* when there are no connected displays (no active power domains) during
|
||||||
|
* command submission.
|
||||||
|
*
|
||||||
|
* This activity has negative impact on the performance of the chip with
|
||||||
|
* huge latencies observed in the interrupt handler and elsewhere.
|
||||||
|
*
|
||||||
|
* Work around it by grabbing a GT IRQ power domain whilst there is any
|
||||||
|
* GT activity, preventing any DC state transitions.
|
||||||
|
*/
|
||||||
|
i915->gt.awake = intel_display_power_get(i915, POWER_DOMAIN_GT_IRQ);
|
||||||
|
GEM_BUG_ON(!i915->gt.awake);
|
||||||
|
|
||||||
|
i915_globals_unpark();
|
||||||
|
|
||||||
|
intel_enable_gt_powersave(i915);
|
||||||
|
i915_update_gfx_val(i915);
|
||||||
|
if (INTEL_GEN(i915) >= 6)
|
||||||
|
gen6_rps_busy(i915);
|
||||||
|
i915_pmu_gt_unparked(i915);
|
||||||
|
|
||||||
|
intel_engines_unpark(i915);
|
||||||
|
|
||||||
|
i915_queue_hangcheck(i915);
|
||||||
|
|
||||||
|
queue_delayed_work(i915->wq,
|
||||||
|
&i915->gem.retire_work,
|
||||||
|
round_jiffies_up_relative(HZ));
|
||||||
|
}
|
||||||
|
|
||||||
|
bool i915_gem_load_power_context(struct drm_i915_private *i915)
|
||||||
|
{
|
||||||
|
/* Force loading the kernel context on all engines */
|
||||||
|
if (!switch_to_kernel_context_sync(i915, ALL_ENGINES))
|
||||||
|
return false;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Immediately park the GPU so that we enable powersaving and
|
||||||
|
* treat it as idle. The next time we issue a request, we will
|
||||||
|
* unpark and start using the engine->pinned_default_state, otherwise
|
||||||
|
* it is in limbo and an early reset may fail.
|
||||||
|
*/
|
||||||
|
__i915_gem_park(i915);
|
||||||
|
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
void i915_gem_suspend(struct drm_i915_private *i915)
|
||||||
|
{
|
||||||
|
intel_wakeref_t wakeref;
|
||||||
|
|
||||||
|
GEM_TRACE("\n");
|
||||||
|
|
||||||
|
wakeref = intel_runtime_pm_get(i915);
|
||||||
|
|
||||||
|
mutex_lock(&i915->drm.struct_mutex);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* We have to flush all the executing contexts to main memory so
|
||||||
|
* that they can saved in the hibernation image. To ensure the last
|
||||||
|
* context image is coherent, we have to switch away from it. That
|
||||||
|
* leaves the i915->kernel_context still active when
|
||||||
|
* we actually suspend, and its image in memory may not match the GPU
|
||||||
|
* state. Fortunately, the kernel_context is disposable and we do
|
||||||
|
* not rely on its state.
|
||||||
|
*/
|
||||||
|
switch_to_kernel_context_sync(i915, i915->gt.active_engines);
|
||||||
|
|
||||||
|
mutex_unlock(&i915->drm.struct_mutex);
|
||||||
|
i915_reset_flush(i915);
|
||||||
|
|
||||||
|
drain_delayed_work(&i915->gem.retire_work);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* As the idle_work is rearming if it detects a race, play safe and
|
||||||
|
* repeat the flush until it is definitely idle.
|
||||||
|
*/
|
||||||
|
drain_delayed_work(&i915->gem.idle_work);
|
||||||
|
|
||||||
|
flush_workqueue(i915->wq);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Assert that we successfully flushed all the work and
|
||||||
|
* reset the GPU back to its idle, low power state.
|
||||||
|
*/
|
||||||
|
GEM_BUG_ON(i915->gt.awake);
|
||||||
|
|
||||||
|
intel_uc_suspend(i915);
|
||||||
|
|
||||||
|
intel_runtime_pm_put(i915, wakeref);
|
||||||
|
}
|
||||||
|
|
||||||
|
void i915_gem_suspend_late(struct drm_i915_private *i915)
|
||||||
|
{
|
||||||
|
struct drm_i915_gem_object *obj;
|
||||||
|
struct list_head *phases[] = {
|
||||||
|
&i915->mm.unbound_list,
|
||||||
|
&i915->mm.bound_list,
|
||||||
|
NULL
|
||||||
|
}, **phase;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Neither the BIOS, ourselves or any other kernel
|
||||||
|
* expects the system to be in execlists mode on startup,
|
||||||
|
* so we need to reset the GPU back to legacy mode. And the only
|
||||||
|
* known way to disable logical contexts is through a GPU reset.
|
||||||
|
*
|
||||||
|
* So in order to leave the system in a known default configuration,
|
||||||
|
* always reset the GPU upon unload and suspend. Afterwards we then
|
||||||
|
* clean up the GEM state tracking, flushing off the requests and
|
||||||
|
* leaving the system in a known idle state.
|
||||||
|
*
|
||||||
|
* Note that is of the upmost importance that the GPU is idle and
|
||||||
|
* all stray writes are flushed *before* we dismantle the backing
|
||||||
|
* storage for the pinned objects.
|
||||||
|
*
|
||||||
|
* However, since we are uncertain that resetting the GPU on older
|
||||||
|
* machines is a good idea, we don't - just in case it leaves the
|
||||||
|
* machine in an unusable condition.
|
||||||
|
*/
|
||||||
|
|
||||||
|
mutex_lock(&i915->drm.struct_mutex);
|
||||||
|
for (phase = phases; *phase; phase++) {
|
||||||
|
list_for_each_entry(obj, *phase, mm.link)
|
||||||
|
WARN_ON(i915_gem_object_set_to_gtt_domain(obj, false));
|
||||||
|
}
|
||||||
|
mutex_unlock(&i915->drm.struct_mutex);
|
||||||
|
|
||||||
|
intel_uc_sanitize(i915);
|
||||||
|
i915_gem_sanitize(i915);
|
||||||
|
}
|
||||||
|
|
||||||
|
void i915_gem_resume(struct drm_i915_private *i915)
|
||||||
|
{
|
||||||
|
GEM_TRACE("\n");
|
||||||
|
|
||||||
|
WARN_ON(i915->gt.awake);
|
||||||
|
|
||||||
|
mutex_lock(&i915->drm.struct_mutex);
|
||||||
|
intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
|
||||||
|
|
||||||
|
i915_gem_restore_gtt_mappings(i915);
|
||||||
|
i915_gem_restore_fences(i915);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* As we didn't flush the kernel context before suspend, we cannot
|
||||||
|
* guarantee that the context image is complete. So let's just reset
|
||||||
|
* it and start again.
|
||||||
|
*/
|
||||||
|
intel_gt_resume(i915);
|
||||||
|
|
||||||
|
if (i915_gem_init_hw(i915))
|
||||||
|
goto err_wedged;
|
||||||
|
|
||||||
|
intel_uc_resume(i915);
|
||||||
|
|
||||||
|
/* Always reload a context for powersaving. */
|
||||||
|
if (!i915_gem_load_power_context(i915))
|
||||||
|
goto err_wedged;
|
||||||
|
|
||||||
|
out_unlock:
|
||||||
|
intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
|
||||||
|
mutex_unlock(&i915->drm.struct_mutex);
|
||||||
|
return;
|
||||||
|
|
||||||
|
err_wedged:
|
||||||
|
if (!i915_reset_failed(i915)) {
|
||||||
|
dev_err(i915->drm.dev,
|
||||||
|
"Failed to re-initialize GPU, declaring it wedged!\n");
|
||||||
|
i915_gem_set_wedged(i915);
|
||||||
|
}
|
||||||
|
goto out_unlock;
|
||||||
|
}
|
||||||
|
|
||||||
|
void i915_gem_init__pm(struct drm_i915_private *i915)
|
||||||
|
{
|
||||||
|
INIT_DELAYED_WORK(&i915->gem.idle_work, idle_work_handler);
|
||||||
|
INIT_DELAYED_WORK(&i915->gem.retire_work, retire_work_handler);
|
||||||
|
}
|
||||||
28
drivers/gpu/drm/i915/i915_gem_pm.h
Normal file
28
drivers/gpu/drm/i915/i915_gem_pm.h
Normal file
@@ -0,0 +1,28 @@
|
|||||||
|
/*
|
||||||
|
* SPDX-License-Identifier: MIT
|
||||||
|
*
|
||||||
|
* Copyright © 2019 Intel Corporation
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef __I915_GEM_PM_H__
|
||||||
|
#define __I915_GEM_PM_H__
|
||||||
|
|
||||||
|
#include <linux/types.h>
|
||||||
|
|
||||||
|
struct drm_i915_private;
|
||||||
|
struct work_struct;
|
||||||
|
|
||||||
|
void i915_gem_init__pm(struct drm_i915_private *i915);
|
||||||
|
|
||||||
|
bool i915_gem_load_power_context(struct drm_i915_private *i915);
|
||||||
|
void i915_gem_resume(struct drm_i915_private *i915);
|
||||||
|
|
||||||
|
void i915_gem_unpark(struct drm_i915_private *i915);
|
||||||
|
void i915_gem_park(struct drm_i915_private *i915);
|
||||||
|
|
||||||
|
void i915_gem_idle_work_handler(struct work_struct *work);
|
||||||
|
|
||||||
|
void i915_gem_suspend(struct drm_i915_private *i915);
|
||||||
|
void i915_gem_suspend_late(struct drm_i915_private *i915);
|
||||||
|
|
||||||
|
#endif /* __I915_GEM_PM_H__ */
|
||||||
@@ -1658,7 +1658,7 @@ static int __igt_switch_to_kernel_context(struct drm_i915_private *i915,
|
|||||||
/* XXX Bonus points for proving we are the kernel context! */
|
/* XXX Bonus points for proving we are the kernel context! */
|
||||||
|
|
||||||
mutex_unlock(&i915->drm.struct_mutex);
|
mutex_unlock(&i915->drm.struct_mutex);
|
||||||
drain_delayed_work(&i915->gt.idle_work);
|
drain_delayed_work(&i915->gem.idle_work);
|
||||||
mutex_lock(&i915->drm.struct_mutex);
|
mutex_lock(&i915->drm.struct_mutex);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -514,8 +514,8 @@ static void disable_retire_worker(struct drm_i915_private *i915)
|
|||||||
}
|
}
|
||||||
mutex_unlock(&i915->drm.struct_mutex);
|
mutex_unlock(&i915->drm.struct_mutex);
|
||||||
|
|
||||||
cancel_delayed_work_sync(&i915->gt.retire_work);
|
cancel_delayed_work_sync(&i915->gem.retire_work);
|
||||||
cancel_delayed_work_sync(&i915->gt.idle_work);
|
cancel_delayed_work_sync(&i915->gem.idle_work);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int igt_mmap_offset_exhaustion(void *arg)
|
static int igt_mmap_offset_exhaustion(void *arg)
|
||||||
@@ -617,9 +617,9 @@ out:
|
|||||||
out_park:
|
out_park:
|
||||||
mutex_lock(&i915->drm.struct_mutex);
|
mutex_lock(&i915->drm.struct_mutex);
|
||||||
if (--i915->gt.active_requests)
|
if (--i915->gt.active_requests)
|
||||||
queue_delayed_work(i915->wq, &i915->gt.retire_work, 0);
|
queue_delayed_work(i915->wq, &i915->gem.retire_work, 0);
|
||||||
else
|
else
|
||||||
queue_delayed_work(i915->wq, &i915->gt.idle_work, 0);
|
queue_delayed_work(i915->wq, &i915->gem.idle_work, 0);
|
||||||
mutex_unlock(&i915->drm.struct_mutex);
|
mutex_unlock(&i915->drm.struct_mutex);
|
||||||
i915_gem_shrinker_register(i915);
|
i915_gem_shrinker_register(i915);
|
||||||
return err;
|
return err;
|
||||||
|
|||||||
@@ -59,8 +59,8 @@ static void mock_device_release(struct drm_device *dev)
|
|||||||
i915_gem_contexts_lost(i915);
|
i915_gem_contexts_lost(i915);
|
||||||
mutex_unlock(&i915->drm.struct_mutex);
|
mutex_unlock(&i915->drm.struct_mutex);
|
||||||
|
|
||||||
drain_delayed_work(&i915->gt.retire_work);
|
drain_delayed_work(&i915->gem.retire_work);
|
||||||
drain_delayed_work(&i915->gt.idle_work);
|
drain_delayed_work(&i915->gem.idle_work);
|
||||||
i915_gem_drain_workqueue(i915);
|
i915_gem_drain_workqueue(i915);
|
||||||
|
|
||||||
mutex_lock(&i915->drm.struct_mutex);
|
mutex_lock(&i915->drm.struct_mutex);
|
||||||
@@ -111,7 +111,7 @@ static void mock_retire_work_handler(struct work_struct *work)
|
|||||||
static void mock_idle_work_handler(struct work_struct *work)
|
static void mock_idle_work_handler(struct work_struct *work)
|
||||||
{
|
{
|
||||||
struct drm_i915_private *i915 =
|
struct drm_i915_private *i915 =
|
||||||
container_of(work, typeof(*i915), gt.idle_work.work);
|
container_of(work, typeof(*i915), gem.idle_work.work);
|
||||||
|
|
||||||
i915->gt.active_engines = 0;
|
i915->gt.active_engines = 0;
|
||||||
}
|
}
|
||||||
@@ -197,8 +197,8 @@ struct drm_i915_private *mock_gem_device(void)
|
|||||||
|
|
||||||
mock_init_contexts(i915);
|
mock_init_contexts(i915);
|
||||||
|
|
||||||
INIT_DELAYED_WORK(&i915->gt.retire_work, mock_retire_work_handler);
|
INIT_DELAYED_WORK(&i915->gem.retire_work, mock_retire_work_handler);
|
||||||
INIT_DELAYED_WORK(&i915->gt.idle_work, mock_idle_work_handler);
|
INIT_DELAYED_WORK(&i915->gem.idle_work, mock_idle_work_handler);
|
||||||
|
|
||||||
i915->gt.awake = true;
|
i915->gt.awake = true;
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user