[PATCH 30/31] drm/i915,intel_ips: Enable GPU wait-boosting with IPS
Chris Wilson
chris at chris-wilson.co.uk
Wed Nov 29 01:00:02 UTC 2017
Refactor the reclocking logic used by RPS on Ironlake to reuse the
infrastructure developed for RPS on Sandybridge+, along with the
waitboosting support for stalled clients and missed frames.
v2: Mark rps interrupts as enabled
Reported-by: dimon at gmx.net
Bugzilla: https://bugs.freedesktop.org/show_bug.cgi?id=90137
Signed-off-by: Chris Wilson <chris at chris-wilson.co.uk>
---
drivers/gpu/drm/i915/i915_debugfs.c | 1 +
drivers/gpu/drm/i915/i915_drv.h | 7 --
drivers/gpu/drm/i915/i915_gem_request.c | 1 -
drivers/gpu/drm/i915/i915_irq.c | 102 ++++++++++++-------------
drivers/gpu/drm/i915/i915_sysfs.c | 10 +++
drivers/gpu/drm/i915/intel_gt_pm.c | 127 +++++++++++++++++++++-----------
drivers/platform/x86/intel_ips.c | 14 +++-
include/drm/i915_drm.h | 1 +
8 files changed, 155 insertions(+), 108 deletions(-)
diff --git a/drivers/gpu/drm/i915/i915_debugfs.c b/drivers/gpu/drm/i915/i915_debugfs.c
index c42e85ded48c..c8a3b9c7a6f1 100644
--- a/drivers/gpu/drm/i915/i915_debugfs.c
+++ b/drivers/gpu/drm/i915/i915_debugfs.c
@@ -1416,6 +1416,7 @@ static int ironlake_drpc_info(struct seq_file *m)
yesno(rgvmodectl & MEMMODE_HWIDLE_EN));
seq_printf(m, "SW control enabled: %s\n",
yesno(rgvmodectl & MEMMODE_SWMODE_EN));
+ seq_printf(m, "RPS active? %s\n", yesno(dev_priv->gt.awake));
seq_printf(m, "Gated voltage change: %s\n",
yesno(rgvmodectl & MEMMODE_RCLK_GATE));
seq_printf(m, "Starting frequency: P%d\n",
diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h
index 33882623d867..c41222487fef 100644
--- a/drivers/gpu/drm/i915/i915_drv.h
+++ b/drivers/gpu/drm/i915/i915_drv.h
@@ -1392,12 +1392,6 @@ struct intel_gt_pm {
extern spinlock_t mchdev_lock;
struct intel_ilk_power_mgmt {
- u8 cur_delay;
- u8 min_delay;
- u8 max_delay;
- u8 fmax;
- u8 fstart;
-
u64 last_count1;
unsigned long last_time1;
unsigned long chipset_power;
@@ -3339,7 +3333,6 @@ extern void intel_hangcheck_init(struct drm_i915_private *dev_priv);
extern unsigned long i915_chipset_val(struct drm_i915_private *dev_priv);
extern unsigned long i915_mch_val(struct drm_i915_private *dev_priv);
extern unsigned long i915_gfx_val(struct drm_i915_private *dev_priv);
-extern void i915_update_gfx_val(struct drm_i915_private *dev_priv);
int vlv_force_gfx_clock(struct drm_i915_private *dev_priv, bool on);
int intel_engines_init_mmio(struct drm_i915_private *dev_priv);
diff --git a/drivers/gpu/drm/i915/i915_gem_request.c b/drivers/gpu/drm/i915/i915_gem_request.c
index 9a84375e0db0..5283014aa16d 100644
--- a/drivers/gpu/drm/i915/i915_gem_request.c
+++ b/drivers/gpu/drm/i915/i915_gem_request.c
@@ -254,7 +254,6 @@ static void mark_busy(struct drm_i915_private *i915)
intel_runtime_pm_get_noresume(i915);
- i915_update_gfx_val(i915);
intel_rps_busy(i915);
i915_pmu_gt_unparked(i915);
diff --git a/drivers/gpu/drm/i915/i915_irq.c b/drivers/gpu/drm/i915/i915_irq.c
index a7e06bd2155c..0e4aa0e82d4e 100644
--- a/drivers/gpu/drm/i915/i915_irq.c
+++ b/drivers/gpu/drm/i915/i915_irq.c
@@ -394,9 +394,13 @@ static void gen6_disable_pm_irq(struct drm_i915_private *dev_priv, u32 disable_m
{
lockdep_assert_held(&dev_priv->irq_lock);
- dev_priv->pm_ier &= ~disable_mask;
- __gen6_mask_pm_irq(dev_priv, disable_mask);
- I915_WRITE(gen6_pm_ier(dev_priv), dev_priv->pm_ier);
+ if (INTEL_INFO(dev_priv)->gen >= 6) {
+ dev_priv->pm_ier &= ~disable_mask;
+ __gen6_mask_pm_irq(dev_priv, disable_mask);
+ I915_WRITE(gen6_pm_ier(dev_priv), dev_priv->pm_ier);
+ } else if (IS_IRONLAKE_M(dev_priv)) {
+ ilk_enable_display_irq(dev_priv, DE_PCU_EVENT);
+ }
/* though a barrier is missing here, but don't really need a one */
}
@@ -434,9 +438,12 @@ void gen6_disable_rps_interrupts(struct drm_i915_private *dev_priv)
spin_lock_irq(&dev_priv->irq_lock);
rps->interrupts_enabled = false;
- I915_WRITE(GEN6_PMINTRMSK, gen6_sanitize_rps_pm_mask(dev_priv, ~0u));
-
- gen6_disable_pm_irq(dev_priv, dev_priv->pm_rps_events);
+ if (INTEL_GEN(dev_priv) >= 6) {
+ I915_WRITE(GEN6_PMINTRMSK, gen6_sanitize_rps_pm_mask(dev_priv, ~0u));
+ gen6_disable_pm_irq(dev_priv, dev_priv->pm_rps_events);
+ } else {
+ ilk_disable_display_irq(dev_priv, DE_PCU_EVENT);
+ }
spin_unlock_irq(&dev_priv->irq_lock);
synchronize_irq(dev_priv->drm.irq);
@@ -1025,44 +1032,6 @@ int intel_get_crtc_scanline(struct intel_crtc *crtc)
return position;
}
-static void ironlake_rps_change_irq_handler(struct drm_i915_private *dev_priv)
-{
- u32 busy_up, busy_down, max_avg, min_avg;
- u8 new_delay;
-
- spin_lock(&mchdev_lock);
-
- I915_WRITE16(MEMINTRSTS, I915_READ(MEMINTRSTS));
-
- new_delay = dev_priv->ips.cur_delay;
-
- I915_WRITE16(MEMINTRSTS, MEMINT_EVAL_CHG);
- busy_up = I915_READ(RCPREVBSYTUPAVG);
- busy_down = I915_READ(RCPREVBSYTDNAVG);
- max_avg = I915_READ(RCBMAXAVG);
- min_avg = I915_READ(RCBMINAVG);
-
- /* Handle RCS change request from hw */
- if (busy_up > max_avg) {
- if (dev_priv->ips.cur_delay != dev_priv->ips.max_delay)
- new_delay = dev_priv->ips.cur_delay - 1;
- if (new_delay < dev_priv->ips.max_delay)
- new_delay = dev_priv->ips.max_delay;
- } else if (busy_down < min_avg) {
- if (dev_priv->ips.cur_delay != dev_priv->ips.min_delay)
- new_delay = dev_priv->ips.cur_delay + 1;
- if (new_delay > dev_priv->ips.min_delay)
- new_delay = dev_priv->ips.min_delay;
- }
-
- if (ironlake_set_drps(dev_priv, new_delay))
- dev_priv->ips.cur_delay = new_delay;
-
- spin_unlock(&mchdev_lock);
-
- return;
-}
-
static void notify_ring(struct intel_engine_cs *engine)
{
struct drm_i915_gem_request *rq = NULL;
@@ -1125,6 +1094,36 @@ static void vlv_c0_read(struct drm_i915_private *dev_priv,
ei->media_c0 = I915_READ(VLV_MEDIA_C0_COUNT);
}
+static u32 ilk_compute_pm_iir(struct drm_i915_private *dev_priv)
+{
+ u32 pm_iir;
+
+ spin_lock(&mchdev_lock);
+ I915_WRITE16(MEMINTRSTS, I915_READ(MEMINTRSTS));
+ I915_WRITE16(MEMINTRSTS, MEMINT_EVAL_CHG);
+
+#define busy_up I915_READ(RCPREVBSYTUPAVG)
+#define busy_down I915_READ(RCPREVBSYTDNAVG)
+#define max_avg I915_READ(RCBMAXAVG)
+#define min_avg I915_READ(RCBMINAVG)
+
+ if (busy_up > max_avg)
+ pm_iir = GEN6_PM_RP_UP_THRESHOLD;
+ else if (busy_down < min_avg)
+ pm_iir = GEN6_PM_RP_DOWN_THRESHOLD;
+ else
+ pm_iir = 0;
+
+#undef busy_up
+#undef busy_down
+#undef max_avg
+#undef min_avg
+
+ spin_unlock(&mchdev_lock);
+
+ return pm_iir;
+}
+
static u32 vlv_wa_c0_ei(struct drm_i915_private *dev_priv, u32 pm_iir)
{
struct intel_rps *rps = &dev_priv->gt_pm.rps;
@@ -1178,6 +1177,8 @@ static void gen6_pm_rps_work(struct work_struct *work)
if (rps->interrupts_enabled) {
pm_iir = fetch_and_zero(&rps->pm_iir);
client_boost = atomic_read(&rps->num_waiters);
+ if (IS_GEN5(dev_priv))
+ pm_iir = ilk_compute_pm_iir(dev_priv);
}
spin_unlock_irq(&dev_priv->irq_lock);
@@ -2404,7 +2405,7 @@ static void ilk_display_irq_handler(struct drm_i915_private *dev_priv,
}
if (IS_GEN5(dev_priv) && de_iir & DE_PCU_EVENT)
- ironlake_rps_change_irq_handler(dev_priv);
+ queue_work(dev_priv->wq, &dev_priv->gt_pm.rps.work);
}
static void ivb_display_irq_handler(struct drm_i915_private *dev_priv,
@@ -3495,17 +3496,6 @@ static int ironlake_irq_postinstall(struct drm_device *dev)
ibx_irq_postinstall(dev);
- if (IS_IRONLAKE_M(dev_priv)) {
- /* Enable PCU event interrupts
- *
- * spinlocking not required here for correctness since interrupt
- * setup is guaranteed to run in single-threaded context. But we
- * need it to make the assert_spin_locked happy. */
- spin_lock_irq(&dev_priv->irq_lock);
- ilk_enable_display_irq(dev_priv, DE_PCU_EVENT);
- spin_unlock_irq(&dev_priv->irq_lock);
- }
-
return 0;
}
diff --git a/drivers/gpu/drm/i915/i915_sysfs.c b/drivers/gpu/drm/i915/i915_sysfs.c
index c74a20b80182..fbff031e85ad 100644
--- a/drivers/gpu/drm/i915/i915_sysfs.c
+++ b/drivers/gpu/drm/i915/i915_sysfs.c
@@ -477,6 +477,14 @@ static ssize_t gt_rp_mhz_show(struct device *kdev, struct device_attribute *attr
return snprintf(buf, PAGE_SIZE, "%d\n", val);
}
+static const struct attribute *gen5_attrs[] = {
+ &dev_attr_gt_cur_freq_mhz.attr,
+ &dev_attr_gt_max_freq_mhz.attr,
+ &dev_attr_gt_min_freq_mhz.attr,
+ &dev_attr_gt_RP0_freq_mhz.attr,
+ &dev_attr_gt_RPn_freq_mhz.attr,
+ NULL,
+};
static const struct attribute *gen6_attrs[] = {
&dev_attr_gt_act_freq_mhz.attr,
&dev_attr_gt_cur_freq_mhz.attr,
@@ -613,6 +621,8 @@ void i915_setup_sysfs(struct drm_i915_private *dev_priv)
ret = sysfs_create_files(&kdev->kobj, vlv_attrs);
else if (INTEL_GEN(dev_priv) >= 6)
ret = sysfs_create_files(&kdev->kobj, gen6_attrs);
+ else if (INTEL_INFO(dev_priv)->gen >= 5)
+ ret = sysfs_create_files(&kdev->kobj, gen5_attrs);
if (ret)
DRM_ERROR("RPS sysfs setup failed\n");
diff --git a/drivers/gpu/drm/i915/intel_gt_pm.c b/drivers/gpu/drm/i915/intel_gt_pm.c
index 7275728dd4bd..e4f7f7d15256 100644
--- a/drivers/gpu/drm/i915/intel_gt_pm.c
+++ b/drivers/gpu/drm/i915/intel_gt_pm.c
@@ -49,26 +49,41 @@
* which brings the most power savings; deeper states save more power, but
* require higher latency to switch to and wake up.
*/
+static void gen5_update_gfx_val(struct drm_i915_private *dev_priv);
/*
* Lock protecting IPS related data structures
*/
DEFINE_SPINLOCK(mchdev_lock);
-bool ironlake_set_drps(struct drm_i915_private *dev_priv, u8 val)
+static bool __ironlake_set_rps(struct drm_i915_private *dev_priv, u8 val)
{
+ struct intel_rps *rps = &dev_priv->gt_pm.rps;
u16 rgvswctl;
lockdep_assert_held(&mchdev_lock);
- rgvswctl = I915_READ16(MEMSWCTL);
- if (rgvswctl & MEMCTL_CMD_STS) {
- DRM_DEBUG("gpu busy, RCS change rejected\n");
+ if (GEM_WARN_ON(val < rps->min_freq))
+ return false;
+ if (GEM_WARN_ON(val > rps->max_freq))
+ return false;
+
+ if (val == rps->cur_freq)
+ return true;
+
+ if (wait_for_atomic((I915_READ(MEMSWCTL) & MEMCTL_CMD_STS) == 0, 10)) {
+ DRM_DEBUG_DRIVER("gpu busy, RCS change rejected\n");
return false; /* still busy with another command */
}
- rgvswctl = (MEMCTL_CMD_CHFREQ << MEMCTL_CMD_SHIFT) |
- (val << MEMCTL_FREQ_SHIFT) | MEMCTL_SFCAVM;
+ trace_intel_gpu_freq_change(val);
+ rps->cur_freq = val;
+ val = rps->max_freq - val + rps->min_freq;
+
+ rgvswctl =
+ (MEMCTL_CMD_CHFREQ << MEMCTL_CMD_SHIFT) |
+ (val << MEMCTL_FREQ_SHIFT) |
+ MEMCTL_SFCAVM;
I915_WRITE16(MEMSWCTL, rgvswctl);
POSTING_READ16(MEMSWCTL);
@@ -78,12 +93,22 @@ bool ironlake_set_drps(struct drm_i915_private *dev_priv, u8 val)
return true;
}
+static int ironlake_set_rps(struct drm_i915_private *dev_priv, u8 val)
+{
+ spin_lock_irq(&mchdev_lock);
+ __ironlake_set_rps(dev_priv, val);
+ spin_unlock_irq(&mchdev_lock);
+
+ return 0;
+}
+
static void ironlake_enable_drps(struct drm_i915_private *dev_priv)
{
+ struct intel_rps *rps = &dev_priv->gt_pm.rps;
u32 rgvmodectl;
u8 fmax, fmin, fstart, vstart;
- spin_lock_irq(&mchdev_lock);
+ spin_lock(&mchdev_lock);
rgvmodectl = I915_READ(MEMMODECTL);
@@ -110,16 +135,18 @@ static void ironlake_enable_drps(struct drm_i915_private *dev_priv)
vstart = (I915_READ(PXVFREQ(fstart)) & PXVFREQ_PX_MASK) >>
PXVFREQ_PX_SHIFT;
- dev_priv->ips.fmax = fmax; /* IPS callback will increase this */
- dev_priv->ips.fstart = fstart;
-
- dev_priv->ips.max_delay = fstart;
- dev_priv->ips.min_delay = fmin;
- dev_priv->ips.cur_delay = fstart;
+ rps->max_freq = fmin;
+ rps->min_freq = fmax;
+ rps->cur_freq = fmin - fstart;
DRM_DEBUG_DRIVER("fmax: %d, fmin: %d, fstart: %d\n",
fmax, fmin, fstart);
+ rps->max_freq_softlimit = rps->min_freq;
+ rps->min_freq_softlimit = rps->min_freq;
+ rps->efficient_freq = rps->cur_freq;
+ rps->idle_freq = rps->min_freq;
+
I915_WRITE(MEMINTREN, MEMINT_CX_SUPR_EN | MEMINT_EVAL_CHG_EN);
/*
@@ -136,7 +163,7 @@ static void ironlake_enable_drps(struct drm_i915_private *dev_priv)
DRM_ERROR("stuck trying to change perf mode\n");
mdelay(1);
- ironlake_set_drps(dev_priv, fstart);
+ __ironlake_set_rps(dev_priv, dev_priv->gt_pm.rps.cur_freq);
dev_priv->ips.last_count1 = I915_READ(DMIEC) +
I915_READ(DDREC) + I915_READ(CSIEC);
@@ -144,14 +171,14 @@ static void ironlake_enable_drps(struct drm_i915_private *dev_priv)
dev_priv->ips.last_count2 = I915_READ(GFXEC);
dev_priv->ips.last_time2 = ktime_get_raw_ns();
- spin_unlock_irq(&mchdev_lock);
+ spin_unlock(&mchdev_lock);
}
static void ironlake_disable_drps(struct drm_i915_private *dev_priv)
{
u16 rgvswctl;
- spin_lock_irq(&mchdev_lock);
+ spin_lock(&mchdev_lock);
rgvswctl = I915_READ16(MEMSWCTL);
@@ -163,13 +190,13 @@ static void ironlake_disable_drps(struct drm_i915_private *dev_priv)
I915_WRITE(DEIMR, I915_READ(DEIMR) | DE_PCU_EVENT);
/* Go back to the starting frequency */
- ironlake_set_drps(dev_priv, dev_priv->ips.fstart);
+ __ironlake_set_rps(dev_priv, dev_priv->gt_pm.rps.efficient_freq);
mdelay(1);
rgvswctl |= MEMCTL_CMD_STS;
I915_WRITE(MEMSWCTL, rgvswctl);
mdelay(1);
- spin_unlock_irq(&mchdev_lock);
+ spin_unlock(&mchdev_lock);
}
/* There's a funny hw issue where the hw returns all 0 when reading from
@@ -416,6 +443,8 @@ void intel_rps_busy(struct drm_i915_private *dev_priv)
if (INTEL_GEN(dev_priv) >= 6)
gen6_enable_rps_interrupts(dev_priv);
+ else if (INTEL_GEN(dev_priv) >= 5)
+ gen5_update_gfx_val(dev_priv);
}
void intel_rps_idle(struct drm_i915_private *dev_priv)
@@ -509,7 +538,9 @@ int intel_set_rps(struct drm_i915_private *dev_priv, u8 val)
return 0;
}
- if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
+ if (IS_GEN5(dev_priv))
+ err = ironlake_set_rps(dev_priv, val);
+ else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
err = valleyview_set_rps(dev_priv, val);
else if (INTEL_GEN(dev_priv) >= 6)
err = gen6_set_rps(dev_priv, val);
@@ -1692,11 +1723,11 @@ unsigned long i915_chipset_val(struct drm_i915_private *dev_priv)
return 0;
intel_runtime_pm_get(dev_priv);
- spin_lock_irq(&mchdev_lock);
+ spin_lock(&mchdev_lock);
val = __i915_chipset_val(dev_priv);
- spin_unlock_irq(&mchdev_lock);
+ spin_unlock(&mchdev_lock);
intel_runtime_pm_put(dev_priv);
return val;
@@ -1772,17 +1803,14 @@ static void __i915_update_gfx_val(struct drm_i915_private *dev_priv)
dev_priv->ips.gfx_power = diff;
}
-void i915_update_gfx_val(struct drm_i915_private *dev_priv)
+static void gen5_update_gfx_val(struct drm_i915_private *dev_priv)
{
- if (INTEL_INFO(dev_priv)->gen != 5)
- return;
-
intel_runtime_pm_get(dev_priv);
- spin_lock_irq(&mchdev_lock);
+ spin_lock(&mchdev_lock);
__i915_update_gfx_val(dev_priv);
- spin_unlock_irq(&mchdev_lock);
+ spin_unlock(&mchdev_lock);
intel_runtime_pm_put(dev_priv);
}
@@ -1831,11 +1859,11 @@ unsigned long i915_gfx_val(struct drm_i915_private *dev_priv)
return 0;
intel_runtime_pm_get(dev_priv);
- spin_lock_irq(&mchdev_lock);
+ spin_lock(&mchdev_lock);
val = __i915_gfx_val(dev_priv);
- spin_unlock_irq(&mchdev_lock);
+ spin_unlock(&mchdev_lock);
intel_runtime_pm_put(dev_priv);
return val;
@@ -1891,15 +1919,15 @@ EXPORT_SYMBOL_GPL(i915_read_mch_val);
bool i915_gpu_raise(void)
{
struct drm_i915_private *i915;
+ struct intel_rps *rps;
i915 = mchdev_get();
if (!i915)
return false;
- spin_lock_irq(&mchdev_lock);
- if (i915->ips.max_delay > i915->ips.fmax)
- i915->ips.max_delay--;
- spin_unlock_irq(&mchdev_lock);
+ rps = &i915->gt_pm.rps;
+ if (rps->max_freq_softlimit < rps->max_freq)
+ rps->max_freq_softlimit++;
drm_dev_put(&i915->drm);
return true;
@@ -1915,15 +1943,17 @@ EXPORT_SYMBOL_GPL(i915_gpu_raise);
bool i915_gpu_lower(void)
{
struct drm_i915_private *i915;
+ struct intel_rps *rps;
i915 = mchdev_get();
if (!i915)
return false;
- spin_lock_irq(&mchdev_lock);
- if (i915->ips.max_delay < i915->ips.min_delay)
- i915->ips.max_delay++;
- spin_unlock_irq(&mchdev_lock);
+ mutex_lock(&i915->pcu_lock);
+ rps = &i915->gt_pm.rps;
+ if (rps->max_freq_softlimit > rps->min_freq)
+ rps->max_freq_softlimit--;
+ mutex_unlock(&i915->pcu_lock);
drm_dev_put(&i915->drm);
return true;
@@ -1960,22 +1990,33 @@ EXPORT_SYMBOL_GPL(i915_gpu_busy);
bool i915_gpu_turbo_disable(void)
{
struct drm_i915_private *i915;
- bool ret;
i915 = mchdev_get();
if (!i915)
return false;
- spin_lock_irq(&mchdev_lock);
- i915->ips.max_delay = i915->ips.fstart;
- ret = ironlake_set_drps(i915, i915->ips.fstart);
- spin_unlock_irq(&mchdev_lock);
+ intel_gt_disable_rps(i915);
drm_dev_put(&i915->drm);
- return ret;
+ return true;
}
EXPORT_SYMBOL_GPL(i915_gpu_turbo_disable);
+bool i915_gpu_turbo_enable(void)
+{
+ struct drm_i915_private *i915;
+
+ i915 = mchdev_get();
+ if (!i915)
+ return false;
+
+ intel_gt_enable_rps(i915);
+
+ drm_dev_put(&i915->drm);
+ return true;
+}
+EXPORT_SYMBOL_GPL(i915_gpu_turbo_enable);
+
/**
* Tells the intel_ips driver that the i915 driver is now loaded, if
* IPS got loaded first.
diff --git a/drivers/platform/x86/intel_ips.c b/drivers/platform/x86/intel_ips.c
index a0c95853fd3f..ad09b3095eec 100644
--- a/drivers/platform/x86/intel_ips.c
+++ b/drivers/platform/x86/intel_ips.c
@@ -336,6 +336,7 @@ struct ips_driver {
bool (*gpu_lower)(void);
bool (*gpu_busy)(void);
bool (*gpu_turbo_disable)(void);
+ bool (*gpu_turbo_enable)(void);
/* For restoration at unload */
u64 orig_turbo_limit;
@@ -575,7 +576,11 @@ static void ips_enable_gpu_turbo(struct ips_driver *ips)
{
if (ips->__gpu_turbo_on)
return;
- ips->__gpu_turbo_on = true;
+
+ if (!ips->gpu_turbo_enable())
+ dev_err(&ips->dev->dev, "failed to enable graphics turbo\n");
+ else
+ ips->__gpu_turbo_on = true;
}
/**
@@ -1432,9 +1437,14 @@ static bool ips_get_i915_syms(struct ips_driver *ips)
ips->gpu_turbo_disable = symbol_get(i915_gpu_turbo_disable);
if (!ips->gpu_turbo_disable)
goto out_put_busy;
+ ips->gpu_turbo_enable = symbol_get(i915_gpu_turbo_enable);
+ if (!ips->gpu_turbo_enable)
+ goto out_put_disable;
return true;
+out_put_disable:
+ symbol_put(i915_gpu_turbo_disable);
out_put_busy:
symbol_put(i915_gpu_busy);
out_put_lower:
@@ -1676,6 +1686,8 @@ static void ips_remove(struct pci_dev *dev)
symbol_put(i915_gpu_busy);
if (ips->gpu_turbo_disable)
symbol_put(i915_gpu_turbo_disable);
+ if (ips->gpu_turbo_enable)
+ symbol_put(i915_gpu_turbo_enable);
rdmsrl(TURBO_POWER_CURRENT_LIMIT, turbo_override);
turbo_override &= ~(TURBO_TDC_OVR_EN | TURBO_TDP_OVR_EN);
diff --git a/include/drm/i915_drm.h b/include/drm/i915_drm.h
index 4e1b274e1164..bf61d8649511 100644
--- a/include/drm/i915_drm.h
+++ b/include/drm/i915_drm.h
@@ -35,6 +35,7 @@ extern bool i915_gpu_raise(void);
extern bool i915_gpu_lower(void);
extern bool i915_gpu_busy(void);
extern bool i915_gpu_turbo_disable(void);
+extern bool i915_gpu_turbo_enable(void);
/*
* The Bridge device's PCI config space has information about the
--
2.15.0
More information about the Intel-gfx-trybot
mailing list