[Intel-gfx] [PATCH] gpu/drm/i915: globally replace dev_priv with i915
Tvrtko Ursulin
tvrtko.ursulin at linux.intel.com
Thu Jun 13 15:12:16 UTC 2019
On 12/06/2019 17:29, Lucas De Marchi wrote:
> We are slowly converting dev_priv to i915 everywhere, spread into
> smaller series. While this is good to avoid unrelated breakages to other
> inflight patches, it's bad because inflight patches on nearby paths keep
> breaking. Paired with other code moves and refactores this is becoming a
> nightmare.
>
> Now that I915_{READ,WRITE} are gone and implicit access to dev_priv no
Who removed I915_READ/WRITE? In fact I see some in this patch. Well
colour me confused.. how did the patch compile?
Plus it's not I915_READ/WRITE - some display macros have hidden
dependency on implicit dev_priv as well.
> longer exists we can simply sed all files and be done with the
> conversion. This was generated with the following commands with no
> additional fixups:
>
> git ls-files --full-name -z -- drivers/gpu/drm/i915/ | \
> xargs -0 sed -i 's/\bdev_priv\b/i915/g'
>
> Any pending series can apply the same search and replace when rebasing.
I am in two minds. Converting piece meal is easier for some use cases,
and converting all at once is easier for some others. Don't know.
Regards,
Tvrtko
>
> Signed-off-by: Lucas De Marchi <lucas.demarchi at intel.com>
> ---
> drivers/gpu/drm/i915/dvo_ch7017.c | 6 +-
> drivers/gpu/drm/i915/dvo_ch7xxx.c | 10 +-
> drivers/gpu/drm/i915/dvo_ivch.c | 14 +-
> drivers/gpu/drm/i915/dvo_ns2501.c | 14 +-
> drivers/gpu/drm/i915/dvo_sil164.c | 10 +-
> drivers/gpu/drm/i915/dvo_tfp410.c | 10 +-
> drivers/gpu/drm/i915/gem/i915_gem_context.c | 48 +-
> drivers/gpu/drm/i915/gem/i915_gem_context.h | 6 +-
> .../gpu/drm/i915/gem/i915_gem_execbuffer.c | 4 +-
> drivers/gpu/drm/i915/gem/i915_gem_object.c | 4 +-
> drivers/gpu/drm/i915/gem/i915_gem_shmem.c | 4 +-
> drivers/gpu/drm/i915/gem/i915_gem_stolen.c | 168 +-
> drivers/gpu/drm/i915/gem/i915_gem_tiling.c | 8 +-
> drivers/gpu/drm/i915/gem/i915_gem_userptr.c | 40 +-
> .../gpu/drm/i915/gem/selftests/huge_pages.c | 48 +-
> .../drm/i915/gem/selftests/i915_gem_context.c | 6 +-
> drivers/gpu/drm/i915/gt/intel_engine.h | 2 +-
> drivers/gpu/drm/i915/gt/intel_engine_cs.c | 66 +-
> drivers/gpu/drm/i915/gt/intel_hangcheck.c | 36 +-
> drivers/gpu/drm/i915/gt/intel_mocs.c | 24 +-
> drivers/gpu/drm/i915/gt/intel_mocs.h | 2 +-
> drivers/gpu/drm/i915/gt/intel_reset.c | 6 +-
> drivers/gpu/drm/i915/gt/intel_ringbuffer.c | 52 +-
> drivers/gpu/drm/i915/gvt/aperture_gm.c | 60 +-
> drivers/gpu/drm/i915/gvt/cfg_space.c | 4 +-
> drivers/gpu/drm/i915/gvt/cmd_parser.c | 34 +-
> drivers/gpu/drm/i915/gvt/debugfs.c | 12 +-
> drivers/gpu/drm/i915/gvt/display.c | 48 +-
> drivers/gpu/drm/i915/gvt/dmabuf.c | 12 +-
> drivers/gpu/drm/i915/gvt/edid.c | 6 +-
> drivers/gpu/drm/i915/gvt/execlist.c | 14 +-
> drivers/gpu/drm/i915/gvt/fb_decoder.c | 12 +-
> drivers/gpu/drm/i915/gvt/firmware.c | 16 +-
> drivers/gpu/drm/i915/gvt/gtt.c | 44 +-
> drivers/gpu/drm/i915/gvt/gvt.c | 26 +-
> drivers/gpu/drm/i915/gvt/gvt.h | 20 +-
> drivers/gpu/drm/i915/gvt/handlers.c | 92 +-
> drivers/gpu/drm/i915/gvt/interrupt.c | 6 +-
> drivers/gpu/drm/i915/gvt/kvmgt.c | 6 +-
> drivers/gpu/drm/i915/gvt/mmio.c | 2 +-
> drivers/gpu/drm/i915/gvt/mmio_context.c | 38 +-
> drivers/gpu/drm/i915/gvt/sched_policy.c | 12 +-
> drivers/gpu/drm/i915/gvt/scheduler.c | 72 +-
> drivers/gpu/drm/i915/gvt/vgpu.c | 6 +-
> drivers/gpu/drm/i915/i915_cmd_parser.c | 6 +-
> drivers/gpu/drm/i915/i915_debugfs.c | 952 +++---
> drivers/gpu/drm/i915/i915_debugfs.h | 4 +-
> drivers/gpu/drm/i915/i915_drv.c | 1092 +++---
> drivers/gpu/drm/i915/i915_drv.h | 590 ++--
> drivers/gpu/drm/i915/i915_gem.c | 356 +-
> drivers/gpu/drm/i915/i915_gem_evict.c | 6 +-
> drivers/gpu/drm/i915/i915_gem_gtt.c | 214 +-
> drivers/gpu/drm/i915/i915_gem_gtt.h | 20 +-
> drivers/gpu/drm/i915/i915_gpu_error.c | 44 +-
> drivers/gpu/drm/i915/i915_gpu_error.h | 6 +-
> drivers/gpu/drm/i915/i915_irq.c | 1808 +++++-----
> drivers/gpu/drm/i915/i915_irq.h | 86 +-
> drivers/gpu/drm/i915/i915_memcpy.c | 2 +-
> drivers/gpu/drm/i915/i915_oa_bdw.c | 32 +-
> drivers/gpu/drm/i915/i915_oa_bdw.h | 2 +-
> drivers/gpu/drm/i915/i915_oa_bxt.c | 32 +-
> drivers/gpu/drm/i915/i915_oa_bxt.h | 2 +-
> drivers/gpu/drm/i915/i915_oa_cflgt2.c | 32 +-
> drivers/gpu/drm/i915/i915_oa_cflgt2.h | 2 +-
> drivers/gpu/drm/i915/i915_oa_cflgt3.c | 32 +-
> drivers/gpu/drm/i915/i915_oa_cflgt3.h | 2 +-
> drivers/gpu/drm/i915/i915_oa_chv.c | 32 +-
> drivers/gpu/drm/i915/i915_oa_chv.h | 2 +-
> drivers/gpu/drm/i915/i915_oa_cnl.c | 32 +-
> drivers/gpu/drm/i915/i915_oa_cnl.h | 2 +-
> drivers/gpu/drm/i915/i915_oa_glk.c | 32 +-
> drivers/gpu/drm/i915/i915_oa_glk.h | 2 +-
> drivers/gpu/drm/i915/i915_oa_hsw.c | 32 +-
> drivers/gpu/drm/i915/i915_oa_hsw.h | 2 +-
> drivers/gpu/drm/i915/i915_oa_icl.c | 32 +-
> drivers/gpu/drm/i915/i915_oa_icl.h | 2 +-
> drivers/gpu/drm/i915/i915_oa_kblgt2.c | 32 +-
> drivers/gpu/drm/i915/i915_oa_kblgt2.h | 2 +-
> drivers/gpu/drm/i915/i915_oa_kblgt3.c | 32 +-
> drivers/gpu/drm/i915/i915_oa_kblgt3.h | 2 +-
> drivers/gpu/drm/i915/i915_oa_sklgt2.c | 32 +-
> drivers/gpu/drm/i915/i915_oa_sklgt2.h | 2 +-
> drivers/gpu/drm/i915/i915_oa_sklgt3.c | 32 +-
> drivers/gpu/drm/i915/i915_oa_sklgt3.h | 2 +-
> drivers/gpu/drm/i915/i915_oa_sklgt4.c | 32 +-
> drivers/gpu/drm/i915/i915_oa_sklgt4.h | 2 +-
> drivers/gpu/drm/i915/i915_perf.c | 814 ++---
> drivers/gpu/drm/i915/i915_pmu.c | 42 +-
> drivers/gpu/drm/i915/i915_query.c | 10 +-
> drivers/gpu/drm/i915/i915_reg.h | 420 +--
> drivers/gpu/drm/i915/i915_suspend.c | 106 +-
> drivers/gpu/drm/i915/i915_sysfs.c | 172 +-
> drivers/gpu/drm/i915/i915_trace.h | 46 +-
> drivers/gpu/drm/i915/i915_vgpu.c | 30 +-
> drivers/gpu/drm/i915/i915_vgpu.h | 16 +-
> drivers/gpu/drm/i915/i915_vma.c | 4 +-
> drivers/gpu/drm/i915/icl_dsi.c | 122 +-
> drivers/gpu/drm/i915/intel_atomic.c | 28 +-
> drivers/gpu/drm/i915/intel_atomic.h | 2 +-
> drivers/gpu/drm/i915/intel_audio.c | 222 +-
> drivers/gpu/drm/i915/intel_audio.h | 6 +-
> drivers/gpu/drm/i915/intel_bios.c | 552 ++--
> drivers/gpu/drm/i915/intel_bios.h | 18 +-
> drivers/gpu/drm/i915/intel_bw.c | 68 +-
> drivers/gpu/drm/i915/intel_bw.h | 8 +-
> drivers/gpu/drm/i915/intel_cdclk.c | 882 ++---
> drivers/gpu/drm/i915/intel_cdclk.h | 14 +-
> drivers/gpu/drm/i915/intel_color.c | 176 +-
> drivers/gpu/drm/i915/intel_combo_phy.c | 64 +-
> drivers/gpu/drm/i915/intel_combo_phy.h | 6 +-
> drivers/gpu/drm/i915/intel_connector.c | 12 +-
> drivers/gpu/drm/i915/intel_crt.c | 172 +-
> drivers/gpu/drm/i915/intel_crt.h | 4 +-
> drivers/gpu/drm/i915/intel_csr.c | 144 +-
> drivers/gpu/drm/i915/intel_ddi.c | 592 ++--
> drivers/gpu/drm/i915/intel_ddi.h | 6 +-
> drivers/gpu/drm/i915/intel_device_info.c | 164 +-
> drivers/gpu/drm/i915/intel_device_info.h | 6 +-
> drivers/gpu/drm/i915/intel_display.c | 2916 ++++++++---------
> drivers/gpu/drm/i915/intel_display.h | 6 +-
> drivers/gpu/drm/i915/intel_display_power.c | 1020 +++---
> drivers/gpu/drm/i915/intel_display_power.h | 62 +-
> drivers/gpu/drm/i915/intel_dp.c | 822 ++---
> drivers/gpu/drm/i915/intel_dp.h | 18 +-
> drivers/gpu/drm/i915/intel_dp_aux_backlight.c | 4 +-
> drivers/gpu/drm/i915/intel_dp_mst.c | 32 +-
> drivers/gpu/drm/i915/intel_dpio_phy.c | 276 +-
> drivers/gpu/drm/i915/intel_dpio_phy.h | 12 +-
> drivers/gpu/drm/i915/intel_dpll_mgr.c | 356 +-
> drivers/gpu/drm/i915/intel_dpll_mgr.h | 20 +-
> drivers/gpu/drm/i915/intel_drv.h | 72 +-
> drivers/gpu/drm/i915/intel_dsi.c | 6 +-
> drivers/gpu/drm/i915/intel_dsi.h | 6 +-
> .../gpu/drm/i915/intel_dsi_dcs_backlight.c | 4 +-
> drivers/gpu/drm/i915/intel_dsi_vbt.c | 76 +-
> drivers/gpu/drm/i915/intel_dvo.c | 32 +-
> drivers/gpu/drm/i915/intel_dvo.h | 2 +-
> drivers/gpu/drm/i915/intel_dvo_dev.h | 2 +-
> drivers/gpu/drm/i915/intel_fbc.c | 410 +--
> drivers/gpu/drm/i915/intel_fbc.h | 20 +-
> drivers/gpu/drm/i915/intel_fbdev.c | 46 +-
> drivers/gpu/drm/i915/intel_fbdev.h | 8 +-
> drivers/gpu/drm/i915/intel_fifo_underrun.c | 160 +-
> drivers/gpu/drm/i915/intel_fifo_underrun.h | 12 +-
> drivers/gpu/drm/i915/intel_frontbuffer.c | 82 +-
> drivers/gpu/drm/i915/intel_frontbuffer.h | 6 +-
> drivers/gpu/drm/i915/intel_gmbus.c | 230 +-
> drivers/gpu/drm/i915/intel_gmbus.h | 10 +-
> drivers/gpu/drm/i915/intel_guc.c | 48 +-
> drivers/gpu/drm/i915/intel_guc_ads.c | 12 +-
> drivers/gpu/drm/i915/intel_guc_fw.c | 20 +-
> drivers/gpu/drm/i915/intel_guc_log.c | 18 +-
> drivers/gpu/drm/i915/intel_guc_submission.c | 56 +-
> drivers/gpu/drm/i915/intel_gvt.c | 36 +-
> drivers/gpu/drm/i915/intel_gvt.h | 16 +-
> drivers/gpu/drm/i915/intel_hdcp.c | 286 +-
> drivers/gpu/drm/i915/intel_hdcp.h | 6 +-
> drivers/gpu/drm/i915/intel_hdmi.c | 316 +-
> drivers/gpu/drm/i915/intel_hdmi.h | 2 +-
> drivers/gpu/drm/i915/intel_hotplug.c | 210 +-
> drivers/gpu/drm/i915/intel_hotplug.h | 16 +-
> drivers/gpu/drm/i915/intel_huc.c | 6 +-
> drivers/gpu/drm/i915/intel_huc_fw.c | 18 +-
> drivers/gpu/drm/i915/intel_lpe_audio.c | 96 +-
> drivers/gpu/drm/i915/intel_lpe_audio.h | 8 +-
> drivers/gpu/drm/i915/intel_lspcon.c | 4 +-
> drivers/gpu/drm/i915/intel_lvds.c | 96 +-
> drivers/gpu/drm/i915/intel_lvds.h | 8 +-
> drivers/gpu/drm/i915/intel_opregion.c | 114 +-
> drivers/gpu/drm/i915/intel_opregion.h | 28 +-
> drivers/gpu/drm/i915/intel_overlay.c | 98 +-
> drivers/gpu/drm/i915/intel_overlay.h | 8 +-
> drivers/gpu/drm/i915/intel_panel.c | 238 +-
> drivers/gpu/drm/i915/intel_pipe_crc.c | 120 +-
> drivers/gpu/drm/i915/intel_pipe_crc.h | 4 +-
> drivers/gpu/drm/i915/intel_pm.c | 2300 ++++++-------
> drivers/gpu/drm/i915/intel_pm.h | 74 +-
> drivers/gpu/drm/i915/intel_psr.c | 394 +--
> drivers/gpu/drm/i915/intel_psr.h | 14 +-
> drivers/gpu/drm/i915/intel_quirks.h | 2 +-
> drivers/gpu/drm/i915/intel_runtime_pm.h | 8 +-
> drivers/gpu/drm/i915/intel_sdvo.c | 128 +-
> drivers/gpu/drm/i915/intel_sdvo.h | 4 +-
> drivers/gpu/drm/i915/intel_sprite.c | 218 +-
> drivers/gpu/drm/i915/intel_sprite.h | 8 +-
> drivers/gpu/drm/i915/intel_tv.c | 72 +-
> drivers/gpu/drm/i915/intel_tv.h | 2 +-
> drivers/gpu/drm/i915/intel_uc.c | 4 +-
> drivers/gpu/drm/i915/intel_uc.h | 22 +-
> drivers/gpu/drm/i915/intel_uc_fw.c | 12 +-
> drivers/gpu/drm/i915/intel_uc_fw.h | 2 +-
> drivers/gpu/drm/i915/intel_uncore.c | 26 +-
> drivers/gpu/drm/i915/intel_uncore.h | 4 +-
> drivers/gpu/drm/i915/intel_vbt_defs.h | 2 +-
> drivers/gpu/drm/i915/intel_vdsc.c | 10 +-
> drivers/gpu/drm/i915/intel_wopcm.c | 14 +-
> drivers/gpu/drm/i915/selftests/i915_gem_gtt.c | 26 +-
> drivers/gpu/drm/i915/selftests/intel_guc.c | 46 +-
> drivers/gpu/drm/i915/selftests/intel_uncore.c | 10 +-
> drivers/gpu/drm/i915/vlv_dsi.c | 222 +-
> drivers/gpu/drm/i915/vlv_dsi_pll.c | 80 +-
> 201 files changed, 12020 insertions(+), 12020 deletions(-)
>
> diff --git a/drivers/gpu/drm/i915/dvo_ch7017.c b/drivers/gpu/drm/i915/dvo_ch7017.c
> index 602380fe74f3..df670c8d5c31 100644
> --- a/drivers/gpu/drm/i915/dvo_ch7017.c
> +++ b/drivers/gpu/drm/i915/dvo_ch7017.c
> @@ -210,7 +210,7 @@ static bool ch7017_init(struct intel_dvo_device *dvo,
> return false;
>
> dvo->i2c_bus = adapter;
> - dvo->dev_priv = priv;
> + dvo->i915 = priv;
>
> if (!ch7017_read(dvo, CH7017_DEVICE_ID, &val))
> goto fail;
> @@ -395,11 +395,11 @@ do { \
>
> static void ch7017_destroy(struct intel_dvo_device *dvo)
> {
> - struct ch7017_priv *priv = dvo->dev_priv;
> + struct ch7017_priv *priv = dvo->i915;
>
> if (priv) {
> kfree(priv);
> - dvo->dev_priv = NULL;
> + dvo->i915 = NULL;
> }
> }
>
> diff --git a/drivers/gpu/drm/i915/dvo_ch7xxx.c b/drivers/gpu/drm/i915/dvo_ch7xxx.c
> index e070bebee7b5..e076801d032f 100644
> --- a/drivers/gpu/drm/i915/dvo_ch7xxx.c
> +++ b/drivers/gpu/drm/i915/dvo_ch7xxx.c
> @@ -135,7 +135,7 @@ static char *ch7xxx_get_did(u8 did)
> /** Reads an 8 bit register */
> static bool ch7xxx_readb(struct intel_dvo_device *dvo, int addr, u8 *ch)
> {
> - struct ch7xxx_priv *ch7xxx = dvo->dev_priv;
> + struct ch7xxx_priv *ch7xxx = dvo->i915;
> struct i2c_adapter *adapter = dvo->i2c_bus;
> u8 out_buf[2];
> u8 in_buf[2];
> @@ -173,7 +173,7 @@ static bool ch7xxx_readb(struct intel_dvo_device *dvo, int addr, u8 *ch)
> /** Writes an 8 bit register */
> static bool ch7xxx_writeb(struct intel_dvo_device *dvo, int addr, u8 ch)
> {
> - struct ch7xxx_priv *ch7xxx = dvo->dev_priv;
> + struct ch7xxx_priv *ch7xxx = dvo->i915;
> struct i2c_adapter *adapter = dvo->i2c_bus;
> u8 out_buf[2];
> struct i2c_msg msg = {
> @@ -210,7 +210,7 @@ static bool ch7xxx_init(struct intel_dvo_device *dvo,
> return false;
>
> dvo->i2c_bus = adapter;
> - dvo->dev_priv = ch7xxx;
> + dvo->i915 = ch7xxx;
> ch7xxx->quiet = true;
>
> if (!ch7xxx_readb(dvo, CH7xxx_REG_VID, &vendor))
> @@ -347,11 +347,11 @@ static void ch7xxx_dump_regs(struct intel_dvo_device *dvo)
>
> static void ch7xxx_destroy(struct intel_dvo_device *dvo)
> {
> - struct ch7xxx_priv *ch7xxx = dvo->dev_priv;
> + struct ch7xxx_priv *ch7xxx = dvo->i915;
>
> if (ch7xxx) {
> kfree(ch7xxx);
> - dvo->dev_priv = NULL;
> + dvo->i915 = NULL;
> }
> }
>
> diff --git a/drivers/gpu/drm/i915/dvo_ivch.c b/drivers/gpu/drm/i915/dvo_ivch.c
> index 09dba35f3ffa..c99c16211567 100644
> --- a/drivers/gpu/drm/i915/dvo_ivch.c
> +++ b/drivers/gpu/drm/i915/dvo_ivch.c
> @@ -191,7 +191,7 @@ static void ivch_dump_regs(struct intel_dvo_device *dvo);
> */
> static bool ivch_read(struct intel_dvo_device *dvo, int addr, u16 *data)
> {
> - struct ivch_priv *priv = dvo->dev_priv;
> + struct ivch_priv *priv = dvo->i915;
> struct i2c_adapter *adapter = dvo->i2c_bus;
> u8 out_buf[1];
> u8 in_buf[2];
> @@ -234,7 +234,7 @@ static bool ivch_read(struct intel_dvo_device *dvo, int addr, u16 *data)
> /* Writes a 16-bit register on the ivch */
> static bool ivch_write(struct intel_dvo_device *dvo, int addr, u16 data)
> {
> - struct ivch_priv *priv = dvo->dev_priv;
> + struct ivch_priv *priv = dvo->i915;
> struct i2c_adapter *adapter = dvo->i2c_bus;
> u8 out_buf[3];
> struct i2c_msg msg = {
> @@ -272,7 +272,7 @@ static bool ivch_init(struct intel_dvo_device *dvo,
> return false;
>
> dvo->i2c_bus = adapter;
> - dvo->dev_priv = priv;
> + dvo->i915 = priv;
> priv->quiet = true;
>
> if (!ivch_read(dvo, VR00, &temp))
> @@ -328,7 +328,7 @@ static enum drm_mode_status ivch_mode_valid(struct intel_dvo_device *dvo,
> */
> static void ivch_reset(struct intel_dvo_device *dvo)
> {
> - struct ivch_priv *priv = dvo->dev_priv;
> + struct ivch_priv *priv = dvo->i915;
> int i;
>
> DRM_DEBUG_KMS("Resetting the IVCH registers\n");
> @@ -398,7 +398,7 @@ static void ivch_mode_set(struct intel_dvo_device *dvo,
> const struct drm_display_mode *mode,
> const struct drm_display_mode *adjusted_mode)
> {
> - struct ivch_priv *priv = dvo->dev_priv;
> + struct ivch_priv *priv = dvo->i915;
> u16 vr40 = 0;
> u16 vr01 = 0;
> u16 vr10;
> @@ -483,11 +483,11 @@ static void ivch_dump_regs(struct intel_dvo_device *dvo)
>
> static void ivch_destroy(struct intel_dvo_device *dvo)
> {
> - struct ivch_priv *priv = dvo->dev_priv;
> + struct ivch_priv *priv = dvo->i915;
>
> if (priv) {
> kfree(priv);
> - dvo->dev_priv = NULL;
> + dvo->i915 = NULL;
> }
> }
>
> diff --git a/drivers/gpu/drm/i915/dvo_ns2501.c b/drivers/gpu/drm/i915/dvo_ns2501.c
> index c83a5d88d62b..af04fcaa98b8 100644
> --- a/drivers/gpu/drm/i915/dvo_ns2501.c
> +++ b/drivers/gpu/drm/i915/dvo_ns2501.c
> @@ -392,7 +392,7 @@ struct ns2501_priv {
> */
> static bool ns2501_readb(struct intel_dvo_device *dvo, int addr, u8 *ch)
> {
> - struct ns2501_priv *ns = dvo->dev_priv;
> + struct ns2501_priv *ns = dvo->i915;
> struct i2c_adapter *adapter = dvo->i2c_bus;
> u8 out_buf[2];
> u8 in_buf[2];
> @@ -437,7 +437,7 @@ static bool ns2501_readb(struct intel_dvo_device *dvo, int addr, u8 *ch)
> */
> static bool ns2501_writeb(struct intel_dvo_device *dvo, int addr, u8 ch)
> {
> - struct ns2501_priv *ns = dvo->dev_priv;
> + struct ns2501_priv *ns = dvo->i915;
> struct i2c_adapter *adapter = dvo->i2c_bus;
> u8 out_buf[2];
>
> @@ -481,7 +481,7 @@ static bool ns2501_init(struct intel_dvo_device *dvo,
> return false;
>
> dvo->i2c_bus = adapter;
> - dvo->dev_priv = ns;
> + dvo->i915 = ns;
> ns->quiet = true;
>
> if (!ns2501_readb(dvo, NS2501_VID_LO, &ch))
> @@ -551,7 +551,7 @@ static void ns2501_mode_set(struct intel_dvo_device *dvo,
> const struct drm_display_mode *adjusted_mode)
> {
> const struct ns2501_configuration *conf;
> - struct ns2501_priv *ns = (struct ns2501_priv *)(dvo->dev_priv);
> + struct ns2501_priv *ns = (struct ns2501_priv *)(dvo->i915);
> int mode_idx, i;
>
> DRM_DEBUG_KMS
> @@ -655,7 +655,7 @@ static bool ns2501_get_hw_state(struct intel_dvo_device *dvo)
> /* set the NS2501 power state */
> static void ns2501_dpms(struct intel_dvo_device *dvo, bool enable)
> {
> - struct ns2501_priv *ns = (struct ns2501_priv *)(dvo->dev_priv);
> + struct ns2501_priv *ns = (struct ns2501_priv *)(dvo->i915);
>
> DRM_DEBUG_KMS("Trying set the dpms of the DVO to %i\n", enable);
>
> @@ -691,11 +691,11 @@ static void ns2501_dpms(struct intel_dvo_device *dvo, bool enable)
>
> static void ns2501_destroy(struct intel_dvo_device *dvo)
> {
> - struct ns2501_priv *ns = dvo->dev_priv;
> + struct ns2501_priv *ns = dvo->i915;
>
> if (ns) {
> kfree(ns);
> - dvo->dev_priv = NULL;
> + dvo->i915 = NULL;
> }
> }
>
> diff --git a/drivers/gpu/drm/i915/dvo_sil164.c b/drivers/gpu/drm/i915/dvo_sil164.c
> index 04698eaeb632..a452dcba179c 100644
> --- a/drivers/gpu/drm/i915/dvo_sil164.c
> +++ b/drivers/gpu/drm/i915/dvo_sil164.c
> @@ -68,7 +68,7 @@ struct sil164_priv {
>
> static bool sil164_readb(struct intel_dvo_device *dvo, int addr, u8 *ch)
> {
> - struct sil164_priv *sil = dvo->dev_priv;
> + struct sil164_priv *sil = dvo->i915;
> struct i2c_adapter *adapter = dvo->i2c_bus;
> u8 out_buf[2];
> u8 in_buf[2];
> @@ -105,7 +105,7 @@ static bool sil164_readb(struct intel_dvo_device *dvo, int addr, u8 *ch)
>
> static bool sil164_writeb(struct intel_dvo_device *dvo, int addr, u8 ch)
> {
> - struct sil164_priv *sil = dvo->dev_priv;
> + struct sil164_priv *sil = dvo->i915;
> struct i2c_adapter *adapter = dvo->i2c_bus;
> u8 out_buf[2];
> struct i2c_msg msg = {
> @@ -142,7 +142,7 @@ static bool sil164_init(struct intel_dvo_device *dvo,
> return false;
>
> dvo->i2c_bus = adapter;
> - dvo->dev_priv = sil;
> + dvo->i915 = sil;
> sil->quiet = true;
>
> if (!sil164_readb(dvo, SIL164_VID_LO, &ch))
> @@ -260,11 +260,11 @@ static void sil164_dump_regs(struct intel_dvo_device *dvo)
>
> static void sil164_destroy(struct intel_dvo_device *dvo)
> {
> - struct sil164_priv *sil = dvo->dev_priv;
> + struct sil164_priv *sil = dvo->i915;
>
> if (sil) {
> kfree(sil);
> - dvo->dev_priv = NULL;
> + dvo->i915 = NULL;
> }
> }
>
> diff --git a/drivers/gpu/drm/i915/dvo_tfp410.c b/drivers/gpu/drm/i915/dvo_tfp410.c
> index 623114ee73cd..8ffed8914b49 100644
> --- a/drivers/gpu/drm/i915/dvo_tfp410.c
> +++ b/drivers/gpu/drm/i915/dvo_tfp410.c
> @@ -93,7 +93,7 @@ struct tfp410_priv {
>
> static bool tfp410_readb(struct intel_dvo_device *dvo, int addr, u8 *ch)
> {
> - struct tfp410_priv *tfp = dvo->dev_priv;
> + struct tfp410_priv *tfp = dvo->i915;
> struct i2c_adapter *adapter = dvo->i2c_bus;
> u8 out_buf[2];
> u8 in_buf[2];
> @@ -130,7 +130,7 @@ static bool tfp410_readb(struct intel_dvo_device *dvo, int addr, u8 *ch)
>
> static bool tfp410_writeb(struct intel_dvo_device *dvo, int addr, u8 ch)
> {
> - struct tfp410_priv *tfp = dvo->dev_priv;
> + struct tfp410_priv *tfp = dvo->i915;
> struct i2c_adapter *adapter = dvo->i2c_bus;
> u8 out_buf[2];
> struct i2c_msg msg = {
> @@ -178,7 +178,7 @@ static bool tfp410_init(struct intel_dvo_device *dvo,
> return false;
>
> dvo->i2c_bus = adapter;
> - dvo->dev_priv = tfp;
> + dvo->i915 = tfp;
> tfp->quiet = true;
>
> if ((id = tfp410_getid(dvo, TFP410_VID_LO)) != TFP410_VID) {
> @@ -299,11 +299,11 @@ static void tfp410_dump_regs(struct intel_dvo_device *dvo)
>
> static void tfp410_destroy(struct intel_dvo_device *dvo)
> {
> - struct tfp410_priv *tfp = dvo->dev_priv;
> + struct tfp410_priv *tfp = dvo->i915;
>
> if (tfp) {
> kfree(tfp);
> - dvo->dev_priv = NULL;
> + dvo->i915 = NULL;
> }
> }
>
> diff --git a/drivers/gpu/drm/i915/gem/i915_gem_context.c b/drivers/gpu/drm/i915/gem/i915_gem_context.c
> index c86ca9f21532..2fa6747c7b6e 100644
> --- a/drivers/gpu/drm/i915/gem/i915_gem_context.c
> +++ b/drivers/gpu/drm/i915/gem/i915_gem_context.c
> @@ -495,27 +495,27 @@ static void __assign_ppgtt(struct i915_gem_context *ctx,
> }
>
> static struct i915_gem_context *
> -i915_gem_create_context(struct drm_i915_private *dev_priv, unsigned int flags)
> +i915_gem_create_context(struct drm_i915_private *i915, unsigned int flags)
> {
> struct i915_gem_context *ctx;
>
> - lockdep_assert_held(&dev_priv->drm.struct_mutex);
> + lockdep_assert_held(&i915->drm.struct_mutex);
>
> if (flags & I915_CONTEXT_CREATE_FLAGS_SINGLE_TIMELINE &&
> - !HAS_EXECLISTS(dev_priv))
> + !HAS_EXECLISTS(i915))
> return ERR_PTR(-EINVAL);
>
> /* Reap the most stale context */
> - contexts_free_first(dev_priv);
> + contexts_free_first(i915);
>
> - ctx = __create_context(dev_priv);
> + ctx = __create_context(i915);
> if (IS_ERR(ctx))
> return ctx;
>
> - if (HAS_FULL_PPGTT(dev_priv)) {
> + if (HAS_FULL_PPGTT(i915)) {
> struct i915_ppgtt *ppgtt;
>
> - ppgtt = i915_ppgtt_create(dev_priv);
> + ppgtt = i915_ppgtt_create(i915);
> if (IS_ERR(ppgtt)) {
> DRM_DEBUG_DRIVER("PPGTT setup failed (%ld)\n",
> PTR_ERR(ppgtt));
> @@ -530,7 +530,7 @@ i915_gem_create_context(struct drm_i915_private *dev_priv, unsigned int flags)
> if (flags & I915_CONTEXT_CREATE_FLAGS_SINGLE_TIMELINE) {
> struct i915_timeline *timeline;
>
> - timeline = i915_timeline_create(dev_priv, NULL);
> + timeline = i915_timeline_create(i915, NULL);
> if (IS_ERR(timeline)) {
> context_close(ctx);
> return ERR_CAST(timeline);
> @@ -649,19 +649,19 @@ static bool needs_preempt_context(struct drm_i915_private *i915)
> return HAS_EXECLISTS(i915);
> }
>
> -int i915_gem_contexts_init(struct drm_i915_private *dev_priv)
> +int i915_gem_contexts_init(struct drm_i915_private *i915)
> {
> struct i915_gem_context *ctx;
>
> /* Reassure ourselves we are only called once */
> - GEM_BUG_ON(dev_priv->kernel_context);
> - GEM_BUG_ON(dev_priv->preempt_context);
> + GEM_BUG_ON(i915->kernel_context);
> + GEM_BUG_ON(i915->preempt_context);
>
> - intel_engine_init_ctx_wa(dev_priv->engine[RCS0]);
> - init_contexts(dev_priv);
> + intel_engine_init_ctx_wa(i915->engine[RCS0]);
> + init_contexts(i915);
>
> /* lowest priority; idle task */
> - ctx = i915_gem_context_create_kernel(dev_priv, I915_PRIORITY_MIN);
> + ctx = i915_gem_context_create_kernel(i915, I915_PRIORITY_MIN);
> if (IS_ERR(ctx)) {
> DRM_ERROR("Failed to create default global context\n");
> return PTR_ERR(ctx);
> @@ -675,31 +675,31 @@ int i915_gem_contexts_init(struct drm_i915_private *dev_priv)
> */
> GEM_BUG_ON(ctx->hw_id);
> GEM_BUG_ON(!atomic_read(&ctx->hw_id_pin_count));
> - dev_priv->kernel_context = ctx;
> + i915->kernel_context = ctx;
>
> /* highest priority; preempting task */
> - if (needs_preempt_context(dev_priv)) {
> - ctx = i915_gem_context_create_kernel(dev_priv, INT_MAX);
> + if (needs_preempt_context(i915)) {
> + ctx = i915_gem_context_create_kernel(i915, INT_MAX);
> if (!IS_ERR(ctx))
> - dev_priv->preempt_context = ctx;
> + i915->preempt_context = ctx;
> else
> DRM_ERROR("Failed to create preempt context; disabling preemption\n");
> }
>
> DRM_DEBUG_DRIVER("%s context support initialized\n",
> - DRIVER_CAPS(dev_priv)->has_logical_contexts ?
> + DRIVER_CAPS(i915)->has_logical_contexts ?
> "logical" : "fake");
> return 0;
> }
>
> -void i915_gem_contexts_lost(struct drm_i915_private *dev_priv)
> +void i915_gem_contexts_lost(struct drm_i915_private *i915)
> {
> struct intel_engine_cs *engine;
> enum intel_engine_id id;
>
> - lockdep_assert_held(&dev_priv->drm.struct_mutex);
> + lockdep_assert_held(&i915->drm.struct_mutex);
>
> - for_each_engine(engine, dev_priv, id)
> + for_each_engine(engine, i915, id)
> intel_engine_lost_context(engine);
> }
>
> @@ -2368,7 +2368,7 @@ int i915_gem_context_setparam_ioctl(struct drm_device *dev, void *data,
> int i915_gem_context_reset_stats_ioctl(struct drm_device *dev,
> void *data, struct drm_file *file)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct drm_i915_reset_stats *args = data;
> struct i915_gem_context *ctx;
> int ret;
> @@ -2390,7 +2390,7 @@ int i915_gem_context_reset_stats_ioctl(struct drm_device *dev,
> */
>
> if (capable(CAP_SYS_ADMIN))
> - args->reset_count = i915_reset_count(&dev_priv->gpu_error);
> + args->reset_count = i915_reset_count(&i915->gpu_error);
> else
> args->reset_count = 0;
>
> diff --git a/drivers/gpu/drm/i915/gem/i915_gem_context.h b/drivers/gpu/drm/i915/gem/i915_gem_context.h
> index 630392c77e48..f55cb8eae87a 100644
> --- a/drivers/gpu/drm/i915/gem/i915_gem_context.h
> +++ b/drivers/gpu/drm/i915/gem/i915_gem_context.h
> @@ -133,9 +133,9 @@ static inline bool i915_gem_context_is_kernel(struct i915_gem_context *ctx)
> }
>
> /* i915_gem_context.c */
> -int __must_check i915_gem_contexts_init(struct drm_i915_private *dev_priv);
> -void i915_gem_contexts_lost(struct drm_i915_private *dev_priv);
> -void i915_gem_contexts_fini(struct drm_i915_private *dev_priv);
> +int __must_check i915_gem_contexts_init(struct drm_i915_private *i915);
> +void i915_gem_contexts_lost(struct drm_i915_private *i915);
> +void i915_gem_contexts_fini(struct drm_i915_private *i915);
>
> int i915_gem_context_open(struct drm_i915_private *i915,
> struct drm_file *file);
> diff --git a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
> index 528eea44dccf..83d64580a2f2 100644
> --- a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
> +++ b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
> @@ -2099,7 +2099,7 @@ static int eb_submit(struct i915_execbuffer *eb)
> * The engine index is returned.
> */
> static unsigned int
> -gen8_dispatch_bsd_engine(struct drm_i915_private *dev_priv,
> +gen8_dispatch_bsd_engine(struct drm_i915_private *i915,
> struct drm_file *file)
> {
> struct drm_i915_file_private *file_priv = file->driver_priv;
> @@ -2107,7 +2107,7 @@ gen8_dispatch_bsd_engine(struct drm_i915_private *dev_priv,
> /* Check whether the file_priv has already selected one ring. */
> if ((int)file_priv->bsd_engine < 0)
> file_priv->bsd_engine = atomic_fetch_xor(1,
> - &dev_priv->mm.bsd_engine_dispatch_index);
> + &i915->mm.bsd_engine_dispatch_index);
>
> return file_priv->bsd_engine;
> }
> diff --git a/drivers/gpu/drm/i915/gem/i915_gem_object.c b/drivers/gpu/drm/i915/gem/i915_gem_object.c
> index 36b76c6a0a9d..a37f0adac16e 100644
> --- a/drivers/gpu/drm/i915/gem/i915_gem_object.c
> +++ b/drivers/gpu/drm/i915/gem/i915_gem_object.c
> @@ -367,7 +367,7 @@ void
> i915_gem_object_flush_write_domain(struct drm_i915_gem_object *obj,
> unsigned int flush_domains)
> {
> - struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
> + struct drm_i915_private *i915 = to_i915(obj->base.dev);
> struct i915_vma *vma;
>
> assert_object_held(obj);
> @@ -377,7 +377,7 @@ i915_gem_object_flush_write_domain(struct drm_i915_gem_object *obj,
>
> switch (obj->write_domain) {
> case I915_GEM_DOMAIN_GTT:
> - i915_gem_flush_ggtt_writes(dev_priv);
> + i915_gem_flush_ggtt_writes(i915);
>
> intel_fb_obj_flush(obj,
> fb_write_origin(obj, I915_GEM_DOMAIN_GTT));
> diff --git a/drivers/gpu/drm/i915/gem/i915_gem_shmem.c b/drivers/gpu/drm/i915/gem/i915_gem_shmem.c
> index 19d9ecdb2894..eb4aae65f6e4 100644
> --- a/drivers/gpu/drm/i915/gem/i915_gem_shmem.c
> +++ b/drivers/gpu/drm/i915/gem/i915_gem_shmem.c
> @@ -521,7 +521,7 @@ i915_gem_object_create_shmem(struct drm_i915_private *i915, u64 size)
>
> /* Allocate a new GEM object and fill it with the supplied data */
> struct drm_i915_gem_object *
> -i915_gem_object_create_shmem_from_data(struct drm_i915_private *dev_priv,
> +i915_gem_object_create_shmem_from_data(struct drm_i915_private *i915,
> const void *data, size_t size)
> {
> struct drm_i915_gem_object *obj;
> @@ -529,7 +529,7 @@ i915_gem_object_create_shmem_from_data(struct drm_i915_private *dev_priv,
> size_t offset;
> int err;
>
> - obj = i915_gem_object_create_shmem(dev_priv, round_up(size, PAGE_SIZE));
> + obj = i915_gem_object_create_shmem(i915, round_up(size, PAGE_SIZE));
> if (IS_ERR(obj))
> return obj;
>
> diff --git a/drivers/gpu/drm/i915/gem/i915_gem_stolen.c b/drivers/gpu/drm/i915/gem/i915_gem_stolen.c
> index de1fab2058ec..24e174f069f3 100644
> --- a/drivers/gpu/drm/i915/gem/i915_gem_stolen.c
> +++ b/drivers/gpu/drm/i915/gem/i915_gem_stolen.c
> @@ -24,48 +24,48 @@
> * for is a boon.
> */
>
> -int i915_gem_stolen_insert_node_in_range(struct drm_i915_private *dev_priv,
> +int i915_gem_stolen_insert_node_in_range(struct drm_i915_private *i915,
> struct drm_mm_node *node, u64 size,
> unsigned alignment, u64 start, u64 end)
> {
> int ret;
>
> - if (!drm_mm_initialized(&dev_priv->mm.stolen))
> + if (!drm_mm_initialized(&i915->mm.stolen))
> return -ENODEV;
>
> /* WaSkipStolenMemoryFirstPage:bdw+ */
> - if (INTEL_GEN(dev_priv) >= 8 && start < 4096)
> + if (INTEL_GEN(i915) >= 8 && start < 4096)
> start = 4096;
>
> - mutex_lock(&dev_priv->mm.stolen_lock);
> - ret = drm_mm_insert_node_in_range(&dev_priv->mm.stolen, node,
> + mutex_lock(&i915->mm.stolen_lock);
> + ret = drm_mm_insert_node_in_range(&i915->mm.stolen, node,
> size, alignment, 0,
> start, end, DRM_MM_INSERT_BEST);
> - mutex_unlock(&dev_priv->mm.stolen_lock);
> + mutex_unlock(&i915->mm.stolen_lock);
>
> return ret;
> }
>
> -int i915_gem_stolen_insert_node(struct drm_i915_private *dev_priv,
> +int i915_gem_stolen_insert_node(struct drm_i915_private *i915,
> struct drm_mm_node *node, u64 size,
> unsigned alignment)
> {
> - return i915_gem_stolen_insert_node_in_range(dev_priv, node, size,
> + return i915_gem_stolen_insert_node_in_range(i915, node, size,
> alignment, 0, U64_MAX);
> }
>
> -void i915_gem_stolen_remove_node(struct drm_i915_private *dev_priv,
> +void i915_gem_stolen_remove_node(struct drm_i915_private *i915,
> struct drm_mm_node *node)
> {
> - mutex_lock(&dev_priv->mm.stolen_lock);
> + mutex_lock(&i915->mm.stolen_lock);
> drm_mm_remove_node(node);
> - mutex_unlock(&dev_priv->mm.stolen_lock);
> + mutex_unlock(&i915->mm.stolen_lock);
> }
>
> -static int i915_adjust_stolen(struct drm_i915_private *dev_priv,
> +static int i915_adjust_stolen(struct drm_i915_private *i915,
> struct resource *dsm)
> {
> - struct i915_ggtt *ggtt = &dev_priv->ggtt;
> + struct i915_ggtt *ggtt = &i915->ggtt;
> struct resource *r;
>
> if (dsm->start == 0 || dsm->end <= dsm->start)
> @@ -77,14 +77,14 @@ static int i915_adjust_stolen(struct drm_i915_private *dev_priv,
> */
>
> /* Make sure we don't clobber the GTT if it's within stolen memory */
> - if (INTEL_GEN(dev_priv) <= 4 &&
> - !IS_G33(dev_priv) && !IS_PINEVIEW(dev_priv) && !IS_G4X(dev_priv)) {
> + if (INTEL_GEN(i915) <= 4 &&
> + !IS_G33(i915) && !IS_PINEVIEW(i915) && !IS_G4X(i915)) {
> struct resource stolen[2] = {*dsm, *dsm};
> struct resource ggtt_res;
> resource_size_t ggtt_start;
>
> ggtt_start = I915_READ(PGTBL_CTL);
> - if (IS_GEN(dev_priv, 4))
> + if (IS_GEN(i915, 4))
> ggtt_start = (ggtt_start & PGTBL_ADDRESS_LO_MASK) |
> (ggtt_start & PGTBL_ADDRESS_HI_MASK) << 28;
> else
> @@ -118,7 +118,7 @@ static int i915_adjust_stolen(struct drm_i915_private *dev_priv,
> * kernel. So if the region is already marked as busy, something
> * is seriously wrong.
> */
> - r = devm_request_mem_region(dev_priv->drm.dev, dsm->start,
> + r = devm_request_mem_region(i915->drm.dev, dsm->start,
> resource_size(dsm),
> "Graphics Stolen Memory");
> if (r == NULL) {
> @@ -131,14 +131,14 @@ static int i915_adjust_stolen(struct drm_i915_private *dev_priv,
> * reservation starting from 1 instead of 0.
> * There's also BIOS with off-by-one on the other end.
> */
> - r = devm_request_mem_region(dev_priv->drm.dev, dsm->start + 1,
> + r = devm_request_mem_region(i915->drm.dev, dsm->start + 1,
> resource_size(dsm) - 2,
> "Graphics Stolen Memory");
> /*
> * GEN3 firmware likes to smash pci bridges into the stolen
> * range. Apparently this works.
> */
> - if (r == NULL && !IS_GEN(dev_priv, 3)) {
> + if (r == NULL && !IS_GEN(i915, 3)) {
> DRM_ERROR("conflict detected with stolen region: %pR\n",
> dsm);
>
> @@ -149,25 +149,25 @@ static int i915_adjust_stolen(struct drm_i915_private *dev_priv,
> return 0;
> }
>
> -void i915_gem_cleanup_stolen(struct drm_i915_private *dev_priv)
> +void i915_gem_cleanup_stolen(struct drm_i915_private *i915)
> {
> - if (!drm_mm_initialized(&dev_priv->mm.stolen))
> + if (!drm_mm_initialized(&i915->mm.stolen))
> return;
>
> - drm_mm_takedown(&dev_priv->mm.stolen);
> + drm_mm_takedown(&i915->mm.stolen);
> }
>
> -static void g4x_get_stolen_reserved(struct drm_i915_private *dev_priv,
> +static void g4x_get_stolen_reserved(struct drm_i915_private *i915,
> resource_size_t *base,
> resource_size_t *size)
> {
> - u32 reg_val = I915_READ(IS_GM45(dev_priv) ?
> + u32 reg_val = I915_READ(IS_GM45(i915) ?
> CTG_STOLEN_RESERVED :
> ELK_STOLEN_RESERVED);
> - resource_size_t stolen_top = dev_priv->dsm.end + 1;
> + resource_size_t stolen_top = i915->dsm.end + 1;
>
> DRM_DEBUG_DRIVER("%s_STOLEN_RESERVED = %08x\n",
> - IS_GM45(dev_priv) ? "CTG" : "ELK", reg_val);
> + IS_GM45(i915) ? "CTG" : "ELK", reg_val);
>
> if ((reg_val & G4X_STOLEN_RESERVED_ENABLE) == 0)
> return;
> @@ -176,7 +176,7 @@ static void g4x_get_stolen_reserved(struct drm_i915_private *dev_priv,
> * Whether ILK really reuses the ELK register for this is unclear.
> * Let's see if we catch anyone with this supposedly enabled on ILK.
> */
> - WARN(IS_GEN(dev_priv, 5), "ILK stolen reserved found? 0x%08x\n",
> + WARN(IS_GEN(i915, 5), "ILK stolen reserved found? 0x%08x\n",
> reg_val);
>
> if (!(reg_val & G4X_STOLEN_RESERVED_ADDR2_MASK))
> @@ -188,7 +188,7 @@ static void g4x_get_stolen_reserved(struct drm_i915_private *dev_priv,
> *size = stolen_top - *base;
> }
>
> -static void gen6_get_stolen_reserved(struct drm_i915_private *dev_priv,
> +static void gen6_get_stolen_reserved(struct drm_i915_private *i915,
> resource_size_t *base,
> resource_size_t *size)
> {
> @@ -220,12 +220,12 @@ static void gen6_get_stolen_reserved(struct drm_i915_private *dev_priv,
> }
> }
>
> -static void vlv_get_stolen_reserved(struct drm_i915_private *dev_priv,
> +static void vlv_get_stolen_reserved(struct drm_i915_private *i915,
> resource_size_t *base,
> resource_size_t *size)
> {
> u32 reg_val = I915_READ(GEN6_STOLEN_RESERVED);
> - resource_size_t stolen_top = dev_priv->dsm.end + 1;
> + resource_size_t stolen_top = i915->dsm.end + 1;
>
> DRM_DEBUG_DRIVER("GEN6_STOLEN_RESERVED = %08x\n", reg_val);
>
> @@ -248,7 +248,7 @@ static void vlv_get_stolen_reserved(struct drm_i915_private *dev_priv,
> *base = stolen_top - *size;
> }
>
> -static void gen7_get_stolen_reserved(struct drm_i915_private *dev_priv,
> +static void gen7_get_stolen_reserved(struct drm_i915_private *i915,
> resource_size_t *base,
> resource_size_t *size)
> {
> @@ -274,7 +274,7 @@ static void gen7_get_stolen_reserved(struct drm_i915_private *dev_priv,
> }
> }
>
> -static void chv_get_stolen_reserved(struct drm_i915_private *dev_priv,
> +static void chv_get_stolen_reserved(struct drm_i915_private *i915,
> resource_size_t *base,
> resource_size_t *size)
> {
> @@ -306,12 +306,12 @@ static void chv_get_stolen_reserved(struct drm_i915_private *dev_priv,
> }
> }
>
> -static void bdw_get_stolen_reserved(struct drm_i915_private *dev_priv,
> +static void bdw_get_stolen_reserved(struct drm_i915_private *i915,
> resource_size_t *base,
> resource_size_t *size)
> {
> u32 reg_val = I915_READ(GEN6_STOLEN_RESERVED);
> - resource_size_t stolen_top = dev_priv->dsm.end + 1;
> + resource_size_t stolen_top = i915->dsm.end + 1;
>
> DRM_DEBUG_DRIVER("GEN6_STOLEN_RESERVED = %08x\n", reg_val);
>
> @@ -354,19 +354,19 @@ static void icl_get_stolen_reserved(struct drm_i915_private *i915,
> }
> }
>
> -int i915_gem_init_stolen(struct drm_i915_private *dev_priv)
> +int i915_gem_init_stolen(struct drm_i915_private *i915)
> {
> resource_size_t reserved_base, stolen_top;
> resource_size_t reserved_total, reserved_size;
>
> - mutex_init(&dev_priv->mm.stolen_lock);
> + mutex_init(&i915->mm.stolen_lock);
>
> - if (intel_vgpu_active(dev_priv)) {
> + if (intel_vgpu_active(i915)) {
> DRM_INFO("iGVT-g active, disabling use of stolen memory\n");
> return 0;
> }
>
> - if (intel_vtd_active() && INTEL_GEN(dev_priv) < 8) {
> + if (intel_vtd_active() && INTEL_GEN(i915) < 8) {
> DRM_INFO("DMAR active, disabling use of stolen memory\n");
> return 0;
> }
> @@ -374,55 +374,55 @@ int i915_gem_init_stolen(struct drm_i915_private *dev_priv)
> if (resource_size(&intel_graphics_stolen_res) == 0)
> return 0;
>
> - dev_priv->dsm = intel_graphics_stolen_res;
> + i915->dsm = intel_graphics_stolen_res;
>
> - if (i915_adjust_stolen(dev_priv, &dev_priv->dsm))
> + if (i915_adjust_stolen(i915, &i915->dsm))
> return 0;
>
> - GEM_BUG_ON(dev_priv->dsm.start == 0);
> - GEM_BUG_ON(dev_priv->dsm.end <= dev_priv->dsm.start);
> + GEM_BUG_ON(i915->dsm.start == 0);
> + GEM_BUG_ON(i915->dsm.end <= i915->dsm.start);
>
> - stolen_top = dev_priv->dsm.end + 1;
> + stolen_top = i915->dsm.end + 1;
> reserved_base = stolen_top;
> reserved_size = 0;
>
> - switch (INTEL_GEN(dev_priv)) {
> + switch (INTEL_GEN(i915)) {
> case 2:
> case 3:
> break;
> case 4:
> - if (!IS_G4X(dev_priv))
> + if (!IS_G4X(i915))
> break;
> /* fall through */
> case 5:
> - g4x_get_stolen_reserved(dev_priv,
> + g4x_get_stolen_reserved(i915,
> &reserved_base, &reserved_size);
> break;
> case 6:
> - gen6_get_stolen_reserved(dev_priv,
> + gen6_get_stolen_reserved(i915,
> &reserved_base, &reserved_size);
> break;
> case 7:
> - if (IS_VALLEYVIEW(dev_priv))
> - vlv_get_stolen_reserved(dev_priv,
> + if (IS_VALLEYVIEW(i915))
> + vlv_get_stolen_reserved(i915,
> &reserved_base, &reserved_size);
> else
> - gen7_get_stolen_reserved(dev_priv,
> + gen7_get_stolen_reserved(i915,
> &reserved_base, &reserved_size);
> break;
> case 8:
> case 9:
> case 10:
> - if (IS_LP(dev_priv))
> - chv_get_stolen_reserved(dev_priv,
> + if (IS_LP(i915))
> + chv_get_stolen_reserved(i915,
> &reserved_base, &reserved_size);
> else
> - bdw_get_stolen_reserved(dev_priv,
> + bdw_get_stolen_reserved(i915,
> &reserved_base, &reserved_size);
> break;
> case 11:
> default:
> - icl_get_stolen_reserved(dev_priv, &reserved_base,
> + icl_get_stolen_reserved(i915, &reserved_base,
> &reserved_size);
> break;
> }
> @@ -439,12 +439,12 @@ int i915_gem_init_stolen(struct drm_i915_private *dev_priv)
> reserved_size = 0;
> }
>
> - dev_priv->dsm_reserved =
> + i915->dsm_reserved =
> (struct resource) DEFINE_RES_MEM(reserved_base, reserved_size);
>
> - if (!resource_contains(&dev_priv->dsm, &dev_priv->dsm_reserved)) {
> + if (!resource_contains(&i915->dsm, &i915->dsm_reserved)) {
> DRM_ERROR("Stolen reserved area %pR outside stolen memory %pR\n",
> - &dev_priv->dsm_reserved, &dev_priv->dsm);
> + &i915->dsm_reserved, &i915->dsm);
> return 0;
> }
>
> @@ -453,14 +453,14 @@ int i915_gem_init_stolen(struct drm_i915_private *dev_priv)
> reserved_total = stolen_top - reserved_base;
>
> DRM_DEBUG_DRIVER("Memory reserved for graphics device: %lluK, usable: %lluK\n",
> - (u64)resource_size(&dev_priv->dsm) >> 10,
> - ((u64)resource_size(&dev_priv->dsm) - reserved_total) >> 10);
> + (u64)resource_size(&i915->dsm) >> 10,
> + ((u64)resource_size(&i915->dsm) - reserved_total) >> 10);
>
> - dev_priv->stolen_usable_size =
> - resource_size(&dev_priv->dsm) - reserved_total;
> + i915->stolen_usable_size =
> + resource_size(&i915->dsm) - reserved_total;
>
> /* Basic memrange allocator for stolen space. */
> - drm_mm_init(&dev_priv->mm.stolen, 0, dev_priv->stolen_usable_size);
> + drm_mm_init(&i915->mm.stolen, 0, i915->stolen_usable_size);
>
> return 0;
> }
> @@ -469,11 +469,11 @@ static struct sg_table *
> i915_pages_create_for_stolen(struct drm_device *dev,
> resource_size_t offset, resource_size_t size)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct sg_table *st;
> struct scatterlist *sg;
>
> - GEM_BUG_ON(range_overflows(offset, size, resource_size(&dev_priv->dsm)));
> + GEM_BUG_ON(range_overflows(offset, size, resource_size(&i915->dsm)));
>
> /* We hide that we have no struct page backing our stolen object
> * by wrapping the contiguous physical allocation with a fake
> @@ -493,7 +493,7 @@ i915_pages_create_for_stolen(struct drm_device *dev,
> sg->offset = 0;
> sg->length = size;
>
> - sg_dma_address(sg) = (dma_addr_t)dev_priv->dsm.start + offset;
> + sg_dma_address(sg) = (dma_addr_t)i915->dsm.start + offset;
> sg_dma_len(sg) = size;
>
> return st;
> @@ -524,14 +524,14 @@ static void i915_gem_object_put_pages_stolen(struct drm_i915_gem_object *obj,
> static void
> i915_gem_object_release_stolen(struct drm_i915_gem_object *obj)
> {
> - struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
> + struct drm_i915_private *i915 = to_i915(obj->base.dev);
> struct drm_mm_node *stolen = fetch_and_zero(&obj->stolen);
>
> GEM_BUG_ON(!stolen);
>
> __i915_gem_object_unpin_pages(obj);
>
> - i915_gem_stolen_remove_node(dev_priv, stolen);
> + i915_gem_stolen_remove_node(i915, stolen);
> kfree(stolen);
> }
>
> @@ -542,7 +542,7 @@ static const struct drm_i915_gem_object_ops i915_gem_object_stolen_ops = {
> };
>
> static struct drm_i915_gem_object *
> -_i915_gem_object_create_stolen(struct drm_i915_private *dev_priv,
> +_i915_gem_object_create_stolen(struct drm_i915_private *i915,
> struct drm_mm_node *stolen)
> {
> struct drm_i915_gem_object *obj;
> @@ -552,12 +552,12 @@ _i915_gem_object_create_stolen(struct drm_i915_private *dev_priv,
> if (obj == NULL)
> return NULL;
>
> - drm_gem_private_object_init(&dev_priv->drm, &obj->base, stolen->size);
> + drm_gem_private_object_init(&i915->drm, &obj->base, stolen->size);
> i915_gem_object_init(obj, &i915_gem_object_stolen_ops);
>
> obj->stolen = stolen;
> obj->read_domains = I915_GEM_DOMAIN_CPU | I915_GEM_DOMAIN_GTT;
> - cache_level = HAS_LLC(dev_priv) ? I915_CACHE_LLC : I915_CACHE_NONE;
> + cache_level = HAS_LLC(i915) ? I915_CACHE_LLC : I915_CACHE_NONE;
> i915_gem_object_set_cache_coherency(obj, cache_level);
>
> if (i915_gem_object_pin_pages(obj))
> @@ -571,14 +571,14 @@ _i915_gem_object_create_stolen(struct drm_i915_private *dev_priv,
> }
>
> struct drm_i915_gem_object *
> -i915_gem_object_create_stolen(struct drm_i915_private *dev_priv,
> +i915_gem_object_create_stolen(struct drm_i915_private *i915,
> resource_size_t size)
> {
> struct drm_i915_gem_object *obj;
> struct drm_mm_node *stolen;
> int ret;
>
> - if (!drm_mm_initialized(&dev_priv->mm.stolen))
> + if (!drm_mm_initialized(&i915->mm.stolen))
> return NULL;
>
> if (size == 0)
> @@ -588,37 +588,37 @@ i915_gem_object_create_stolen(struct drm_i915_private *dev_priv,
> if (!stolen)
> return NULL;
>
> - ret = i915_gem_stolen_insert_node(dev_priv, stolen, size, 4096);
> + ret = i915_gem_stolen_insert_node(i915, stolen, size, 4096);
> if (ret) {
> kfree(stolen);
> return NULL;
> }
>
> - obj = _i915_gem_object_create_stolen(dev_priv, stolen);
> + obj = _i915_gem_object_create_stolen(i915, stolen);
> if (obj)
> return obj;
>
> - i915_gem_stolen_remove_node(dev_priv, stolen);
> + i915_gem_stolen_remove_node(i915, stolen);
> kfree(stolen);
> return NULL;
> }
>
> struct drm_i915_gem_object *
> -i915_gem_object_create_stolen_for_preallocated(struct drm_i915_private *dev_priv,
> +i915_gem_object_create_stolen_for_preallocated(struct drm_i915_private *i915,
> resource_size_t stolen_offset,
> resource_size_t gtt_offset,
> resource_size_t size)
> {
> - struct i915_ggtt *ggtt = &dev_priv->ggtt;
> + struct i915_ggtt *ggtt = &i915->ggtt;
> struct drm_i915_gem_object *obj;
> struct drm_mm_node *stolen;
> struct i915_vma *vma;
> int ret;
>
> - if (!drm_mm_initialized(&dev_priv->mm.stolen))
> + if (!drm_mm_initialized(&i915->mm.stolen))
> return NULL;
>
> - lockdep_assert_held(&dev_priv->drm.struct_mutex);
> + lockdep_assert_held(&i915->drm.struct_mutex);
>
> DRM_DEBUG_DRIVER("creating preallocated stolen object: stolen_offset=%pa, gtt_offset=%pa, size=%pa\n",
> &stolen_offset, >t_offset, &size);
> @@ -635,19 +635,19 @@ i915_gem_object_create_stolen_for_preallocated(struct drm_i915_private *dev_priv
>
> stolen->start = stolen_offset;
> stolen->size = size;
> - mutex_lock(&dev_priv->mm.stolen_lock);
> - ret = drm_mm_reserve_node(&dev_priv->mm.stolen, stolen);
> - mutex_unlock(&dev_priv->mm.stolen_lock);
> + mutex_lock(&i915->mm.stolen_lock);
> + ret = drm_mm_reserve_node(&i915->mm.stolen, stolen);
> + mutex_unlock(&i915->mm.stolen_lock);
> if (ret) {
> DRM_DEBUG_DRIVER("failed to allocate stolen space\n");
> kfree(stolen);
> return NULL;
> }
>
> - obj = _i915_gem_object_create_stolen(dev_priv, stolen);
> + obj = _i915_gem_object_create_stolen(i915, stolen);
> if (obj == NULL) {
> DRM_DEBUG_DRIVER("failed to allocate stolen object\n");
> - i915_gem_stolen_remove_node(dev_priv, stolen);
> + i915_gem_stolen_remove_node(i915, stolen);
> kfree(stolen);
> return NULL;
> }
> diff --git a/drivers/gpu/drm/i915/gem/i915_gem_tiling.c b/drivers/gpu/drm/i915/gem/i915_gem_tiling.c
> index ca0c2f451742..46b523d4ed84 100644
> --- a/drivers/gpu/drm/i915/gem/i915_gem_tiling.c
> +++ b/drivers/gpu/drm/i915/gem/i915_gem_tiling.c
> @@ -398,7 +398,7 @@ i915_gem_get_tiling_ioctl(struct drm_device *dev, void *data,
> struct drm_file *file)
> {
> struct drm_i915_gem_get_tiling *args = data;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct drm_i915_gem_object *obj;
> int err = -ENOENT;
>
> @@ -415,10 +415,10 @@ i915_gem_get_tiling_ioctl(struct drm_device *dev, void *data,
>
> switch (args->tiling_mode) {
> case I915_TILING_X:
> - args->swizzle_mode = dev_priv->mm.bit_6_swizzle_x;
> + args->swizzle_mode = i915->mm.bit_6_swizzle_x;
> break;
> case I915_TILING_Y:
> - args->swizzle_mode = dev_priv->mm.bit_6_swizzle_y;
> + args->swizzle_mode = i915->mm.bit_6_swizzle_y;
> break;
> default:
> case I915_TILING_NONE:
> @@ -427,7 +427,7 @@ i915_gem_get_tiling_ioctl(struct drm_device *dev, void *data,
> }
>
> /* Hide bit 17 from the user -- see comment in i915_gem_set_tiling */
> - if (dev_priv->quirks & QUIRK_PIN_SWIZZLED_PAGES)
> + if (i915->quirks & QUIRK_PIN_SWIZZLED_PAGES)
> args->phys_swizzle_mode = I915_BIT_6_SWIZZLE_UNKNOWN;
> else
> args->phys_swizzle_mode = args->swizzle_mode;
> diff --git a/drivers/gpu/drm/i915/gem/i915_gem_userptr.c b/drivers/gpu/drm/i915/gem/i915_gem_userptr.c
> index 528b61678334..f22e10e8895a 100644
> --- a/drivers/gpu/drm/i915/gem/i915_gem_userptr.c
> +++ b/drivers/gpu/drm/i915/gem/i915_gem_userptr.c
> @@ -327,12 +327,12 @@ i915_mmu_notifier_free(struct i915_mmu_notifier *mn,
> #endif
>
> static struct i915_mm_struct *
> -__i915_mm_struct_find(struct drm_i915_private *dev_priv, struct mm_struct *real)
> +__i915_mm_struct_find(struct drm_i915_private *i915, struct mm_struct *real)
> {
> struct i915_mm_struct *mm;
>
> - /* Protected by dev_priv->mm_lock */
> - hash_for_each_possible(dev_priv->mm_structs, mm, node, (unsigned long)real)
> + /* Protected by i915->mm_lock */
> + hash_for_each_possible(i915->mm_structs, mm, node, (unsigned long)real)
> if (mm->mm == real)
> return mm;
>
> @@ -342,7 +342,7 @@ __i915_mm_struct_find(struct drm_i915_private *dev_priv, struct mm_struct *real)
> static int
> i915_gem_userptr_init__mm_struct(struct drm_i915_gem_object *obj)
> {
> - struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
> + struct drm_i915_private *i915 = to_i915(obj->base.dev);
> struct i915_mm_struct *mm;
> int ret = 0;
>
> @@ -356,8 +356,8 @@ i915_gem_userptr_init__mm_struct(struct drm_i915_gem_object *obj)
> * struct_mutex, i.e. we need to schedule a worker to do the clean
> * up.
> */
> - mutex_lock(&dev_priv->mm_lock);
> - mm = __i915_mm_struct_find(dev_priv, current->mm);
> + mutex_lock(&i915->mm_lock);
> + mm = __i915_mm_struct_find(i915, current->mm);
> if (mm == NULL) {
> mm = kmalloc(sizeof(*mm), GFP_KERNEL);
> if (mm == NULL) {
> @@ -373,15 +373,15 @@ i915_gem_userptr_init__mm_struct(struct drm_i915_gem_object *obj)
>
> mm->mn = NULL;
>
> - /* Protected by dev_priv->mm_lock */
> - hash_add(dev_priv->mm_structs,
> + /* Protected by i915->mm_lock */
> + hash_add(i915->mm_structs,
> &mm->node, (unsigned long)mm->mm);
> } else
> kref_get(&mm->kref);
>
> obj->userptr.mm = mm;
> out:
> - mutex_unlock(&dev_priv->mm_lock);
> + mutex_unlock(&i915->mm_lock);
> return ret;
> }
>
> @@ -399,7 +399,7 @@ __i915_mm_struct_free(struct kref *kref)
> {
> struct i915_mm_struct *mm = container_of(kref, typeof(*mm), kref);
>
> - /* Protected by dev_priv->mm_lock */
> + /* Protected by i915->mm_lock */
> hash_del(&mm->node);
> mutex_unlock(&mm->i915->mm_lock);
>
> @@ -741,13 +741,13 @@ i915_gem_userptr_ioctl(struct drm_device *dev,
> void *data,
> struct drm_file *file)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct drm_i915_gem_userptr *args = data;
> struct drm_i915_gem_object *obj;
> int ret;
> u32 handle;
>
> - if (!HAS_LLC(dev_priv) && !HAS_SNOOP(dev_priv)) {
> + if (!HAS_LLC(i915) && !HAS_SNOOP(i915)) {
> /* We cannot support coherent userptr objects on hw without
> * LLC and broken snooping.
> */
> @@ -774,7 +774,7 @@ i915_gem_userptr_ioctl(struct drm_device *dev,
> * On almost all of the older hw, we cannot tell the GPU that
> * a page is readonly.
> */
> - vm = dev_priv->kernel_context->vm;
> + vm = i915->kernel_context->vm;
> if (!vm || !vm->has_read_only)
> return -ENODEV;
> }
> @@ -812,22 +812,22 @@ i915_gem_userptr_ioctl(struct drm_device *dev,
> return 0;
> }
>
> -int i915_gem_init_userptr(struct drm_i915_private *dev_priv)
> +int i915_gem_init_userptr(struct drm_i915_private *i915)
> {
> - mutex_init(&dev_priv->mm_lock);
> - hash_init(dev_priv->mm_structs);
> + mutex_init(&i915->mm_lock);
> + hash_init(i915->mm_structs);
>
> - dev_priv->mm.userptr_wq =
> + i915->mm.userptr_wq =
> alloc_workqueue("i915-userptr-acquire",
> WQ_HIGHPRI | WQ_UNBOUND,
> 0);
> - if (!dev_priv->mm.userptr_wq)
> + if (!i915->mm.userptr_wq)
> return -ENOMEM;
>
> return 0;
> }
>
> -void i915_gem_cleanup_userptr(struct drm_i915_private *dev_priv)
> +void i915_gem_cleanup_userptr(struct drm_i915_private *i915)
> {
> - destroy_workqueue(dev_priv->mm.userptr_wq);
> + destroy_workqueue(i915->mm.userptr_wq);
> }
> diff --git a/drivers/gpu/drm/i915/gem/selftests/huge_pages.c b/drivers/gpu/drm/i915/gem/selftests/huge_pages.c
> index 73e667b31cc4..78797699feb0 100644
> --- a/drivers/gpu/drm/i915/gem/selftests/huge_pages.c
> +++ b/drivers/gpu/drm/i915/gem/selftests/huge_pages.c
> @@ -1415,8 +1415,8 @@ static int igt_ppgtt_gemfs_huge(void *arg)
> static int igt_ppgtt_pin_update(void *arg)
> {
> struct i915_gem_context *ctx = arg;
> - struct drm_i915_private *dev_priv = ctx->i915;
> - unsigned long supported = INTEL_INFO(dev_priv)->page_sizes;
> + struct drm_i915_private *i915 = ctx->i915;
> + unsigned long supported = INTEL_INFO(i915)->page_sizes;
> struct i915_address_space *vm = ctx->vm;
> struct drm_i915_gem_object *obj;
> struct i915_vma *vma;
> @@ -1443,7 +1443,7 @@ static int igt_ppgtt_pin_update(void *arg)
> for_each_set_bit_from(first, &supported, last + 1) {
> unsigned int page_size = BIT(first);
>
> - obj = i915_gem_object_create_internal(dev_priv, page_size);
> + obj = i915_gem_object_create_internal(i915, page_size);
> if (IS_ERR(obj))
> return PTR_ERR(obj);
>
> @@ -1497,7 +1497,7 @@ static int igt_ppgtt_pin_update(void *arg)
> i915_gem_object_put(obj);
> }
>
> - obj = i915_gem_object_create_internal(dev_priv, PAGE_SIZE);
> + obj = i915_gem_object_create_internal(i915, PAGE_SIZE);
> if (IS_ERR(obj))
> return PTR_ERR(obj);
>
> @@ -1518,7 +1518,7 @@ static int igt_ppgtt_pin_update(void *arg)
> * land in the now stale 2M page.
> */
>
> - err = gpu_write(vma, ctx, dev_priv->engine[RCS0], 0, 0xdeadbeaf);
> + err = gpu_write(vma, ctx, i915->engine[RCS0], 0, 0xdeadbeaf);
> if (err)
> goto out_unpin;
>
> @@ -1682,20 +1682,20 @@ int i915_gem_huge_page_mock_selftests(void)
> SUBTEST(igt_mock_ppgtt_huge_fill),
> SUBTEST(igt_mock_ppgtt_64K),
> };
> - struct drm_i915_private *dev_priv;
> + struct drm_i915_private *i915;
> struct i915_ppgtt *ppgtt;
> int err;
>
> - dev_priv = mock_gem_device();
> - if (!dev_priv)
> + i915 = mock_gem_device();
> + if (!i915)
> return -ENOMEM;
>
> /* Pretend to be a device which supports the 48b PPGTT */
> - mkwrite_device_info(dev_priv)->ppgtt_type = INTEL_PPGTT_FULL;
> - mkwrite_device_info(dev_priv)->ppgtt_size = 48;
> + mkwrite_device_info(i915)->ppgtt_type = INTEL_PPGTT_FULL;
> + mkwrite_device_info(i915)->ppgtt_size = 48;
>
> - mutex_lock(&dev_priv->drm.struct_mutex);
> - ppgtt = i915_ppgtt_create(dev_priv);
> + mutex_lock(&i915->drm.struct_mutex);
> + ppgtt = i915_ppgtt_create(i915);
> if (IS_ERR(ppgtt)) {
> err = PTR_ERR(ppgtt);
> goto out_unlock;
> @@ -1720,13 +1720,13 @@ int i915_gem_huge_page_mock_selftests(void)
> i915_vm_put(&ppgtt->vm);
>
> out_unlock:
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> - drm_dev_put(&dev_priv->drm);
> + mutex_unlock(&i915->drm.struct_mutex);
> + drm_dev_put(&i915->drm);
>
> return err;
> }
>
> -int i915_gem_huge_page_live_selftests(struct drm_i915_private *dev_priv)
> +int i915_gem_huge_page_live_selftests(struct drm_i915_private *i915)
> {
> static const struct i915_subtest tests[] = {
> SUBTEST(igt_shrink_thp),
> @@ -1741,22 +1741,22 @@ int i915_gem_huge_page_live_selftests(struct drm_i915_private *dev_priv)
> intel_wakeref_t wakeref;
> int err;
>
> - if (!HAS_PPGTT(dev_priv)) {
> + if (!HAS_PPGTT(i915)) {
> pr_info("PPGTT not supported, skipping live-selftests\n");
> return 0;
> }
>
> - if (i915_terminally_wedged(dev_priv))
> + if (i915_terminally_wedged(i915))
> return 0;
>
> - file = mock_file(dev_priv);
> + file = mock_file(i915);
> if (IS_ERR(file))
> return PTR_ERR(file);
>
> - mutex_lock(&dev_priv->drm.struct_mutex);
> - wakeref = intel_runtime_pm_get(dev_priv);
> + mutex_lock(&i915->drm.struct_mutex);
> + wakeref = intel_runtime_pm_get(i915);
>
> - ctx = live_context(dev_priv, file);
> + ctx = live_context(i915, file);
> if (IS_ERR(ctx)) {
> err = PTR_ERR(ctx);
> goto out_unlock;
> @@ -1768,10 +1768,10 @@ int i915_gem_huge_page_live_selftests(struct drm_i915_private *dev_priv)
> err = i915_subtests(tests, ctx);
>
> out_unlock:
> - intel_runtime_pm_put(dev_priv, wakeref);
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + intel_runtime_pm_put(i915, wakeref);
> + mutex_unlock(&i915->drm.struct_mutex);
>
> - mock_file_free(dev_priv, file);
> + mock_file_free(i915, file);
>
> return err;
> }
> diff --git a/drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c b/drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
> index 74b0e5871c4b..2fdcaa1667bf 100644
> --- a/drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
> +++ b/drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
> @@ -1748,7 +1748,7 @@ int i915_gem_context_mock_selftests(void)
> return err;
> }
>
> -int i915_gem_context_live_selftests(struct drm_i915_private *dev_priv)
> +int i915_gem_context_live_selftests(struct drm_i915_private *i915)
> {
> static const struct i915_subtest tests[] = {
> SUBTEST(live_nop_switch),
> @@ -1759,8 +1759,8 @@ int i915_gem_context_live_selftests(struct drm_i915_private *dev_priv)
> SUBTEST(igt_vm_isolation),
> };
>
> - if (i915_terminally_wedged(dev_priv))
> + if (i915_terminally_wedged(i915))
> return 0;
>
> - return i915_subtests(tests, dev_priv);
> + return i915_subtests(tests, i915);
> }
> diff --git a/drivers/gpu/drm/i915/gt/intel_engine.h b/drivers/gpu/drm/i915/gt/intel_engine.h
> index 1439fa4093ac..74f9ad6648eb 100644
> --- a/drivers/gpu/drm/i915/gt/intel_engine.h
> +++ b/drivers/gpu/drm/i915/gt/intel_engine.h
> @@ -465,7 +465,7 @@ static inline void intel_engine_reset(struct intel_engine_cs *engine,
> }
>
> bool intel_engine_is_idle(struct intel_engine_cs *engine);
> -bool intel_engines_are_idle(struct drm_i915_private *dev_priv);
> +bool intel_engines_are_idle(struct drm_i915_private *i915);
>
> void intel_engine_lost_context(struct intel_engine_cs *engine);
>
> diff --git a/drivers/gpu/drm/i915/gt/intel_engine_cs.c b/drivers/gpu/drm/i915/gt/intel_engine_cs.c
> index c0d986db5a75..a048b8743ce6 100644
> --- a/drivers/gpu/drm/i915/gt/intel_engine_cs.c
> +++ b/drivers/gpu/drm/i915/gt/intel_engine_cs.c
> @@ -160,7 +160,7 @@ static const struct engine_info intel_engines[] = {
>
> /**
> * intel_engine_context_size() - return the size of the context for an engine
> - * @dev_priv: i915 device private
> + * @i915: i915 device private
> * @class: engine class
> *
> * Each engine class may require a different amount of space for a context
> @@ -172,7 +172,7 @@ static const struct engine_info intel_engines[] = {
> * in LRC mode, but does not include the "shared data page" used with
> * GuC submission. The caller should account for this if using the GuC.
> */
> -u32 intel_engine_context_size(struct drm_i915_private *dev_priv, u8 class)
> +u32 intel_engine_context_size(struct drm_i915_private *i915, u8 class)
> {
> u32 cxt_size;
>
> @@ -180,9 +180,9 @@ u32 intel_engine_context_size(struct drm_i915_private *dev_priv, u8 class)
>
> switch (class) {
> case RENDER_CLASS:
> - switch (INTEL_GEN(dev_priv)) {
> + switch (INTEL_GEN(i915)) {
> default:
> - MISSING_CASE(INTEL_GEN(dev_priv));
> + MISSING_CASE(INTEL_GEN(i915));
> return DEFAULT_LR_CONTEXT_RENDER_SIZE;
> case 11:
> return GEN11_LR_CONTEXT_RENDER_SIZE;
> @@ -193,7 +193,7 @@ u32 intel_engine_context_size(struct drm_i915_private *dev_priv, u8 class)
> case 8:
> return GEN8_LR_CONTEXT_RENDER_SIZE;
> case 7:
> - if (IS_HASWELL(dev_priv))
> + if (IS_HASWELL(i915))
> return HSW_CXT_TOTAL_SIZE;
>
> cxt_size = I915_READ(GEN7_CXT_SIZE);
> @@ -217,7 +217,7 @@ u32 intel_engine_context_size(struct drm_i915_private *dev_priv, u8 class)
> */
> cxt_size = I915_READ(CXT_SIZE) + 1;
> DRM_DEBUG_DRIVER("gen%d CXT_SIZE = %d bytes [0x%08x]\n",
> - INTEL_GEN(dev_priv),
> + INTEL_GEN(i915),
> cxt_size * 64,
> cxt_size - 1);
> return round_up(cxt_size * 64, PAGE_SIZE);
> @@ -234,7 +234,7 @@ u32 intel_engine_context_size(struct drm_i915_private *dev_priv, u8 class)
> case VIDEO_DECODE_CLASS:
> case VIDEO_ENHANCEMENT_CLASS:
> case COPY_ENGINE_CLASS:
> - if (INTEL_GEN(dev_priv) < 8)
> + if (INTEL_GEN(i915) < 8)
> return 0;
> return GEN8_LR_CONTEXT_OTHER_SIZE;
> }
> @@ -284,7 +284,7 @@ static void intel_engine_sanitize_mmio(struct intel_engine_cs *engine)
> }
>
> static int
> -intel_engine_setup(struct drm_i915_private *dev_priv,
> +intel_engine_setup(struct drm_i915_private *i915,
> enum intel_engine_id id)
> {
> const struct engine_info *info = &intel_engines[id];
> @@ -301,10 +301,10 @@ intel_engine_setup(struct drm_i915_private *dev_priv,
> if (GEM_DEBUG_WARN_ON(info->instance > MAX_ENGINE_INSTANCE))
> return -EINVAL;
>
> - if (GEM_DEBUG_WARN_ON(dev_priv->engine_class[info->class][info->instance]))
> + if (GEM_DEBUG_WARN_ON(i915->engine_class[info->class][info->instance]))
> return -EINVAL;
>
> - GEM_BUG_ON(dev_priv->engine[id]);
> + GEM_BUG_ON(i915->engine[id]);
> engine = kzalloc(sizeof(*engine), GFP_KERNEL);
> if (!engine)
> return -ENOMEM;
> @@ -313,11 +313,11 @@ intel_engine_setup(struct drm_i915_private *dev_priv,
>
> engine->id = id;
> engine->mask = BIT(id);
> - engine->i915 = dev_priv;
> - engine->uncore = &dev_priv->uncore;
> + engine->i915 = i915;
> + engine->uncore = &i915->uncore;
> __sprint_engine_name(engine->name, info);
> engine->hw_id = engine->guc_id = info->hw_id;
> - engine->mmio_base = __engine_mmio_base(dev_priv, info->mmio_bases);
> + engine->mmio_base = __engine_mmio_base(i915, info->mmio_bases);
> engine->class = info->class;
> engine->instance = info->instance;
>
> @@ -329,12 +329,12 @@ intel_engine_setup(struct drm_i915_private *dev_priv,
>
> engine->uabi_class = intel_engine_classes[info->class].uabi_class;
>
> - engine->context_size = intel_engine_context_size(dev_priv,
> + engine->context_size = intel_engine_context_size(i915,
> engine->class);
> if (WARN_ON(engine->context_size > BIT(20)))
> engine->context_size = 0;
> if (engine->context_size)
> - DRIVER_CAPS(dev_priv)->has_logical_contexts = true;
> + DRIVER_CAPS(i915)->has_logical_contexts = true;
>
> /* Nothing to do here, execute in order of dependencies */
> engine->schedule = NULL;
> @@ -346,8 +346,8 @@ intel_engine_setup(struct drm_i915_private *dev_priv,
> /* Scrub mmio state on takeover */
> intel_engine_sanitize_mmio(engine);
>
> - dev_priv->engine_class[info->class][info->instance] = engine;
> - dev_priv->engine[id] = engine;
> + i915->engine_class[info->class][info->instance] = engine;
> + i915->engine[id] = engine;
> return 0;
> }
>
> @@ -953,17 +953,17 @@ const char *i915_cache_level_str(struct drm_i915_private *i915, int type)
> }
> }
>
> -u32 intel_calculate_mcr_s_ss_select(struct drm_i915_private *dev_priv)
> +u32 intel_calculate_mcr_s_ss_select(struct drm_i915_private *i915)
> {
> - const struct sseu_dev_info *sseu = &RUNTIME_INFO(dev_priv)->sseu;
> + const struct sseu_dev_info *sseu = &RUNTIME_INFO(i915)->sseu;
> u32 mcr_s_ss_select;
> u32 slice = fls(sseu->slice_mask);
> u32 subslice = fls(sseu->subslice_mask[slice]);
>
> - if (IS_GEN(dev_priv, 10))
> + if (IS_GEN(i915, 10))
> mcr_s_ss_select = GEN8_MCR_SLICE(slice) |
> GEN8_MCR_SUBSLICE(subslice);
> - else if (INTEL_GEN(dev_priv) >= 11)
> + else if (INTEL_GEN(i915) >= 11)
> mcr_s_ss_select = GEN11_MCR_SLICE(slice) |
> GEN11_MCR_SUBSLICE(subslice);
> else
> @@ -1095,7 +1095,7 @@ void intel_engine_get_instdone(struct intel_engine_cs *engine,
>
> static bool ring_is_idle(struct intel_engine_cs *engine)
> {
> - struct drm_i915_private *dev_priv = engine->i915;
> + struct drm_i915_private *i915 = engine->i915;
> intel_wakeref_t wakeref;
> bool idle = true;
>
> @@ -1103,7 +1103,7 @@ static bool ring_is_idle(struct intel_engine_cs *engine)
> return true;
>
> /* If the whole device is asleep, the engine must be idle */
> - wakeref = intel_runtime_pm_get_if_in_use(dev_priv);
> + wakeref = intel_runtime_pm_get_if_in_use(i915);
> if (!wakeref)
> return true;
>
> @@ -1113,11 +1113,11 @@ static bool ring_is_idle(struct intel_engine_cs *engine)
> idle = false;
>
> /* No bit for gen2, so assume the CS parser is idle */
> - if (INTEL_GEN(dev_priv) > 2 &&
> + if (INTEL_GEN(i915) > 2 &&
> !(ENGINE_READ(engine, RING_MI_MODE) & MODE_IDLE))
> idle = false;
>
> - intel_runtime_pm_put(dev_priv, wakeref);
> + intel_runtime_pm_put(i915, wakeref);
>
> return idle;
> }
> @@ -1321,12 +1321,12 @@ static void hexdump(struct drm_printer *m, const void *buf, size_t len)
> static void intel_engine_print_registers(const struct intel_engine_cs *engine,
> struct drm_printer *m)
> {
> - struct drm_i915_private *dev_priv = engine->i915;
> + struct drm_i915_private *i915 = engine->i915;
> const struct intel_engine_execlists * const execlists =
> &engine->execlists;
> u64 addr;
>
> - if (engine->id == RCS0 && IS_GEN_RANGE(dev_priv, 4, 7))
> + if (engine->id == RCS0 && IS_GEN_RANGE(i915, 4, 7))
> drm_printf(m, "\tCCID: 0x%08x\n", ENGINE_READ(engine, CCID));
> drm_printf(m, "\tRING_START: 0x%08x\n",
> ENGINE_READ(engine, RING_START));
> @@ -1343,7 +1343,7 @@ static void intel_engine_print_registers(const struct intel_engine_cs *engine,
> ENGINE_READ(engine, RING_MI_MODE) & (MODE_IDLE) ? " [idle]" : "");
> }
>
> - if (INTEL_GEN(dev_priv) >= 6) {
> + if (INTEL_GEN(i915) >= 6) {
> drm_printf(m, "\tRING_IMR: %08x\n",
> ENGINE_READ(engine, RING_IMR));
> }
> @@ -1354,15 +1354,15 @@ static void intel_engine_print_registers(const struct intel_engine_cs *engine,
> addr = intel_engine_get_last_batch_head(engine);
> drm_printf(m, "\tBBADDR: 0x%08x_%08x\n",
> upper_32_bits(addr), lower_32_bits(addr));
> - if (INTEL_GEN(dev_priv) >= 8)
> + if (INTEL_GEN(i915) >= 8)
> addr = ENGINE_READ64(engine, RING_DMA_FADD, RING_DMA_FADD_UDW);
> - else if (INTEL_GEN(dev_priv) >= 4)
> + else if (INTEL_GEN(i915) >= 4)
> addr = ENGINE_READ(engine, RING_DMA_FADD);
> else
> addr = ENGINE_READ(engine, DMA_FADD_I8XX);
> drm_printf(m, "\tDMA_FADDR: 0x%08x_%08x\n",
> upper_32_bits(addr), lower_32_bits(addr));
> - if (INTEL_GEN(dev_priv) >= 4) {
> + if (INTEL_GEN(i915) >= 4) {
> drm_printf(m, "\tIPEIR: 0x%08x\n",
> ENGINE_READ(engine, RING_IPEIR));
> drm_printf(m, "\tIPEHR: 0x%08x\n",
> @@ -1372,7 +1372,7 @@ static void intel_engine_print_registers(const struct intel_engine_cs *engine,
> drm_printf(m, "\tIPEHR: 0x%08x\n", ENGINE_READ(engine, IPEHR));
> }
>
> - if (HAS_EXECLISTS(dev_priv)) {
> + if (HAS_EXECLISTS(i915)) {
> const u32 *hws =
> &engine->status_page.addr[I915_HWS_CSB_BUF0_INDEX];
> const u8 num_entries = execlists->csb_size;
> @@ -1426,7 +1426,7 @@ static void intel_engine_print_registers(const struct intel_engine_cs *engine,
> }
> drm_printf(m, "\t\tHW active? 0x%x\n", execlists->active);
> rcu_read_unlock();
> - } else if (INTEL_GEN(dev_priv) > 6) {
> + } else if (INTEL_GEN(i915) > 6) {
> drm_printf(m, "\tPP_DIR_BASE: 0x%08x\n",
> ENGINE_READ(engine, RING_PP_DIR_BASE));
> drm_printf(m, "\tPP_DIR_BASE_READ: 0x%08x\n",
> diff --git a/drivers/gpu/drm/i915/gt/intel_hangcheck.c b/drivers/gpu/drm/i915/gt/intel_hangcheck.c
> index 174bb0a60309..948423c5f67d 100644
> --- a/drivers/gpu/drm/i915/gt/intel_hangcheck.c
> +++ b/drivers/gpu/drm/i915/gt/intel_hangcheck.c
> @@ -50,7 +50,7 @@ static bool instdone_unchanged(u32 current_instdone, u32 *old_instdone)
>
> static bool subunits_stuck(struct intel_engine_cs *engine)
> {
> - struct drm_i915_private *dev_priv = engine->i915;
> + struct drm_i915_private *i915 = engine->i915;
> struct intel_instdone instdone;
> struct intel_instdone *accu_instdone = &engine->hangcheck.instdone;
> bool stuck;
> @@ -72,7 +72,7 @@ static bool subunits_stuck(struct intel_engine_cs *engine)
> stuck &= instdone_unchanged(instdone.slice_common,
> &accu_instdone->slice_common);
>
> - for_each_instdone_slice_subslice(dev_priv, slice, subslice) {
> + for_each_instdone_slice_subslice(i915, slice, subslice) {
> stuck &= instdone_unchanged(instdone.sampler[slice][subslice],
> &accu_instdone->sampler[slice][subslice]);
> stuck &= instdone_unchanged(instdone.row[slice][subslice],
> @@ -103,7 +103,7 @@ head_stuck(struct intel_engine_cs *engine, u64 acthd)
> static enum intel_engine_hangcheck_action
> engine_stuck(struct intel_engine_cs *engine, u64 acthd)
> {
> - struct drm_i915_private *dev_priv = engine->i915;
> + struct drm_i915_private *i915 = engine->i915;
> enum intel_engine_hangcheck_action ha;
> u32 tmp;
>
> @@ -111,7 +111,7 @@ engine_stuck(struct intel_engine_cs *engine, u64 acthd)
> if (ha != ENGINE_DEAD)
> return ha;
>
> - if (IS_GEN(dev_priv, 2))
> + if (IS_GEN(i915, 2))
> return ENGINE_DEAD;
>
> /* Is the chip hanging on a WAIT_FOR_EVENT?
> @@ -121,7 +121,7 @@ engine_stuck(struct intel_engine_cs *engine, u64 acthd)
> */
> tmp = ENGINE_READ(engine, RING_CTL);
> if (tmp & RING_WAIT) {
> - i915_handle_error(dev_priv, engine->mask, 0,
> + i915_handle_error(i915, engine->mask, 0,
> "stuck wait on %s", engine->name);
> ENGINE_WRITE(engine, RING_CTL, tmp);
> return ENGINE_WAIT_KICK;
> @@ -256,8 +256,8 @@ static void hangcheck_declare_hang(struct drm_i915_private *i915,
> */
> static void i915_hangcheck_elapsed(struct work_struct *work)
> {
> - struct drm_i915_private *dev_priv =
> - container_of(work, typeof(*dev_priv),
> + struct drm_i915_private *i915 =
> + container_of(work, typeof(*i915),
> gpu_error.hangcheck_work.work);
> intel_engine_mask_t hung = 0, stuck = 0, wedged = 0;
> struct intel_engine_cs *engine;
> @@ -267,13 +267,13 @@ static void i915_hangcheck_elapsed(struct work_struct *work)
> if (!i915_modparams.enable_hangcheck)
> return;
>
> - if (!READ_ONCE(dev_priv->gt.awake))
> + if (!READ_ONCE(i915->gt.awake))
> return;
>
> - if (i915_terminally_wedged(dev_priv))
> + if (i915_terminally_wedged(i915))
> return;
>
> - wakeref = intel_runtime_pm_get_if_in_use(dev_priv);
> + wakeref = intel_runtime_pm_get_if_in_use(i915);
> if (!wakeref)
> return;
>
> @@ -281,9 +281,9 @@ static void i915_hangcheck_elapsed(struct work_struct *work)
> * periodically arm the mmio checker to see if we are triggering
> * any invalid access.
> */
> - intel_uncore_arm_unclaimed_mmio_detection(&dev_priv->uncore);
> + intel_uncore_arm_unclaimed_mmio_detection(&i915->uncore);
>
> - for_each_engine(engine, dev_priv, id) {
> + for_each_engine(engine, i915, id) {
> struct hangcheck hc;
>
> intel_engine_signal_breadcrumbs(engine);
> @@ -305,7 +305,7 @@ static void i915_hangcheck_elapsed(struct work_struct *work)
> if (GEM_SHOW_DEBUG() && (hung | stuck)) {
> struct drm_printer p = drm_debug_printer("hangcheck");
>
> - for_each_engine(engine, dev_priv, id) {
> + for_each_engine(engine, i915, id) {
> if (intel_engine_is_idle(engine))
> continue;
>
> @@ -314,20 +314,20 @@ static void i915_hangcheck_elapsed(struct work_struct *work)
> }
>
> if (wedged) {
> - dev_err(dev_priv->drm.dev,
> + dev_err(i915->drm.dev,
> "GPU recovery timed out,"
> " cancelling all in-flight rendering.\n");
> GEM_TRACE_DUMP();
> - i915_gem_set_wedged(dev_priv);
> + i915_gem_set_wedged(i915);
> }
>
> if (hung)
> - hangcheck_declare_hang(dev_priv, hung, stuck);
> + hangcheck_declare_hang(i915, hung, stuck);
>
> - intel_runtime_pm_put(dev_priv, wakeref);
> + intel_runtime_pm_put(i915, wakeref);
>
> /* Reset timer in case GPU hangs without another request being added */
> - i915_queue_hangcheck(dev_priv);
> + i915_queue_hangcheck(i915);
> }
>
> void intel_engine_init_hangcheck(struct intel_engine_cs *engine)
> diff --git a/drivers/gpu/drm/i915/gt/intel_mocs.c b/drivers/gpu/drm/i915/gt/intel_mocs.c
> index 79df66022d3a..9d6f0893f528 100644
> --- a/drivers/gpu/drm/i915/gt/intel_mocs.c
> +++ b/drivers/gpu/drm/i915/gt/intel_mocs.c
> @@ -239,7 +239,7 @@ static const struct drm_i915_mocs_entry icelake_mocs_table[] = {
>
> /**
> * get_mocs_settings()
> - * @dev_priv: i915 device.
> + * @i915: i915 device.
> * @table: Output table that will be made to point at appropriate
> * MOCS values for the device.
> *
> @@ -249,33 +249,33 @@ static const struct drm_i915_mocs_entry icelake_mocs_table[] = {
> *
> * Return: true if there are applicable MOCS settings for the device.
> */
> -static bool get_mocs_settings(struct drm_i915_private *dev_priv,
> +static bool get_mocs_settings(struct drm_i915_private *i915,
> struct drm_i915_mocs_table *table)
> {
> bool result = false;
>
> - if (INTEL_GEN(dev_priv) >= 11) {
> + if (INTEL_GEN(i915) >= 11) {
> table->size = ARRAY_SIZE(icelake_mocs_table);
> table->table = icelake_mocs_table;
> table->n_entries = GEN11_NUM_MOCS_ENTRIES;
> result = true;
> - } else if (IS_GEN9_BC(dev_priv) || IS_CANNONLAKE(dev_priv)) {
> + } else if (IS_GEN9_BC(i915) || IS_CANNONLAKE(i915)) {
> table->size = ARRAY_SIZE(skylake_mocs_table);
> table->n_entries = GEN9_NUM_MOCS_ENTRIES;
> table->table = skylake_mocs_table;
> result = true;
> - } else if (IS_GEN9_LP(dev_priv)) {
> + } else if (IS_GEN9_LP(i915)) {
> table->size = ARRAY_SIZE(broxton_mocs_table);
> table->n_entries = GEN9_NUM_MOCS_ENTRIES;
> table->table = broxton_mocs_table;
> result = true;
> } else {
> - WARN_ONCE(INTEL_GEN(dev_priv) >= 9,
> + WARN_ONCE(INTEL_GEN(i915) >= 9,
> "Platform that should have a MOCS table does not.\n");
> }
>
> /* WaDisableSkipCaching:skl,bxt,kbl,glk */
> - if (IS_GEN(dev_priv, 9)) {
> + if (IS_GEN(i915, 9)) {
> int i;
>
> for (i = 0; i < table->size; i++)
> @@ -330,12 +330,12 @@ static u32 get_entry_control(const struct drm_i915_mocs_table *table,
> */
> void intel_mocs_init_engine(struct intel_engine_cs *engine)
> {
> - struct drm_i915_private *dev_priv = engine->i915;
> + struct drm_i915_private *i915 = engine->i915;
> struct drm_i915_mocs_table table;
> unsigned int index;
> u32 unused_value;
>
> - if (!get_mocs_settings(dev_priv, &table))
> + if (!get_mocs_settings(i915, &table))
> return;
>
> /* Set unused values to PTE */
> @@ -482,7 +482,7 @@ static int emit_mocs_l3cc_table(struct i915_request *rq,
>
> /**
> * intel_mocs_init_l3cc_table() - program the mocs control table
> - * @dev_priv: i915 device private
> + * @i915: i915 device private
> *
> * This function simply programs the mocs registers for the given table
> * starting at the given address. This register set is programmed in pairs.
> @@ -494,13 +494,13 @@ static int emit_mocs_l3cc_table(struct i915_request *rq,
> *
> * Return: Nothing.
> */
> -void intel_mocs_init_l3cc_table(struct drm_i915_private *dev_priv)
> +void intel_mocs_init_l3cc_table(struct drm_i915_private *i915)
> {
> struct drm_i915_mocs_table table;
> unsigned int i;
> u16 unused_value;
>
> - if (!get_mocs_settings(dev_priv, &table))
> + if (!get_mocs_settings(i915, &table))
> return;
>
> /* Set unused values to PTE */
> diff --git a/drivers/gpu/drm/i915/gt/intel_mocs.h b/drivers/gpu/drm/i915/gt/intel_mocs.h
> index 0913704a1af2..6f8973fbcb3e 100644
> --- a/drivers/gpu/drm/i915/gt/intel_mocs.h
> +++ b/drivers/gpu/drm/i915/gt/intel_mocs.h
> @@ -54,7 +54,7 @@ struct i915_request;
> struct intel_engine_cs;
>
> int intel_rcs_context_init_mocs(struct i915_request *rq);
> -void intel_mocs_init_l3cc_table(struct drm_i915_private *dev_priv);
> +void intel_mocs_init_l3cc_table(struct drm_i915_private *i915);
> void intel_mocs_init_engine(struct intel_engine_cs *engine);
>
> #endif
> diff --git a/drivers/gpu/drm/i915/gt/intel_reset.c b/drivers/gpu/drm/i915/gt/intel_reset.c
> index 6368b37f26d1..d5cd3cccb407 100644
> --- a/drivers/gpu/drm/i915/gt/intel_reset.c
> +++ b/drivers/gpu/drm/i915/gt/intel_reset.c
> @@ -1208,13 +1208,13 @@ static void clear_error_registers(struct drm_i915_private *i915,
> }
> }
>
> -static void gen6_check_faults(struct drm_i915_private *dev_priv)
> +static void gen6_check_faults(struct drm_i915_private *i915)
> {
> struct intel_engine_cs *engine;
> enum intel_engine_id id;
> u32 fault;
>
> - for_each_engine(engine, dev_priv, id) {
> + for_each_engine(engine, i915, id) {
> fault = GEN6_RING_FAULT_REG_READ(engine);
> if (fault & RING_FAULT_VALID) {
> DRM_DEBUG_DRIVER("Unexpected fault\n"
> @@ -1230,7 +1230,7 @@ static void gen6_check_faults(struct drm_i915_private *dev_priv)
> }
> }
>
> -static void gen8_check_faults(struct drm_i915_private *dev_priv)
> +static void gen8_check_faults(struct drm_i915_private *i915)
> {
> u32 fault = I915_READ(GEN8_RING_FAULT_REG);
>
> diff --git a/drivers/gpu/drm/i915/gt/intel_ringbuffer.c b/drivers/gpu/drm/i915/gt/intel_ringbuffer.c
> index b3bf47e8162f..8ec2d332e206 100644
> --- a/drivers/gpu/drm/i915/gt/intel_ringbuffer.c
> +++ b/drivers/gpu/drm/i915/gt/intel_ringbuffer.c
> @@ -489,11 +489,11 @@ static void set_hwstam(struct intel_engine_cs *engine, u32 mask)
>
> static void set_hws_pga(struct intel_engine_cs *engine, phys_addr_t phys)
> {
> - struct drm_i915_private *dev_priv = engine->i915;
> + struct drm_i915_private *i915 = engine->i915;
> u32 addr;
>
> addr = lower_32_bits(phys);
> - if (INTEL_GEN(dev_priv) >= 4)
> + if (INTEL_GEN(i915) >= 4)
> addr |= (phys >> 28) & 0xf0;
>
> I915_WRITE(HWS_PGA, addr);
> @@ -515,14 +515,14 @@ static void ring_setup_phys_status_page(struct intel_engine_cs *engine)
>
> static void set_hwsp(struct intel_engine_cs *engine, u32 offset)
> {
> - struct drm_i915_private *dev_priv = engine->i915;
> + struct drm_i915_private *i915 = engine->i915;
> i915_reg_t hwsp;
>
> /*
> * The ring status page addresses are no longer next to the rest of
> * the ring registers as of gen7.
> */
> - if (IS_GEN(dev_priv, 7)) {
> + if (IS_GEN(i915, 7)) {
> switch (engine->id) {
> /*
> * No more rings exist on Gen7. Default case is only to shut up
> @@ -544,7 +544,7 @@ static void set_hwsp(struct intel_engine_cs *engine, u32 offset)
> hwsp = VEBOX_HWS_PGA_GEN7;
> break;
> }
> - } else if (IS_GEN(dev_priv, 6)) {
> + } else if (IS_GEN(i915, 6)) {
> hwsp = RING_HWS_PGA_GEN6(engine->mmio_base);
> } else {
> hwsp = RING_HWS_PGA(engine->mmio_base);
> @@ -556,9 +556,9 @@ static void set_hwsp(struct intel_engine_cs *engine, u32 offset)
>
> static void flush_cs_tlb(struct intel_engine_cs *engine)
> {
> - struct drm_i915_private *dev_priv = engine->i915;
> + struct drm_i915_private *i915 = engine->i915;
>
> - if (!IS_GEN_RANGE(dev_priv, 6, 7))
> + if (!IS_GEN_RANGE(i915, 6, 7))
> return;
>
> /* ring should be idle before issuing a sync flush*/
> @@ -585,9 +585,9 @@ static void ring_setup_status_page(struct intel_engine_cs *engine)
>
> static bool stop_ring(struct intel_engine_cs *engine)
> {
> - struct drm_i915_private *dev_priv = engine->i915;
> + struct drm_i915_private *i915 = engine->i915;
>
> - if (INTEL_GEN(dev_priv) > 2) {
> + if (INTEL_GEN(i915) > 2) {
> ENGINE_WRITE(engine,
> RING_MI_MODE, _MASKED_BIT_ENABLE(STOP_RING));
> if (intel_wait_for_register(engine->uncore,
> @@ -622,7 +622,7 @@ static bool stop_ring(struct intel_engine_cs *engine)
>
> static int xcs_resume(struct intel_engine_cs *engine)
> {
> - struct drm_i915_private *dev_priv = engine->i915;
> + struct drm_i915_private *i915 = engine->i915;
> struct intel_ring *ring = engine->buffer;
> int ret = 0;
>
> @@ -654,7 +654,7 @@ static int xcs_resume(struct intel_engine_cs *engine)
> }
> }
>
> - if (HWS_NEEDS_PHYSICAL(dev_priv))
> + if (HWS_NEEDS_PHYSICAL(i915))
> ring_setup_phys_status_page(engine);
> else
> ring_setup_status_page(engine);
> @@ -705,7 +705,7 @@ static int xcs_resume(struct intel_engine_cs *engine)
> goto out;
> }
>
> - if (INTEL_GEN(dev_priv) > 2)
> + if (INTEL_GEN(i915) > 2)
> ENGINE_WRITE(engine,
> RING_MI_MODE, _MASKED_BIT_DISABLE(STOP_RING));
>
> @@ -815,7 +815,7 @@ static int intel_rcs_ctx_init(struct i915_request *rq)
>
> static int rcs_resume(struct intel_engine_cs *engine)
> {
> - struct drm_i915_private *dev_priv = engine->i915;
> + struct drm_i915_private *i915 = engine->i915;
>
> /*
> * Disable CONSTANT_BUFFER before it is loaded from the context
> @@ -827,12 +827,12 @@ static int rcs_resume(struct intel_engine_cs *engine)
> * they are already accustomed to from before contexts were
> * enabled.
> */
> - if (IS_GEN(dev_priv, 4))
> + if (IS_GEN(i915, 4))
> I915_WRITE(ECOSKPD,
> _MASKED_BIT_ENABLE(ECO_CONSTANT_BUFFER_SR_DISABLE));
>
> /* WaTimedSingleVertexDispatch:cl,bw,ctg,elk,ilk,snb */
> - if (IS_GEN_RANGE(dev_priv, 4, 6))
> + if (IS_GEN_RANGE(i915, 4, 6))
> I915_WRITE(MI_MODE, _MASKED_BIT_ENABLE(VS_TIMER_DISPATCH));
>
> /* We need to disable the AsyncFlip performance optimisations in order
> @@ -841,22 +841,22 @@ static int rcs_resume(struct intel_engine_cs *engine)
> *
> * WaDisableAsyncFlipPerfMode:snb,ivb,hsw,vlv
> */
> - if (IS_GEN_RANGE(dev_priv, 6, 7))
> + if (IS_GEN_RANGE(i915, 6, 7))
> I915_WRITE(MI_MODE, _MASKED_BIT_ENABLE(ASYNC_FLIP_PERF_DISABLE));
>
> /* Required for the hardware to program scanline values for waiting */
> /* WaEnableFlushTlbInvalidationMode:snb */
> - if (IS_GEN(dev_priv, 6))
> + if (IS_GEN(i915, 6))
> I915_WRITE(GFX_MODE,
> _MASKED_BIT_ENABLE(GFX_TLB_INVALIDATE_EXPLICIT));
>
> /* WaBCSVCSTlbInvalidationMode:ivb,vlv,hsw */
> - if (IS_GEN(dev_priv, 7))
> + if (IS_GEN(i915, 7))
> I915_WRITE(GFX_MODE_GEN7,
> _MASKED_BIT_ENABLE(GFX_TLB_INVALIDATE_EXPLICIT) |
> _MASKED_BIT_ENABLE(GFX_REPLAY_MODE));
>
> - if (IS_GEN(dev_priv, 6)) {
> + if (IS_GEN(i915, 6)) {
> /* From the Sandybridge PRM, volume 1 part 3, page 24:
> * "If this bit is set, STCunit will have LRA as replacement
> * policy. [...] This bit must be reset. LRA replacement
> @@ -866,7 +866,7 @@ static int rcs_resume(struct intel_engine_cs *engine)
> _MASKED_BIT_DISABLE(CM0_STC_EVICT_DISABLE_LRA_SNB));
> }
>
> - if (IS_GEN_RANGE(dev_priv, 6, 7))
> + if (IS_GEN_RANGE(i915, 6, 7))
> I915_WRITE(INSTPM, _MASKED_BIT_ENABLE(INSTPM_FORCE_ORDERING));
>
> return xcs_resume(engine);
> @@ -1227,15 +1227,15 @@ void intel_ring_unpin(struct intel_ring *ring)
> }
>
> static struct i915_vma *
> -intel_ring_create_vma(struct drm_i915_private *dev_priv, int size)
> +intel_ring_create_vma(struct drm_i915_private *i915, int size)
> {
> - struct i915_address_space *vm = &dev_priv->ggtt.vm;
> + struct i915_address_space *vm = &i915->ggtt.vm;
> struct drm_i915_gem_object *obj;
> struct i915_vma *vma;
>
> - obj = i915_gem_object_create_stolen(dev_priv, size);
> + obj = i915_gem_object_create_stolen(i915, size);
> if (!obj)
> - obj = i915_gem_object_create_internal(dev_priv, size);
> + obj = i915_gem_object_create_internal(i915, size);
> if (IS_ERR(obj))
> return ERR_CAST(obj);
>
> @@ -2115,9 +2115,9 @@ static void gen6_bsd_set_default_submission(struct intel_engine_cs *engine)
>
> static void ring_destroy(struct intel_engine_cs *engine)
> {
> - struct drm_i915_private *dev_priv = engine->i915;
> + struct drm_i915_private *i915 = engine->i915;
>
> - WARN_ON(INTEL_GEN(dev_priv) > 2 &&
> + WARN_ON(INTEL_GEN(i915) > 2 &&
> (ENGINE_READ(engine, RING_MI_MODE) & MODE_IDLE) == 0);
>
> intel_ring_unpin(engine->buffer);
> diff --git a/drivers/gpu/drm/i915/gvt/aperture_gm.c b/drivers/gpu/drm/i915/gvt/aperture_gm.c
> index 1fa2f65c3cd1..8807de566361 100644
> --- a/drivers/gpu/drm/i915/gvt/aperture_gm.c
> +++ b/drivers/gpu/drm/i915/gvt/aperture_gm.c
> @@ -40,7 +40,7 @@
> static int alloc_gm(struct intel_vgpu *vgpu, bool high_gm)
> {
> struct intel_gvt *gvt = vgpu->gvt;
> - struct drm_i915_private *dev_priv = gvt->dev_priv;
> + struct drm_i915_private *i915 = gvt->i915;
> unsigned int flags;
> u64 start, end, size;
> struct drm_mm_node *node;
> @@ -60,14 +60,14 @@ static int alloc_gm(struct intel_vgpu *vgpu, bool high_gm)
> flags = PIN_MAPPABLE;
> }
>
> - mutex_lock(&dev_priv->drm.struct_mutex);
> - mmio_hw_access_pre(dev_priv);
> - ret = i915_gem_gtt_insert(&dev_priv->ggtt.vm, node,
> + mutex_lock(&i915->drm.struct_mutex);
> + mmio_hw_access_pre(i915);
> + ret = i915_gem_gtt_insert(&i915->ggtt.vm, node,
> size, I915_GTT_PAGE_SIZE,
> I915_COLOR_UNEVICTABLE,
> start, end, flags);
> - mmio_hw_access_post(dev_priv);
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + mmio_hw_access_post(i915);
> + mutex_unlock(&i915->drm.struct_mutex);
> if (ret)
> gvt_err("fail to alloc %s gm space from host\n",
> high_gm ? "high" : "low");
> @@ -78,7 +78,7 @@ static int alloc_gm(struct intel_vgpu *vgpu, bool high_gm)
> static int alloc_vgpu_gm(struct intel_vgpu *vgpu)
> {
> struct intel_gvt *gvt = vgpu->gvt;
> - struct drm_i915_private *dev_priv = gvt->dev_priv;
> + struct drm_i915_private *i915 = gvt->i915;
> int ret;
>
> ret = alloc_gm(vgpu, false);
> @@ -97,20 +97,20 @@ static int alloc_vgpu_gm(struct intel_vgpu *vgpu)
>
> return 0;
> out_free_aperture:
> - mutex_lock(&dev_priv->drm.struct_mutex);
> + mutex_lock(&i915->drm.struct_mutex);
> drm_mm_remove_node(&vgpu->gm.low_gm_node);
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + mutex_unlock(&i915->drm.struct_mutex);
> return ret;
> }
>
> static void free_vgpu_gm(struct intel_vgpu *vgpu)
> {
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
>
> - mutex_lock(&dev_priv->drm.struct_mutex);
> + mutex_lock(&i915->drm.struct_mutex);
> drm_mm_remove_node(&vgpu->gm.low_gm_node);
> drm_mm_remove_node(&vgpu->gm.high_gm_node);
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + mutex_unlock(&i915->drm.struct_mutex);
> }
>
> /**
> @@ -127,11 +127,11 @@ void intel_vgpu_write_fence(struct intel_vgpu *vgpu,
> u32 fence, u64 value)
> {
> struct intel_gvt *gvt = vgpu->gvt;
> - struct drm_i915_private *dev_priv = gvt->dev_priv;
> + struct drm_i915_private *i915 = gvt->i915;
> struct drm_i915_fence_reg *reg;
> i915_reg_t fence_reg_lo, fence_reg_hi;
>
> - assert_rpm_wakelock_held(dev_priv);
> + assert_rpm_wakelock_held(i915);
>
> if (WARN_ON(fence >= vgpu_fence_sz(vgpu)))
> return;
> @@ -162,41 +162,41 @@ static void _clear_vgpu_fence(struct intel_vgpu *vgpu)
> static void free_vgpu_fence(struct intel_vgpu *vgpu)
> {
> struct intel_gvt *gvt = vgpu->gvt;
> - struct drm_i915_private *dev_priv = gvt->dev_priv;
> + struct drm_i915_private *i915 = gvt->i915;
> struct drm_i915_fence_reg *reg;
> u32 i;
>
> if (WARN_ON(!vgpu_fence_sz(vgpu)))
> return;
>
> - intel_runtime_pm_get(dev_priv);
> + intel_runtime_pm_get(i915);
>
> - mutex_lock(&dev_priv->drm.struct_mutex);
> + mutex_lock(&i915->drm.struct_mutex);
> _clear_vgpu_fence(vgpu);
> for (i = 0; i < vgpu_fence_sz(vgpu); i++) {
> reg = vgpu->fence.regs[i];
> i915_unreserve_fence(reg);
> vgpu->fence.regs[i] = NULL;
> }
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + mutex_unlock(&i915->drm.struct_mutex);
>
> - intel_runtime_pm_put_unchecked(dev_priv);
> + intel_runtime_pm_put_unchecked(i915);
> }
>
> static int alloc_vgpu_fence(struct intel_vgpu *vgpu)
> {
> struct intel_gvt *gvt = vgpu->gvt;
> - struct drm_i915_private *dev_priv = gvt->dev_priv;
> + struct drm_i915_private *i915 = gvt->i915;
> struct drm_i915_fence_reg *reg;
> int i;
>
> - intel_runtime_pm_get(dev_priv);
> + intel_runtime_pm_get(i915);
>
> /* Request fences from host */
> - mutex_lock(&dev_priv->drm.struct_mutex);
> + mutex_lock(&i915->drm.struct_mutex);
>
> for (i = 0; i < vgpu_fence_sz(vgpu); i++) {
> - reg = i915_reserve_fence(dev_priv);
> + reg = i915_reserve_fence(i915);
> if (IS_ERR(reg))
> goto out_free_fence;
>
> @@ -205,8 +205,8 @@ static int alloc_vgpu_fence(struct intel_vgpu *vgpu)
>
> _clear_vgpu_fence(vgpu);
>
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> - intel_runtime_pm_put_unchecked(dev_priv);
> + mutex_unlock(&i915->drm.struct_mutex);
> + intel_runtime_pm_put_unchecked(i915);
> return 0;
> out_free_fence:
> gvt_vgpu_err("Failed to alloc fences\n");
> @@ -218,8 +218,8 @@ static int alloc_vgpu_fence(struct intel_vgpu *vgpu)
> i915_unreserve_fence(reg);
> vgpu->fence.regs[i] = NULL;
> }
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> - intel_runtime_pm_put_unchecked(dev_priv);
> + mutex_unlock(&i915->drm.struct_mutex);
> + intel_runtime_pm_put_unchecked(i915);
> return -ENOSPC;
> }
>
> @@ -313,11 +313,11 @@ void intel_vgpu_free_resource(struct intel_vgpu *vgpu)
> */
> void intel_vgpu_reset_resource(struct intel_vgpu *vgpu)
> {
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
>
> - intel_runtime_pm_get(dev_priv);
> + intel_runtime_pm_get(i915);
> _clear_vgpu_fence(vgpu);
> - intel_runtime_pm_put_unchecked(dev_priv);
> + intel_runtime_pm_put_unchecked(i915);
> }
>
> /**
> diff --git a/drivers/gpu/drm/i915/gvt/cfg_space.c b/drivers/gpu/drm/i915/gvt/cfg_space.c
> index 19cf1bbe059d..d3a744e7c959 100644
> --- a/drivers/gpu/drm/i915/gvt/cfg_space.c
> +++ b/drivers/gpu/drm/i915/gvt/cfg_space.c
> @@ -391,9 +391,9 @@ void intel_vgpu_init_cfg_space(struct intel_vgpu *vgpu,
> memset(vgpu_cfg_space(vgpu) + INTEL_GVT_PCI_OPREGION, 0, 4);
>
> vgpu->cfg_space.bar[INTEL_GVT_PCI_BAR_GTTMMIO].size =
> - pci_resource_len(gvt->dev_priv->drm.pdev, 0);
> + pci_resource_len(gvt->i915->drm.pdev, 0);
> vgpu->cfg_space.bar[INTEL_GVT_PCI_BAR_APERTURE].size =
> - pci_resource_len(gvt->dev_priv->drm.pdev, 2);
> + pci_resource_len(gvt->i915->drm.pdev, 2);
>
> memset(vgpu_cfg_space(vgpu) + PCI_ROM_ADDRESS, 0, 4);
> }
> diff --git a/drivers/gpu/drm/i915/gvt/cmd_parser.c b/drivers/gpu/drm/i915/gvt/cmd_parser.c
> index 6ea88270c818..2b09acc0502f 100644
> --- a/drivers/gpu/drm/i915/gvt/cmd_parser.c
> +++ b/drivers/gpu/drm/i915/gvt/cmd_parser.c
> @@ -819,7 +819,7 @@ static int force_nonpriv_reg_handler(struct parser_exec_state *s,
> unsigned int data;
> u32 ring_base;
> u32 nopid;
> - struct drm_i915_private *dev_priv = s->vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = s->vgpu->gvt->i915;
>
> if (!strcmp(cmd, "lri"))
> data = cmd_val(s, index + 1);
> @@ -829,7 +829,7 @@ static int force_nonpriv_reg_handler(struct parser_exec_state *s,
> return -EINVAL;
> }
>
> - ring_base = dev_priv->engine[s->ring_id]->mmio_base;
> + ring_base = i915->engine[s->ring_id]->mmio_base;
> nopid = i915_mmio_reg_offset(RING_NOPID(ring_base));
>
> if (!intel_gvt_in_force_nonpriv_whitelist(gvt, data) &&
> @@ -905,7 +905,7 @@ static int cmd_reg_handler(struct parser_exec_state *s,
> * update reg values in it into vregs, so LRIs in workload with
> * inhibit context will restore with correct values
> */
> - if (IS_GEN(gvt->dev_priv, 9) &&
> + if (IS_GEN(gvt->i915, 9) &&
> intel_gvt_mmio_is_in_ctx(gvt, offset) &&
> !strncmp(cmd, "lri", 3)) {
> intel_gvt_hypervisor_read_gpa(s->vgpu,
> @@ -946,7 +946,7 @@ static int cmd_handler_lri(struct parser_exec_state *s)
> struct intel_gvt *gvt = s->vgpu->gvt;
>
> for (i = 1; i < cmd_len; i += 2) {
> - if (IS_BROADWELL(gvt->dev_priv) && s->ring_id != RCS0) {
> + if (IS_BROADWELL(gvt->i915) && s->ring_id != RCS0) {
> if (s->ring_id == BCS0 &&
> cmd_reg(s, i) == i915_mmio_reg_offset(DERRMR))
> ret |= 0;
> @@ -968,7 +968,7 @@ static int cmd_handler_lrr(struct parser_exec_state *s)
> int cmd_len = cmd_length(s);
>
> for (i = 1; i < cmd_len; i += 2) {
> - if (IS_BROADWELL(s->vgpu->gvt->dev_priv))
> + if (IS_BROADWELL(s->vgpu->gvt->i915))
> ret |= ((cmd_reg_inhibit(s, i) ||
> (cmd_reg_inhibit(s, i + 1)))) ?
> -EBADRQC : 0;
> @@ -996,7 +996,7 @@ static int cmd_handler_lrm(struct parser_exec_state *s)
> int cmd_len = cmd_length(s);
>
> for (i = 1; i < cmd_len;) {
> - if (IS_BROADWELL(gvt->dev_priv))
> + if (IS_BROADWELL(gvt->i915))
> ret |= (cmd_reg_inhibit(s, i)) ? -EBADRQC : 0;
> if (ret)
> break;
> @@ -1180,7 +1180,7 @@ struct plane_code_mapping {
> static int gen8_decode_mi_display_flip(struct parser_exec_state *s,
> struct mi_display_flip_command_info *info)
> {
> - struct drm_i915_private *dev_priv = s->vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = s->vgpu->gvt->i915;
> struct plane_code_mapping gen8_plane_code[] = {
> [0] = {PIPE_A, PLANE_A, PRIMARY_A_FLIP_DONE},
> [1] = {PIPE_B, PLANE_A, PRIMARY_B_FLIP_DONE},
> @@ -1226,7 +1226,7 @@ static int gen8_decode_mi_display_flip(struct parser_exec_state *s,
> static int skl_decode_mi_display_flip(struct parser_exec_state *s,
> struct mi_display_flip_command_info *info)
> {
> - struct drm_i915_private *dev_priv = s->vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = s->vgpu->gvt->i915;
> struct intel_vgpu *vgpu = s->vgpu;
> u32 dword0 = cmd_val(s, 0);
> u32 dword1 = cmd_val(s, 1);
> @@ -1285,13 +1285,13 @@ static int skl_decode_mi_display_flip(struct parser_exec_state *s,
> static int gen8_check_mi_display_flip(struct parser_exec_state *s,
> struct mi_display_flip_command_info *info)
> {
> - struct drm_i915_private *dev_priv = s->vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = s->vgpu->gvt->i915;
> u32 stride, tile;
>
> if (!info->async_flip)
> return 0;
>
> - if (INTEL_GEN(dev_priv) >= 9) {
> + if (INTEL_GEN(i915) >= 9) {
> stride = vgpu_vreg_t(s->vgpu, info->stride_reg) & GENMASK(9, 0);
> tile = (vgpu_vreg_t(s->vgpu, info->ctrl_reg) &
> GENMASK(12, 10)) >> 10;
> @@ -1314,12 +1314,12 @@ static int gen8_update_plane_mmio_from_mi_display_flip(
> struct parser_exec_state *s,
> struct mi_display_flip_command_info *info)
> {
> - struct drm_i915_private *dev_priv = s->vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = s->vgpu->gvt->i915;
> struct intel_vgpu *vgpu = s->vgpu;
>
> set_mask_bits(&vgpu_vreg_t(vgpu, info->surf_reg), GENMASK(31, 12),
> info->surf_val << 12);
> - if (INTEL_GEN(dev_priv) >= 9) {
> + if (INTEL_GEN(i915) >= 9) {
> set_mask_bits(&vgpu_vreg_t(vgpu, info->stride_reg), GENMASK(9, 0),
> info->stride_val);
> set_mask_bits(&vgpu_vreg_t(vgpu, info->ctrl_reg), GENMASK(12, 10),
> @@ -1345,11 +1345,11 @@ static int gen8_update_plane_mmio_from_mi_display_flip(
> static int decode_mi_display_flip(struct parser_exec_state *s,
> struct mi_display_flip_command_info *info)
> {
> - struct drm_i915_private *dev_priv = s->vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = s->vgpu->gvt->i915;
>
> - if (IS_BROADWELL(dev_priv))
> + if (IS_BROADWELL(i915))
> return gen8_decode_mi_display_flip(s, info);
> - if (INTEL_GEN(dev_priv) >= 9)
> + if (INTEL_GEN(i915) >= 9)
> return skl_decode_mi_display_flip(s, info);
>
> return -ENODEV;
> @@ -1756,7 +1756,7 @@ static int perform_bb_shadow(struct parser_exec_state *s)
> if (bb->ppgtt)
> start_offset = gma & ~I915_GTT_PAGE_MASK;
>
> - bb->obj = i915_gem_object_create_shmem(s->vgpu->gvt->dev_priv,
> + bb->obj = i915_gem_object_create_shmem(s->vgpu->gvt->i915,
> round_up(bb_size + start_offset,
> PAGE_SIZE));
> if (IS_ERR(bb->obj)) {
> @@ -2830,7 +2830,7 @@ static int shadow_indirect_ctx(struct intel_shadow_wa_ctx *wa_ctx)
> int ret = 0;
> void *map;
>
> - obj = i915_gem_object_create_shmem(workload->vgpu->gvt->dev_priv,
> + obj = i915_gem_object_create_shmem(workload->vgpu->gvt->i915,
> roundup(ctx_size + CACHELINE_BYTES,
> PAGE_SIZE));
> if (IS_ERR(obj))
> diff --git a/drivers/gpu/drm/i915/gvt/debugfs.c b/drivers/gpu/drm/i915/gvt/debugfs.c
> index 2fb7b73b260d..1ff7f11c89a2 100644
> --- a/drivers/gpu/drm/i915/gvt/debugfs.c
> +++ b/drivers/gpu/drm/i915/gvt/debugfs.c
> @@ -58,7 +58,7 @@ static int mmio_offset_compare(void *priv,
> static inline int mmio_diff_handler(struct intel_gvt *gvt,
> u32 offset, void *data)
> {
> - struct drm_i915_private *i915 = gvt->dev_priv;
> + struct drm_i915_private *i915 = gvt->i915;
> struct mmio_diff_param *param = data;
> struct diff_mmio *node;
> u32 preg, vreg;
> @@ -98,10 +98,10 @@ static int vgpu_mmio_diff_show(struct seq_file *s, void *unused)
> mutex_lock(&gvt->lock);
> spin_lock_bh(&gvt->scheduler.mmio_context_lock);
>
> - mmio_hw_access_pre(gvt->dev_priv);
> + mmio_hw_access_pre(gvt->i915);
> /* Recognize all the diff mmios to list. */
> intel_gvt_for_each_tracked_mmio(gvt, mmio_diff_handler, ¶m);
> - mmio_hw_access_post(gvt->dev_priv);
> + mmio_hw_access_post(gvt->i915);
>
> spin_unlock_bh(&gvt->scheduler.mmio_context_lock);
> mutex_unlock(&gvt->lock);
> @@ -142,7 +142,7 @@ static int
> vgpu_scan_nonprivbb_set(void *data, u64 val)
> {
> struct intel_vgpu *vgpu = (struct intel_vgpu *)data;
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
> enum intel_engine_id id;
> char buf[128], *s;
> int len;
> @@ -164,7 +164,7 @@ vgpu_scan_nonprivbb_set(void *data, u64 val)
> for (id = 0; id < I915_NUM_ENGINES; id++) {
> struct intel_engine_cs *engine;
>
> - engine = dev_priv->engine[id];
> + engine = i915->engine[id];
> if (engine && (val & (1 << id))) {
> len = snprintf(s, 4, "%d, ", engine->id);
> s += len;
> @@ -240,7 +240,7 @@ void intel_gvt_debugfs_remove_vgpu(struct intel_vgpu *vgpu)
> */
> int intel_gvt_debugfs_init(struct intel_gvt *gvt)
> {
> - struct drm_minor *minor = gvt->dev_priv->drm.primary;
> + struct drm_minor *minor = gvt->i915->drm.primary;
> struct dentry *ent;
>
> gvt->debugfs_root = debugfs_create_dir("gvt", minor->debugfs_root);
> diff --git a/drivers/gpu/drm/i915/gvt/display.c b/drivers/gpu/drm/i915/gvt/display.c
> index e1c313da6c00..b037aeed2f40 100644
> --- a/drivers/gpu/drm/i915/gvt/display.c
> +++ b/drivers/gpu/drm/i915/gvt/display.c
> @@ -57,7 +57,7 @@ static int get_edp_pipe(struct intel_vgpu *vgpu)
>
> static int edp_pipe_is_enabled(struct intel_vgpu *vgpu)
> {
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
>
> if (!(vgpu_vreg_t(vgpu, PIPECONF(_PIPE_EDP)) & PIPECONF_ENABLE))
> return 0;
> @@ -69,7 +69,7 @@ static int edp_pipe_is_enabled(struct intel_vgpu *vgpu)
>
> int pipe_is_enabled(struct intel_vgpu *vgpu, int pipe)
> {
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
>
> if (WARN_ON(pipe < PIPE_A || pipe >= I915_MAX_PIPES))
> return -EINVAL;
> @@ -168,10 +168,10 @@ static u8 dpcd_fix_data[DPCD_HEADER_SIZE] = {
>
> static void emulate_monitor_status_change(struct intel_vgpu *vgpu)
> {
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
> int pipe;
>
> - if (IS_BROXTON(dev_priv)) {
> + if (IS_BROXTON(i915)) {
> vgpu_vreg_t(vgpu, GEN8_DE_PORT_ISR) &= ~(BXT_DE_PORT_HP_DDIA |
> BXT_DE_PORT_HP_DDIB |
> BXT_DE_PORT_HP_DDIC);
> @@ -198,8 +198,8 @@ static void emulate_monitor_status_change(struct intel_vgpu *vgpu)
> SDE_PORTC_HOTPLUG_CPT |
> SDE_PORTD_HOTPLUG_CPT);
>
> - if (IS_SKYLAKE(dev_priv) || IS_KABYLAKE(dev_priv) ||
> - IS_COFFEELAKE(dev_priv)) {
> + if (IS_SKYLAKE(i915) || IS_KABYLAKE(i915) ||
> + IS_COFFEELAKE(i915)) {
> vgpu_vreg_t(vgpu, SDEISR) &= ~(SDE_PORTA_HOTPLUG_SPT |
> SDE_PORTE_HOTPLUG_SPT);
> vgpu_vreg_t(vgpu, SKL_FUSE_STATUS) |=
> @@ -223,7 +223,7 @@ static void emulate_monitor_status_change(struct intel_vgpu *vgpu)
> (TRANS_DDI_BPC_8 | TRANS_DDI_MODE_SELECT_DVI |
> (PORT_B << TRANS_DDI_PORT_SHIFT) |
> TRANS_DDI_FUNC_ENABLE);
> - if (IS_BROADWELL(dev_priv)) {
> + if (IS_BROADWELL(i915)) {
> vgpu_vreg_t(vgpu, PORT_CLK_SEL(PORT_B)) &=
> ~PORT_CLK_SEL_MASK;
> vgpu_vreg_t(vgpu, PORT_CLK_SEL(PORT_B)) |=
> @@ -243,7 +243,7 @@ static void emulate_monitor_status_change(struct intel_vgpu *vgpu)
> (TRANS_DDI_BPC_8 | TRANS_DDI_MODE_SELECT_DVI |
> (PORT_C << TRANS_DDI_PORT_SHIFT) |
> TRANS_DDI_FUNC_ENABLE);
> - if (IS_BROADWELL(dev_priv)) {
> + if (IS_BROADWELL(i915)) {
> vgpu_vreg_t(vgpu, PORT_CLK_SEL(PORT_C)) &=
> ~PORT_CLK_SEL_MASK;
> vgpu_vreg_t(vgpu, PORT_CLK_SEL(PORT_C)) |=
> @@ -263,7 +263,7 @@ static void emulate_monitor_status_change(struct intel_vgpu *vgpu)
> (TRANS_DDI_BPC_8 | TRANS_DDI_MODE_SELECT_DVI |
> (PORT_D << TRANS_DDI_PORT_SHIFT) |
> TRANS_DDI_FUNC_ENABLE);
> - if (IS_BROADWELL(dev_priv)) {
> + if (IS_BROADWELL(i915)) {
> vgpu_vreg_t(vgpu, PORT_CLK_SEL(PORT_D)) &=
> ~PORT_CLK_SEL_MASK;
> vgpu_vreg_t(vgpu, PORT_CLK_SEL(PORT_D)) |=
> @@ -274,14 +274,14 @@ static void emulate_monitor_status_change(struct intel_vgpu *vgpu)
> vgpu_vreg_t(vgpu, SFUSE_STRAP) |= SFUSE_STRAP_DDID_DETECTED;
> }
>
> - if ((IS_SKYLAKE(dev_priv) || IS_KABYLAKE(dev_priv) ||
> - IS_COFFEELAKE(dev_priv)) &&
> + if ((IS_SKYLAKE(i915) || IS_KABYLAKE(i915) ||
> + IS_COFFEELAKE(i915)) &&
> intel_vgpu_has_monitor_on_port(vgpu, PORT_E)) {
> vgpu_vreg_t(vgpu, SDEISR) |= SDE_PORTE_HOTPLUG_SPT;
> }
>
> if (intel_vgpu_has_monitor_on_port(vgpu, PORT_A)) {
> - if (IS_BROADWELL(dev_priv))
> + if (IS_BROADWELL(i915))
> vgpu_vreg_t(vgpu, GEN8_DE_PORT_ISR) |=
> GEN8_PORT_DP_A_HOTPLUG;
> else
> @@ -291,11 +291,11 @@ static void emulate_monitor_status_change(struct intel_vgpu *vgpu)
> }
>
> /* Clear host CRT status, so guest couldn't detect this host CRT. */
> - if (IS_BROADWELL(dev_priv))
> + if (IS_BROADWELL(i915))
> vgpu_vreg_t(vgpu, PCH_ADPA) &= ~ADPA_CRT_HOTPLUG_MONITOR_MASK;
>
> /* Disable Primary/Sprite/Cursor plane */
> - for_each_pipe(dev_priv, pipe) {
> + for_each_pipe(i915, pipe) {
> vgpu_vreg_t(vgpu, DSPCNTR(pipe)) &= ~DISPLAY_PLANE_ENABLE;
> vgpu_vreg_t(vgpu, SPRCTL(pipe)) &= ~SPRITE_ENABLE;
> vgpu_vreg_t(vgpu, CURCNTR(pipe)) &= ~MCURSOR_MODE;
> @@ -389,7 +389,7 @@ void intel_gvt_check_vblank_emulation(struct intel_gvt *gvt)
>
> static void emulate_vblank_on_pipe(struct intel_vgpu *vgpu, int pipe)
> {
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
> struct intel_vgpu_irq *irq = &vgpu->irq;
> int vblank_event[] = {
> [PIPE_A] = PIPE_A_VBLANK,
> @@ -421,7 +421,7 @@ static void emulate_vblank(struct intel_vgpu *vgpu)
> int pipe;
>
> mutex_lock(&vgpu->vgpu_lock);
> - for_each_pipe(vgpu->gvt->dev_priv, pipe)
> + for_each_pipe(vgpu->gvt->i915, pipe)
> emulate_vblank_on_pipe(vgpu, pipe);
> mutex_unlock(&vgpu->vgpu_lock);
> }
> @@ -454,10 +454,10 @@ void intel_gvt_emulate_vblank(struct intel_gvt *gvt)
> */
> void intel_vgpu_emulate_hotplug(struct intel_vgpu *vgpu, bool connected)
> {
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
>
> /* TODO: add more platforms support */
> - if (IS_SKYLAKE(dev_priv) || IS_KABYLAKE(dev_priv)) {
> + if (IS_SKYLAKE(i915) || IS_KABYLAKE(i915)) {
> if (connected) {
> vgpu_vreg_t(vgpu, SFUSE_STRAP) |=
> SFUSE_STRAP_DDID_DETECTED;
> @@ -483,10 +483,10 @@ void intel_vgpu_emulate_hotplug(struct intel_vgpu *vgpu, bool connected)
> */
> void intel_vgpu_clean_display(struct intel_vgpu *vgpu)
> {
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
>
> - if (IS_SKYLAKE(dev_priv) || IS_KABYLAKE(dev_priv) ||
> - IS_COFFEELAKE(dev_priv))
> + if (IS_SKYLAKE(i915) || IS_KABYLAKE(i915) ||
> + IS_COFFEELAKE(i915))
> clean_virtual_dp_monitor(vgpu, PORT_D);
> else
> clean_virtual_dp_monitor(vgpu, PORT_B);
> @@ -505,12 +505,12 @@ void intel_vgpu_clean_display(struct intel_vgpu *vgpu)
> */
> int intel_vgpu_init_display(struct intel_vgpu *vgpu, u64 resolution)
> {
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
>
> intel_vgpu_init_i2c_edid(vgpu);
>
> - if (IS_SKYLAKE(dev_priv) || IS_KABYLAKE(dev_priv) ||
> - IS_COFFEELAKE(dev_priv))
> + if (IS_SKYLAKE(i915) || IS_KABYLAKE(i915) ||
> + IS_COFFEELAKE(i915))
> return setup_virtual_dp_monitor(vgpu, PORT_D, GVT_DP_D,
> resolution);
> else
> diff --git a/drivers/gpu/drm/i915/gvt/dmabuf.c b/drivers/gpu/drm/i915/gvt/dmabuf.c
> index 41c8ebc60c63..523038d224a1 100644
> --- a/drivers/gpu/drm/i915/gvt/dmabuf.c
> +++ b/drivers/gpu/drm/i915/gvt/dmabuf.c
> @@ -39,7 +39,7 @@
> static int vgpu_gem_get_pages(
> struct drm_i915_gem_object *obj)
> {
> - struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
> + struct drm_i915_private *i915 = to_i915(obj->base.dev);
> struct sg_table *st;
> struct scatterlist *sg;
> int i, ret;
> @@ -61,7 +61,7 @@ static int vgpu_gem_get_pages(
> kfree(st);
> return ret;
> }
> - gtt_entries = (gen8_pte_t __iomem *)dev_priv->ggtt.gsm +
> + gtt_entries = (gen8_pte_t __iomem *)i915->ggtt.gsm +
> (fb_info->start >> PAGE_SHIFT);
> for_each_sg(st->sgl, sg, page_num, i) {
> sg->offset = 0;
> @@ -152,7 +152,7 @@ static const struct drm_i915_gem_object_ops intel_vgpu_gem_ops = {
> static struct drm_i915_gem_object *vgpu_create_gem(struct drm_device *dev,
> struct intel_vgpu_fb_info *info)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct drm_i915_gem_object *obj;
>
> obj = i915_gem_object_alloc();
> @@ -165,7 +165,7 @@ static struct drm_i915_gem_object *vgpu_create_gem(struct drm_device *dev,
>
> obj->read_domains = I915_GEM_DOMAIN_GTT;
> obj->write_domain = 0;
> - if (INTEL_GEN(dev_priv) >= 9) {
> + if (INTEL_GEN(i915) >= 9) {
> unsigned int tiling_mode = 0;
> unsigned int stride = 0;
>
> @@ -360,7 +360,7 @@ static void update_fb_info(struct vfio_device_gfx_plane_info *gvt_dmabuf,
>
> int intel_vgpu_query_plane(struct intel_vgpu *vgpu, void *args)
> {
> - struct drm_device *dev = &vgpu->gvt->dev_priv->drm;
> + struct drm_device *dev = &vgpu->gvt->i915->drm;
> struct vfio_device_gfx_plane_info *gfx_plane_info = args;
> struct intel_vgpu_dmabuf_obj *dmabuf_obj;
> struct intel_vgpu_fb_info fb_info;
> @@ -466,7 +466,7 @@ int intel_vgpu_query_plane(struct intel_vgpu *vgpu, void *args)
> /* To associate an exposed dmabuf with the dmabuf_obj */
> int intel_vgpu_get_dmabuf(struct intel_vgpu *vgpu, unsigned int dmabuf_id)
> {
> - struct drm_device *dev = &vgpu->gvt->dev_priv->drm;
> + struct drm_device *dev = &vgpu->gvt->i915->drm;
> struct intel_vgpu_dmabuf_obj *dmabuf_obj;
> struct drm_i915_gem_object *obj;
> struct dma_buf *dmabuf;
> diff --git a/drivers/gpu/drm/i915/gvt/edid.c b/drivers/gpu/drm/i915/gvt/edid.c
> index 1fe6124918f1..9e45cbac5eed 100644
> --- a/drivers/gpu/drm/i915/gvt/edid.c
> +++ b/drivers/gpu/drm/i915/gvt/edid.c
> @@ -135,7 +135,7 @@ static void reset_gmbus_controller(struct intel_vgpu *vgpu)
> static int gmbus0_mmio_write(struct intel_vgpu *vgpu,
> unsigned int offset, void *p_data, unsigned int bytes)
> {
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
> int port, pin_select;
>
> memcpy(&vgpu_vreg(vgpu, offset), p_data, bytes);
> @@ -147,9 +147,9 @@ static int gmbus0_mmio_write(struct intel_vgpu *vgpu,
> if (pin_select == 0)
> return 0;
>
> - if (IS_BROXTON(dev_priv))
> + if (IS_BROXTON(i915))
> port = bxt_get_port_from_gmbus0(pin_select);
> - else if (IS_COFFEELAKE(dev_priv))
> + else if (IS_COFFEELAKE(i915))
> port = cnp_get_port_from_gmbus0(pin_select);
> else
> port = get_port_from_gmbus0(pin_select);
> diff --git a/drivers/gpu/drm/i915/gvt/execlist.c b/drivers/gpu/drm/i915/gvt/execlist.c
> index f21b8fb5b37e..7a7a925565a0 100644
> --- a/drivers/gpu/drm/i915/gvt/execlist.c
> +++ b/drivers/gpu/drm/i915/gvt/execlist.c
> @@ -40,7 +40,7 @@
> #define _EL_OFFSET_STATUS_PTR 0x3A0
>
> #define execlist_ring_mmio(gvt, ring_id, offset) \
> - (gvt->dev_priv->engine[ring_id]->mmio_base + (offset))
> + (gvt->i915->engine[ring_id]->mmio_base + (offset))
>
> #define valid_context(ctx) ((ctx)->valid)
> #define same_context(a, b) (((a)->context_id == (b)->context_id) && \
> @@ -133,7 +133,7 @@ static void emulate_csb_update(struct intel_vgpu_execlist *execlist,
> u32 write_pointer;
> u32 ctx_status_ptr_reg, ctx_status_buf_reg, offset;
> unsigned long hwsp_gpa;
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
>
> ctx_status_ptr_reg = execlist_ring_mmio(vgpu->gvt, ring_id,
> _EL_OFFSET_STATUS_PTR);
> @@ -169,7 +169,7 @@ static void emulate_csb_update(struct intel_vgpu_execlist *execlist,
> status, 8);
> intel_gvt_hypervisor_write_gpa(vgpu,
> hwsp_gpa +
> - intel_hws_csb_write_index(dev_priv) * 4,
> + intel_hws_csb_write_index(i915) * 4,
> &write_pointer, 4);
> }
>
> @@ -529,12 +529,12 @@ static void init_vgpu_execlist(struct intel_vgpu *vgpu, int ring_id)
> static void clean_execlist(struct intel_vgpu *vgpu,
> intel_engine_mask_t engine_mask)
> {
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
> struct intel_engine_cs *engine;
> struct intel_vgpu_submission *s = &vgpu->submission;
> intel_engine_mask_t tmp;
>
> - for_each_engine_masked(engine, dev_priv, engine_mask, tmp) {
> + for_each_engine_masked(engine, i915, engine_mask, tmp) {
> kfree(s->ring_scan_buffer[engine->id]);
> s->ring_scan_buffer[engine->id] = NULL;
> s->ring_scan_buffer_size[engine->id] = 0;
> @@ -544,11 +544,11 @@ static void clean_execlist(struct intel_vgpu *vgpu,
> static void reset_execlist(struct intel_vgpu *vgpu,
> intel_engine_mask_t engine_mask)
> {
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
> struct intel_engine_cs *engine;
> intel_engine_mask_t tmp;
>
> - for_each_engine_masked(engine, dev_priv, engine_mask, tmp)
> + for_each_engine_masked(engine, i915, engine_mask, tmp)
> init_vgpu_execlist(vgpu, engine->id);
> }
>
> diff --git a/drivers/gpu/drm/i915/gvt/fb_decoder.c b/drivers/gpu/drm/i915/gvt/fb_decoder.c
> index 65e847392aea..76c346eecda3 100644
> --- a/drivers/gpu/drm/i915/gvt/fb_decoder.c
> +++ b/drivers/gpu/drm/i915/gvt/fb_decoder.c
> @@ -146,12 +146,12 @@ static int skl_format_to_drm(int format, bool rgb_order, bool alpha,
> static u32 intel_vgpu_get_stride(struct intel_vgpu *vgpu, int pipe,
> u32 tiled, int stride_mask, int bpp)
> {
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
>
> u32 stride_reg = vgpu_vreg_t(vgpu, DSPSTRIDE(pipe)) & stride_mask;
> u32 stride = stride_reg;
>
> - if (INTEL_GEN(dev_priv) >= 9) {
> + if (INTEL_GEN(i915) >= 9) {
> switch (tiled) {
> case PLANE_CTL_TILED_LINEAR:
> stride = stride_reg * 64;
> @@ -203,7 +203,7 @@ int intel_vgpu_decode_primary_plane(struct intel_vgpu *vgpu,
> struct intel_vgpu_primary_plane_format *plane)
> {
> u32 val, fmt;
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
> int pipe;
>
> pipe = get_active_pipe(vgpu);
> @@ -215,7 +215,7 @@ int intel_vgpu_decode_primary_plane(struct intel_vgpu *vgpu,
> if (!plane->enabled)
> return -ENODEV;
>
> - if (INTEL_GEN(dev_priv) >= 9) {
> + if (INTEL_GEN(i915) >= 9) {
> plane->tiled = val & PLANE_CTL_TILED_MASK;
> fmt = skl_format_to_drm(
> val & PLANE_CTL_FORMAT_MASK,
> @@ -256,7 +256,7 @@ int intel_vgpu_decode_primary_plane(struct intel_vgpu *vgpu,
> }
>
> plane->stride = intel_vgpu_get_stride(vgpu, pipe, plane->tiled,
> - (INTEL_GEN(dev_priv) >= 9) ?
> + (INTEL_GEN(i915) >= 9) ?
> (_PRI_PLANE_STRIDE_MASK >> 6) :
> _PRI_PLANE_STRIDE_MASK, plane->bpp);
>
> @@ -334,7 +334,7 @@ int intel_vgpu_decode_cursor_plane(struct intel_vgpu *vgpu,
> {
> u32 val, mode, index;
> u32 alpha_plane, alpha_force;
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
> int pipe;
>
> pipe = get_active_pipe(vgpu);
> diff --git a/drivers/gpu/drm/i915/gvt/firmware.c b/drivers/gpu/drm/i915/gvt/firmware.c
> index 049775e8e350..efdc9df4f731 100644
> --- a/drivers/gpu/drm/i915/gvt/firmware.c
> +++ b/drivers/gpu/drm/i915/gvt/firmware.c
> @@ -68,7 +68,7 @@ static struct bin_attribute firmware_attr = {
>
> static int mmio_snapshot_handler(struct intel_gvt *gvt, u32 offset, void *data)
> {
> - struct drm_i915_private *i915 = gvt->dev_priv;
> + struct drm_i915_private *i915 = gvt->i915;
>
> *(u32 *)(data + offset) = intel_uncore_read_notrace(&i915->uncore,
> _MMIO(offset));
> @@ -78,7 +78,7 @@ static int mmio_snapshot_handler(struct intel_gvt *gvt, u32 offset, void *data)
> static int expose_firmware_sysfs(struct intel_gvt *gvt)
> {
> struct intel_gvt_device_info *info = &gvt->device_info;
> - struct pci_dev *pdev = gvt->dev_priv->drm.pdev;
> + struct pci_dev *pdev = gvt->i915->drm.pdev;
> struct gvt_firmware_header *h;
> void *firmware;
> void *p;
> @@ -129,7 +129,7 @@ static int expose_firmware_sysfs(struct intel_gvt *gvt)
>
> static void clean_firmware_sysfs(struct intel_gvt *gvt)
> {
> - struct pci_dev *pdev = gvt->dev_priv->drm.pdev;
> + struct pci_dev *pdev = gvt->i915->drm.pdev;
>
> device_remove_bin_file(&pdev->dev, &firmware_attr);
> vfree(firmware_attr.private);
> @@ -153,8 +153,8 @@ static int verify_firmware(struct intel_gvt *gvt,
> const struct firmware *fw)
> {
> struct intel_gvt_device_info *info = &gvt->device_info;
> - struct drm_i915_private *dev_priv = gvt->dev_priv;
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct drm_i915_private *i915 = gvt->i915;
> + struct pci_dev *pdev = i915->drm.pdev;
> struct gvt_firmware_header *h;
> unsigned long id, crc32_start;
> const void *mem;
> @@ -208,8 +208,8 @@ static int verify_firmware(struct intel_gvt *gvt,
> int intel_gvt_load_firmware(struct intel_gvt *gvt)
> {
> struct intel_gvt_device_info *info = &gvt->device_info;
> - struct drm_i915_private *dev_priv = gvt->dev_priv;
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct drm_i915_private *i915 = gvt->i915;
> + struct pci_dev *pdev = i915->drm.pdev;
> struct intel_gvt_firmware *firmware = &gvt->firmware;
> struct gvt_firmware_header *h;
> const struct firmware *fw;
> @@ -244,7 +244,7 @@ int intel_gvt_load_firmware(struct intel_gvt *gvt)
>
> gvt_dbg_core("request hw state firmware %s...\n", path);
>
> - ret = request_firmware(&fw, path, &dev_priv->drm.pdev->dev);
> + ret = request_firmware(&fw, path, &i915->drm.pdev->dev);
> kfree(path);
>
> if (ret)
> diff --git a/drivers/gpu/drm/i915/gvt/gtt.c b/drivers/gpu/drm/i915/gvt/gtt.c
> index 53115bdae12b..390562fe3223 100644
> --- a/drivers/gpu/drm/i915/gvt/gtt.c
> +++ b/drivers/gpu/drm/i915/gvt/gtt.c
> @@ -275,24 +275,24 @@ static inline int get_pse_type(int type)
> return gtt_type_table[type].pse_entry_type;
> }
>
> -static u64 read_pte64(struct drm_i915_private *dev_priv, unsigned long index)
> +static u64 read_pte64(struct drm_i915_private *i915, unsigned long index)
> {
> - void __iomem *addr = (gen8_pte_t __iomem *)dev_priv->ggtt.gsm + index;
> + void __iomem *addr = (gen8_pte_t __iomem *)i915->ggtt.gsm + index;
>
> return readq(addr);
> }
>
> -static void ggtt_invalidate(struct drm_i915_private *dev_priv)
> +static void ggtt_invalidate(struct drm_i915_private *i915)
> {
> - mmio_hw_access_pre(dev_priv);
> + mmio_hw_access_pre(i915);
> I915_WRITE(GFX_FLSH_CNTL_GEN6, GFX_FLSH_CNTL_EN);
> - mmio_hw_access_post(dev_priv);
> + mmio_hw_access_post(i915);
> }
>
> -static void write_pte64(struct drm_i915_private *dev_priv,
> +static void write_pte64(struct drm_i915_private *i915,
> unsigned long index, u64 pte)
> {
> - void __iomem *addr = (gen8_pte_t __iomem *)dev_priv->ggtt.gsm + index;
> + void __iomem *addr = (gen8_pte_t __iomem *)i915->ggtt.gsm + index;
>
> writeq(pte, addr);
> }
> @@ -315,7 +315,7 @@ static inline int gtt_get_entry64(void *pt,
> if (WARN_ON(ret))
> return ret;
> } else if (!pt) {
> - e->val64 = read_pte64(vgpu->gvt->dev_priv, index);
> + e->val64 = read_pte64(vgpu->gvt->i915, index);
> } else {
> e->val64 = *((u64 *)pt + index);
> }
> @@ -340,7 +340,7 @@ static inline int gtt_set_entry64(void *pt,
> if (WARN_ON(ret))
> return ret;
> } else if (!pt) {
> - write_pte64(vgpu->gvt->dev_priv, index, e->val64);
> + write_pte64(vgpu->gvt->i915, index, e->val64);
> } else {
> *((u64 *)pt + index) = e->val64;
> }
> @@ -734,7 +734,7 @@ static int detach_oos_page(struct intel_vgpu *vgpu,
>
> static void ppgtt_free_spt(struct intel_vgpu_ppgtt_spt *spt)
> {
> - struct device *kdev = &spt->vgpu->gvt->dev_priv->drm.pdev->dev;
> + struct device *kdev = &spt->vgpu->gvt->i915->drm.pdev->dev;
>
> trace_spt_free(spt->vgpu->id, spt, spt->guest_page.type);
>
> @@ -819,7 +819,7 @@ static int reclaim_one_ppgtt_mm(struct intel_gvt *gvt);
> static struct intel_vgpu_ppgtt_spt *ppgtt_alloc_spt(
> struct intel_vgpu *vgpu, enum intel_gvt_gtt_type type)
> {
> - struct device *kdev = &vgpu->gvt->dev_priv->drm.pdev->dev;
> + struct device *kdev = &vgpu->gvt->i915->drm.pdev->dev;
> struct intel_vgpu_ppgtt_spt *spt = NULL;
> dma_addr_t daddr;
> int ret;
> @@ -1044,14 +1044,14 @@ static int ppgtt_invalidate_spt(struct intel_vgpu_ppgtt_spt *spt)
>
> static bool vgpu_ips_enabled(struct intel_vgpu *vgpu)
> {
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
>
> - if (INTEL_GEN(dev_priv) == 9 || INTEL_GEN(dev_priv) == 10) {
> + if (INTEL_GEN(i915) == 9 || INTEL_GEN(i915) == 10) {
> u32 ips = vgpu_vreg_t(vgpu, GEN8_GAMW_ECO_DEV_RW_IA) &
> GAMW_ECO_ENABLE_64K_IPS_FIELD;
>
> return ips == GAMW_ECO_ENABLE_64K_IPS_FIELD;
> - } else if (INTEL_GEN(dev_priv) >= 11) {
> + } else if (INTEL_GEN(i915) >= 11) {
> /* 64K paging only controlled by IPS bit in PTE now. */
> return true;
> } else
> @@ -1153,7 +1153,7 @@ static int is_2MB_gtt_possible(struct intel_vgpu *vgpu,
> struct intel_gvt_gtt_pte_ops *ops = vgpu->gvt->gtt.pte_ops;
> unsigned long pfn;
>
> - if (!HAS_PAGE_SIZES(vgpu->gvt->dev_priv, I915_GTT_PAGE_SIZE_2M))
> + if (!HAS_PAGE_SIZES(vgpu->gvt->i915, I915_GTT_PAGE_SIZE_2M))
> return 0;
>
> pfn = intel_gvt_hypervisor_gfn_to_mfn(vgpu, ops->get_pfn(entry));
> @@ -2301,7 +2301,7 @@ static int emulate_ggtt_mmio_write(struct intel_vgpu *vgpu, unsigned int off,
> ggtt_invalidate_pte(vgpu, &e);
>
> ggtt_set_host_entry(ggtt_mm, &m, g_gtt_index);
> - ggtt_invalidate(gvt->dev_priv);
> + ggtt_invalidate(gvt->i915);
> return 0;
> }
>
> @@ -2340,7 +2340,7 @@ static int alloc_scratch_pages(struct intel_vgpu *vgpu,
> vgpu->gvt->device_info.gtt_entry_size_shift;
> void *scratch_pt;
> int i;
> - struct device *dev = &vgpu->gvt->dev_priv->drm.pdev->dev;
> + struct device *dev = &vgpu->gvt->i915->drm.pdev->dev;
> dma_addr_t daddr;
>
> if (WARN_ON(type < GTT_TYPE_PPGTT_PTE_PT || type >= GTT_TYPE_MAX))
> @@ -2397,7 +2397,7 @@ static int alloc_scratch_pages(struct intel_vgpu *vgpu,
> static int release_scratch_page_tree(struct intel_vgpu *vgpu)
> {
> int i;
> - struct device *dev = &vgpu->gvt->dev_priv->drm.pdev->dev;
> + struct device *dev = &vgpu->gvt->i915->drm.pdev->dev;
> dma_addr_t daddr;
>
> for (i = GTT_TYPE_PPGTT_PTE_PT; i < GTT_TYPE_MAX; i++) {
> @@ -2669,7 +2669,7 @@ int intel_gvt_init_gtt(struct intel_gvt *gvt)
> {
> int ret;
> void *page;
> - struct device *dev = &gvt->dev_priv->drm.pdev->dev;
> + struct device *dev = &gvt->i915->drm.pdev->dev;
> dma_addr_t daddr;
>
> gvt_dbg_core("init gtt\n");
> @@ -2718,7 +2718,7 @@ int intel_gvt_init_gtt(struct intel_gvt *gvt)
> */
> void intel_gvt_clean_gtt(struct intel_gvt *gvt)
> {
> - struct device *dev = &gvt->dev_priv->drm.pdev->dev;
> + struct device *dev = &gvt->i915->drm.pdev->dev;
> dma_addr_t daddr = (dma_addr_t)(gvt->gtt.scratch_mfn <<
> I915_GTT_PAGE_SHIFT);
>
> @@ -2766,7 +2766,7 @@ void intel_vgpu_invalidate_ppgtt(struct intel_vgpu *vgpu)
> void intel_vgpu_reset_ggtt(struct intel_vgpu *vgpu, bool invalidate_old)
> {
> struct intel_gvt *gvt = vgpu->gvt;
> - struct drm_i915_private *dev_priv = gvt->dev_priv;
> + struct drm_i915_private *i915 = gvt->i915;
> struct intel_gvt_gtt_pte_ops *pte_ops = vgpu->gvt->gtt.pte_ops;
> struct intel_gvt_gtt_entry entry = {.type = GTT_TYPE_GGTT_PTE};
> struct intel_gvt_gtt_entry old_entry;
> @@ -2796,7 +2796,7 @@ void intel_vgpu_reset_ggtt(struct intel_vgpu *vgpu, bool invalidate_old)
> ggtt_set_host_entry(vgpu->gtt.ggtt_mm, &entry, index++);
> }
>
> - ggtt_invalidate(dev_priv);
> + ggtt_invalidate(i915);
> }
>
> /**
> diff --git a/drivers/gpu/drm/i915/gvt/gvt.c b/drivers/gpu/drm/i915/gvt/gvt.c
> index 43f4242062dd..b84cdd8d8c98 100644
> --- a/drivers/gpu/drm/i915/gvt/gvt.c
> +++ b/drivers/gpu/drm/i915/gvt/gvt.c
> @@ -52,7 +52,7 @@ static struct intel_vgpu_type *intel_gvt_find_vgpu_type(struct intel_gvt *gvt,
> int i;
> struct intel_vgpu_type *t;
> const char *driver_name = dev_driver_string(
> - &gvt->dev_priv->drm.pdev->dev);
> + &gvt->i915->drm.pdev->dev);
>
> for (i = 0; i < gvt->num_types; i++) {
> t = &gvt->types[i];
> @@ -191,7 +191,7 @@ static const struct intel_gvt_ops intel_gvt_ops = {
> static void init_device_info(struct intel_gvt *gvt)
> {
> struct intel_gvt_device_info *info = &gvt->device_info;
> - struct pci_dev *pdev = gvt->dev_priv->drm.pdev;
> + struct pci_dev *pdev = gvt->i915->drm.pdev;
>
> info->max_support_vgpus = 8;
> info->cfg_space_size = PCI_CFG_SPACE_EXP_SIZE;
> @@ -257,15 +257,15 @@ static int init_service_thread(struct intel_gvt *gvt)
>
> /**
> * intel_gvt_clean_device - clean a GVT device
> - * @dev_priv: i915 private
> + * @i915: i915 private
> *
> * This function is called at the driver unloading stage, to free the
> * resources owned by a GVT device.
> *
> */
> -void intel_gvt_clean_device(struct drm_i915_private *dev_priv)
> +void intel_gvt_clean_device(struct drm_i915_private *i915)
> {
> - struct intel_gvt *gvt = to_gvt(dev_priv);
> + struct intel_gvt *gvt = to_gvt(i915);
>
> if (WARN_ON(!gvt))
> return;
> @@ -285,13 +285,13 @@ void intel_gvt_clean_device(struct drm_i915_private *dev_priv)
> intel_gvt_clean_mmio_info(gvt);
> idr_destroy(&gvt->vgpu_idr);
>
> - kfree(dev_priv->gvt);
> - dev_priv->gvt = NULL;
> + kfree(i915->gvt);
> + i915->gvt = NULL;
> }
>
> /**
> * intel_gvt_init_device - initialize a GVT device
> - * @dev_priv: drm i915 private data
> + * @i915: drm i915 private data
> *
> * This function is called at the initialization stage, to initialize
> * necessary GVT components.
> @@ -300,13 +300,13 @@ void intel_gvt_clean_device(struct drm_i915_private *dev_priv)
> * Zero on success, negative error code if failed.
> *
> */
> -int intel_gvt_init_device(struct drm_i915_private *dev_priv)
> +int intel_gvt_init_device(struct drm_i915_private *i915)
> {
> struct intel_gvt *gvt;
> struct intel_vgpu *vgpu;
> int ret;
>
> - if (WARN_ON(dev_priv->gvt))
> + if (WARN_ON(i915->gvt))
> return -EEXIST;
>
> gvt = kzalloc(sizeof(struct intel_gvt), GFP_KERNEL);
> @@ -319,7 +319,7 @@ int intel_gvt_init_device(struct drm_i915_private *dev_priv)
> spin_lock_init(&gvt->scheduler.mmio_context_lock);
> mutex_init(&gvt->lock);
> mutex_init(&gvt->sched_lock);
> - gvt->dev_priv = dev_priv;
> + gvt->i915 = i915;
>
> init_device_info(gvt);
>
> @@ -380,8 +380,8 @@ int intel_gvt_init_device(struct drm_i915_private *dev_priv)
> gvt_err("debugfs registration failed, go on.\n");
>
> gvt_dbg_core("gvt device initialization is done\n");
> - dev_priv->gvt = gvt;
> - intel_gvt_host.dev = &dev_priv->drm.pdev->dev;
> + i915->gvt = gvt;
> + intel_gvt_host.dev = &i915->drm.pdev->dev;
> intel_gvt_host.initialized = true;
> return 0;
>
> diff --git a/drivers/gpu/drm/i915/gvt/gvt.h b/drivers/gpu/drm/i915/gvt/gvt.h
> index b54f2bdc13a4..fbebe0ae4096 100644
> --- a/drivers/gpu/drm/i915/gvt/gvt.h
> +++ b/drivers/gpu/drm/i915/gvt/gvt.h
> @@ -306,7 +306,7 @@ struct intel_gvt {
> /* scheduler scope lock, protect gvt and vgpu schedule related data */
> struct mutex sched_lock;
>
> - struct drm_i915_private *dev_priv;
> + struct drm_i915_private *i915;
> struct idr vgpu_idr; /* vGPU IDR pool */
>
> struct intel_gvt_device_info device_info;
> @@ -373,12 +373,12 @@ int intel_gvt_load_firmware(struct intel_gvt *gvt);
> #define HOST_FENCE 4
>
> /* Aperture/GM space definitions for GVT device */
> -#define gvt_aperture_sz(gvt) (gvt->dev_priv->ggtt.mappable_end)
> -#define gvt_aperture_pa_base(gvt) (gvt->dev_priv->ggtt.gmadr.start)
> +#define gvt_aperture_sz(gvt) (gvt->i915->ggtt.mappable_end)
> +#define gvt_aperture_pa_base(gvt) (gvt->i915->ggtt.gmadr.start)
>
> -#define gvt_ggtt_gm_sz(gvt) (gvt->dev_priv->ggtt.vm.total)
> +#define gvt_ggtt_gm_sz(gvt) (gvt->i915->ggtt.vm.total)
> #define gvt_ggtt_sz(gvt) \
> - ((gvt->dev_priv->ggtt.vm.total >> PAGE_SHIFT) << 3)
> + ((gvt->i915->ggtt.vm.total >> PAGE_SHIFT) << 3)
> #define gvt_hidden_sz(gvt) (gvt_ggtt_gm_sz(gvt) - gvt_aperture_sz(gvt))
>
> #define gvt_aperture_gmadr_base(gvt) (0)
> @@ -390,7 +390,7 @@ int intel_gvt_load_firmware(struct intel_gvt *gvt);
> #define gvt_hidden_gmadr_end(gvt) (gvt_hidden_gmadr_base(gvt) \
> + gvt_hidden_sz(gvt) - 1)
>
> -#define gvt_fence_sz(gvt) (gvt->dev_priv->num_fence_regs)
> +#define gvt_fence_sz(gvt) (gvt->i915->num_fence_regs)
>
> /* Aperture/GM space definitions for vGPU */
> #define vgpu_aperture_offset(vgpu) ((vgpu)->gm.low_gm_node.start)
> @@ -582,14 +582,14 @@ enum {
> GVT_FAILSAFE_GUEST_ERR,
> };
>
> -static inline void mmio_hw_access_pre(struct drm_i915_private *dev_priv)
> +static inline void mmio_hw_access_pre(struct drm_i915_private *i915)
> {
> - intel_runtime_pm_get(dev_priv);
> + intel_runtime_pm_get(i915);
> }
>
> -static inline void mmio_hw_access_post(struct drm_i915_private *dev_priv)
> +static inline void mmio_hw_access_post(struct drm_i915_private *i915)
> {
> - intel_runtime_pm_put_unchecked(dev_priv);
> + intel_runtime_pm_put_unchecked(i915);
> }
>
> /**
> diff --git a/drivers/gpu/drm/i915/gvt/handlers.c b/drivers/gpu/drm/i915/gvt/handlers.c
> index a6ade66349bd..9acb01b0b7c7 100644
> --- a/drivers/gpu/drm/i915/gvt/handlers.c
> +++ b/drivers/gpu/drm/i915/gvt/handlers.c
> @@ -49,15 +49,15 @@
>
> unsigned long intel_gvt_get_device_type(struct intel_gvt *gvt)
> {
> - if (IS_BROADWELL(gvt->dev_priv))
> + if (IS_BROADWELL(gvt->i915))
> return D_BDW;
> - else if (IS_SKYLAKE(gvt->dev_priv))
> + else if (IS_SKYLAKE(gvt->i915))
> return D_SKL;
> - else if (IS_KABYLAKE(gvt->dev_priv))
> + else if (IS_KABYLAKE(gvt->i915))
> return D_KBL;
> - else if (IS_BROXTON(gvt->dev_priv))
> + else if (IS_BROXTON(gvt->i915))
> return D_BXT;
> - else if (IS_COFFEELAKE(gvt->dev_priv))
> + else if (IS_COFFEELAKE(gvt->i915))
> return D_CFL;
>
> return 0;
> @@ -156,7 +156,7 @@ int intel_gvt_render_mmio_to_ring_id(struct intel_gvt *gvt,
> struct intel_engine_cs *engine;
>
> offset &= ~GENMASK(11, 0);
> - for_each_engine(engine, gvt->dev_priv, id) {
> + for_each_engine(engine, gvt->i915, id) {
> if (engine->mmio_base == offset)
> return id;
> }
> @@ -217,7 +217,7 @@ static int gamw_echo_dev_rw_ia_write(struct intel_vgpu *vgpu,
> {
> u32 ips = (*(u32 *)p_data) & GAMW_ECO_ENABLE_64K_IPS_FIELD;
>
> - if (INTEL_GEN(vgpu->gvt->dev_priv) <= 10) {
> + if (INTEL_GEN(vgpu->gvt->i915) <= 10) {
> if (ips == GAMW_ECO_ENABLE_64K_IPS_FIELD)
> gvt_dbg_core("vgpu%d: ips enabled\n", vgpu->id);
> else if (!ips)
> @@ -253,7 +253,7 @@ static int fence_mmio_read(struct intel_vgpu *vgpu, unsigned int off,
> static int fence_mmio_write(struct intel_vgpu *vgpu, unsigned int off,
> void *p_data, unsigned int bytes)
> {
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
> unsigned int fence_num = offset_to_fence_num(off);
> int ret;
>
> @@ -262,10 +262,10 @@ static int fence_mmio_write(struct intel_vgpu *vgpu, unsigned int off,
> return ret;
> write_vreg(vgpu, off, p_data, bytes);
>
> - mmio_hw_access_pre(dev_priv);
> + mmio_hw_access_pre(i915);
> intel_vgpu_write_fence(vgpu, fence_num,
> vgpu_vreg64(vgpu, fence_num_to_offset(fence_num)));
> - mmio_hw_access_post(dev_priv);
> + mmio_hw_access_post(i915);
> return 0;
> }
>
> @@ -283,7 +283,7 @@ static int mul_force_wake_write(struct intel_vgpu *vgpu,
> old = vgpu_vreg(vgpu, offset);
> new = CALC_MODE_MASK_REG(old, *(u32 *)p_data);
>
> - if (INTEL_GEN(vgpu->gvt->dev_priv) >= 9) {
> + if (INTEL_GEN(vgpu->gvt->i915) >= 9) {
> switch (offset) {
> case FORCEWAKE_RENDER_GEN9_REG:
> ack_reg_offset = FORCEWAKE_ACK_RENDER_GEN9_REG;
> @@ -341,7 +341,7 @@ static int gdrst_mmio_write(struct intel_vgpu *vgpu, unsigned int offset,
> gvt_dbg_mmio("vgpu%d: request VCS2 Reset\n", vgpu->id);
> engine_mask |= BIT(VCS1);
> }
> - engine_mask &= INTEL_INFO(vgpu->gvt->dev_priv)->engine_mask;
> + engine_mask &= INTEL_INFO(vgpu->gvt->i915)->engine_mask;
> }
>
> /* vgpu_lock already hold by emulate mmio r/w */
> @@ -511,7 +511,7 @@ static int force_nonpriv_write(struct intel_vgpu *vgpu,
> u32 reg_nonpriv = *(u32 *)p_data;
> int ring_id = intel_gvt_render_mmio_to_ring_id(vgpu->gvt, offset);
> u32 ring_base;
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
> int ret = -EINVAL;
>
> if ((bytes != 4) || ((offset & (bytes - 1)) != 0) || ring_id < 0) {
> @@ -520,7 +520,7 @@ static int force_nonpriv_write(struct intel_vgpu *vgpu,
> return ret;
> }
>
> - ring_base = dev_priv->engine[ring_id]->mmio_base;
> + ring_base = i915->engine[ring_id]->mmio_base;
>
> if (in_whitelist(reg_nonpriv) ||
> reg_nonpriv == i915_mmio_reg_offset(RING_NOPID(ring_base))) {
> @@ -751,7 +751,7 @@ static int south_chicken2_mmio_write(struct intel_vgpu *vgpu,
> static int pri_surf_mmio_write(struct intel_vgpu *vgpu, unsigned int offset,
> void *p_data, unsigned int bytes)
> {
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
> u32 pipe = DSPSURF_TO_PIPE(offset);
> int event = SKL_FLIP_EVENT(pipe, PLANE_PRIMARY);
>
> @@ -792,7 +792,7 @@ static int reg50080_mmio_write(struct intel_vgpu *vgpu,
> unsigned int offset, void *p_data,
> unsigned int bytes)
> {
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
> enum pipe pipe = REG_50080_TO_PIPE(offset);
> enum plane_id plane = REG_50080_TO_PLANE(offset);
> int event = SKL_FLIP_EVENT(pipe, plane);
> @@ -816,7 +816,7 @@ static int reg50080_mmio_write(struct intel_vgpu *vgpu,
> static int trigger_aux_channel_interrupt(struct intel_vgpu *vgpu,
> unsigned int reg)
> {
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
> enum intel_gvt_event_type event;
>
> if (reg == _DPA_AUX_CH_CTL)
> @@ -916,11 +916,11 @@ static int dp_aux_ch_ctl_mmio_write(struct intel_vgpu *vgpu,
> write_vreg(vgpu, offset, p_data, bytes);
> data = vgpu_vreg(vgpu, offset);
>
> - if ((INTEL_GEN(vgpu->gvt->dev_priv) >= 9)
> + if ((INTEL_GEN(vgpu->gvt->i915) >= 9)
> && offset != _REG_SKL_DP_AUX_CH_CTL(port_index)) {
> /* SKL DPB/C/D aux ctl register changed */
> return 0;
> - } else if (IS_BROADWELL(vgpu->gvt->dev_priv) &&
> + } else if (IS_BROADWELL(vgpu->gvt->i915) &&
> offset != _REG_HSW_DP_AUX_CH_CTL(port_index)) {
> /* write to the data registers */
> return 0;
> @@ -1236,8 +1236,8 @@ static int handle_g2v_notification(struct intel_vgpu *vgpu, int notification)
>
> static int send_display_ready_uevent(struct intel_vgpu *vgpu, int ready)
> {
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> - struct kobject *kobj = &dev_priv->drm.primary->kdev->kobj;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
> + struct kobject *kobj = &i915->drm.primary->kdev->kobj;
> char *env[3] = {NULL, NULL, NULL};
> char vmid_str[20];
> char display_ready_str[20];
> @@ -1416,9 +1416,9 @@ static int mailbox_write(struct intel_vgpu *vgpu, unsigned int offset,
>
> switch (cmd) {
> case GEN9_PCODE_READ_MEM_LATENCY:
> - if (IS_SKYLAKE(vgpu->gvt->dev_priv)
> - || IS_KABYLAKE(vgpu->gvt->dev_priv)
> - || IS_COFFEELAKE(vgpu->gvt->dev_priv)) {
> + if (IS_SKYLAKE(vgpu->gvt->i915)
> + || IS_KABYLAKE(vgpu->gvt->i915)
> + || IS_COFFEELAKE(vgpu->gvt->i915)) {
> /**
> * "Read memory latency" command on gen9.
> * Below memory latency values are read
> @@ -1428,7 +1428,7 @@ static int mailbox_write(struct intel_vgpu *vgpu, unsigned int offset,
> *data0 = 0x1e1a1100;
> else
> *data0 = 0x61514b3d;
> - } else if (IS_BROXTON(vgpu->gvt->dev_priv)) {
> + } else if (IS_BROXTON(vgpu->gvt->i915)) {
> /**
> * "Read memory latency" command on gen9.
> * Below memory latency values are read
> @@ -1441,9 +1441,9 @@ static int mailbox_write(struct intel_vgpu *vgpu, unsigned int offset,
> }
> break;
> case SKL_PCODE_CDCLK_CONTROL:
> - if (IS_SKYLAKE(vgpu->gvt->dev_priv)
> - || IS_KABYLAKE(vgpu->gvt->dev_priv)
> - || IS_COFFEELAKE(vgpu->gvt->dev_priv))
> + if (IS_SKYLAKE(vgpu->gvt->i915)
> + || IS_KABYLAKE(vgpu->gvt->i915)
> + || IS_COFFEELAKE(vgpu->gvt->i915))
> *data0 = SKL_CDCLK_READY_FOR_CHANGE;
> break;
> case GEN6_PCODE_READ_RC6VIDS:
> @@ -1496,7 +1496,7 @@ static int skl_power_well_ctl_write(struct intel_vgpu *vgpu,
> {
> u32 v = *(u32 *)p_data;
>
> - if (IS_BROXTON(vgpu->gvt->dev_priv))
> + if (IS_BROXTON(vgpu->gvt->i915))
> v &= (1 << 31) | (1 << 29);
> else
> v &= (1 << 31) | (1 << 29) | (1 << 9) |
> @@ -1633,7 +1633,7 @@ static int mmio_read_from_hw(struct intel_vgpu *vgpu,
> unsigned int offset, void *p_data, unsigned int bytes)
> {
> struct intel_gvt *gvt = vgpu->gvt;
> - struct drm_i915_private *dev_priv = gvt->dev_priv;
> + struct drm_i915_private *i915 = gvt->i915;
> int ring_id;
> u32 ring_base;
>
> @@ -1645,14 +1645,14 @@ static int mmio_read_from_hw(struct intel_vgpu *vgpu,
> * c. the offset is ring time stamp mmio
> */
> if (ring_id >= 0)
> - ring_base = dev_priv->engine[ring_id]->mmio_base;
> + ring_base = i915->engine[ring_id]->mmio_base;
>
> if (ring_id < 0 || vgpu == gvt->scheduler.engine_owner[ring_id] ||
> offset == i915_mmio_reg_offset(RING_TIMESTAMP(ring_base)) ||
> offset == i915_mmio_reg_offset(RING_TIMESTAMP_UDW(ring_base))) {
> - mmio_hw_access_pre(dev_priv);
> + mmio_hw_access_pre(i915);
> vgpu_vreg(vgpu, offset) = I915_READ(_MMIO(offset));
> - mmio_hw_access_post(dev_priv);
> + mmio_hw_access_post(i915);
> }
>
> return intel_vgpu_default_mmio_read(vgpu, offset, p_data, bytes);
> @@ -1693,7 +1693,7 @@ static int ring_mode_mmio_write(struct intel_vgpu *vgpu, unsigned int offset,
> int ret;
>
> (*(u32 *)p_data) &= ~_MASKED_BIT_ENABLE(1);
> - if (IS_COFFEELAKE(vgpu->gvt->dev_priv))
> + if (IS_COFFEELAKE(vgpu->gvt->i915))
> (*(u32 *)p_data) &= ~_MASKED_BIT_ENABLE(2);
> write_vreg(vgpu, offset, p_data, bytes);
>
> @@ -1702,7 +1702,7 @@ static int ring_mode_mmio_write(struct intel_vgpu *vgpu, unsigned int offset,
> return 0;
> }
>
> - if (IS_COFFEELAKE(vgpu->gvt->dev_priv) &&
> + if (IS_COFFEELAKE(vgpu->gvt->i915) &&
> data & _MASKED_BIT_ENABLE(2)) {
> enter_failsafe_mode(vgpu, GVT_FAILSAFE_UNSUPPORTED_GUEST);
> return 0;
> @@ -1834,7 +1834,7 @@ static int csfe_chicken1_mmio_write(struct intel_vgpu *vgpu,
> MMIO_F(prefix(BLT_RING_BASE), s, f, am, rm, d, r, w); \
> MMIO_F(prefix(GEN6_BSD_RING_BASE), s, f, am, rm, d, r, w); \
> MMIO_F(prefix(VEBOX_RING_BASE), s, f, am, rm, d, r, w); \
> - if (HAS_ENGINE(dev_priv, VCS1)) \
> + if (HAS_ENGINE(i915, VCS1)) \
> MMIO_F(prefix(GEN8_BSD2_RING_BASE), s, f, am, rm, d, r, w); \
> } while (0)
>
> @@ -1855,7 +1855,7 @@ static int csfe_chicken1_mmio_write(struct intel_vgpu *vgpu,
>
> static int init_generic_mmio_info(struct intel_gvt *gvt)
> {
> - struct drm_i915_private *dev_priv = gvt->dev_priv;
> + struct drm_i915_private *i915 = gvt->i915;
> int ret;
>
> MMIO_RING_DFH(RING_IMR, D_ALL, F_CMD_ACCESS, NULL,
> @@ -2670,7 +2670,7 @@ static int init_generic_mmio_info(struct intel_gvt *gvt)
>
> static int init_broadwell_mmio_info(struct intel_gvt *gvt)
> {
> - struct drm_i915_private *dev_priv = gvt->dev_priv;
> + struct drm_i915_private *i915 = gvt->i915;
> int ret;
>
> MMIO_DH(GEN8_GT_IMR(0), D_BDW_PLUS, NULL, intel_vgpu_reg_imr_handler);
> @@ -2859,7 +2859,7 @@ static int init_broadwell_mmio_info(struct intel_gvt *gvt)
>
> static int init_skl_mmio_info(struct intel_gvt *gvt)
> {
> - struct drm_i915_private *dev_priv = gvt->dev_priv;
> + struct drm_i915_private *i915 = gvt->i915;
> int ret;
>
> MMIO_DH(FORCEWAKE_RENDER_GEN9, D_SKL_PLUS, NULL, mul_force_wake_write);
> @@ -3108,7 +3108,7 @@ static int init_skl_mmio_info(struct intel_gvt *gvt)
>
> static int init_bxt_mmio_info(struct intel_gvt *gvt)
> {
> - struct drm_i915_private *dev_priv = gvt->dev_priv;
> + struct drm_i915_private *i915 = gvt->i915;
> int ret;
>
> MMIO_F(_MMIO(0x80000), 0x3000, 0, 0, 0, D_BXT, NULL, NULL);
> @@ -3344,7 +3344,7 @@ static struct gvt_mmio_block mmio_blocks[] = {
> int intel_gvt_setup_mmio_info(struct intel_gvt *gvt)
> {
> struct intel_gvt_device_info *info = &gvt->device_info;
> - struct drm_i915_private *dev_priv = gvt->dev_priv;
> + struct drm_i915_private *i915 = gvt->i915;
> int size = info->mmio_size / 4 * sizeof(*gvt->mmio.mmio_attribute);
> int ret;
>
> @@ -3356,20 +3356,20 @@ int intel_gvt_setup_mmio_info(struct intel_gvt *gvt)
> if (ret)
> goto err;
>
> - if (IS_BROADWELL(dev_priv)) {
> + if (IS_BROADWELL(i915)) {
> ret = init_broadwell_mmio_info(gvt);
> if (ret)
> goto err;
> - } else if (IS_SKYLAKE(dev_priv)
> - || IS_KABYLAKE(dev_priv)
> - || IS_COFFEELAKE(dev_priv)) {
> + } else if (IS_SKYLAKE(i915)
> + || IS_KABYLAKE(i915)
> + || IS_COFFEELAKE(i915)) {
> ret = init_broadwell_mmio_info(gvt);
> if (ret)
> goto err;
> ret = init_skl_mmio_info(gvt);
> if (ret)
> goto err;
> - } else if (IS_BROXTON(dev_priv)) {
> + } else if (IS_BROXTON(i915)) {
> ret = init_broadwell_mmio_info(gvt);
> if (ret)
> goto err;
> diff --git a/drivers/gpu/drm/i915/gvt/interrupt.c b/drivers/gpu/drm/i915/gvt/interrupt.c
> index 951681813230..37e2e72bfacc 100644
> --- a/drivers/gpu/drm/i915/gvt/interrupt.c
> +++ b/drivers/gpu/drm/i915/gvt/interrupt.c
> @@ -536,7 +536,7 @@ static void gen8_init_irq(
> SET_BIT_INFO(irq, 4, VCS_MI_FLUSH_DW, INTEL_GVT_IRQ_INFO_GT1);
> SET_BIT_INFO(irq, 8, VCS_AS_CONTEXT_SWITCH, INTEL_GVT_IRQ_INFO_GT1);
>
> - if (HAS_ENGINE(gvt->dev_priv, VCS1)) {
> + if (HAS_ENGINE(gvt->i915, VCS1)) {
> SET_BIT_INFO(irq, 16, VCS2_MI_USER_INTERRUPT,
> INTEL_GVT_IRQ_INFO_GT1);
> SET_BIT_INFO(irq, 20, VCS2_MI_FLUSH_DW,
> @@ -568,7 +568,7 @@ static void gen8_init_irq(
> SET_BIT_INFO(irq, 22, DP_C_HOTPLUG, INTEL_GVT_IRQ_INFO_PCH);
> SET_BIT_INFO(irq, 23, DP_D_HOTPLUG, INTEL_GVT_IRQ_INFO_PCH);
>
> - if (IS_BROADWELL(gvt->dev_priv)) {
> + if (IS_BROADWELL(gvt->i915)) {
> SET_BIT_INFO(irq, 25, AUX_CHANNEL_B, INTEL_GVT_IRQ_INFO_PCH);
> SET_BIT_INFO(irq, 26, AUX_CHANNEL_C, INTEL_GVT_IRQ_INFO_PCH);
> SET_BIT_INFO(irq, 27, AUX_CHANNEL_D, INTEL_GVT_IRQ_INFO_PCH);
> @@ -581,7 +581,7 @@ static void gen8_init_irq(
>
> SET_BIT_INFO(irq, 4, PRIMARY_C_FLIP_DONE, INTEL_GVT_IRQ_INFO_DE_PIPE_C);
> SET_BIT_INFO(irq, 5, SPRITE_C_FLIP_DONE, INTEL_GVT_IRQ_INFO_DE_PIPE_C);
> - } else if (INTEL_GEN(gvt->dev_priv) >= 9) {
> + } else if (INTEL_GEN(gvt->i915) >= 9) {
> SET_BIT_INFO(irq, 25, AUX_CHANNEL_B, INTEL_GVT_IRQ_INFO_DE_PORT);
> SET_BIT_INFO(irq, 26, AUX_CHANNEL_C, INTEL_GVT_IRQ_INFO_DE_PORT);
> SET_BIT_INFO(irq, 27, AUX_CHANNEL_D, INTEL_GVT_IRQ_INFO_DE_PORT);
> diff --git a/drivers/gpu/drm/i915/gvt/kvmgt.c b/drivers/gpu/drm/i915/gvt/kvmgt.c
> index 144301b778df..615a7959b762 100644
> --- a/drivers/gpu/drm/i915/gvt/kvmgt.c
> +++ b/drivers/gpu/drm/i915/gvt/kvmgt.c
> @@ -187,7 +187,7 @@ static int gvt_pin_guest_page(struct intel_vgpu *vgpu, unsigned long gfn,
> static int gvt_dma_map_page(struct intel_vgpu *vgpu, unsigned long gfn,
> dma_addr_t *dma_addr, unsigned long size)
> {
> - struct device *dev = &vgpu->gvt->dev_priv->drm.pdev->dev;
> + struct device *dev = &vgpu->gvt->i915->drm.pdev->dev;
> struct page *page = NULL;
> int ret;
>
> @@ -210,7 +210,7 @@ static int gvt_dma_map_page(struct intel_vgpu *vgpu, unsigned long gfn,
> static void gvt_dma_unmap_page(struct intel_vgpu *vgpu, unsigned long gfn,
> dma_addr_t dma_addr, unsigned long size)
> {
> - struct device *dev = &vgpu->gvt->dev_priv->drm.pdev->dev;
> + struct device *dev = &vgpu->gvt->i915->drm.pdev->dev;
>
> dma_unmap_page(dev, dma_addr, size, PCI_DMA_BIDIRECTIONAL);
> gvt_unpin_guest_page(vgpu, gfn, size);
> @@ -913,7 +913,7 @@ static int intel_vgpu_aperture_rw(struct intel_vgpu *vgpu, u64 off,
> return -EINVAL;
> }
>
> - aperture_va = io_mapping_map_wc(&vgpu->gvt->dev_priv->ggtt.iomap,
> + aperture_va = io_mapping_map_wc(&vgpu->gvt->i915->ggtt.iomap,
> ALIGN_DOWN(off, PAGE_SIZE),
> count + offset_in_page(off));
> if (!aperture_va)
> diff --git a/drivers/gpu/drm/i915/gvt/mmio.c b/drivers/gpu/drm/i915/gvt/mmio.c
> index a55178884d67..572b8328af61 100644
> --- a/drivers/gpu/drm/i915/gvt/mmio.c
> +++ b/drivers/gpu/drm/i915/gvt/mmio.c
> @@ -245,7 +245,7 @@ void intel_vgpu_reset_mmio(struct intel_vgpu *vgpu, bool dmlr)
> /* set the bit 0:2(Core C-State ) to C0 */
> vgpu_vreg_t(vgpu, GEN6_GT_CORE_STATUS) = 0;
>
> - if (IS_BROXTON(vgpu->gvt->dev_priv)) {
> + if (IS_BROXTON(vgpu->gvt->i915)) {
> vgpu_vreg_t(vgpu, BXT_P_CR_GT_DISP_PWRON) &=
> ~(BIT(0) | BIT(1));
> vgpu_vreg_t(vgpu, BXT_PORT_CL1CM_DW0(DPIO_PHY0)) &=
> diff --git a/drivers/gpu/drm/i915/gvt/mmio_context.c b/drivers/gpu/drm/i915/gvt/mmio_context.c
> index 2998999e8568..44fd6a0a698c 100644
> --- a/drivers/gpu/drm/i915/gvt/mmio_context.c
> +++ b/drivers/gpu/drm/i915/gvt/mmio_context.c
> @@ -148,7 +148,7 @@ static struct {
> u32 l3cc_table[GEN9_MOCS_SIZE / 2];
> } gen9_render_mocs;
>
> -static void load_render_mocs(struct drm_i915_private *dev_priv)
> +static void load_render_mocs(struct drm_i915_private *i915)
> {
> i915_reg_t offset;
> u32 regs[] = {
> @@ -161,7 +161,7 @@ static void load_render_mocs(struct drm_i915_private *dev_priv)
> int ring_id, i;
>
> for (ring_id = 0; ring_id < ARRAY_SIZE(regs); ring_id++) {
> - if (!HAS_ENGINE(dev_priv, ring_id))
> + if (!HAS_ENGINE(i915, ring_id))
> continue;
> offset.reg = regs[ring_id];
> for (i = 0; i < GEN9_MOCS_SIZE; i++) {
> @@ -329,8 +329,8 @@ int intel_vgpu_restore_inhibit_context(struct intel_vgpu *vgpu,
>
> static void handle_tlb_pending_event(struct intel_vgpu *vgpu, int ring_id)
> {
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
> + struct intel_uncore *uncore = &i915->uncore;
> struct intel_vgpu_submission *s = &vgpu->submission;
> enum forcewake_domains fw;
> i915_reg_t reg;
> @@ -357,7 +357,7 @@ static void handle_tlb_pending_event(struct intel_vgpu *vgpu, int ring_id)
> */
> fw = intel_uncore_forcewake_for_reg(uncore, reg,
> FW_REG_READ | FW_REG_WRITE);
> - if (ring_id == RCS0 && INTEL_GEN(dev_priv) >= 9)
> + if (ring_id == RCS0 && INTEL_GEN(i915) >= 9)
> fw |= FORCEWAKE_RENDER;
>
> intel_uncore_forcewake_get(uncore, fw);
> @@ -377,7 +377,7 @@ static void handle_tlb_pending_event(struct intel_vgpu *vgpu, int ring_id)
> static void switch_mocs(struct intel_vgpu *pre, struct intel_vgpu *next,
> int ring_id)
> {
> - struct drm_i915_private *dev_priv;
> + struct drm_i915_private *i915;
> i915_reg_t offset, l3_offset;
> u32 old_v, new_v;
>
> @@ -390,15 +390,15 @@ static void switch_mocs(struct intel_vgpu *pre, struct intel_vgpu *next,
> };
> int i;
>
> - dev_priv = pre ? pre->gvt->dev_priv : next->gvt->dev_priv;
> + i915 = pre ? pre->gvt->i915 : next->gvt->i915;
> if (WARN_ON(ring_id >= ARRAY_SIZE(regs)))
> return;
>
> - if (ring_id == RCS0 && IS_GEN(dev_priv, 9))
> + if (ring_id == RCS0 && IS_GEN(i915, 9))
> return;
>
> if (!pre && !gen9_render_mocs.initialized)
> - load_render_mocs(dev_priv);
> + load_render_mocs(i915);
>
> offset.reg = regs[ring_id];
> for (i = 0; i < GEN9_MOCS_SIZE; i++) {
> @@ -454,16 +454,16 @@ static void switch_mmio(struct intel_vgpu *pre,
> struct intel_vgpu *next,
> int ring_id)
> {
> - struct drm_i915_private *dev_priv;
> + struct drm_i915_private *i915;
> struct intel_vgpu_submission *s;
> struct engine_mmio *mmio;
> u32 old_v, new_v;
>
> - dev_priv = pre ? pre->gvt->dev_priv : next->gvt->dev_priv;
> - if (INTEL_GEN(dev_priv) >= 9)
> + i915 = pre ? pre->gvt->i915 : next->gvt->i915;
> + if (INTEL_GEN(i915) >= 9)
> switch_mocs(pre, next, ring_id);
>
> - for (mmio = dev_priv->gvt->engine_mmio_list.mmio;
> + for (mmio = i915->gvt->engine_mmio_list.mmio;
> i915_mmio_reg_valid(mmio->reg); mmio++) {
> if (mmio->ring_id != ring_id)
> continue;
> @@ -472,7 +472,7 @@ static void switch_mmio(struct intel_vgpu *pre,
> * state image on gen9, it's initialized by lri command and
> * save or restore with context together.
> */
> - if (IS_GEN(dev_priv, 9) && mmio->in_context)
> + if (IS_GEN(i915, 9) && mmio->in_context)
> continue;
>
> // save
> @@ -536,7 +536,7 @@ static void switch_mmio(struct intel_vgpu *pre,
> void intel_gvt_switch_mmio(struct intel_vgpu *pre,
> struct intel_vgpu *next, int ring_id)
> {
> - struct drm_i915_private *dev_priv;
> + struct drm_i915_private *i915;
>
> if (WARN_ON(!pre && !next))
> return;
> @@ -544,16 +544,16 @@ void intel_gvt_switch_mmio(struct intel_vgpu *pre,
> gvt_dbg_render("switch ring %d from %s to %s\n", ring_id,
> pre ? "vGPU" : "host", next ? "vGPU" : "HOST");
>
> - dev_priv = pre ? pre->gvt->dev_priv : next->gvt->dev_priv;
> + i915 = pre ? pre->gvt->i915 : next->gvt->i915;
>
> /**
> * We are using raw mmio access wrapper to improve the
> * performace for batch mmio read/write, so we need
> * handle forcewake mannually.
> */
> - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
> switch_mmio(pre, next, ring_id);
> - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> }
>
> /**
> @@ -565,7 +565,7 @@ void intel_gvt_init_engine_mmio_context(struct intel_gvt *gvt)
> {
> struct engine_mmio *mmio;
>
> - if (INTEL_GEN(gvt->dev_priv) >= 9)
> + if (INTEL_GEN(gvt->i915) >= 9)
> gvt->engine_mmio_list.mmio = gen9_engine_mmio_list;
> else
> gvt->engine_mmio_list.mmio = gen8_engine_mmio_list;
> diff --git a/drivers/gpu/drm/i915/gvt/sched_policy.c b/drivers/gpu/drm/i915/gvt/sched_policy.c
> index 1c763a27a412..64a5d4688c89 100644
> --- a/drivers/gpu/drm/i915/gvt/sched_policy.c
> +++ b/drivers/gpu/drm/i915/gvt/sched_policy.c
> @@ -39,7 +39,7 @@ static bool vgpu_has_pending_workload(struct intel_vgpu *vgpu)
> enum intel_engine_id i;
> struct intel_engine_cs *engine;
>
> - for_each_engine(engine, vgpu->gvt->dev_priv, i) {
> + for_each_engine(engine, vgpu->gvt->i915, i) {
> if (!list_empty(workload_q_head(vgpu, i)))
> return true;
> }
> @@ -152,7 +152,7 @@ static void try_to_schedule_next_vgpu(struct intel_gvt *gvt)
> scheduler->need_reschedule = true;
>
> /* still have uncompleted workload? */
> - for_each_engine(engine, gvt->dev_priv, i) {
> + for_each_engine(engine, gvt->i915, i) {
> if (scheduler->current_workload[i])
> return;
> }
> @@ -169,7 +169,7 @@ static void try_to_schedule_next_vgpu(struct intel_gvt *gvt)
> scheduler->need_reschedule = false;
>
> /* wake up workload dispatch thread */
> - for_each_engine(engine, gvt->dev_priv, i)
> + for_each_engine(engine, gvt->i915, i)
> wake_up(&scheduler->waitq[i]);
> }
>
> @@ -446,7 +446,7 @@ void intel_vgpu_stop_schedule(struct intel_vgpu *vgpu)
> &vgpu->gvt->scheduler;
> int ring_id;
> struct vgpu_sched_data *vgpu_data = vgpu->sched_data;
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
>
> if (!vgpu_data->active)
> return;
> @@ -465,7 +465,7 @@ void intel_vgpu_stop_schedule(struct intel_vgpu *vgpu)
> scheduler->current_vgpu = NULL;
> }
>
> - intel_runtime_pm_get(dev_priv);
> + intel_runtime_pm_get(i915);
> spin_lock_bh(&scheduler->mmio_context_lock);
> for (ring_id = 0; ring_id < I915_NUM_ENGINES; ring_id++) {
> if (scheduler->engine_owner[ring_id] == vgpu) {
> @@ -474,6 +474,6 @@ void intel_vgpu_stop_schedule(struct intel_vgpu *vgpu)
> }
> }
> spin_unlock_bh(&scheduler->mmio_context_lock);
> - intel_runtime_pm_put_unchecked(dev_priv);
> + intel_runtime_pm_put_unchecked(i915);
> mutex_unlock(&vgpu->gvt->sched_lock);
> }
> diff --git a/drivers/gpu/drm/i915/gvt/scheduler.c b/drivers/gpu/drm/i915/gvt/scheduler.c
> index e301efb18d45..ec7299d8a738 100644
> --- a/drivers/gpu/drm/i915/gvt/scheduler.c
> +++ b/drivers/gpu/drm/i915/gvt/scheduler.c
> @@ -83,9 +83,9 @@ static void update_shadow_pdps(struct intel_vgpu_workload *workload)
> static void sr_oa_regs(struct intel_vgpu_workload *workload,
> u32 *reg_state, bool save)
> {
> - struct drm_i915_private *dev_priv = workload->vgpu->gvt->dev_priv;
> - u32 ctx_oactxctrl = dev_priv->perf.oa.ctx_oactxctrl_offset;
> - u32 ctx_flexeu0 = dev_priv->perf.oa.ctx_flexeu0_offset;
> + struct drm_i915_private *i915 = workload->vgpu->gvt->i915;
> + u32 ctx_oactxctrl = i915->perf.oa.ctx_oactxctrl_offset;
> + u32 ctx_flexeu0 = i915->perf.oa.ctx_flexeu0_offset;
> int i = 0;
> u32 flex_mmio[] = {
> i915_mmio_reg_offset(EU_PERF_CNTL0),
> @@ -177,11 +177,11 @@ static int populate_shadow_context(struct intel_vgpu_workload *workload)
> gvt_dbg_sched("ring id %d workload lrca %x", ring_id,
> workload->ctx_desc.lrca);
>
> - context_page_num = gvt->dev_priv->engine[ring_id]->context_size;
> + context_page_num = gvt->i915->engine[ring_id]->context_size;
>
> context_page_num = context_page_num >> PAGE_SHIFT;
>
> - if (IS_BROADWELL(gvt->dev_priv) && ring_id == RCS0)
> + if (IS_BROADWELL(gvt->i915) && ring_id == RCS0)
> context_page_num = 19;
>
> i = 2;
> @@ -211,8 +211,8 @@ static inline bool is_gvt_request(struct i915_request *req)
>
> static void save_ring_hw_state(struct intel_vgpu *vgpu, int ring_id)
> {
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> - u32 ring_base = dev_priv->engine[ring_id]->mmio_base;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
> + u32 ring_base = i915->engine[ring_id]->mmio_base;
> i915_reg_t reg;
>
> reg = RING_INSTDONE(ring_base);
> @@ -391,10 +391,10 @@ intel_gvt_workload_req_alloc(struct intel_vgpu_workload *workload)
> {
> struct intel_vgpu *vgpu = workload->vgpu;
> struct intel_vgpu_submission *s = &vgpu->submission;
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
> struct i915_request *rq;
>
> - lockdep_assert_held(&dev_priv->drm.struct_mutex);
> + lockdep_assert_held(&i915->drm.struct_mutex);
>
> if (workload->req)
> return 0;
> @@ -421,10 +421,10 @@ int intel_gvt_scan_and_shadow_workload(struct intel_vgpu_workload *workload)
> {
> struct intel_vgpu *vgpu = workload->vgpu;
> struct intel_vgpu_submission *s = &vgpu->submission;
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
> int ret;
>
> - lockdep_assert_held(&dev_priv->drm.struct_mutex);
> + lockdep_assert_held(&i915->drm.struct_mutex);
>
> if (workload->shadow)
> return 0;
> @@ -577,7 +577,7 @@ static int prepare_shadow_wa_ctx(struct intel_shadow_wa_ctx *wa_ctx)
> static void release_shadow_batch_buffer(struct intel_vgpu_workload *workload)
> {
> struct intel_vgpu *vgpu = workload->vgpu;
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
> struct intel_vgpu_shadow_bb *bb, *pos;
>
> if (list_empty(&workload->shadow_bb))
> @@ -586,7 +586,7 @@ static void release_shadow_batch_buffer(struct intel_vgpu_workload *workload)
> bb = list_first_entry(&workload->shadow_bb,
> struct intel_vgpu_shadow_bb, list);
>
> - mutex_lock(&dev_priv->drm.struct_mutex);
> + mutex_lock(&i915->drm.struct_mutex);
>
> list_for_each_entry_safe(bb, pos, &workload->shadow_bb, list) {
> if (bb->obj) {
> @@ -606,7 +606,7 @@ static void release_shadow_batch_buffer(struct intel_vgpu_workload *workload)
> kfree(bb);
> }
>
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + mutex_unlock(&i915->drm.struct_mutex);
> }
>
> static int prepare_workload(struct intel_vgpu_workload *workload)
> @@ -671,7 +671,7 @@ static int prepare_workload(struct intel_vgpu_workload *workload)
> static int dispatch_workload(struct intel_vgpu_workload *workload)
> {
> struct intel_vgpu *vgpu = workload->vgpu;
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
> struct intel_vgpu_submission *s = &vgpu->submission;
> struct i915_request *rq;
> int ring_id = workload->ring_id;
> @@ -681,7 +681,7 @@ static int dispatch_workload(struct intel_vgpu_workload *workload)
> ring_id, workload);
>
> mutex_lock(&vgpu->vgpu_lock);
> - mutex_lock(&dev_priv->drm.struct_mutex);
> + mutex_lock(&i915->drm.struct_mutex);
>
> ret = set_context_ppgtt_from_shadow(workload,
> s->shadow[ring_id]->gem_context);
> @@ -723,7 +723,7 @@ static int dispatch_workload(struct intel_vgpu_workload *workload)
> err_req:
> if (ret)
> workload->status = ret;
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + mutex_unlock(&i915->drm.struct_mutex);
> mutex_unlock(&vgpu->vgpu_lock);
> return ret;
> }
> @@ -796,7 +796,7 @@ static void update_guest_context(struct intel_vgpu_workload *workload)
> void *src;
> unsigned long context_gpa, context_page_num;
> int i;
> - struct drm_i915_private *dev_priv = gvt->dev_priv;
> + struct drm_i915_private *i915 = gvt->i915;
> u32 ring_base;
> u32 head, tail;
> u16 wrap_count;
> @@ -817,14 +817,14 @@ static void update_guest_context(struct intel_vgpu_workload *workload)
>
> head = (wrap_count << RB_HEAD_WRAP_CNT_OFF) | tail;
>
> - ring_base = dev_priv->engine[workload->ring_id]->mmio_base;
> + ring_base = i915->engine[workload->ring_id]->mmio_base;
> vgpu_vreg_t(vgpu, RING_TAIL(ring_base)) = tail;
> vgpu_vreg_t(vgpu, RING_HEAD(ring_base)) = head;
>
> context_page_num = rq->engine->context_size;
> context_page_num = context_page_num >> PAGE_SHIFT;
>
> - if (IS_BROADWELL(gvt->dev_priv) && rq->engine->id == RCS0)
> + if (IS_BROADWELL(gvt->i915) && rq->engine->id == RCS0)
> context_page_num = 19;
>
> i = 2;
> @@ -875,13 +875,13 @@ void intel_vgpu_clean_workloads(struct intel_vgpu *vgpu,
> intel_engine_mask_t engine_mask)
> {
> struct intel_vgpu_submission *s = &vgpu->submission;
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
> struct intel_engine_cs *engine;
> struct intel_vgpu_workload *pos, *n;
> intel_engine_mask_t tmp;
>
> /* free the unsubmited workloads in the queues. */
> - for_each_engine_masked(engine, dev_priv, engine_mask, tmp) {
> + for_each_engine_masked(engine, i915, engine_mask, tmp) {
> list_for_each_entry_safe(pos, n,
> &s->workload_q_head[engine->id], list) {
> list_del_init(&pos->list);
> @@ -986,7 +986,7 @@ static int workload_thread(void *priv)
> struct intel_vgpu_workload *workload = NULL;
> struct intel_vgpu *vgpu = NULL;
> int ret;
> - bool need_force_wake = (INTEL_GEN(gvt->dev_priv) >= 9);
> + bool need_force_wake = (INTEL_GEN(gvt->i915) >= 9);
> DEFINE_WAIT_FUNC(wait, woken_wake_function);
>
> kfree(p);
> @@ -1015,7 +1015,7 @@ static int workload_thread(void *priv)
> workload->ring_id, workload);
>
> if (need_force_wake)
> - intel_uncore_forcewake_get(&gvt->dev_priv->uncore,
> + intel_uncore_forcewake_get(&gvt->i915->uncore,
> FORCEWAKE_ALL);
>
> ret = dispatch_workload(workload);
> @@ -1037,7 +1037,7 @@ static int workload_thread(void *priv)
> complete_current_workload(gvt, ring_id);
>
> if (need_force_wake)
> - intel_uncore_forcewake_put(&gvt->dev_priv->uncore,
> + intel_uncore_forcewake_put(&gvt->i915->uncore,
> FORCEWAKE_ALL);
>
> if (ret && (vgpu_is_vm_unhealthy(ret)))
> @@ -1068,7 +1068,7 @@ void intel_gvt_clean_workload_scheduler(struct intel_gvt *gvt)
>
> gvt_dbg_core("clean workload scheduler\n");
>
> - for_each_engine(engine, gvt->dev_priv, i) {
> + for_each_engine(engine, gvt->i915, i) {
> atomic_notifier_chain_unregister(
> &engine->context_status_notifier,
> &gvt->shadow_ctx_notifier_block[i]);
> @@ -1088,7 +1088,7 @@ int intel_gvt_init_workload_scheduler(struct intel_gvt *gvt)
>
> init_waitqueue_head(&scheduler->workload_complete_wq);
>
> - for_each_engine(engine, gvt->dev_priv, i) {
> + for_each_engine(engine, gvt->i915, i) {
> init_waitqueue_head(&scheduler->waitq[i]);
>
> param = kzalloc(sizeof(*param), GFP_KERNEL);
> @@ -1152,7 +1152,7 @@ void intel_vgpu_clean_submission(struct intel_vgpu *vgpu)
> intel_vgpu_select_submission_ops(vgpu, ALL_ENGINES, 0);
>
> i915_context_ppgtt_root_restore(s, i915_vm_to_ppgtt(s->shadow[0]->gem_context->vm));
> - for_each_engine(engine, vgpu->gvt->dev_priv, id)
> + for_each_engine(engine, vgpu->gvt->i915, id)
> intel_context_unpin(s->shadow[id]);
>
> kmem_cache_destroy(s->workloads);
> @@ -1212,13 +1212,13 @@ int intel_vgpu_setup_submission(struct intel_vgpu *vgpu)
> enum intel_engine_id i;
> int ret;
>
> - ctx = i915_gem_context_create_gvt(&vgpu->gvt->dev_priv->drm);
> + ctx = i915_gem_context_create_gvt(&vgpu->gvt->i915->drm);
> if (IS_ERR(ctx))
> return PTR_ERR(ctx);
>
> i915_context_ppgtt_root_save(s, i915_vm_to_ppgtt(ctx->vm));
>
> - for_each_engine(engine, vgpu->gvt->dev_priv, i) {
> + for_each_engine(engine, vgpu->gvt->i915, i) {
> struct intel_context *ce;
>
> INIT_LIST_HEAD(&s->workload_q_head[i]);
> @@ -1260,7 +1260,7 @@ int intel_vgpu_setup_submission(struct intel_vgpu *vgpu)
>
> out_shadow_ctx:
> i915_context_ppgtt_root_restore(s, i915_vm_to_ppgtt(ctx->vm));
> - for_each_engine(engine, vgpu->gvt->dev_priv, i) {
> + for_each_engine(engine, vgpu->gvt->i915, i) {
> if (IS_ERR(s->shadow[i]))
> break;
>
> @@ -1439,7 +1439,7 @@ intel_vgpu_create_workload(struct intel_vgpu *vgpu, int ring_id,
> struct list_head *q = workload_q_head(vgpu, ring_id);
> struct intel_vgpu_workload *last_workload = get_last_workload(q);
> struct intel_vgpu_workload *workload = NULL;
> - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> + struct drm_i915_private *i915 = vgpu->gvt->i915;
> u64 ring_context_gpa;
> u32 head, tail, start, ctl, ctx_ctl, per_ctx, indirect_ctx;
> u32 guest_head;
> @@ -1526,11 +1526,11 @@ intel_vgpu_create_workload(struct intel_vgpu *vgpu, int ring_id,
> * as there is only one pre-allocated buf-obj for shadow.
> */
> if (list_empty(workload_q_head(vgpu, ring_id))) {
> - intel_runtime_pm_get(dev_priv);
> - mutex_lock(&dev_priv->drm.struct_mutex);
> + intel_runtime_pm_get(i915);
> + mutex_lock(&i915->drm.struct_mutex);
> ret = intel_gvt_scan_and_shadow_workload(workload);
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> - intel_runtime_pm_put_unchecked(dev_priv);
> + mutex_unlock(&i915->drm.struct_mutex);
> + intel_runtime_pm_put_unchecked(i915);
> }
>
> if (ret) {
> diff --git a/drivers/gpu/drm/i915/gvt/vgpu.c b/drivers/gpu/drm/i915/gvt/vgpu.c
> index 44ce3c2b9ac1..a311737c9633 100644
> --- a/drivers/gpu/drm/i915/gvt/vgpu.c
> +++ b/drivers/gpu/drm/i915/gvt/vgpu.c
> @@ -148,10 +148,10 @@ int intel_gvt_init_vgpu_types(struct intel_gvt *gvt)
> gvt->types[i].avail_instance = min(low_avail / vgpu_types[i].low_mm,
> high_avail / vgpu_types[i].high_mm);
>
> - if (IS_GEN(gvt->dev_priv, 8))
> + if (IS_GEN(gvt->i915, 8))
> sprintf(gvt->types[i].name, "GVTg_V4_%s",
> vgpu_types[i].name);
> - else if (IS_GEN(gvt->dev_priv, 9))
> + else if (IS_GEN(gvt->i915, 9))
> sprintf(gvt->types[i].name, "GVTg_V5_%s",
> vgpu_types[i].name);
>
> @@ -429,7 +429,7 @@ static struct intel_vgpu *__intel_gvt_create_vgpu(struct intel_gvt *gvt,
> goto out_clean_sched_policy;
>
> /*TODO: add more platforms support */
> - if (IS_SKYLAKE(gvt->dev_priv) || IS_KABYLAKE(gvt->dev_priv))
> + if (IS_SKYLAKE(gvt->i915) || IS_KABYLAKE(gvt->i915))
> ret = intel_gvt_hypervisor_set_edid(vgpu, PORT_D);
> if (ret)
> goto out_clean_sched_policy;
> diff --git a/drivers/gpu/drm/i915/i915_cmd_parser.c b/drivers/gpu/drm/i915/i915_cmd_parser.c
> index a28bcd2d7c09..e9d076167b24 100644
> --- a/drivers/gpu/drm/i915/i915_cmd_parser.c
> +++ b/drivers/gpu/drm/i915/i915_cmd_parser.c
> @@ -1342,21 +1342,21 @@ int intel_engine_cmd_parser(struct intel_engine_cs *engine,
>
> /**
> * i915_cmd_parser_get_version() - get the cmd parser version number
> - * @dev_priv: i915 device private
> + * @i915: i915 device private
> *
> * The cmd parser maintains a simple increasing integer version number suitable
> * for passing to userspace clients to determine what operations are permitted.
> *
> * Return: the current version number of the cmd parser
> */
> -int i915_cmd_parser_get_version(struct drm_i915_private *dev_priv)
> +int i915_cmd_parser_get_version(struct drm_i915_private *i915)
> {
> struct intel_engine_cs *engine;
> enum intel_engine_id id;
> bool active = false;
>
> /* If the command parser is not enabled, report 0 - unsupported */
> - for_each_engine(engine, dev_priv, id) {
> + for_each_engine(engine, i915, id) {
> if (intel_engine_needs_cmd_parser(engine)) {
> active = true;
> break;
> diff --git a/drivers/gpu/drm/i915/i915_debugfs.c b/drivers/gpu/drm/i915/i915_debugfs.c
> index 026fb46020f6..fec01858685d 100644
> --- a/drivers/gpu/drm/i915/i915_debugfs.c
> +++ b/drivers/gpu/drm/i915/i915_debugfs.c
> @@ -55,17 +55,17 @@ static inline struct drm_i915_private *node_to_i915(struct drm_info_node *node)
>
> static int i915_capabilities(struct seq_file *m, void *data)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> - const struct intel_device_info *info = INTEL_INFO(dev_priv);
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> + const struct intel_device_info *info = INTEL_INFO(i915);
> struct drm_printer p = drm_seq_file_printer(m);
>
> - seq_printf(m, "gen: %d\n", INTEL_GEN(dev_priv));
> + seq_printf(m, "gen: %d\n", INTEL_GEN(i915));
> seq_printf(m, "platform: %s\n", intel_platform_name(info->platform));
> - seq_printf(m, "pch: %d\n", INTEL_PCH_TYPE(dev_priv));
> + seq_printf(m, "pch: %d\n", INTEL_PCH_TYPE(i915));
>
> intel_device_info_dump_flags(info, &p);
> - intel_device_info_dump_runtime(RUNTIME_INFO(dev_priv), &p);
> - intel_driver_caps_print(&dev_priv->caps, &p);
> + intel_device_info_dump_runtime(RUNTIME_INFO(i915), &p);
> + intel_driver_caps_print(&i915->caps, &p);
>
> kernel_param_lock(THIS_MODULE);
> i915_params_dump(&i915_modparams, &p);
> @@ -137,7 +137,7 @@ stringify_page_sizes(unsigned int page_sizes, char *buf, size_t len)
> static void
> describe_obj(struct seq_file *m, struct drm_i915_gem_object *obj)
> {
> - struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
> + struct drm_i915_private *i915 = to_i915(obj->base.dev);
> struct intel_engine_cs *engine;
> struct i915_vma *vma;
> unsigned int frontbuffer_bits;
> @@ -155,7 +155,7 @@ describe_obj(struct seq_file *m, struct drm_i915_gem_object *obj)
> obj->base.size / 1024,
> obj->read_domains,
> obj->write_domain,
> - i915_cache_level_str(dev_priv, obj->cache_level),
> + i915_cache_level_str(i915, obj->cache_level),
> obj->mm.dirty ? " dirty" : "",
> obj->mm.madv == I915_MADV_DONTNEED ? " purgeable" : "");
> if (obj->base.name)
> @@ -296,7 +296,7 @@ static int per_file_stats(int id, void *ptr, void *data)
> } while (0)
>
> static void print_batch_pool_stats(struct seq_file *m,
> - struct drm_i915_private *dev_priv)
> + struct drm_i915_private *i915)
> {
> struct drm_i915_gem_object *obj;
> struct intel_engine_cs *engine;
> @@ -304,7 +304,7 @@ static void print_batch_pool_stats(struct seq_file *m,
> enum intel_engine_id id;
> int j;
>
> - for_each_engine(engine, dev_priv, id) {
> + for_each_engine(engine, i915, id) {
> for (j = 0; j < ARRAY_SIZE(engine->batch_pool.cache_list); j++) {
> list_for_each_entry(obj,
> &engine->batch_pool.cache_list[j],
> @@ -382,8 +382,8 @@ static int i915_gem_object_info(struct seq_file *m, void *data)
>
> static int i915_gem_batch_pool_info(struct seq_file *m, void *data)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> + struct drm_device *dev = &i915->drm;
> struct drm_i915_gem_object *obj;
> struct intel_engine_cs *engine;
> enum intel_engine_id id;
> @@ -394,7 +394,7 @@ static int i915_gem_batch_pool_info(struct seq_file *m, void *data)
> if (ret)
> return ret;
>
> - for_each_engine(engine, dev_priv, id) {
> + for_each_engine(engine, i915, id) {
> for (j = 0; j < ARRAY_SIZE(engine->batch_pool.cache_list); j++) {
> int count;
>
> @@ -427,15 +427,15 @@ static int i915_gem_batch_pool_info(struct seq_file *m, void *data)
>
> static void gen8_display_interrupt_info(struct seq_file *m)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> int pipe;
>
> - for_each_pipe(dev_priv, pipe) {
> + for_each_pipe(i915, pipe) {
> enum intel_display_power_domain power_domain;
> intel_wakeref_t wakeref;
>
> power_domain = POWER_DOMAIN_PIPE(pipe);
> - wakeref = intel_display_power_get_if_enabled(dev_priv,
> + wakeref = intel_display_power_get_if_enabled(i915,
> power_domain);
> if (!wakeref) {
> seq_printf(m, "Pipe %c power disabled\n",
> @@ -452,7 +452,7 @@ static void gen8_display_interrupt_info(struct seq_file *m)
> pipe_name(pipe),
> I915_READ(GEN8_DE_PIPE_IER(pipe)));
>
> - intel_display_power_put(dev_priv, power_domain, wakeref);
> + intel_display_power_put(i915, power_domain, wakeref);
> }
>
> seq_printf(m, "Display Engine port interrupt mask:\t%08x\n",
> @@ -479,15 +479,15 @@ static void gen8_display_interrupt_info(struct seq_file *m)
>
> static int i915_interrupt_info(struct seq_file *m, void *data)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> struct intel_engine_cs *engine;
> enum intel_engine_id id;
> intel_wakeref_t wakeref;
> int i, pipe;
>
> - wakeref = intel_runtime_pm_get(dev_priv);
> + wakeref = intel_runtime_pm_get(i915);
>
> - if (IS_CHERRYVIEW(dev_priv)) {
> + if (IS_CHERRYVIEW(i915)) {
> intel_wakeref_t pref;
>
> seq_printf(m, "Master Interrupt Control:\t%08x\n",
> @@ -501,11 +501,11 @@ static int i915_interrupt_info(struct seq_file *m, void *data)
> I915_READ(VLV_IIR_RW));
> seq_printf(m, "Display IMR:\t%08x\n",
> I915_READ(VLV_IMR));
> - for_each_pipe(dev_priv, pipe) {
> + for_each_pipe(i915, pipe) {
> enum intel_display_power_domain power_domain;
>
> power_domain = POWER_DOMAIN_PIPE(pipe);
> - pref = intel_display_power_get_if_enabled(dev_priv,
> + pref = intel_display_power_get_if_enabled(i915,
> power_domain);
> if (!pref) {
> seq_printf(m, "Pipe %c power disabled\n",
> @@ -517,17 +517,17 @@ static int i915_interrupt_info(struct seq_file *m, void *data)
> pipe_name(pipe),
> I915_READ(PIPESTAT(pipe)));
>
> - intel_display_power_put(dev_priv, power_domain, pref);
> + intel_display_power_put(i915, power_domain, pref);
> }
>
> - pref = intel_display_power_get(dev_priv, POWER_DOMAIN_INIT);
> + pref = intel_display_power_get(i915, POWER_DOMAIN_INIT);
> seq_printf(m, "Port hotplug:\t%08x\n",
> I915_READ(PORT_HOTPLUG_EN));
> seq_printf(m, "DPFLIPSTAT:\t%08x\n",
> I915_READ(VLV_DPFLIPSTAT));
> seq_printf(m, "DPINVGTT:\t%08x\n",
> I915_READ(DPINVGTT));
> - intel_display_power_put(dev_priv, POWER_DOMAIN_INIT, pref);
> + intel_display_power_put(i915, POWER_DOMAIN_INIT, pref);
>
> for (i = 0; i < 4; i++) {
> seq_printf(m, "GT Interrupt IMR %d:\t%08x\n",
> @@ -544,7 +544,7 @@ static int i915_interrupt_info(struct seq_file *m, void *data)
> I915_READ(GEN8_PCU_IIR));
> seq_printf(m, "PCU interrupt enable:\t%08x\n",
> I915_READ(GEN8_PCU_IER));
> - } else if (INTEL_GEN(dev_priv) >= 11) {
> + } else if (INTEL_GEN(i915) >= 11) {
> seq_printf(m, "Master Interrupt Control: %08x\n",
> I915_READ(GEN11_GFX_MSTR_IRQ));
>
> @@ -565,7 +565,7 @@ static int i915_interrupt_info(struct seq_file *m, void *data)
> I915_READ(GEN11_DISPLAY_INT_CTL));
>
> gen8_display_interrupt_info(m);
> - } else if (INTEL_GEN(dev_priv) >= 8) {
> + } else if (INTEL_GEN(i915) >= 8) {
> seq_printf(m, "Master Interrupt Control:\t%08x\n",
> I915_READ(GEN8_MASTER_IRQ));
>
> @@ -579,7 +579,7 @@ static int i915_interrupt_info(struct seq_file *m, void *data)
> }
>
> gen8_display_interrupt_info(m);
> - } else if (IS_VALLEYVIEW(dev_priv)) {
> + } else if (IS_VALLEYVIEW(i915)) {
> seq_printf(m, "Display IER:\t%08x\n",
> I915_READ(VLV_IER));
> seq_printf(m, "Display IIR:\t%08x\n",
> @@ -588,12 +588,12 @@ static int i915_interrupt_info(struct seq_file *m, void *data)
> I915_READ(VLV_IIR_RW));
> seq_printf(m, "Display IMR:\t%08x\n",
> I915_READ(VLV_IMR));
> - for_each_pipe(dev_priv, pipe) {
> + for_each_pipe(i915, pipe) {
> enum intel_display_power_domain power_domain;
> intel_wakeref_t pref;
>
> power_domain = POWER_DOMAIN_PIPE(pipe);
> - pref = intel_display_power_get_if_enabled(dev_priv,
> + pref = intel_display_power_get_if_enabled(i915,
> power_domain);
> if (!pref) {
> seq_printf(m, "Pipe %c power disabled\n",
> @@ -604,7 +604,7 @@ static int i915_interrupt_info(struct seq_file *m, void *data)
> seq_printf(m, "Pipe %c stat:\t%08x\n",
> pipe_name(pipe),
> I915_READ(PIPESTAT(pipe)));
> - intel_display_power_put(dev_priv, power_domain, pref);
> + intel_display_power_put(i915, power_domain, pref);
> }
>
> seq_printf(m, "Master IER:\t%08x\n",
> @@ -631,14 +631,14 @@ static int i915_interrupt_info(struct seq_file *m, void *data)
> seq_printf(m, "DPINVGTT:\t%08x\n",
> I915_READ(DPINVGTT));
>
> - } else if (!HAS_PCH_SPLIT(dev_priv)) {
> + } else if (!HAS_PCH_SPLIT(i915)) {
> seq_printf(m, "Interrupt enable: %08x\n",
> I915_READ(GEN2_IER));
> seq_printf(m, "Interrupt identity: %08x\n",
> I915_READ(GEN2_IIR));
> seq_printf(m, "Interrupt mask: %08x\n",
> I915_READ(GEN2_IMR));
> - for_each_pipe(dev_priv, pipe)
> + for_each_pipe(i915, pipe)
> seq_printf(m, "Pipe %c stat: %08x\n",
> pipe_name(pipe),
> I915_READ(PIPESTAT(pipe)));
> @@ -663,7 +663,7 @@ static int i915_interrupt_info(struct seq_file *m, void *data)
> I915_READ(GTIMR));
> }
>
> - if (INTEL_GEN(dev_priv) >= 11) {
> + if (INTEL_GEN(i915) >= 11) {
> seq_printf(m, "RCS Intr Mask:\t %08x\n",
> I915_READ(GEN11_RCS0_RSVD_INTR_MASK));
> seq_printf(m, "BCS Intr Mask:\t %08x\n",
> @@ -683,35 +683,35 @@ static int i915_interrupt_info(struct seq_file *m, void *data)
> seq_printf(m, "Gunit/CSME Intr Mask:\t %08x\n",
> I915_READ(GEN11_GUNIT_CSME_INTR_MASK));
>
> - } else if (INTEL_GEN(dev_priv) >= 6) {
> - for_each_engine(engine, dev_priv, id) {
> + } else if (INTEL_GEN(i915) >= 6) {
> + for_each_engine(engine, i915, id) {
> seq_printf(m,
> "Graphics Interrupt mask (%s): %08x\n",
> engine->name, ENGINE_READ(engine, RING_IMR));
> }
> }
>
> - intel_runtime_pm_put(dev_priv, wakeref);
> + intel_runtime_pm_put(i915, wakeref);
>
> return 0;
> }
>
> static int i915_gem_fence_regs_info(struct seq_file *m, void *data)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> + struct drm_device *dev = &i915->drm;
> int i, ret;
>
> ret = mutex_lock_interruptible(&dev->struct_mutex);
> if (ret)
> return ret;
>
> - seq_printf(m, "Total fences = %d\n", dev_priv->num_fence_regs);
> - for (i = 0; i < dev_priv->num_fence_regs; i++) {
> - struct i915_vma *vma = dev_priv->fence_regs[i].vma;
> + seq_printf(m, "Total fences = %d\n", i915->num_fence_regs);
> + for (i = 0; i < i915->num_fence_regs; i++) {
> + struct i915_vma *vma = i915->fence_regs[i].vma;
>
> seq_printf(m, "Fence %d, pin count = %d, object = ",
> - i, dev_priv->fence_regs[i].pin_count);
> + i, i915->fence_regs[i].pin_count);
> if (!vma)
> seq_puts(m, "unused");
> else
> @@ -825,15 +825,15 @@ static const struct file_operations i915_error_state_fops = {
>
> static int i915_frequency_info(struct seq_file *m, void *unused)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> - struct intel_uncore *uncore = &dev_priv->uncore;
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> + struct intel_uncore *uncore = &i915->uncore;
> + struct intel_rps *rps = &i915->gt_pm.rps;
> intel_wakeref_t wakeref;
> int ret = 0;
>
> - wakeref = intel_runtime_pm_get(dev_priv);
> + wakeref = intel_runtime_pm_get(i915);
>
> - if (IS_GEN(dev_priv, 5)) {
> + if (IS_GEN(i915, 5)) {
> u16 rgvswctl = intel_uncore_read16(uncore, MEMSWCTL);
> u16 rgvstat = intel_uncore_read16(uncore, MEMSTAT_ILK);
>
> @@ -843,7 +843,7 @@ static int i915_frequency_info(struct seq_file *m, void *unused)
> MEMSTAT_VID_SHIFT);
> seq_printf(m, "Current P-state: %d\n",
> (rgvstat & MEMSTAT_PSTATE_MASK) >> MEMSTAT_PSTATE_SHIFT);
> - } else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> + } else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> u32 rpmodectl, freq_sts;
>
> rpmodectl = I915_READ(GEN6_RP_CONTROL);
> @@ -855,32 +855,32 @@ static int i915_frequency_info(struct seq_file *m, void *unused)
> yesno((rpmodectl & GEN6_RP_MEDIA_MODE_MASK) ==
> GEN6_RP_MEDIA_SW_MODE));
>
> - vlv_punit_get(dev_priv);
> - freq_sts = vlv_punit_read(dev_priv, PUNIT_REG_GPU_FREQ_STS);
> - vlv_punit_put(dev_priv);
> + vlv_punit_get(i915);
> + freq_sts = vlv_punit_read(i915, PUNIT_REG_GPU_FREQ_STS);
> + vlv_punit_put(i915);
>
> seq_printf(m, "PUNIT_REG_GPU_FREQ_STS: 0x%08x\n", freq_sts);
> - seq_printf(m, "DDR freq: %d MHz\n", dev_priv->mem_freq);
> + seq_printf(m, "DDR freq: %d MHz\n", i915->mem_freq);
>
> seq_printf(m, "actual GPU freq: %d MHz\n",
> - intel_gpu_freq(dev_priv, (freq_sts >> 8) & 0xff));
> + intel_gpu_freq(i915, (freq_sts >> 8) & 0xff));
>
> seq_printf(m, "current GPU freq: %d MHz\n",
> - intel_gpu_freq(dev_priv, rps->cur_freq));
> + intel_gpu_freq(i915, rps->cur_freq));
>
> seq_printf(m, "max GPU freq: %d MHz\n",
> - intel_gpu_freq(dev_priv, rps->max_freq));
> + intel_gpu_freq(i915, rps->max_freq));
>
> seq_printf(m, "min GPU freq: %d MHz\n",
> - intel_gpu_freq(dev_priv, rps->min_freq));
> + intel_gpu_freq(i915, rps->min_freq));
>
> seq_printf(m, "idle GPU freq: %d MHz\n",
> - intel_gpu_freq(dev_priv, rps->idle_freq));
> + intel_gpu_freq(i915, rps->idle_freq));
>
> seq_printf(m,
> "efficient (RPe) frequency: %d MHz\n",
> - intel_gpu_freq(dev_priv, rps->efficient_freq));
> - } else if (INTEL_GEN(dev_priv) >= 6) {
> + intel_gpu_freq(i915, rps->efficient_freq));
> + } else if (INTEL_GEN(i915) >= 6) {
> u32 rp_state_limits;
> u32 gt_perf_status;
> u32 rp_state_cap;
> @@ -892,7 +892,7 @@ static int i915_frequency_info(struct seq_file *m, void *unused)
> int max_freq;
>
> rp_state_limits = I915_READ(GEN6_RP_STATE_LIMITS);
> - if (IS_GEN9_LP(dev_priv)) {
> + if (IS_GEN9_LP(i915)) {
> rp_state_cap = I915_READ(BXT_RP_STATE_CAP);
> gt_perf_status = I915_READ(BXT_GT_PERF_STATUS);
> } else {
> @@ -901,19 +901,19 @@ static int i915_frequency_info(struct seq_file *m, void *unused)
> }
>
> /* RPSTAT1 is in the GT power well */
> - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
>
> reqf = I915_READ(GEN6_RPNSWREQ);
> - if (INTEL_GEN(dev_priv) >= 9)
> + if (INTEL_GEN(i915) >= 9)
> reqf >>= 23;
> else {
> reqf &= ~GEN6_TURBO_DISABLE;
> - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> + if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> reqf >>= 24;
> else
> reqf >>= 25;
> }
> - reqf = intel_gpu_freq(dev_priv, reqf);
> + reqf = intel_gpu_freq(i915, reqf);
>
> rpmodectl = I915_READ(GEN6_RP_CONTROL);
> rpinclimit = I915_READ(GEN6_RP_UP_THRESHOLD);
> @@ -926,12 +926,12 @@ static int i915_frequency_info(struct seq_file *m, void *unused)
> rpdownei = I915_READ(GEN6_RP_CUR_DOWN_EI) & GEN6_CURIAVG_MASK;
> rpcurdown = I915_READ(GEN6_RP_CUR_DOWN) & GEN6_CURBSYTAVG_MASK;
> rpprevdown = I915_READ(GEN6_RP_PREV_DOWN) & GEN6_CURBSYTAVG_MASK;
> - cagf = intel_gpu_freq(dev_priv,
> - intel_get_cagf(dev_priv, rpstat));
> + cagf = intel_gpu_freq(i915,
> + intel_get_cagf(i915, rpstat));
>
> - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
>
> - if (INTEL_GEN(dev_priv) >= 11) {
> + if (INTEL_GEN(i915) >= 11) {
> pm_ier = I915_READ(GEN11_GPM_WGBOXPERF_INTR_ENABLE);
> pm_imr = I915_READ(GEN11_GPM_WGBOXPERF_INTR_MASK);
> /*
> @@ -940,7 +940,7 @@ static int i915_frequency_info(struct seq_file *m, void *unused)
> */
> pm_isr = 0;
> pm_iir = 0;
> - } else if (INTEL_GEN(dev_priv) >= 8) {
> + } else if (INTEL_GEN(i915) >= 8) {
> pm_ier = I915_READ(GEN8_GT_IER(2));
> pm_imr = I915_READ(GEN8_GT_IMR(2));
> pm_isr = I915_READ(GEN8_GT_ISR(2));
> @@ -963,14 +963,14 @@ static int i915_frequency_info(struct seq_file *m, void *unused)
>
> seq_printf(m, "PM IER=0x%08x IMR=0x%08x, MASK=0x%08x\n",
> pm_ier, pm_imr, pm_mask);
> - if (INTEL_GEN(dev_priv) <= 10)
> + if (INTEL_GEN(i915) <= 10)
> seq_printf(m, "PM ISR=0x%08x IIR=0x%08x\n",
> pm_isr, pm_iir);
> seq_printf(m, "pm_intrmsk_mbz: 0x%08x\n",
> rps->pm_intrmsk_mbz);
> seq_printf(m, "GT_PERF_STATUS: 0x%08x\n", gt_perf_status);
> seq_printf(m, "Render p-state ratio: %d\n",
> - (gt_perf_status & (INTEL_GEN(dev_priv) >= 9 ? 0x1ff00 : 0xff00)) >> 8);
> + (gt_perf_status & (INTEL_GEN(i915) >= 9 ? 0x1ff00 : 0xff00)) >> 8);
> seq_printf(m, "Render p-state VID: %d\n",
> gt_perf_status & 0xff);
> seq_printf(m, "Render p-state limit: %d\n",
> @@ -982,72 +982,72 @@ static int i915_frequency_info(struct seq_file *m, void *unused)
> seq_printf(m, "RPNSWREQ: %dMHz\n", reqf);
> seq_printf(m, "CAGF: %dMHz\n", cagf);
> seq_printf(m, "RP CUR UP EI: %d (%dus)\n",
> - rpupei, GT_PM_INTERVAL_TO_US(dev_priv, rpupei));
> + rpupei, GT_PM_INTERVAL_TO_US(i915, rpupei));
> seq_printf(m, "RP CUR UP: %d (%dus)\n",
> - rpcurup, GT_PM_INTERVAL_TO_US(dev_priv, rpcurup));
> + rpcurup, GT_PM_INTERVAL_TO_US(i915, rpcurup));
> seq_printf(m, "RP PREV UP: %d (%dus)\n",
> - rpprevup, GT_PM_INTERVAL_TO_US(dev_priv, rpprevup));
> + rpprevup, GT_PM_INTERVAL_TO_US(i915, rpprevup));
> seq_printf(m, "Up threshold: %d%%\n",
> rps->power.up_threshold);
>
> seq_printf(m, "RP CUR DOWN EI: %d (%dus)\n",
> - rpdownei, GT_PM_INTERVAL_TO_US(dev_priv, rpdownei));
> + rpdownei, GT_PM_INTERVAL_TO_US(i915, rpdownei));
> seq_printf(m, "RP CUR DOWN: %d (%dus)\n",
> - rpcurdown, GT_PM_INTERVAL_TO_US(dev_priv, rpcurdown));
> + rpcurdown, GT_PM_INTERVAL_TO_US(i915, rpcurdown));
> seq_printf(m, "RP PREV DOWN: %d (%dus)\n",
> - rpprevdown, GT_PM_INTERVAL_TO_US(dev_priv, rpprevdown));
> + rpprevdown, GT_PM_INTERVAL_TO_US(i915, rpprevdown));
> seq_printf(m, "Down threshold: %d%%\n",
> rps->power.down_threshold);
>
> - max_freq = (IS_GEN9_LP(dev_priv) ? rp_state_cap >> 0 :
> + max_freq = (IS_GEN9_LP(i915) ? rp_state_cap >> 0 :
> rp_state_cap >> 16) & 0xff;
> - max_freq *= (IS_GEN9_BC(dev_priv) ||
> - INTEL_GEN(dev_priv) >= 10 ? GEN9_FREQ_SCALER : 1);
> + max_freq *= (IS_GEN9_BC(i915) ||
> + INTEL_GEN(i915) >= 10 ? GEN9_FREQ_SCALER : 1);
> seq_printf(m, "Lowest (RPN) frequency: %dMHz\n",
> - intel_gpu_freq(dev_priv, max_freq));
> + intel_gpu_freq(i915, max_freq));
>
> max_freq = (rp_state_cap & 0xff00) >> 8;
> - max_freq *= (IS_GEN9_BC(dev_priv) ||
> - INTEL_GEN(dev_priv) >= 10 ? GEN9_FREQ_SCALER : 1);
> + max_freq *= (IS_GEN9_BC(i915) ||
> + INTEL_GEN(i915) >= 10 ? GEN9_FREQ_SCALER : 1);
> seq_printf(m, "Nominal (RP1) frequency: %dMHz\n",
> - intel_gpu_freq(dev_priv, max_freq));
> + intel_gpu_freq(i915, max_freq));
>
> - max_freq = (IS_GEN9_LP(dev_priv) ? rp_state_cap >> 16 :
> + max_freq = (IS_GEN9_LP(i915) ? rp_state_cap >> 16 :
> rp_state_cap >> 0) & 0xff;
> - max_freq *= (IS_GEN9_BC(dev_priv) ||
> - INTEL_GEN(dev_priv) >= 10 ? GEN9_FREQ_SCALER : 1);
> + max_freq *= (IS_GEN9_BC(i915) ||
> + INTEL_GEN(i915) >= 10 ? GEN9_FREQ_SCALER : 1);
> seq_printf(m, "Max non-overclocked (RP0) frequency: %dMHz\n",
> - intel_gpu_freq(dev_priv, max_freq));
> + intel_gpu_freq(i915, max_freq));
> seq_printf(m, "Max overclocked frequency: %dMHz\n",
> - intel_gpu_freq(dev_priv, rps->max_freq));
> + intel_gpu_freq(i915, rps->max_freq));
>
> seq_printf(m, "Current freq: %d MHz\n",
> - intel_gpu_freq(dev_priv, rps->cur_freq));
> + intel_gpu_freq(i915, rps->cur_freq));
> seq_printf(m, "Actual freq: %d MHz\n", cagf);
> seq_printf(m, "Idle freq: %d MHz\n",
> - intel_gpu_freq(dev_priv, rps->idle_freq));
> + intel_gpu_freq(i915, rps->idle_freq));
> seq_printf(m, "Min freq: %d MHz\n",
> - intel_gpu_freq(dev_priv, rps->min_freq));
> + intel_gpu_freq(i915, rps->min_freq));
> seq_printf(m, "Boost freq: %d MHz\n",
> - intel_gpu_freq(dev_priv, rps->boost_freq));
> + intel_gpu_freq(i915, rps->boost_freq));
> seq_printf(m, "Max freq: %d MHz\n",
> - intel_gpu_freq(dev_priv, rps->max_freq));
> + intel_gpu_freq(i915, rps->max_freq));
> seq_printf(m,
> "efficient (RPe) frequency: %d MHz\n",
> - intel_gpu_freq(dev_priv, rps->efficient_freq));
> + intel_gpu_freq(i915, rps->efficient_freq));
> } else {
> seq_puts(m, "no P-state info available\n");
> }
>
> - seq_printf(m, "Current CD clock frequency: %d kHz\n", dev_priv->cdclk.hw.cdclk);
> - seq_printf(m, "Max CD clock frequency: %d kHz\n", dev_priv->max_cdclk_freq);
> - seq_printf(m, "Max pixel clock frequency: %d kHz\n", dev_priv->max_dotclk_freq);
> + seq_printf(m, "Current CD clock frequency: %d kHz\n", i915->cdclk.hw.cdclk);
> + seq_printf(m, "Max CD clock frequency: %d kHz\n", i915->max_cdclk_freq);
> + seq_printf(m, "Max pixel clock frequency: %d kHz\n", i915->max_dotclk_freq);
>
> - intel_runtime_pm_put(dev_priv, wakeref);
> + intel_runtime_pm_put(i915, wakeref);
> return ret;
> }
>
> -static void i915_instdone_info(struct drm_i915_private *dev_priv,
> +static void i915_instdone_info(struct drm_i915_private *i915,
> struct seq_file *m,
> struct intel_instdone *instdone)
> {
> @@ -1057,37 +1057,37 @@ static void i915_instdone_info(struct drm_i915_private *dev_priv,
> seq_printf(m, "\t\tINSTDONE: 0x%08x\n",
> instdone->instdone);
>
> - if (INTEL_GEN(dev_priv) <= 3)
> + if (INTEL_GEN(i915) <= 3)
> return;
>
> seq_printf(m, "\t\tSC_INSTDONE: 0x%08x\n",
> instdone->slice_common);
>
> - if (INTEL_GEN(dev_priv) <= 6)
> + if (INTEL_GEN(i915) <= 6)
> return;
>
> - for_each_instdone_slice_subslice(dev_priv, slice, subslice)
> + for_each_instdone_slice_subslice(i915, slice, subslice)
> seq_printf(m, "\t\tSAMPLER_INSTDONE[%d][%d]: 0x%08x\n",
> slice, subslice, instdone->sampler[slice][subslice]);
>
> - for_each_instdone_slice_subslice(dev_priv, slice, subslice)
> + for_each_instdone_slice_subslice(i915, slice, subslice)
> seq_printf(m, "\t\tROW_INSTDONE[%d][%d]: 0x%08x\n",
> slice, subslice, instdone->row[slice][subslice]);
> }
>
> static int i915_hangcheck_info(struct seq_file *m, void *unused)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> struct intel_engine_cs *engine;
> u64 acthd[I915_NUM_ENGINES];
> struct intel_instdone instdone;
> intel_wakeref_t wakeref;
> enum intel_engine_id id;
>
> - seq_printf(m, "Reset flags: %lx\n", dev_priv->gpu_error.flags);
> - if (test_bit(I915_WEDGED, &dev_priv->gpu_error.flags))
> + seq_printf(m, "Reset flags: %lx\n", i915->gpu_error.flags);
> + if (test_bit(I915_WEDGED, &i915->gpu_error.flags))
> seq_puts(m, "\tWedged\n");
> - if (test_bit(I915_RESET_BACKOFF, &dev_priv->gpu_error.flags))
> + if (test_bit(I915_RESET_BACKOFF, &i915->gpu_error.flags))
> seq_puts(m, "\tDevice (global) reset in progress\n");
>
> if (!i915_modparams.enable_hangcheck) {
> @@ -1095,25 +1095,25 @@ static int i915_hangcheck_info(struct seq_file *m, void *unused)
> return 0;
> }
>
> - with_intel_runtime_pm(dev_priv, wakeref) {
> - for_each_engine(engine, dev_priv, id)
> + with_intel_runtime_pm(i915, wakeref) {
> + for_each_engine(engine, i915, id)
> acthd[id] = intel_engine_get_active_head(engine);
>
> - intel_engine_get_instdone(dev_priv->engine[RCS0], &instdone);
> + intel_engine_get_instdone(i915->engine[RCS0], &instdone);
> }
>
> - if (timer_pending(&dev_priv->gpu_error.hangcheck_work.timer))
> + if (timer_pending(&i915->gpu_error.hangcheck_work.timer))
> seq_printf(m, "Hangcheck active, timer fires in %dms\n",
> - jiffies_to_msecs(dev_priv->gpu_error.hangcheck_work.timer.expires -
> + jiffies_to_msecs(i915->gpu_error.hangcheck_work.timer.expires -
> jiffies));
> - else if (delayed_work_pending(&dev_priv->gpu_error.hangcheck_work))
> + else if (delayed_work_pending(&i915->gpu_error.hangcheck_work))
> seq_puts(m, "Hangcheck active, work pending\n");
> else
> seq_puts(m, "Hangcheck inactive\n");
>
> - seq_printf(m, "GT active? %s\n", yesno(dev_priv->gt.awake));
> + seq_printf(m, "GT active? %s\n", yesno(i915->gt.awake));
>
> - for_each_engine(engine, dev_priv, id) {
> + for_each_engine(engine, i915, id) {
> seq_printf(m, "%s: %d ms ago\n",
> engine->name,
> jiffies_to_msecs(jiffies -
> @@ -1126,11 +1126,11 @@ static int i915_hangcheck_info(struct seq_file *m, void *unused)
> if (engine->id == RCS0) {
> seq_puts(m, "\tinstdone read =\n");
>
> - i915_instdone_info(dev_priv, m, &instdone);
> + i915_instdone_info(i915, m, &instdone);
>
> seq_puts(m, "\tinstdone accu =\n");
>
> - i915_instdone_info(dev_priv, m,
> + i915_instdone_info(i915, m,
> &engine->hangcheck.instdone);
> }
> }
> @@ -1140,14 +1140,14 @@ static int i915_hangcheck_info(struct seq_file *m, void *unused)
>
> static int i915_reset_info(struct seq_file *m, void *unused)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> - struct i915_gpu_error *error = &dev_priv->gpu_error;
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> + struct i915_gpu_error *error = &i915->gpu_error;
> struct intel_engine_cs *engine;
> enum intel_engine_id id;
>
> seq_printf(m, "full gpu reset = %u\n", i915_reset_count(error));
>
> - for_each_engine(engine, dev_priv, id) {
> + for_each_engine(engine, i915, id) {
> seq_printf(m, "%s = %u\n", engine->name,
> i915_reset_engine_count(error, engine));
> }
> @@ -1235,16 +1235,16 @@ static void print_rc6_res(struct seq_file *m,
> const char *title,
> const i915_reg_t reg)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> + struct drm_i915_private *i915 = node_to_i915(m->private);
>
> seq_printf(m, "%s %u (%llu us)\n",
> title, I915_READ(reg),
> - intel_rc6_residency_us(dev_priv, reg));
> + intel_rc6_residency_us(i915, reg));
> }
>
> static int vlv_drpc_info(struct seq_file *m)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> u32 rcctl1, pw_status;
>
> pw_status = I915_READ(VLV_GTLC_PW_STATUS);
> @@ -1266,7 +1266,7 @@ static int vlv_drpc_info(struct seq_file *m)
>
> static int gen6_drpc_info(struct seq_file *m)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> u32 gt_core_status, rcctl1, rc6vids = 0;
> u32 gen9_powergate_enable = 0, gen9_powergate_status = 0;
>
> @@ -1274,20 +1274,20 @@ static int gen6_drpc_info(struct seq_file *m)
> trace_i915_reg_rw(false, GEN6_GT_CORE_STATUS, gt_core_status, 4, true);
>
> rcctl1 = I915_READ(GEN6_RC_CONTROL);
> - if (INTEL_GEN(dev_priv) >= 9) {
> + if (INTEL_GEN(i915) >= 9) {
> gen9_powergate_enable = I915_READ(GEN9_PG_ENABLE);
> gen9_powergate_status = I915_READ(GEN9_PWRGT_DOMAIN_STATUS);
> }
>
> - if (INTEL_GEN(dev_priv) <= 7)
> - sandybridge_pcode_read(dev_priv, GEN6_PCODE_READ_RC6VIDS,
> + if (INTEL_GEN(i915) <= 7)
> + sandybridge_pcode_read(i915, GEN6_PCODE_READ_RC6VIDS,
> &rc6vids, NULL);
>
> seq_printf(m, "RC1e Enabled: %s\n",
> yesno(rcctl1 & GEN6_RC_CTL_RC1e_ENABLE));
> seq_printf(m, "RC6 Enabled: %s\n",
> yesno(rcctl1 & GEN6_RC_CTL_RC6_ENABLE));
> - if (INTEL_GEN(dev_priv) >= 9) {
> + if (INTEL_GEN(i915) >= 9) {
> seq_printf(m, "Render Well Gating Enabled: %s\n",
> yesno(gen9_powergate_enable & GEN9_RENDER_PG_ENABLE));
> seq_printf(m, "Media Well Gating Enabled: %s\n",
> @@ -1321,7 +1321,7 @@ static int gen6_drpc_info(struct seq_file *m)
>
> seq_printf(m, "Core Power Down: %s\n",
> yesno(gt_core_status & GEN6_CORE_CPD_STATE_MASK));
> - if (INTEL_GEN(dev_priv) >= 9) {
> + if (INTEL_GEN(i915) >= 9) {
> seq_printf(m, "Render Power Well: %s\n",
> (gen9_powergate_status &
> GEN9_PWRGT_RENDER_STATUS_MASK) ? "Up" : "Down");
> @@ -1337,7 +1337,7 @@ static int gen6_drpc_info(struct seq_file *m)
> print_rc6_res(m, "RC6+ residency since boot:", GEN6_GT_GFX_RC6p);
> print_rc6_res(m, "RC6++ residency since boot:", GEN6_GT_GFX_RC6pp);
>
> - if (INTEL_GEN(dev_priv) <= 7) {
> + if (INTEL_GEN(i915) <= 7) {
> seq_printf(m, "RC6 voltage: %dmV\n",
> GEN6_DECODE_RC6_VID(((rc6vids >> 0) & 0xff)));
> seq_printf(m, "RC6+ voltage: %dmV\n",
> @@ -1351,14 +1351,14 @@ static int gen6_drpc_info(struct seq_file *m)
>
> static int i915_drpc_info(struct seq_file *m, void *unused)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> intel_wakeref_t wakeref;
> int err = -ENODEV;
>
> - with_intel_runtime_pm(dev_priv, wakeref) {
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> + with_intel_runtime_pm(i915, wakeref) {
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> err = vlv_drpc_info(m);
> - else if (INTEL_GEN(dev_priv) >= 6)
> + else if (INTEL_GEN(i915) >= 6)
> err = gen6_drpc_info(m);
> else
> err = ironlake_drpc_info(m);
> @@ -1369,44 +1369,44 @@ static int i915_drpc_info(struct seq_file *m, void *unused)
>
> static int i915_frontbuffer_tracking(struct seq_file *m, void *unused)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> + struct drm_i915_private *i915 = node_to_i915(m->private);
>
> seq_printf(m, "FB tracking busy bits: 0x%08x\n",
> - dev_priv->fb_tracking.busy_bits);
> + i915->fb_tracking.busy_bits);
>
> seq_printf(m, "FB tracking flip bits: 0x%08x\n",
> - dev_priv->fb_tracking.flip_bits);
> + i915->fb_tracking.flip_bits);
>
> return 0;
> }
>
> static int i915_fbc_status(struct seq_file *m, void *unused)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> - struct intel_fbc *fbc = &dev_priv->fbc;
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> + struct intel_fbc *fbc = &i915->fbc;
> intel_wakeref_t wakeref;
>
> - if (!HAS_FBC(dev_priv))
> + if (!HAS_FBC(i915))
> return -ENODEV;
>
> - wakeref = intel_runtime_pm_get(dev_priv);
> + wakeref = intel_runtime_pm_get(i915);
> mutex_lock(&fbc->lock);
>
> - if (intel_fbc_is_active(dev_priv))
> + if (intel_fbc_is_active(i915))
> seq_puts(m, "FBC enabled\n");
> else
> seq_printf(m, "FBC disabled: %s\n", fbc->no_fbc_reason);
>
> - if (intel_fbc_is_active(dev_priv)) {
> + if (intel_fbc_is_active(i915)) {
> u32 mask;
>
> - if (INTEL_GEN(dev_priv) >= 8)
> + if (INTEL_GEN(i915) >= 8)
> mask = I915_READ(IVB_FBC_STATUS2) & BDW_FBC_COMP_SEG_MASK;
> - else if (INTEL_GEN(dev_priv) >= 7)
> + else if (INTEL_GEN(i915) >= 7)
> mask = I915_READ(IVB_FBC_STATUS2) & IVB_FBC_COMP_SEG_MASK;
> - else if (INTEL_GEN(dev_priv) >= 5)
> + else if (INTEL_GEN(i915) >= 5)
> mask = I915_READ(ILK_DPFC_STATUS) & ILK_DPFC_COMP_SEG_MASK;
> - else if (IS_G4X(dev_priv))
> + else if (IS_G4X(i915))
> mask = I915_READ(DPFC_STATUS) & DPFC_COMP_SEG_MASK;
> else
> mask = I915_READ(FBC_STATUS) & (FBC_STAT_COMPRESSING |
> @@ -1416,41 +1416,41 @@ static int i915_fbc_status(struct seq_file *m, void *unused)
> }
>
> mutex_unlock(&fbc->lock);
> - intel_runtime_pm_put(dev_priv, wakeref);
> + intel_runtime_pm_put(i915, wakeref);
>
> return 0;
> }
>
> static int i915_fbc_false_color_get(void *data, u64 *val)
> {
> - struct drm_i915_private *dev_priv = data;
> + struct drm_i915_private *i915 = data;
>
> - if (INTEL_GEN(dev_priv) < 7 || !HAS_FBC(dev_priv))
> + if (INTEL_GEN(i915) < 7 || !HAS_FBC(i915))
> return -ENODEV;
>
> - *val = dev_priv->fbc.false_color;
> + *val = i915->fbc.false_color;
>
> return 0;
> }
>
> static int i915_fbc_false_color_set(void *data, u64 val)
> {
> - struct drm_i915_private *dev_priv = data;
> + struct drm_i915_private *i915 = data;
> u32 reg;
>
> - if (INTEL_GEN(dev_priv) < 7 || !HAS_FBC(dev_priv))
> + if (INTEL_GEN(i915) < 7 || !HAS_FBC(i915))
> return -ENODEV;
>
> - mutex_lock(&dev_priv->fbc.lock);
> + mutex_lock(&i915->fbc.lock);
>
> reg = I915_READ(ILK_DPFC_CONTROL);
> - dev_priv->fbc.false_color = val;
> + i915->fbc.false_color = val;
>
> I915_WRITE(ILK_DPFC_CONTROL, val ?
> (reg | FBC_CTL_FALSE_COLOR) :
> (reg & ~FBC_CTL_FALSE_COLOR));
>
> - mutex_unlock(&dev_priv->fbc.lock);
> + mutex_unlock(&i915->fbc.lock);
> return 0;
> }
>
> @@ -1460,18 +1460,18 @@ DEFINE_SIMPLE_ATTRIBUTE(i915_fbc_false_color_fops,
>
> static int i915_ips_status(struct seq_file *m, void *unused)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> intel_wakeref_t wakeref;
>
> - if (!HAS_IPS(dev_priv))
> + if (!HAS_IPS(i915))
> return -ENODEV;
>
> - wakeref = intel_runtime_pm_get(dev_priv);
> + wakeref = intel_runtime_pm_get(i915);
>
> seq_printf(m, "Enabled by kernel parameter: %s\n",
> yesno(i915_modparams.enable_ips));
>
> - if (INTEL_GEN(dev_priv) >= 8) {
> + if (INTEL_GEN(i915) >= 8) {
> seq_puts(m, "Currently: unknown\n");
> } else {
> if (I915_READ(IPS_CTL) & IPS_ENABLE)
> @@ -1480,34 +1480,34 @@ static int i915_ips_status(struct seq_file *m, void *unused)
> seq_puts(m, "Currently: disabled\n");
> }
>
> - intel_runtime_pm_put(dev_priv, wakeref);
> + intel_runtime_pm_put(i915, wakeref);
>
> return 0;
> }
>
> static int i915_sr_status(struct seq_file *m, void *unused)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> intel_wakeref_t wakeref;
> bool sr_enabled = false;
>
> - wakeref = intel_display_power_get(dev_priv, POWER_DOMAIN_INIT);
> + wakeref = intel_display_power_get(i915, POWER_DOMAIN_INIT);
>
> - if (INTEL_GEN(dev_priv) >= 9)
> + if (INTEL_GEN(i915) >= 9)
> /* no global SR status; inspect per-plane WM */;
> - else if (HAS_PCH_SPLIT(dev_priv))
> + else if (HAS_PCH_SPLIT(i915))
> sr_enabled = I915_READ(WM1_LP_ILK) & WM1_LP_SR_EN;
> - else if (IS_I965GM(dev_priv) || IS_G4X(dev_priv) ||
> - IS_I945G(dev_priv) || IS_I945GM(dev_priv))
> + else if (IS_I965GM(i915) || IS_G4X(i915) ||
> + IS_I945G(i915) || IS_I945GM(i915))
> sr_enabled = I915_READ(FW_BLC_SELF) & FW_BLC_SELF_EN;
> - else if (IS_I915GM(dev_priv))
> + else if (IS_I915GM(i915))
> sr_enabled = I915_READ(INSTPM) & INSTPM_SELF_EN;
> - else if (IS_PINEVIEW(dev_priv))
> + else if (IS_PINEVIEW(i915))
> sr_enabled = I915_READ(DSPFW3) & PINEVIEW_SELF_REFRESH_EN;
> - else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> + else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> sr_enabled = I915_READ(FW_BLC_SELF_VLV) & FW_CSPWRDWNEN;
>
> - intel_display_power_put(dev_priv, POWER_DOMAIN_INIT, wakeref);
> + intel_display_power_put(i915, POWER_DOMAIN_INIT, wakeref);
>
> seq_printf(m, "self-refresh: %s\n", enableddisabled(sr_enabled));
>
> @@ -1540,18 +1540,18 @@ static int i915_emon_status(struct seq_file *m, void *unused)
>
> static int i915_ring_freq_table(struct seq_file *m, void *unused)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> + struct intel_rps *rps = &i915->gt_pm.rps;
> unsigned int max_gpu_freq, min_gpu_freq;
> intel_wakeref_t wakeref;
> int gpu_freq, ia_freq;
>
> - if (!HAS_LLC(dev_priv))
> + if (!HAS_LLC(i915))
> return -ENODEV;
>
> min_gpu_freq = rps->min_freq;
> max_gpu_freq = rps->max_freq;
> - if (IS_GEN9_BC(dev_priv) || INTEL_GEN(dev_priv) >= 10) {
> + if (IS_GEN9_BC(i915) || INTEL_GEN(i915) >= 10) {
> /* Convert GT frequency to 50 HZ units */
> min_gpu_freq /= GEN9_FREQ_SCALER;
> max_gpu_freq /= GEN9_FREQ_SCALER;
> @@ -1559,30 +1559,30 @@ static int i915_ring_freq_table(struct seq_file *m, void *unused)
>
> seq_puts(m, "GPU freq (MHz)\tEffective CPU freq (MHz)\tEffective Ring freq (MHz)\n");
>
> - wakeref = intel_runtime_pm_get(dev_priv);
> + wakeref = intel_runtime_pm_get(i915);
> for (gpu_freq = min_gpu_freq; gpu_freq <= max_gpu_freq; gpu_freq++) {
> ia_freq = gpu_freq;
> - sandybridge_pcode_read(dev_priv,
> + sandybridge_pcode_read(i915,
> GEN6_PCODE_READ_MIN_FREQ_TABLE,
> &ia_freq, NULL);
> seq_printf(m, "%d\t\t%d\t\t\t\t%d\n",
> - intel_gpu_freq(dev_priv, (gpu_freq *
> - (IS_GEN9_BC(dev_priv) ||
> - INTEL_GEN(dev_priv) >= 10 ?
> + intel_gpu_freq(i915, (gpu_freq *
> + (IS_GEN9_BC(i915) ||
> + INTEL_GEN(i915) >= 10 ?
> GEN9_FREQ_SCALER : 1))),
> ((ia_freq >> 0) & 0xff) * 100,
> ((ia_freq >> 8) & 0xff) * 100);
> }
> - intel_runtime_pm_put(dev_priv, wakeref);
> + intel_runtime_pm_put(i915, wakeref);
>
> return 0;
> }
>
> static int i915_opregion(struct seq_file *m, void *unused)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> - struct drm_device *dev = &dev_priv->drm;
> - struct intel_opregion *opregion = &dev_priv->opregion;
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> + struct drm_device *dev = &i915->drm;
> + struct intel_opregion *opregion = &i915->opregion;
> int ret;
>
> ret = mutex_lock_interruptible(&dev->struct_mutex);
> @@ -1610,8 +1610,8 @@ static int i915_vbt(struct seq_file *m, void *unused)
>
> static int i915_gem_framebuffer_info(struct seq_file *m, void *data)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> + struct drm_device *dev = &i915->drm;
> struct intel_framebuffer *fbdev_fb = NULL;
> struct drm_framebuffer *drm_fb;
> int ret;
> @@ -1621,8 +1621,8 @@ static int i915_gem_framebuffer_info(struct seq_file *m, void *data)
> return ret;
>
> #ifdef CONFIG_DRM_FBDEV_EMULATION
> - if (dev_priv->fbdev && dev_priv->fbdev->helper.fb) {
> - fbdev_fb = to_intel_framebuffer(dev_priv->fbdev->helper.fb);
> + if (i915->fbdev && i915->fbdev->helper.fb) {
> + fbdev_fb = to_intel_framebuffer(i915->fbdev->helper.fb);
>
> seq_printf(m, "fbcon size: %d x %d, depth %d, %d bpp, modifier 0x%llx, refcount %d, obj ",
> fbdev_fb->base.width,
> @@ -1666,8 +1666,8 @@ static void describe_ctx_ring(struct seq_file *m, struct intel_ring *ring)
>
> static int i915_context_status(struct seq_file *m, void *unused)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> + struct drm_device *dev = &i915->drm;
> struct i915_gem_context *ctx;
> int ret;
>
> @@ -1675,7 +1675,7 @@ static int i915_context_status(struct seq_file *m, void *unused)
> if (ret)
> return ret;
>
> - list_for_each_entry(ctx, &dev_priv->contexts.list, link) {
> + list_for_each_entry(ctx, &i915->contexts.list, link) {
> struct i915_gem_engines_iter it;
> struct intel_context *ce;
>
> @@ -1746,18 +1746,18 @@ static const char *swizzle_string(unsigned swizzle)
>
> static int i915_swizzle_info(struct seq_file *m, void *data)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> + struct intel_uncore *uncore = &i915->uncore;
> intel_wakeref_t wakeref;
>
> - wakeref = intel_runtime_pm_get(dev_priv);
> + wakeref = intel_runtime_pm_get(i915);
>
> seq_printf(m, "bit6 swizzle for X-tiling = %s\n",
> - swizzle_string(dev_priv->mm.bit_6_swizzle_x));
> + swizzle_string(i915->mm.bit_6_swizzle_x));
> seq_printf(m, "bit6 swizzle for Y-tiling = %s\n",
> - swizzle_string(dev_priv->mm.bit_6_swizzle_y));
> + swizzle_string(i915->mm.bit_6_swizzle_y));
>
> - if (IS_GEN_RANGE(dev_priv, 3, 4)) {
> + if (IS_GEN_RANGE(i915, 3, 4)) {
> seq_printf(m, "DDC = 0x%08x\n",
> intel_uncore_read(uncore, DCC));
> seq_printf(m, "DDC2 = 0x%08x\n",
> @@ -1766,7 +1766,7 @@ static int i915_swizzle_info(struct seq_file *m, void *data)
> intel_uncore_read16(uncore, C0DRB3));
> seq_printf(m, "C1DRB3 = 0x%04x\n",
> intel_uncore_read16(uncore, C1DRB3));
> - } else if (INTEL_GEN(dev_priv) >= 6) {
> + } else if (INTEL_GEN(i915) >= 6) {
> seq_printf(m, "MAD_DIMM_C0 = 0x%08x\n",
> intel_uncore_read(uncore, MAD_DIMM_C0));
> seq_printf(m, "MAD_DIMM_C1 = 0x%08x\n",
> @@ -1775,7 +1775,7 @@ static int i915_swizzle_info(struct seq_file *m, void *data)
> intel_uncore_read(uncore, MAD_DIMM_C2));
> seq_printf(m, "TILECTL = 0x%08x\n",
> intel_uncore_read(uncore, TILECTL));
> - if (INTEL_GEN(dev_priv) >= 8)
> + if (INTEL_GEN(i915) >= 8)
> seq_printf(m, "GAMTARBMODE = 0x%08x\n",
> intel_uncore_read(uncore, GAMTARBMODE));
> else
> @@ -1785,10 +1785,10 @@ static int i915_swizzle_info(struct seq_file *m, void *data)
> intel_uncore_read(uncore, DISP_ARB_CTL));
> }
>
> - if (dev_priv->quirks & QUIRK_PIN_SWIZZLED_PAGES)
> + if (i915->quirks & QUIRK_PIN_SWIZZLED_PAGES)
> seq_puts(m, "L-shaped memory detected\n");
>
> - intel_runtime_pm_put(dev_priv, wakeref);
> + intel_runtime_pm_put(i915, wakeref);
>
> return 0;
> }
> @@ -1809,54 +1809,54 @@ static const char *rps_power_to_str(unsigned int power)
>
> static int i915_rps_boost_info(struct seq_file *m, void *data)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> + struct intel_rps *rps = &i915->gt_pm.rps;
> u32 act_freq = rps->cur_freq;
> intel_wakeref_t wakeref;
>
> - with_intel_runtime_pm_if_in_use(dev_priv, wakeref) {
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> - vlv_punit_get(dev_priv);
> - act_freq = vlv_punit_read(dev_priv,
> + with_intel_runtime_pm_if_in_use(i915, wakeref) {
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> + vlv_punit_get(i915);
> + act_freq = vlv_punit_read(i915,
> PUNIT_REG_GPU_FREQ_STS);
> - vlv_punit_put(dev_priv);
> + vlv_punit_put(i915);
> act_freq = (act_freq >> 8) & 0xff;
> } else {
> - act_freq = intel_get_cagf(dev_priv,
> + act_freq = intel_get_cagf(i915,
> I915_READ(GEN6_RPSTAT1));
> }
> }
>
> seq_printf(m, "RPS enabled? %d\n", rps->enabled);
> - seq_printf(m, "GPU busy? %s\n", yesno(dev_priv->gt.awake));
> + seq_printf(m, "GPU busy? %s\n", yesno(i915->gt.awake));
> seq_printf(m, "Boosts outstanding? %d\n",
> atomic_read(&rps->num_waiters));
> seq_printf(m, "Interactive? %d\n", READ_ONCE(rps->power.interactive));
> seq_printf(m, "Frequency requested %d, actual %d\n",
> - intel_gpu_freq(dev_priv, rps->cur_freq),
> - intel_gpu_freq(dev_priv, act_freq));
> + intel_gpu_freq(i915, rps->cur_freq),
> + intel_gpu_freq(i915, act_freq));
> seq_printf(m, " min hard:%d, soft:%d; max soft:%d, hard:%d\n",
> - intel_gpu_freq(dev_priv, rps->min_freq),
> - intel_gpu_freq(dev_priv, rps->min_freq_softlimit),
> - intel_gpu_freq(dev_priv, rps->max_freq_softlimit),
> - intel_gpu_freq(dev_priv, rps->max_freq));
> + intel_gpu_freq(i915, rps->min_freq),
> + intel_gpu_freq(i915, rps->min_freq_softlimit),
> + intel_gpu_freq(i915, rps->max_freq_softlimit),
> + intel_gpu_freq(i915, rps->max_freq));
> seq_printf(m, " idle:%d, efficient:%d, boost:%d\n",
> - intel_gpu_freq(dev_priv, rps->idle_freq),
> - intel_gpu_freq(dev_priv, rps->efficient_freq),
> - intel_gpu_freq(dev_priv, rps->boost_freq));
> + intel_gpu_freq(i915, rps->idle_freq),
> + intel_gpu_freq(i915, rps->efficient_freq),
> + intel_gpu_freq(i915, rps->boost_freq));
>
> seq_printf(m, "Wait boosts: %d\n", atomic_read(&rps->boosts));
>
> - if (INTEL_GEN(dev_priv) >= 6 && rps->enabled && dev_priv->gt.awake) {
> + if (INTEL_GEN(i915) >= 6 && rps->enabled && i915->gt.awake) {
> u32 rpup, rpupei;
> u32 rpdown, rpdownei;
>
> - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
> rpup = I915_READ_FW(GEN6_RP_CUR_UP) & GEN6_RP_EI_MASK;
> rpupei = I915_READ_FW(GEN6_RP_CUR_UP_EI) & GEN6_RP_EI_MASK;
> rpdown = I915_READ_FW(GEN6_RP_CUR_DOWN) & GEN6_RP_EI_MASK;
> rpdownei = I915_READ_FW(GEN6_RP_CUR_DOWN_EI) & GEN6_RP_EI_MASK;
> - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
>
> seq_printf(m, "\nRPS Autotuning (current \"%s\" window):\n",
> rps_power_to_str(rps->power.mode));
> @@ -1875,29 +1875,29 @@ static int i915_rps_boost_info(struct seq_file *m, void *data)
>
> static int i915_llc(struct seq_file *m, void *data)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> - const bool edram = INTEL_GEN(dev_priv) > 8;
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> + const bool edram = INTEL_GEN(i915) > 8;
>
> - seq_printf(m, "LLC: %s\n", yesno(HAS_LLC(dev_priv)));
> + seq_printf(m, "LLC: %s\n", yesno(HAS_LLC(i915)));
> seq_printf(m, "%s: %uMB\n", edram ? "eDRAM" : "eLLC",
> - dev_priv->edram_size_mb);
> + i915->edram_size_mb);
>
> return 0;
> }
>
> static int i915_huc_load_status_info(struct seq_file *m, void *data)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> intel_wakeref_t wakeref;
> struct drm_printer p;
>
> - if (!HAS_HUC(dev_priv))
> + if (!HAS_HUC(i915))
> return -ENODEV;
>
> p = drm_seq_file_printer(m);
> - intel_uc_fw_dump(&dev_priv->huc.fw, &p);
> + intel_uc_fw_dump(&i915->huc.fw, &p);
>
> - with_intel_runtime_pm(dev_priv, wakeref)
> + with_intel_runtime_pm(i915, wakeref)
> seq_printf(m, "\nHuC status 0x%08x:\n", I915_READ(HUC_STATUS2));
>
> return 0;
> @@ -1905,17 +1905,17 @@ static int i915_huc_load_status_info(struct seq_file *m, void *data)
>
> static int i915_guc_load_status_info(struct seq_file *m, void *data)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> intel_wakeref_t wakeref;
> struct drm_printer p;
>
> - if (!HAS_GUC(dev_priv))
> + if (!HAS_GUC(i915))
> return -ENODEV;
>
> p = drm_seq_file_printer(m);
> - intel_uc_fw_dump(&dev_priv->guc.fw, &p);
> + intel_uc_fw_dump(&i915->guc.fw, &p);
>
> - with_intel_runtime_pm(dev_priv, wakeref) {
> + with_intel_runtime_pm(i915, wakeref) {
> u32 tmp = I915_READ(GUC_STATUS);
> u32 i;
>
> @@ -1954,9 +1954,9 @@ stringify_guc_log_type(enum guc_log_buffer_type type)
> }
>
> static void i915_guc_log_info(struct seq_file *m,
> - struct drm_i915_private *dev_priv)
> + struct drm_i915_private *i915)
> {
> - struct intel_guc_log *log = &dev_priv->guc.log;
> + struct intel_guc_log *log = &i915->guc.log;
> enum guc_log_buffer_type type;
>
> if (!intel_guc_log_relay_enabled(log)) {
> @@ -1978,7 +1978,7 @@ static void i915_guc_log_info(struct seq_file *m,
> }
>
> static void i915_guc_client_info(struct seq_file *m,
> - struct drm_i915_private *dev_priv,
> + struct drm_i915_private *i915,
> struct intel_guc_client *client)
> {
> struct intel_engine_cs *engine;
> @@ -1990,7 +1990,7 @@ static void i915_guc_client_info(struct seq_file *m,
> seq_printf(m, "\tDoorbell id %d, offset: 0x%lx\n",
> client->doorbell_id, client->doorbell_offset);
>
> - for_each_engine(engine, dev_priv, id) {
> + for_each_engine(engine, i915, id) {
> u64 submissions = client->submissions[id];
> tot += submissions;
> seq_printf(m, "\tSubmissions: %llu %s\n",
> @@ -2001,15 +2001,15 @@ static void i915_guc_client_info(struct seq_file *m,
>
> static int i915_guc_info(struct seq_file *m, void *data)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> - const struct intel_guc *guc = &dev_priv->guc;
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> + const struct intel_guc *guc = &i915->guc;
>
> - if (!USES_GUC(dev_priv))
> + if (!USES_GUC(i915))
> return -ENODEV;
>
> - i915_guc_log_info(m, dev_priv);
> + i915_guc_log_info(m, i915);
>
> - if (!USES_GUC_SUBMISSION(dev_priv))
> + if (!USES_GUC_SUBMISSION(i915))
> return 0;
>
> GEM_BUG_ON(!guc->execbuf_client);
> @@ -2019,11 +2019,11 @@ static int i915_guc_info(struct seq_file *m, void *data)
> seq_printf(m, "Doorbell next cacheline: 0x%x\n", guc->db_cacheline);
>
> seq_printf(m, "\nGuC execbuf client @ %p:\n", guc->execbuf_client);
> - i915_guc_client_info(m, dev_priv, guc->execbuf_client);
> + i915_guc_client_info(m, i915, guc->execbuf_client);
> if (guc->preempt_client) {
> seq_printf(m, "\nGuC preempt client @ %p:\n",
> guc->preempt_client);
> - i915_guc_client_info(m, dev_priv, guc->preempt_client);
> + i915_guc_client_info(m, i915, guc->preempt_client);
> }
>
> /* Add more as required ... */
> @@ -2033,14 +2033,14 @@ static int i915_guc_info(struct seq_file *m, void *data)
>
> static int i915_guc_stage_pool(struct seq_file *m, void *data)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> - const struct intel_guc *guc = &dev_priv->guc;
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> + const struct intel_guc *guc = &i915->guc;
> struct guc_stage_desc *desc = guc->stage_desc_pool_vaddr;
> struct intel_guc_client *client = guc->execbuf_client;
> intel_engine_mask_t tmp;
> int index;
>
> - if (!USES_GUC_SUBMISSION(dev_priv))
> + if (!USES_GUC_SUBMISSION(i915))
> return -ENODEV;
>
> for (index = 0; index < GUC_MAX_STAGE_DESCRIPTORS; index++, desc++) {
> @@ -2066,7 +2066,7 @@ static int i915_guc_stage_pool(struct seq_file *m, void *data)
> desc->wq_addr, desc->wq_size);
> seq_putc(m, '\n');
>
> - for_each_engine_masked(engine, dev_priv, client->engines, tmp) {
> + for_each_engine_masked(engine, i915, client->engines, tmp) {
> u32 guc_engine_id = engine->guc_id;
> struct guc_execlist_context *lrc =
> &desc->lrc[guc_engine_id];
> @@ -2088,19 +2088,19 @@ static int i915_guc_stage_pool(struct seq_file *m, void *data)
> static int i915_guc_log_dump(struct seq_file *m, void *data)
> {
> struct drm_info_node *node = m->private;
> - struct drm_i915_private *dev_priv = node_to_i915(node);
> + struct drm_i915_private *i915 = node_to_i915(node);
> bool dump_load_err = !!node->info_ent->data;
> struct drm_i915_gem_object *obj = NULL;
> u32 *log;
> int i = 0;
>
> - if (!HAS_GUC(dev_priv))
> + if (!HAS_GUC(i915))
> return -ENODEV;
>
> if (dump_load_err)
> - obj = dev_priv->guc.load_err_log;
> - else if (dev_priv->guc.log.vma)
> - obj = dev_priv->guc.log.vma->obj;
> + obj = i915->guc.load_err_log;
> + else if (i915->guc.log.vma)
> + obj = i915->guc.log.vma->obj;
>
> if (!obj)
> return 0;
> @@ -2126,24 +2126,24 @@ static int i915_guc_log_dump(struct seq_file *m, void *data)
>
> static int i915_guc_log_level_get(void *data, u64 *val)
> {
> - struct drm_i915_private *dev_priv = data;
> + struct drm_i915_private *i915 = data;
>
> - if (!USES_GUC(dev_priv))
> + if (!USES_GUC(i915))
> return -ENODEV;
>
> - *val = intel_guc_log_get_level(&dev_priv->guc.log);
> + *val = intel_guc_log_get_level(&i915->guc.log);
>
> return 0;
> }
>
> static int i915_guc_log_level_set(void *data, u64 val)
> {
> - struct drm_i915_private *dev_priv = data;
> + struct drm_i915_private *i915 = data;
>
> - if (!USES_GUC(dev_priv))
> + if (!USES_GUC(i915))
> return -ENODEV;
>
> - return intel_guc_log_set_level(&dev_priv->guc.log, val);
> + return intel_guc_log_set_level(&i915->guc.log, val);
> }
>
> DEFINE_SIMPLE_ATTRIBUTE(i915_guc_log_level_fops,
> @@ -2152,14 +2152,14 @@ DEFINE_SIMPLE_ATTRIBUTE(i915_guc_log_level_fops,
>
> static int i915_guc_log_relay_open(struct inode *inode, struct file *file)
> {
> - struct drm_i915_private *dev_priv = inode->i_private;
> + struct drm_i915_private *i915 = inode->i_private;
>
> - if (!USES_GUC(dev_priv))
> + if (!USES_GUC(i915))
> return -ENODEV;
>
> - file->private_data = &dev_priv->guc.log;
> + file->private_data = &i915->guc.log;
>
> - return intel_guc_log_relay_open(&dev_priv->guc.log);
> + return intel_guc_log_relay_open(&i915->guc.log);
> }
>
> static ssize_t
> @@ -2177,9 +2177,9 @@ i915_guc_log_relay_write(struct file *filp,
>
> static int i915_guc_log_relay_release(struct inode *inode, struct file *file)
> {
> - struct drm_i915_private *dev_priv = inode->i_private;
> + struct drm_i915_private *i915 = inode->i_private;
>
> - intel_guc_log_relay_close(&dev_priv->guc.log);
> + intel_guc_log_relay_close(&i915->guc.log);
>
> return 0;
> }
> @@ -2205,12 +2205,12 @@ static int i915_psr_sink_status_show(struct seq_file *m, void *data)
> "sink internal error",
> };
> struct drm_connector *connector = m->private;
> - struct drm_i915_private *dev_priv = to_i915(connector->dev);
> + struct drm_i915_private *i915 = to_i915(connector->dev);
> struct intel_dp *intel_dp =
> enc_to_intel_dp(&intel_attached_encoder(connector)->base);
> int ret;
>
> - if (!CAN_PSR(dev_priv)) {
> + if (!CAN_PSR(i915)) {
> seq_puts(m, "PSR Unsupported\n");
> return -ENODEV;
> }
> @@ -2236,12 +2236,12 @@ static int i915_psr_sink_status_show(struct seq_file *m, void *data)
> DEFINE_SHOW_ATTRIBUTE(i915_psr_sink_status);
>
> static void
> -psr_source_status(struct drm_i915_private *dev_priv, struct seq_file *m)
> +psr_source_status(struct drm_i915_private *i915, struct seq_file *m)
> {
> u32 val, status_val;
> const char *status = "unknown";
>
> - if (dev_priv->psr.psr2_enabled) {
> + if (i915->psr.psr2_enabled) {
> static const char * const live_status[] = {
> "IDLE",
> "CAPTURE",
> @@ -2283,14 +2283,14 @@ psr_source_status(struct drm_i915_private *dev_priv, struct seq_file *m)
>
> static int i915_edp_psr_status(struct seq_file *m, void *data)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> - struct i915_psr *psr = &dev_priv->psr;
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> + struct i915_psr *psr = &i915->psr;
> intel_wakeref_t wakeref;
> const char *status;
> bool enabled;
> u32 val;
>
> - if (!HAS_PSR(dev_priv))
> + if (!HAS_PSR(i915))
> return -ENODEV;
>
> seq_printf(m, "Sink support: %s", yesno(psr->sink_support));
> @@ -2301,7 +2301,7 @@ static int i915_edp_psr_status(struct seq_file *m, void *data)
> if (!psr->sink_support)
> return 0;
>
> - wakeref = intel_runtime_pm_get(dev_priv);
> + wakeref = intel_runtime_pm_get(i915);
> mutex_lock(&psr->lock);
>
> if (psr->enabled)
> @@ -2322,14 +2322,14 @@ static int i915_edp_psr_status(struct seq_file *m, void *data)
> }
> seq_printf(m, "Source PSR ctl: %s [0x%08x]\n",
> enableddisabled(enabled), val);
> - psr_source_status(dev_priv, m);
> + psr_source_status(i915, m);
> seq_printf(m, "Busy frontbuffer bits: 0x%08x\n",
> psr->busy_frontbuffer_bits);
>
> /*
> * SKL+ Perf counter is reset to 0 everytime DC state is entered
> */
> - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) {
> + if (IS_HASWELL(i915) || IS_BROADWELL(i915)) {
> val = I915_READ(EDP_PSR_PERF_CNT) & EDP_PSR_PERF_CNT_MASK;
> seq_printf(m, "Performance counter: %u\n", val);
> }
> @@ -2365,7 +2365,7 @@ static int i915_edp_psr_status(struct seq_file *m, void *data)
>
> unlock:
> mutex_unlock(&psr->lock);
> - intel_runtime_pm_put(dev_priv, wakeref);
> + intel_runtime_pm_put(i915, wakeref);
>
> return 0;
> }
> @@ -2373,20 +2373,20 @@ static int i915_edp_psr_status(struct seq_file *m, void *data)
> static int
> i915_edp_psr_debug_set(void *data, u64 val)
> {
> - struct drm_i915_private *dev_priv = data;
> + struct drm_i915_private *i915 = data;
> intel_wakeref_t wakeref;
> int ret;
>
> - if (!CAN_PSR(dev_priv))
> + if (!CAN_PSR(i915))
> return -ENODEV;
>
> DRM_DEBUG_KMS("Setting PSR debug to %llx\n", val);
>
> - wakeref = intel_runtime_pm_get(dev_priv);
> + wakeref = intel_runtime_pm_get(i915);
>
> - ret = intel_psr_debug_set(dev_priv, val);
> + ret = intel_psr_debug_set(i915, val);
>
> - intel_runtime_pm_put(dev_priv, wakeref);
> + intel_runtime_pm_put(i915, wakeref);
>
> return ret;
> }
> @@ -2394,12 +2394,12 @@ i915_edp_psr_debug_set(void *data, u64 val)
> static int
> i915_edp_psr_debug_get(void *data, u64 *val)
> {
> - struct drm_i915_private *dev_priv = data;
> + struct drm_i915_private *i915 = data;
>
> - if (!CAN_PSR(dev_priv))
> + if (!CAN_PSR(i915))
> return -ENODEV;
>
> - *val = READ_ONCE(dev_priv->psr.debug);
> + *val = READ_ONCE(i915->psr.debug);
> return 0;
> }
>
> @@ -2409,19 +2409,19 @@ DEFINE_SIMPLE_ATTRIBUTE(i915_edp_psr_debug_fops,
>
> static int i915_energy_uJ(struct seq_file *m, void *data)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> unsigned long long power;
> intel_wakeref_t wakeref;
> u32 units;
>
> - if (INTEL_GEN(dev_priv) < 6)
> + if (INTEL_GEN(i915) < 6)
> return -ENODEV;
>
> if (rdmsrl_safe(MSR_RAPL_POWER_UNIT, &power))
> return -ENODEV;
>
> units = (power & 0x1f00) >> 8;
> - with_intel_runtime_pm(dev_priv, wakeref)
> + with_intel_runtime_pm(i915, wakeref)
> power = I915_READ(MCH_SECP_NRG_STTS);
>
> power = (1000000 * power) >> units; /* convert to uJ */
> @@ -2432,21 +2432,21 @@ static int i915_energy_uJ(struct seq_file *m, void *data)
>
> static int i915_runtime_pm_status(struct seq_file *m, void *unused)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> + struct pci_dev *pdev = i915->drm.pdev;
>
> - if (!HAS_RUNTIME_PM(dev_priv))
> + if (!HAS_RUNTIME_PM(i915))
> seq_puts(m, "Runtime power management not supported\n");
>
> seq_printf(m, "Runtime power status: %s\n",
> - enableddisabled(!dev_priv->power_domains.wakeref));
> + enableddisabled(!i915->power_domains.wakeref));
>
> - seq_printf(m, "GPU idle: %s\n", yesno(!dev_priv->gt.awake));
> + seq_printf(m, "GPU idle: %s\n", yesno(!i915->gt.awake));
> seq_printf(m, "IRQs disabled: %s\n",
> - yesno(!intel_irqs_enabled(dev_priv)));
> + yesno(!intel_irqs_enabled(i915)));
> #ifdef CONFIG_PM
> seq_printf(m, "Usage count: %d\n",
> - atomic_read(&dev_priv->drm.dev->power.usage_count));
> + atomic_read(&i915->drm.dev->power.usage_count));
> #else
> seq_printf(m, "Device Power Management (CONFIG_PM) disabled\n");
> #endif
> @@ -2457,7 +2457,7 @@ static int i915_runtime_pm_status(struct seq_file *m, void *unused)
> if (IS_ENABLED(CONFIG_DRM_I915_DEBUG_RUNTIME_PM)) {
> struct drm_printer p = drm_seq_file_printer(m);
>
> - print_intel_runtime_pm_wakeref(dev_priv, &p);
> + print_intel_runtime_pm_wakeref(i915, &p);
> }
>
> return 0;
> @@ -2465,8 +2465,8 @@ static int i915_runtime_pm_status(struct seq_file *m, void *unused)
>
> static int i915_power_domain_info(struct seq_file *m, void *unused)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> - struct i915_power_domains *power_domains = &dev_priv->power_domains;
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> + struct i915_power_domains *power_domains = &i915->power_domains;
> int i;
>
> mutex_lock(&power_domains->lock);
> @@ -2493,16 +2493,16 @@ static int i915_power_domain_info(struct seq_file *m, void *unused)
>
> static int i915_dmc_info(struct seq_file *m, void *unused)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> intel_wakeref_t wakeref;
> struct intel_csr *csr;
>
> - if (!HAS_CSR(dev_priv))
> + if (!HAS_CSR(i915))
> return -ENODEV;
>
> - csr = &dev_priv->csr;
> + csr = &i915->csr;
>
> - wakeref = intel_runtime_pm_get(dev_priv);
> + wakeref = intel_runtime_pm_get(i915);
>
> seq_printf(m, "fw loaded: %s\n", yesno(csr->dmc_payload != NULL));
> seq_printf(m, "path: %s\n", csr->fw_path);
> @@ -2513,13 +2513,13 @@ static int i915_dmc_info(struct seq_file *m, void *unused)
> seq_printf(m, "version: %d.%d\n", CSR_VERSION_MAJOR(csr->version),
> CSR_VERSION_MINOR(csr->version));
>
> - if (WARN_ON(INTEL_GEN(dev_priv) > 11))
> + if (WARN_ON(INTEL_GEN(i915) > 11))
> goto out;
>
> seq_printf(m, "DC3 -> DC5 count: %d\n",
> - I915_READ(IS_BROXTON(dev_priv) ? BXT_CSR_DC3_DC5_COUNT :
> + I915_READ(IS_BROXTON(i915) ? BXT_CSR_DC3_DC5_COUNT :
> SKL_CSR_DC3_DC5_COUNT));
> - if (!IS_GEN9_LP(dev_priv))
> + if (!IS_GEN9_LP(i915))
> seq_printf(m, "DC5 -> DC6 count: %d\n",
> I915_READ(SKL_CSR_DC5_DC6_COUNT));
>
> @@ -2528,7 +2528,7 @@ static int i915_dmc_info(struct seq_file *m, void *unused)
> seq_printf(m, "ssp base: 0x%08x\n", I915_READ(CSR_SSP_BASE));
> seq_printf(m, "htp: 0x%08x\n", I915_READ(CSR_HTP_SKL));
>
> - intel_runtime_pm_put(dev_priv, wakeref);
> + intel_runtime_pm_put(i915, wakeref);
>
> return 0;
> }
> @@ -2548,8 +2548,8 @@ static void intel_encoder_info(struct seq_file *m,
> struct intel_crtc *intel_crtc,
> struct intel_encoder *intel_encoder)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> + struct drm_device *dev = &i915->drm;
> struct drm_crtc *crtc = &intel_crtc->base;
> struct intel_connector *intel_connector;
> struct drm_encoder *encoder;
> @@ -2575,8 +2575,8 @@ static void intel_encoder_info(struct seq_file *m,
>
> static void intel_crtc_info(struct seq_file *m, struct intel_crtc *intel_crtc)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> + struct drm_device *dev = &i915->drm;
> struct drm_crtc *crtc = &intel_crtc->base;
> struct intel_encoder *intel_encoder;
> struct drm_plane_state *plane_state = crtc->primary->state;
> @@ -2731,8 +2731,8 @@ static void plane_rotation(char *buf, size_t bufsize, unsigned int rotation)
>
> static void intel_plane_info(struct seq_file *m, struct intel_crtc *intel_crtc)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> + struct drm_device *dev = &i915->drm;
> struct intel_plane *intel_plane;
>
> for_each_intel_plane_on_crtc(dev, intel_crtc, intel_plane) {
> @@ -2805,14 +2805,14 @@ static void intel_scaler_info(struct seq_file *m, struct intel_crtc *intel_crtc)
>
> static int i915_display_info(struct seq_file *m, void *unused)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> + struct drm_device *dev = &i915->drm;
> struct intel_crtc *crtc;
> struct drm_connector *connector;
> struct drm_connector_list_iter conn_iter;
> intel_wakeref_t wakeref;
>
> - wakeref = intel_runtime_pm_get(dev_priv);
> + wakeref = intel_runtime_pm_get(i915);
>
> seq_printf(m, "CRTC info\n");
> seq_printf(m, "---------\n");
> @@ -2861,42 +2861,42 @@ static int i915_display_info(struct seq_file *m, void *unused)
> drm_connector_list_iter_end(&conn_iter);
> mutex_unlock(&dev->mode_config.mutex);
>
> - intel_runtime_pm_put(dev_priv, wakeref);
> + intel_runtime_pm_put(i915, wakeref);
>
> return 0;
> }
>
> static int i915_engine_info(struct seq_file *m, void *unused)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> struct intel_engine_cs *engine;
> intel_wakeref_t wakeref;
> enum intel_engine_id id;
> struct drm_printer p;
>
> - wakeref = intel_runtime_pm_get(dev_priv);
> + wakeref = intel_runtime_pm_get(i915);
>
> seq_printf(m, "GT awake? %s [%d]\n",
> - yesno(dev_priv->gt.awake),
> - atomic_read(&dev_priv->gt.wakeref.count));
> + yesno(i915->gt.awake),
> + atomic_read(&i915->gt.wakeref.count));
> seq_printf(m, "CS timestamp frequency: %u kHz\n",
> - RUNTIME_INFO(dev_priv)->cs_timestamp_frequency_khz);
> + RUNTIME_INFO(i915)->cs_timestamp_frequency_khz);
>
> p = drm_seq_file_printer(m);
> - for_each_engine(engine, dev_priv, id)
> + for_each_engine(engine, i915, id)
> intel_engine_dump(engine, &p, "%s\n", engine->name);
>
> - intel_runtime_pm_put(dev_priv, wakeref);
> + intel_runtime_pm_put(i915, wakeref);
>
> return 0;
> }
>
> static int i915_rcs_topology(struct seq_file *m, void *unused)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> struct drm_printer p = drm_seq_file_printer(m);
>
> - intel_device_info_dump_topology(&RUNTIME_INFO(dev_priv)->sseu, &p);
> + intel_device_info_dump_topology(&RUNTIME_INFO(i915)->sseu, &p);
>
> return 0;
> }
> @@ -2913,13 +2913,13 @@ static int i915_shrinker_info(struct seq_file *m, void *unused)
>
> static int i915_shared_dplls_info(struct seq_file *m, void *unused)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> + struct drm_device *dev = &i915->drm;
> int i;
>
> drm_modeset_lock_all(dev);
> - for (i = 0; i < dev_priv->num_shared_dpll; i++) {
> - struct intel_shared_dpll *pll = &dev_priv->shared_dplls[i];
> + for (i = 0; i < i915->num_shared_dpll; i++) {
> + struct intel_shared_dpll *pll = &i915->shared_dplls[i];
>
> seq_printf(m, "DPLL%i: %s, id: %i\n", i, pll->info->name,
> pll->info->id);
> @@ -2977,28 +2977,28 @@ static int i915_wa_registers(struct seq_file *m, void *unused)
>
> static int i915_ipc_status_show(struct seq_file *m, void *data)
> {
> - struct drm_i915_private *dev_priv = m->private;
> + struct drm_i915_private *i915 = m->private;
>
> seq_printf(m, "Isochronous Priority Control: %s\n",
> - yesno(dev_priv->ipc_enabled));
> + yesno(i915->ipc_enabled));
> return 0;
> }
>
> static int i915_ipc_status_open(struct inode *inode, struct file *file)
> {
> - struct drm_i915_private *dev_priv = inode->i_private;
> + struct drm_i915_private *i915 = inode->i_private;
>
> - if (!HAS_IPC(dev_priv))
> + if (!HAS_IPC(i915))
> return -ENODEV;
>
> - return single_open(file, i915_ipc_status_show, dev_priv);
> + return single_open(file, i915_ipc_status_show, i915);
> }
>
> static ssize_t i915_ipc_status_write(struct file *file, const char __user *ubuf,
> size_t len, loff_t *offp)
> {
> struct seq_file *m = file->private_data;
> - struct drm_i915_private *dev_priv = m->private;
> + struct drm_i915_private *i915 = m->private;
> intel_wakeref_t wakeref;
> bool enable;
> int ret;
> @@ -3007,12 +3007,12 @@ static ssize_t i915_ipc_status_write(struct file *file, const char __user *ubuf,
> if (ret < 0)
> return ret;
>
> - with_intel_runtime_pm(dev_priv, wakeref) {
> - if (!dev_priv->ipc_enabled && enable)
> + with_intel_runtime_pm(i915, wakeref) {
> + if (!i915->ipc_enabled && enable)
> DRM_INFO("Enabling IPC: WM will be proper only after next commit\n");
> - dev_priv->wm.distrust_bios_wm = true;
> - dev_priv->ipc_enabled = enable;
> - intel_enable_ipc(dev_priv);
> + i915->wm.distrust_bios_wm = true;
> + i915->ipc_enabled = enable;
> + intel_enable_ipc(i915);
> }
>
> return len;
> @@ -3029,19 +3029,19 @@ static const struct file_operations i915_ipc_status_fops = {
>
> static int i915_ddb_info(struct seq_file *m, void *unused)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> + struct drm_device *dev = &i915->drm;
> struct skl_ddb_entry *entry;
> struct intel_crtc *crtc;
>
> - if (INTEL_GEN(dev_priv) < 9)
> + if (INTEL_GEN(i915) < 9)
> return -ENODEV;
>
> drm_modeset_lock_all(dev);
>
> seq_printf(m, "%-15s%8s%8s%8s\n", "", "Start", "End", "Size");
>
> - for_each_intel_crtc(&dev_priv->drm, crtc) {
> + for_each_intel_crtc(&i915->drm, crtc) {
> struct intel_crtc_state *crtc_state =
> to_intel_crtc_state(crtc->base.state);
> enum pipe pipe = crtc->pipe;
> @@ -3070,8 +3070,8 @@ static void drrs_status_per_crtc(struct seq_file *m,
> struct drm_device *dev,
> struct intel_crtc *intel_crtc)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct i915_drrs *drrs = &dev_priv->drrs;
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct i915_drrs *drrs = &i915->drrs;
> int vrefresh = 0;
> struct drm_connector *connector;
> struct drm_connector_list_iter conn_iter;
> @@ -3085,11 +3085,11 @@ static void drrs_status_per_crtc(struct seq_file *m,
> }
> drm_connector_list_iter_end(&conn_iter);
>
> - if (dev_priv->vbt.drrs_type == STATIC_DRRS_SUPPORT)
> + if (i915->vbt.drrs_type == STATIC_DRRS_SUPPORT)
> seq_puts(m, "\tVBT: DRRS_type: Static");
> - else if (dev_priv->vbt.drrs_type == SEAMLESS_DRRS_SUPPORT)
> + else if (i915->vbt.drrs_type == SEAMLESS_DRRS_SUPPORT)
> seq_puts(m, "\tVBT: DRRS_type: Seamless");
> - else if (dev_priv->vbt.drrs_type == DRRS_NOT_SUPPORTED)
> + else if (i915->vbt.drrs_type == DRRS_NOT_SUPPORTED)
> seq_puts(m, "\tVBT: DRRS_type: None");
> else
> seq_puts(m, "\tVBT: DRRS_type: FIXME: Unrecognized Value");
> @@ -3106,7 +3106,7 @@ static void drrs_status_per_crtc(struct seq_file *m,
> /* disable_drrs() will make drrs->dp NULL */
> if (!drrs->dp) {
> seq_puts(m, "Idleness DRRS: Disabled\n");
> - if (dev_priv->psr.enabled)
> + if (i915->psr.enabled)
> seq_puts(m,
> "\tAs PSR is enabled, DRRS is not enabled\n");
> mutex_unlock(&drrs->mutex);
> @@ -3143,8 +3143,8 @@ static void drrs_status_per_crtc(struct seq_file *m,
>
> static int i915_drrs_status(struct seq_file *m, void *unused)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> + struct drm_device *dev = &i915->drm;
> struct intel_crtc *intel_crtc;
> int active_crtc_cnt = 0;
>
> @@ -3167,8 +3167,8 @@ static int i915_drrs_status(struct seq_file *m, void *unused)
>
> static int i915_dp_mst_info(struct seq_file *m, void *unused)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> + struct drm_device *dev = &i915->drm;
> struct intel_encoder *intel_encoder;
> struct intel_digital_port *intel_dig_port;
> struct drm_connector *connector;
> @@ -3257,8 +3257,8 @@ static ssize_t i915_displayport_test_active_write(struct file *file,
>
> static int i915_displayport_test_active_show(struct seq_file *m, void *data)
> {
> - struct drm_i915_private *dev_priv = m->private;
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_i915_private *i915 = m->private;
> + struct drm_device *dev = &i915->drm;
> struct drm_connector *connector;
> struct drm_connector_list_iter conn_iter;
> struct intel_dp *intel_dp;
> @@ -3307,8 +3307,8 @@ static const struct file_operations i915_displayport_test_active_fops = {
>
> static int i915_displayport_test_data_show(struct seq_file *m, void *data)
> {
> - struct drm_i915_private *dev_priv = m->private;
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_i915_private *i915 = m->private;
> + struct drm_device *dev = &i915->drm;
> struct drm_connector *connector;
> struct drm_connector_list_iter conn_iter;
> struct intel_dp *intel_dp;
> @@ -3351,8 +3351,8 @@ DEFINE_SHOW_ATTRIBUTE(i915_displayport_test_data);
>
> static int i915_displayport_test_type_show(struct seq_file *m, void *data)
> {
> - struct drm_i915_private *dev_priv = m->private;
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_i915_private *i915 = m->private;
> + struct drm_device *dev = &i915->drm;
> struct drm_connector *connector;
> struct drm_connector_list_iter conn_iter;
> struct intel_dp *intel_dp;
> @@ -3383,19 +3383,19 @@ DEFINE_SHOW_ATTRIBUTE(i915_displayport_test_type);
>
> static void wm_latency_show(struct seq_file *m, const u16 wm[8])
> {
> - struct drm_i915_private *dev_priv = m->private;
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_i915_private *i915 = m->private;
> + struct drm_device *dev = &i915->drm;
> int level;
> int num_levels;
>
> - if (IS_CHERRYVIEW(dev_priv))
> + if (IS_CHERRYVIEW(i915))
> num_levels = 3;
> - else if (IS_VALLEYVIEW(dev_priv))
> + else if (IS_VALLEYVIEW(i915))
> num_levels = 1;
> - else if (IS_G4X(dev_priv))
> + else if (IS_G4X(i915))
> num_levels = 3;
> else
> - num_levels = ilk_wm_max_level(dev_priv) + 1;
> + num_levels = ilk_wm_max_level(i915) + 1;
>
> drm_modeset_lock_all(dev);
>
> @@ -3406,10 +3406,10 @@ static void wm_latency_show(struct seq_file *m, const u16 wm[8])
> * - WM1+ latency values in 0.5us units
> * - latencies are in us on gen9/vlv/chv
> */
> - if (INTEL_GEN(dev_priv) >= 9 ||
> - IS_VALLEYVIEW(dev_priv) ||
> - IS_CHERRYVIEW(dev_priv) ||
> - IS_G4X(dev_priv))
> + if (INTEL_GEN(i915) >= 9 ||
> + IS_VALLEYVIEW(i915) ||
> + IS_CHERRYVIEW(i915) ||
> + IS_G4X(i915))
> latency *= 10;
> else if (level > 0)
> latency *= 5;
> @@ -3423,13 +3423,13 @@ static void wm_latency_show(struct seq_file *m, const u16 wm[8])
>
> static int pri_wm_latency_show(struct seq_file *m, void *data)
> {
> - struct drm_i915_private *dev_priv = m->private;
> + struct drm_i915_private *i915 = m->private;
> const u16 *latencies;
>
> - if (INTEL_GEN(dev_priv) >= 9)
> - latencies = dev_priv->wm.skl_latency;
> + if (INTEL_GEN(i915) >= 9)
> + latencies = i915->wm.skl_latency;
> else
> - latencies = dev_priv->wm.pri_latency;
> + latencies = i915->wm.pri_latency;
>
> wm_latency_show(m, latencies);
>
> @@ -3438,13 +3438,13 @@ static int pri_wm_latency_show(struct seq_file *m, void *data)
>
> static int spr_wm_latency_show(struct seq_file *m, void *data)
> {
> - struct drm_i915_private *dev_priv = m->private;
> + struct drm_i915_private *i915 = m->private;
> const u16 *latencies;
>
> - if (INTEL_GEN(dev_priv) >= 9)
> - latencies = dev_priv->wm.skl_latency;
> + if (INTEL_GEN(i915) >= 9)
> + latencies = i915->wm.skl_latency;
> else
> - latencies = dev_priv->wm.spr_latency;
> + latencies = i915->wm.spr_latency;
>
> wm_latency_show(m, latencies);
>
> @@ -3453,13 +3453,13 @@ static int spr_wm_latency_show(struct seq_file *m, void *data)
>
> static int cur_wm_latency_show(struct seq_file *m, void *data)
> {
> - struct drm_i915_private *dev_priv = m->private;
> + struct drm_i915_private *i915 = m->private;
> const u16 *latencies;
>
> - if (INTEL_GEN(dev_priv) >= 9)
> - latencies = dev_priv->wm.skl_latency;
> + if (INTEL_GEN(i915) >= 9)
> + latencies = i915->wm.skl_latency;
> else
> - latencies = dev_priv->wm.cur_latency;
> + latencies = i915->wm.cur_latency;
>
> wm_latency_show(m, latencies);
>
> @@ -3468,54 +3468,54 @@ static int cur_wm_latency_show(struct seq_file *m, void *data)
>
> static int pri_wm_latency_open(struct inode *inode, struct file *file)
> {
> - struct drm_i915_private *dev_priv = inode->i_private;
> + struct drm_i915_private *i915 = inode->i_private;
>
> - if (INTEL_GEN(dev_priv) < 5 && !IS_G4X(dev_priv))
> + if (INTEL_GEN(i915) < 5 && !IS_G4X(i915))
> return -ENODEV;
>
> - return single_open(file, pri_wm_latency_show, dev_priv);
> + return single_open(file, pri_wm_latency_show, i915);
> }
>
> static int spr_wm_latency_open(struct inode *inode, struct file *file)
> {
> - struct drm_i915_private *dev_priv = inode->i_private;
> + struct drm_i915_private *i915 = inode->i_private;
>
> - if (HAS_GMCH(dev_priv))
> + if (HAS_GMCH(i915))
> return -ENODEV;
>
> - return single_open(file, spr_wm_latency_show, dev_priv);
> + return single_open(file, spr_wm_latency_show, i915);
> }
>
> static int cur_wm_latency_open(struct inode *inode, struct file *file)
> {
> - struct drm_i915_private *dev_priv = inode->i_private;
> + struct drm_i915_private *i915 = inode->i_private;
>
> - if (HAS_GMCH(dev_priv))
> + if (HAS_GMCH(i915))
> return -ENODEV;
>
> - return single_open(file, cur_wm_latency_show, dev_priv);
> + return single_open(file, cur_wm_latency_show, i915);
> }
>
> static ssize_t wm_latency_write(struct file *file, const char __user *ubuf,
> size_t len, loff_t *offp, u16 wm[8])
> {
> struct seq_file *m = file->private_data;
> - struct drm_i915_private *dev_priv = m->private;
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_i915_private *i915 = m->private;
> + struct drm_device *dev = &i915->drm;
> u16 new[8] = { 0 };
> int num_levels;
> int level;
> int ret;
> char tmp[32];
>
> - if (IS_CHERRYVIEW(dev_priv))
> + if (IS_CHERRYVIEW(i915))
> num_levels = 3;
> - else if (IS_VALLEYVIEW(dev_priv))
> + else if (IS_VALLEYVIEW(i915))
> num_levels = 1;
> - else if (IS_G4X(dev_priv))
> + else if (IS_G4X(i915))
> num_levels = 3;
> else
> - num_levels = ilk_wm_max_level(dev_priv) + 1;
> + num_levels = ilk_wm_max_level(i915) + 1;
>
> if (len >= sizeof(tmp))
> return -EINVAL;
> @@ -3546,13 +3546,13 @@ static ssize_t pri_wm_latency_write(struct file *file, const char __user *ubuf,
> size_t len, loff_t *offp)
> {
> struct seq_file *m = file->private_data;
> - struct drm_i915_private *dev_priv = m->private;
> + struct drm_i915_private *i915 = m->private;
> u16 *latencies;
>
> - if (INTEL_GEN(dev_priv) >= 9)
> - latencies = dev_priv->wm.skl_latency;
> + if (INTEL_GEN(i915) >= 9)
> + latencies = i915->wm.skl_latency;
> else
> - latencies = dev_priv->wm.pri_latency;
> + latencies = i915->wm.pri_latency;
>
> return wm_latency_write(file, ubuf, len, offp, latencies);
> }
> @@ -3561,13 +3561,13 @@ static ssize_t spr_wm_latency_write(struct file *file, const char __user *ubuf,
> size_t len, loff_t *offp)
> {
> struct seq_file *m = file->private_data;
> - struct drm_i915_private *dev_priv = m->private;
> + struct drm_i915_private *i915 = m->private;
> u16 *latencies;
>
> - if (INTEL_GEN(dev_priv) >= 9)
> - latencies = dev_priv->wm.skl_latency;
> + if (INTEL_GEN(i915) >= 9)
> + latencies = i915->wm.skl_latency;
> else
> - latencies = dev_priv->wm.spr_latency;
> + latencies = i915->wm.spr_latency;
>
> return wm_latency_write(file, ubuf, len, offp, latencies);
> }
> @@ -3576,13 +3576,13 @@ static ssize_t cur_wm_latency_write(struct file *file, const char __user *ubuf,
> size_t len, loff_t *offp)
> {
> struct seq_file *m = file->private_data;
> - struct drm_i915_private *dev_priv = m->private;
> + struct drm_i915_private *i915 = m->private;
> u16 *latencies;
>
> - if (INTEL_GEN(dev_priv) >= 9)
> - latencies = dev_priv->wm.skl_latency;
> + if (INTEL_GEN(i915) >= 9)
> + latencies = i915->wm.skl_latency;
> else
> - latencies = dev_priv->wm.cur_latency;
> + latencies = i915->wm.cur_latency;
>
> return wm_latency_write(file, ubuf, len, offp, latencies);
> }
> @@ -3752,14 +3752,14 @@ DEFINE_SIMPLE_ATTRIBUTE(i915_drop_caches_fops,
> static int
> i915_cache_sharing_get(void *data, u64 *val)
> {
> - struct drm_i915_private *dev_priv = data;
> + struct drm_i915_private *i915 = data;
> intel_wakeref_t wakeref;
> u32 snpcr = 0;
>
> - if (!(IS_GEN_RANGE(dev_priv, 6, 7)))
> + if (!(IS_GEN_RANGE(i915, 6, 7)))
> return -ENODEV;
>
> - with_intel_runtime_pm(dev_priv, wakeref)
> + with_intel_runtime_pm(i915, wakeref)
> snpcr = I915_READ(GEN6_MBCUNIT_SNPCR);
>
> *val = (snpcr & GEN6_MBC_SNPCR_MASK) >> GEN6_MBC_SNPCR_SHIFT;
> @@ -3770,17 +3770,17 @@ i915_cache_sharing_get(void *data, u64 *val)
> static int
> i915_cache_sharing_set(void *data, u64 val)
> {
> - struct drm_i915_private *dev_priv = data;
> + struct drm_i915_private *i915 = data;
> intel_wakeref_t wakeref;
>
> - if (!(IS_GEN_RANGE(dev_priv, 6, 7)))
> + if (!(IS_GEN_RANGE(i915, 6, 7)))
> return -ENODEV;
>
> if (val > 3)
> return -EINVAL;
>
> DRM_DEBUG_DRIVER("Manually setting uncore sharing to %llu\n", val);
> - with_intel_runtime_pm(dev_priv, wakeref) {
> + with_intel_runtime_pm(i915, wakeref) {
> u32 snpcr;
>
> /* Update the cache sharing policy here as well */
> @@ -3797,7 +3797,7 @@ DEFINE_SIMPLE_ATTRIBUTE(i915_cache_sharing_fops,
> i915_cache_sharing_get, i915_cache_sharing_set,
> "%llu\n");
>
> -static void cherryview_sseu_device_status(struct drm_i915_private *dev_priv,
> +static void cherryview_sseu_device_status(struct drm_i915_private *i915,
> struct sseu_dev_info *sseu)
> {
> #define SS_MAX 2
> @@ -3830,11 +3830,11 @@ static void cherryview_sseu_device_status(struct drm_i915_private *dev_priv,
> #undef SS_MAX
> }
>
> -static void gen10_sseu_device_status(struct drm_i915_private *dev_priv,
> +static void gen10_sseu_device_status(struct drm_i915_private *i915,
> struct sseu_dev_info *sseu)
> {
> #define SS_MAX 6
> - const struct intel_runtime_info *info = RUNTIME_INFO(dev_priv);
> + const struct intel_runtime_info *info = RUNTIME_INFO(i915);
> u32 s_reg[SS_MAX], eu_reg[2 * SS_MAX], eu_mask[2];
> int s, ss;
>
> @@ -3886,11 +3886,11 @@ static void gen10_sseu_device_status(struct drm_i915_private *dev_priv,
> #undef SS_MAX
> }
>
> -static void gen9_sseu_device_status(struct drm_i915_private *dev_priv,
> +static void gen9_sseu_device_status(struct drm_i915_private *i915,
> struct sseu_dev_info *sseu)
> {
> #define SS_MAX 3
> - const struct intel_runtime_info *info = RUNTIME_INFO(dev_priv);
> + const struct intel_runtime_info *info = RUNTIME_INFO(i915);
> u32 s_reg[SS_MAX], eu_reg[2 * SS_MAX], eu_mask[2];
> int s, ss;
>
> @@ -3916,14 +3916,14 @@ static void gen9_sseu_device_status(struct drm_i915_private *dev_priv,
>
> sseu->slice_mask |= BIT(s);
>
> - if (IS_GEN9_BC(dev_priv))
> + if (IS_GEN9_BC(i915))
> sseu->subslice_mask[s] =
> - RUNTIME_INFO(dev_priv)->sseu.subslice_mask[s];
> + RUNTIME_INFO(i915)->sseu.subslice_mask[s];
>
> for (ss = 0; ss < info->sseu.max_subslices; ss++) {
> unsigned int eu_cnt;
>
> - if (IS_GEN9_LP(dev_priv)) {
> + if (IS_GEN9_LP(i915)) {
> if (!(s_reg[s] & (GEN9_PGCTL_SS_ACK(ss))))
> /* skip disabled subslice */
> continue;
> @@ -3942,7 +3942,7 @@ static void gen9_sseu_device_status(struct drm_i915_private *dev_priv,
> #undef SS_MAX
> }
>
> -static void broadwell_sseu_device_status(struct drm_i915_private *dev_priv,
> +static void broadwell_sseu_device_status(struct drm_i915_private *i915,
> struct sseu_dev_info *sseu)
> {
> u32 slice_info = I915_READ(GEN8_GT_SLICE_INFO);
> @@ -3952,10 +3952,10 @@ static void broadwell_sseu_device_status(struct drm_i915_private *dev_priv,
>
> if (sseu->slice_mask) {
> sseu->eu_per_subslice =
> - RUNTIME_INFO(dev_priv)->sseu.eu_per_subslice;
> + RUNTIME_INFO(i915)->sseu.eu_per_subslice;
> for (s = 0; s < fls(sseu->slice_mask); s++) {
> sseu->subslice_mask[s] =
> - RUNTIME_INFO(dev_priv)->sseu.subslice_mask[s];
> + RUNTIME_INFO(i915)->sseu.subslice_mask[s];
> }
> sseu->eu_total = sseu->eu_per_subslice *
> intel_sseu_subslice_total(sseu);
> @@ -3963,7 +3963,7 @@ static void broadwell_sseu_device_status(struct drm_i915_private *dev_priv,
> /* subtract fused off EU(s) from enabled slice(s) */
> for (s = 0; s < fls(sseu->slice_mask); s++) {
> u8 subslice_7eu =
> - RUNTIME_INFO(dev_priv)->sseu.subslice_7eu[s];
> + RUNTIME_INFO(i915)->sseu.subslice_7eu[s];
>
> sseu->eu_total -= hweight8(subslice_7eu);
> }
> @@ -3973,7 +3973,7 @@ static void broadwell_sseu_device_status(struct drm_i915_private *dev_priv,
> static void i915_print_sseu_info(struct seq_file *m, bool is_available_info,
> const struct sseu_dev_info *sseu)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> const char *type = is_available_info ? "Available" : "Enabled";
> int s;
>
> @@ -3995,8 +3995,8 @@ static void i915_print_sseu_info(struct seq_file *m, bool is_available_info,
> if (!is_available_info)
> return;
>
> - seq_printf(m, " Has Pooled EU: %s\n", yesno(HAS_POOLED_EU(dev_priv)));
> - if (HAS_POOLED_EU(dev_priv))
> + seq_printf(m, " Has Pooled EU: %s\n", yesno(HAS_POOLED_EU(i915)));
> + if (HAS_POOLED_EU(i915))
> seq_printf(m, " Min EU in pool: %u\n", sseu->min_eu_in_pool);
>
> seq_printf(m, " Has Slice Power Gating: %s\n",
> @@ -4009,32 +4009,32 @@ static void i915_print_sseu_info(struct seq_file *m, bool is_available_info,
>
> static int i915_sseu_status(struct seq_file *m, void *unused)
> {
> - struct drm_i915_private *dev_priv = node_to_i915(m->private);
> + struct drm_i915_private *i915 = node_to_i915(m->private);
> struct sseu_dev_info sseu;
> intel_wakeref_t wakeref;
>
> - if (INTEL_GEN(dev_priv) < 8)
> + if (INTEL_GEN(i915) < 8)
> return -ENODEV;
>
> seq_puts(m, "SSEU Device Info\n");
> - i915_print_sseu_info(m, true, &RUNTIME_INFO(dev_priv)->sseu);
> + i915_print_sseu_info(m, true, &RUNTIME_INFO(i915)->sseu);
>
> seq_puts(m, "SSEU Device Status\n");
> memset(&sseu, 0, sizeof(sseu));
> - sseu.max_slices = RUNTIME_INFO(dev_priv)->sseu.max_slices;
> - sseu.max_subslices = RUNTIME_INFO(dev_priv)->sseu.max_subslices;
> + sseu.max_slices = RUNTIME_INFO(i915)->sseu.max_slices;
> + sseu.max_subslices = RUNTIME_INFO(i915)->sseu.max_subslices;
> sseu.max_eus_per_subslice =
> - RUNTIME_INFO(dev_priv)->sseu.max_eus_per_subslice;
> -
> - with_intel_runtime_pm(dev_priv, wakeref) {
> - if (IS_CHERRYVIEW(dev_priv))
> - cherryview_sseu_device_status(dev_priv, &sseu);
> - else if (IS_BROADWELL(dev_priv))
> - broadwell_sseu_device_status(dev_priv, &sseu);
> - else if (IS_GEN(dev_priv, 9))
> - gen9_sseu_device_status(dev_priv, &sseu);
> - else if (INTEL_GEN(dev_priv) >= 10)
> - gen10_sseu_device_status(dev_priv, &sseu);
> + RUNTIME_INFO(i915)->sseu.max_eus_per_subslice;
> +
> + with_intel_runtime_pm(i915, wakeref) {
> + if (IS_CHERRYVIEW(i915))
> + cherryview_sseu_device_status(i915, &sseu);
> + else if (IS_BROADWELL(i915))
> + broadwell_sseu_device_status(i915, &sseu);
> + else if (IS_GEN(i915, 9))
> + gen9_sseu_device_status(i915, &sseu);
> + else if (INTEL_GEN(i915) >= 10)
> + gen10_sseu_device_status(i915, &sseu);
> }
>
> i915_print_sseu_info(m, false, &sseu);
> @@ -4077,15 +4077,15 @@ static const struct file_operations i915_forcewake_fops = {
>
> static int i915_hpd_storm_ctl_show(struct seq_file *m, void *data)
> {
> - struct drm_i915_private *dev_priv = m->private;
> - struct i915_hotplug *hotplug = &dev_priv->hotplug;
> + struct drm_i915_private *i915 = m->private;
> + struct i915_hotplug *hotplug = &i915->hotplug;
>
> /* Synchronize with everything first in case there's been an HPD
> * storm, but we haven't finished handling it in the kernel yet
> */
> - synchronize_irq(dev_priv->drm.irq);
> - flush_work(&dev_priv->hotplug.dig_port_work);
> - flush_work(&dev_priv->hotplug.hotplug_work);
> + synchronize_irq(i915->drm.irq);
> + flush_work(&i915->hotplug.dig_port_work);
> + flush_work(&i915->hotplug.hotplug_work);
>
> seq_printf(m, "Threshold: %d\n", hotplug->hpd_storm_threshold);
> seq_printf(m, "Detected: %s\n",
> @@ -4099,8 +4099,8 @@ static ssize_t i915_hpd_storm_ctl_write(struct file *file,
> loff_t *offp)
> {
> struct seq_file *m = file->private_data;
> - struct drm_i915_private *dev_priv = m->private;
> - struct i915_hotplug *hotplug = &dev_priv->hotplug;
> + struct drm_i915_private *i915 = m->private;
> + struct i915_hotplug *hotplug = &i915->hotplug;
> unsigned int new_threshold;
> int i;
> char *newline;
> @@ -4130,15 +4130,15 @@ static ssize_t i915_hpd_storm_ctl_write(struct file *file,
> else
> DRM_DEBUG_KMS("Disabling HPD storm detection\n");
>
> - spin_lock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
> hotplug->hpd_storm_threshold = new_threshold;
> /* Reset the HPD storm stats so we don't accidentally trigger a storm */
> for_each_hpd_pin(i)
> hotplug->stats[i].count = 0;
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_unlock_irq(&i915->irq_lock);
>
> /* Re-enable hpd immediately if we were in an irq storm */
> - flush_delayed_work(&dev_priv->hotplug.reenable_work);
> + flush_delayed_work(&i915->hotplug.reenable_work);
>
> return len;
> }
> @@ -4159,10 +4159,10 @@ static const struct file_operations i915_hpd_storm_ctl_fops = {
>
> static int i915_hpd_short_storm_ctl_show(struct seq_file *m, void *data)
> {
> - struct drm_i915_private *dev_priv = m->private;
> + struct drm_i915_private *i915 = m->private;
>
> seq_printf(m, "Enabled: %s\n",
> - yesno(dev_priv->hotplug.hpd_short_storm_enabled));
> + yesno(i915->hotplug.hpd_short_storm_enabled));
>
> return 0;
> }
> @@ -4179,8 +4179,8 @@ static ssize_t i915_hpd_short_storm_ctl_write(struct file *file,
> size_t len, loff_t *offp)
> {
> struct seq_file *m = file->private_data;
> - struct drm_i915_private *dev_priv = m->private;
> - struct i915_hotplug *hotplug = &dev_priv->hotplug;
> + struct drm_i915_private *i915 = m->private;
> + struct i915_hotplug *hotplug = &i915->hotplug;
> char *newline;
> char tmp[16];
> int i;
> @@ -4201,22 +4201,22 @@ static ssize_t i915_hpd_short_storm_ctl_write(struct file *file,
>
> /* Reset to the "default" state for this system */
> if (strcmp(tmp, "reset") == 0)
> - new_state = !HAS_DP_MST(dev_priv);
> + new_state = !HAS_DP_MST(i915);
> else if (kstrtobool(tmp, &new_state) != 0)
> return -EINVAL;
>
> DRM_DEBUG_KMS("%sabling HPD short storm detection\n",
> new_state ? "En" : "Dis");
>
> - spin_lock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
> hotplug->hpd_short_storm_enabled = new_state;
> /* Reset the HPD storm stats so we don't accidentally trigger a storm */
> for_each_hpd_pin(i)
> hotplug->stats[i].count = 0;
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_unlock_irq(&i915->irq_lock);
>
> /* Re-enable hpd immediately if we were in an irq storm */
> - flush_delayed_work(&dev_priv->hotplug.reenable_work);
> + flush_delayed_work(&i915->hotplug.reenable_work);
>
> return len;
> }
> @@ -4232,11 +4232,11 @@ static const struct file_operations i915_hpd_short_storm_ctl_fops = {
>
> static int i915_drrs_ctl_set(void *data, u64 val)
> {
> - struct drm_i915_private *dev_priv = data;
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_i915_private *i915 = data;
> + struct drm_device *dev = &i915->drm;
> struct intel_crtc *crtc;
>
> - if (INTEL_GEN(dev_priv) < 7)
> + if (INTEL_GEN(i915) < 7)
> return -ENODEV;
>
> for_each_intel_crtc(dev, crtc) {
> @@ -4305,9 +4305,9 @@ i915_fifo_underrun_reset_write(struct file *filp,
> const char __user *ubuf,
> size_t cnt, loff_t *ppos)
> {
> - struct drm_i915_private *dev_priv = filp->private_data;
> + struct drm_i915_private *i915 = filp->private_data;
> struct intel_crtc *intel_crtc;
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_device *dev = &i915->drm;
> int ret;
> bool reset;
>
> @@ -4347,7 +4347,7 @@ i915_fifo_underrun_reset_write(struct file *filp,
> return ret;
> }
>
> - ret = intel_fbc_reset_underrun(dev_priv);
> + ret = intel_fbc_reset_underrun(i915);
> if (ret)
> return ret;
>
> @@ -4437,9 +4437,9 @@ static const struct i915_debugfs_files {
> {"i915_edp_psr_debug", &i915_edp_psr_debug_fops}
> };
>
> -int i915_debugfs_register(struct drm_i915_private *dev_priv)
> +int i915_debugfs_register(struct drm_i915_private *i915)
> {
> - struct drm_minor *minor = dev_priv->drm.primary;
> + struct drm_minor *minor = i915->drm.primary;
> struct dentry *ent;
> int i;
>
> @@ -4692,7 +4692,7 @@ static const struct file_operations i915_dsc_fec_support_fops = {
> int i915_debugfs_connector_add(struct drm_connector *connector)
> {
> struct dentry *root = connector->debugfs_entry;
> - struct drm_i915_private *dev_priv = to_i915(connector->dev);
> + struct drm_i915_private *i915 = to_i915(connector->dev);
>
> /* The connector must have been registered beforehands. */
> if (!root)
> @@ -4717,7 +4717,7 @@ int i915_debugfs_connector_add(struct drm_connector *connector)
> connector, &i915_hdcp_sink_capability_fops);
> }
>
> - if (INTEL_GEN(dev_priv) >= 10 &&
> + if (INTEL_GEN(i915) >= 10 &&
> (connector->connector_type == DRM_MODE_CONNECTOR_DisplayPort ||
> connector->connector_type == DRM_MODE_CONNECTOR_eDP))
> debugfs_create_file("i915_dsc_fec_support", S_IRUGO, root,
> diff --git a/drivers/gpu/drm/i915/i915_debugfs.h b/drivers/gpu/drm/i915/i915_debugfs.h
> index c0cd22eb916d..3e99eaf8de02 100644
> --- a/drivers/gpu/drm/i915/i915_debugfs.h
> +++ b/drivers/gpu/drm/i915/i915_debugfs.h
> @@ -10,10 +10,10 @@ struct drm_i915_private;
> struct drm_connector;
>
> #ifdef CONFIG_DEBUG_FS
> -int i915_debugfs_register(struct drm_i915_private *dev_priv);
> +int i915_debugfs_register(struct drm_i915_private *i915);
> int i915_debugfs_connector_add(struct drm_connector *connector);
> #else
> -static inline int i915_debugfs_register(struct drm_i915_private *dev_priv) { return 0; }
> +static inline int i915_debugfs_register(struct drm_i915_private *i915) { return 0; }
> static inline int i915_debugfs_connector_add(struct drm_connector *connector) { return 0; }
> #endif
>
> diff --git a/drivers/gpu/drm/i915/i915_drv.c b/drivers/gpu/drm/i915/i915_drv.c
> index 81ff2c78fd55..a09d028ec8d8 100644
> --- a/drivers/gpu/drm/i915/i915_drv.c
> +++ b/drivers/gpu/drm/i915/i915_drv.c
> @@ -108,11 +108,11 @@ bool i915_error_injected(void)
> "providing the dmesg log by booting with drm.debug=0xf"
>
> void
> -__i915_printk(struct drm_i915_private *dev_priv, const char *level,
> +__i915_printk(struct drm_i915_private *i915, const char *level,
> const char *fmt, ...)
> {
> static bool shown_bug_once;
> - struct device *kdev = dev_priv->drm.dev;
> + struct device *kdev = i915->drm.dev;
> bool is_error = level[1] <= KERN_ERR[1];
> bool is_debug = level[1] == KERN_DEBUG[1];
> struct va_format vaf;
> @@ -148,74 +148,74 @@ __i915_printk(struct drm_i915_private *dev_priv, const char *level,
>
> /* Map PCH device id to PCH type, or PCH_NONE if unknown. */
> static enum intel_pch
> -intel_pch_type(const struct drm_i915_private *dev_priv, unsigned short id)
> +intel_pch_type(const struct drm_i915_private *i915, unsigned short id)
> {
> switch (id) {
> case INTEL_PCH_IBX_DEVICE_ID_TYPE:
> DRM_DEBUG_KMS("Found Ibex Peak PCH\n");
> - WARN_ON(!IS_GEN(dev_priv, 5));
> + WARN_ON(!IS_GEN(i915, 5));
> return PCH_IBX;
> case INTEL_PCH_CPT_DEVICE_ID_TYPE:
> DRM_DEBUG_KMS("Found CougarPoint PCH\n");
> - WARN_ON(!IS_GEN(dev_priv, 6) && !IS_IVYBRIDGE(dev_priv));
> + WARN_ON(!IS_GEN(i915, 6) && !IS_IVYBRIDGE(i915));
> return PCH_CPT;
> case INTEL_PCH_PPT_DEVICE_ID_TYPE:
> DRM_DEBUG_KMS("Found PantherPoint PCH\n");
> - WARN_ON(!IS_GEN(dev_priv, 6) && !IS_IVYBRIDGE(dev_priv));
> + WARN_ON(!IS_GEN(i915, 6) && !IS_IVYBRIDGE(i915));
> /* PantherPoint is CPT compatible */
> return PCH_CPT;
> case INTEL_PCH_LPT_DEVICE_ID_TYPE:
> DRM_DEBUG_KMS("Found LynxPoint PCH\n");
> - WARN_ON(!IS_HASWELL(dev_priv) && !IS_BROADWELL(dev_priv));
> - WARN_ON(IS_HSW_ULT(dev_priv) || IS_BDW_ULT(dev_priv));
> + WARN_ON(!IS_HASWELL(i915) && !IS_BROADWELL(i915));
> + WARN_ON(IS_HSW_ULT(i915) || IS_BDW_ULT(i915));
> return PCH_LPT;
> case INTEL_PCH_LPT_LP_DEVICE_ID_TYPE:
> DRM_DEBUG_KMS("Found LynxPoint LP PCH\n");
> - WARN_ON(!IS_HASWELL(dev_priv) && !IS_BROADWELL(dev_priv));
> - WARN_ON(!IS_HSW_ULT(dev_priv) && !IS_BDW_ULT(dev_priv));
> + WARN_ON(!IS_HASWELL(i915) && !IS_BROADWELL(i915));
> + WARN_ON(!IS_HSW_ULT(i915) && !IS_BDW_ULT(i915));
> return PCH_LPT;
> case INTEL_PCH_WPT_DEVICE_ID_TYPE:
> DRM_DEBUG_KMS("Found WildcatPoint PCH\n");
> - WARN_ON(!IS_HASWELL(dev_priv) && !IS_BROADWELL(dev_priv));
> - WARN_ON(IS_HSW_ULT(dev_priv) || IS_BDW_ULT(dev_priv));
> + WARN_ON(!IS_HASWELL(i915) && !IS_BROADWELL(i915));
> + WARN_ON(IS_HSW_ULT(i915) || IS_BDW_ULT(i915));
> /* WildcatPoint is LPT compatible */
> return PCH_LPT;
> case INTEL_PCH_WPT_LP_DEVICE_ID_TYPE:
> DRM_DEBUG_KMS("Found WildcatPoint LP PCH\n");
> - WARN_ON(!IS_HASWELL(dev_priv) && !IS_BROADWELL(dev_priv));
> - WARN_ON(!IS_HSW_ULT(dev_priv) && !IS_BDW_ULT(dev_priv));
> + WARN_ON(!IS_HASWELL(i915) && !IS_BROADWELL(i915));
> + WARN_ON(!IS_HSW_ULT(i915) && !IS_BDW_ULT(i915));
> /* WildcatPoint is LPT compatible */
> return PCH_LPT;
> case INTEL_PCH_SPT_DEVICE_ID_TYPE:
> DRM_DEBUG_KMS("Found SunrisePoint PCH\n");
> - WARN_ON(!IS_SKYLAKE(dev_priv) && !IS_KABYLAKE(dev_priv));
> + WARN_ON(!IS_SKYLAKE(i915) && !IS_KABYLAKE(i915));
> return PCH_SPT;
> case INTEL_PCH_SPT_LP_DEVICE_ID_TYPE:
> DRM_DEBUG_KMS("Found SunrisePoint LP PCH\n");
> - WARN_ON(!IS_SKYLAKE(dev_priv) && !IS_KABYLAKE(dev_priv));
> + WARN_ON(!IS_SKYLAKE(i915) && !IS_KABYLAKE(i915));
> return PCH_SPT;
> case INTEL_PCH_KBP_DEVICE_ID_TYPE:
> DRM_DEBUG_KMS("Found Kaby Lake PCH (KBP)\n");
> - WARN_ON(!IS_SKYLAKE(dev_priv) && !IS_KABYLAKE(dev_priv) &&
> - !IS_COFFEELAKE(dev_priv));
> + WARN_ON(!IS_SKYLAKE(i915) && !IS_KABYLAKE(i915) &&
> + !IS_COFFEELAKE(i915));
> /* KBP is SPT compatible */
> return PCH_SPT;
> case INTEL_PCH_CNP_DEVICE_ID_TYPE:
> DRM_DEBUG_KMS("Found Cannon Lake PCH (CNP)\n");
> - WARN_ON(!IS_CANNONLAKE(dev_priv) && !IS_COFFEELAKE(dev_priv));
> + WARN_ON(!IS_CANNONLAKE(i915) && !IS_COFFEELAKE(i915));
> return PCH_CNP;
> case INTEL_PCH_CNP_LP_DEVICE_ID_TYPE:
> DRM_DEBUG_KMS("Found Cannon Lake LP PCH (CNP-LP)\n");
> - WARN_ON(!IS_CANNONLAKE(dev_priv) && !IS_COFFEELAKE(dev_priv));
> + WARN_ON(!IS_CANNONLAKE(i915) && !IS_COFFEELAKE(i915));
> return PCH_CNP;
> case INTEL_PCH_CMP_DEVICE_ID_TYPE:
> DRM_DEBUG_KMS("Found Comet Lake PCH (CMP)\n");
> - WARN_ON(!IS_COFFEELAKE(dev_priv));
> + WARN_ON(!IS_COFFEELAKE(i915));
> /* CometPoint is CNP Compatible */
> return PCH_CNP;
> case INTEL_PCH_ICP_DEVICE_ID_TYPE:
> DRM_DEBUG_KMS("Found Ice Lake PCH\n");
> - WARN_ON(!IS_ICELAKE(dev_priv));
> + WARN_ON(!IS_ICELAKE(i915));
> return PCH_ICP;
> default:
> return PCH_NONE;
> @@ -233,7 +233,7 @@ static bool intel_is_virt_pch(unsigned short id,
> }
>
> static unsigned short
> -intel_virt_detect_pch(const struct drm_i915_private *dev_priv)
> +intel_virt_detect_pch(const struct drm_i915_private *i915)
> {
> unsigned short id = 0;
>
> @@ -244,19 +244,19 @@ intel_virt_detect_pch(const struct drm_i915_private *dev_priv)
> * make an educated guess as to which PCH is really there.
> */
>
> - if (IS_ICELAKE(dev_priv))
> + if (IS_ICELAKE(i915))
> id = INTEL_PCH_ICP_DEVICE_ID_TYPE;
> - else if (IS_CANNONLAKE(dev_priv) || IS_COFFEELAKE(dev_priv))
> + else if (IS_CANNONLAKE(i915) || IS_COFFEELAKE(i915))
> id = INTEL_PCH_CNP_DEVICE_ID_TYPE;
> - else if (IS_KABYLAKE(dev_priv) || IS_SKYLAKE(dev_priv))
> + else if (IS_KABYLAKE(i915) || IS_SKYLAKE(i915))
> id = INTEL_PCH_SPT_DEVICE_ID_TYPE;
> - else if (IS_HSW_ULT(dev_priv) || IS_BDW_ULT(dev_priv))
> + else if (IS_HSW_ULT(i915) || IS_BDW_ULT(i915))
> id = INTEL_PCH_LPT_LP_DEVICE_ID_TYPE;
> - else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> + else if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> id = INTEL_PCH_LPT_DEVICE_ID_TYPE;
> - else if (IS_GEN(dev_priv, 6) || IS_IVYBRIDGE(dev_priv))
> + else if (IS_GEN(i915, 6) || IS_IVYBRIDGE(i915))
> id = INTEL_PCH_CPT_DEVICE_ID_TYPE;
> - else if (IS_GEN(dev_priv, 5))
> + else if (IS_GEN(i915, 5))
> id = INTEL_PCH_IBX_DEVICE_ID_TYPE;
>
> if (id)
> @@ -267,7 +267,7 @@ intel_virt_detect_pch(const struct drm_i915_private *dev_priv)
> return id;
> }
>
> -static void intel_detect_pch(struct drm_i915_private *dev_priv)
> +static void intel_detect_pch(struct drm_i915_private *i915)
> {
> struct pci_dev *pch = NULL;
>
> @@ -291,22 +291,22 @@ static void intel_detect_pch(struct drm_i915_private *dev_priv)
>
> id = pch->device & INTEL_PCH_DEVICE_ID_MASK;
>
> - pch_type = intel_pch_type(dev_priv, id);
> + pch_type = intel_pch_type(i915, id);
> if (pch_type != PCH_NONE) {
> - dev_priv->pch_type = pch_type;
> - dev_priv->pch_id = id;
> + i915->pch_type = pch_type;
> + i915->pch_id = id;
> break;
> } else if (intel_is_virt_pch(id, pch->subsystem_vendor,
> pch->subsystem_device)) {
> - id = intel_virt_detect_pch(dev_priv);
> - pch_type = intel_pch_type(dev_priv, id);
> + id = intel_virt_detect_pch(i915);
> + pch_type = intel_pch_type(i915, id);
>
> /* Sanity check virtual PCH id */
> if (WARN_ON(id && pch_type == PCH_NONE))
> id = 0;
>
> - dev_priv->pch_type = pch_type;
> - dev_priv->pch_id = id;
> + i915->pch_type = pch_type;
> + i915->pch_id = id;
> break;
> }
> }
> @@ -315,10 +315,10 @@ static void intel_detect_pch(struct drm_i915_private *dev_priv)
> * Use PCH_NOP (PCH but no South Display) for PCH platforms without
> * display.
> */
> - if (pch && !HAS_DISPLAY(dev_priv)) {
> + if (pch && !HAS_DISPLAY(i915)) {
> DRM_DEBUG_KMS("Display disabled, reverting to NOP PCH\n");
> - dev_priv->pch_type = PCH_NOP;
> - dev_priv->pch_id = 0;
> + i915->pch_type = PCH_NOP;
> + i915->pch_id = 0;
> }
>
> if (!pch)
> @@ -330,9 +330,9 @@ static void intel_detect_pch(struct drm_i915_private *dev_priv)
> static int i915_getparam_ioctl(struct drm_device *dev, void *data,
> struct drm_file *file_priv)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> - const struct sseu_dev_info *sseu = &RUNTIME_INFO(dev_priv)->sseu;
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct pci_dev *pdev = i915->drm.pdev;
> + const struct sseu_dev_info *sseu = &RUNTIME_INFO(i915)->sseu;
> drm_i915_getparam_t *param = data;
> int value;
>
> @@ -350,40 +350,40 @@ static int i915_getparam_ioctl(struct drm_device *dev, void *data,
> value = pdev->revision;
> break;
> case I915_PARAM_NUM_FENCES_AVAIL:
> - value = dev_priv->num_fence_regs;
> + value = i915->num_fence_regs;
> break;
> case I915_PARAM_HAS_OVERLAY:
> - value = dev_priv->overlay ? 1 : 0;
> + value = i915->overlay ? 1 : 0;
> break;
> case I915_PARAM_HAS_BSD:
> - value = !!dev_priv->engine[VCS0];
> + value = !!i915->engine[VCS0];
> break;
> case I915_PARAM_HAS_BLT:
> - value = !!dev_priv->engine[BCS0];
> + value = !!i915->engine[BCS0];
> break;
> case I915_PARAM_HAS_VEBOX:
> - value = !!dev_priv->engine[VECS0];
> + value = !!i915->engine[VECS0];
> break;
> case I915_PARAM_HAS_BSD2:
> - value = !!dev_priv->engine[VCS1];
> + value = !!i915->engine[VCS1];
> break;
> case I915_PARAM_HAS_LLC:
> - value = HAS_LLC(dev_priv);
> + value = HAS_LLC(i915);
> break;
> case I915_PARAM_HAS_WT:
> - value = HAS_WT(dev_priv);
> + value = HAS_WT(i915);
> break;
> case I915_PARAM_HAS_ALIASING_PPGTT:
> - value = INTEL_PPGTT(dev_priv);
> + value = INTEL_PPGTT(i915);
> break;
> case I915_PARAM_HAS_SEMAPHORES:
> - value = !!(dev_priv->caps.scheduler & I915_SCHEDULER_CAP_SEMAPHORES);
> + value = !!(i915->caps.scheduler & I915_SCHEDULER_CAP_SEMAPHORES);
> break;
> case I915_PARAM_HAS_SECURE_BATCHES:
> value = capable(CAP_SYS_ADMIN);
> break;
> case I915_PARAM_CMD_PARSER_VERSION:
> - value = i915_cmd_parser_get_version(dev_priv);
> + value = i915_cmd_parser_get_version(i915);
> break;
> case I915_PARAM_SUBSLICE_TOTAL:
> value = intel_sseu_subslice_total(sseu);
> @@ -397,21 +397,21 @@ static int i915_getparam_ioctl(struct drm_device *dev, void *data,
> break;
> case I915_PARAM_HAS_GPU_RESET:
> value = i915_modparams.enable_hangcheck &&
> - intel_has_gpu_reset(dev_priv);
> - if (value && intel_has_reset_engine(dev_priv))
> + intel_has_gpu_reset(i915);
> + if (value && intel_has_reset_engine(i915))
> value = 2;
> break;
> case I915_PARAM_HAS_RESOURCE_STREAMER:
> value = 0;
> break;
> case I915_PARAM_HAS_POOLED_EU:
> - value = HAS_POOLED_EU(dev_priv);
> + value = HAS_POOLED_EU(i915);
> break;
> case I915_PARAM_MIN_EU_IN_POOL:
> value = sseu->min_eu_in_pool;
> break;
> case I915_PARAM_HUC_STATUS:
> - value = intel_huc_check_status(&dev_priv->huc);
> + value = intel_huc_check_status(&i915->huc);
> if (value < 0)
> return value;
> break;
> @@ -423,7 +423,7 @@ static int i915_getparam_ioctl(struct drm_device *dev, void *data,
> value = i915_gem_mmap_gtt_version();
> break;
> case I915_PARAM_HAS_SCHEDULER:
> - value = dev_priv->caps.scheduler;
> + value = i915->caps.scheduler;
> break;
>
> case I915_PARAM_MMAP_VERSION:
> @@ -456,7 +456,7 @@ static int i915_getparam_ioctl(struct drm_device *dev, void *data,
> value = 1;
> break;
> case I915_PARAM_HAS_CONTEXT_ISOLATION:
> - value = intel_engines_has_context_isolation(dev_priv);
> + value = intel_engines_has_context_isolation(i915);
> break;
> case I915_PARAM_SLICE_MASK:
> value = sseu->slice_mask;
> @@ -469,10 +469,10 @@ static int i915_getparam_ioctl(struct drm_device *dev, void *data,
> return -ENODEV;
> break;
> case I915_PARAM_CS_TIMESTAMP_FREQUENCY:
> - value = 1000 * RUNTIME_INFO(dev_priv)->cs_timestamp_frequency_khz;
> + value = 1000 * RUNTIME_INFO(i915)->cs_timestamp_frequency_khz;
> break;
> case I915_PARAM_MMAP_GTT_COHERENT:
> - value = INTEL_INFO(dev_priv)->has_coherent_ggtt;
> + value = INTEL_INFO(i915)->has_coherent_ggtt;
> break;
> default:
> DRM_DEBUG("Unknown parameter %d\n", param->param);
> @@ -485,13 +485,13 @@ static int i915_getparam_ioctl(struct drm_device *dev, void *data,
> return 0;
> }
>
> -static int i915_get_bridge_dev(struct drm_i915_private *dev_priv)
> +static int i915_get_bridge_dev(struct drm_i915_private *i915)
> {
> - int domain = pci_domain_nr(dev_priv->drm.pdev->bus);
> + int domain = pci_domain_nr(i915->drm.pdev->bus);
>
> - dev_priv->bridge_dev =
> + i915->bridge_dev =
> pci_get_domain_bus_and_slot(domain, 0, PCI_DEVFN(0, 0));
> - if (!dev_priv->bridge_dev) {
> + if (!i915->bridge_dev) {
> DRM_ERROR("bridge device not found\n");
> return -1;
> }
> @@ -500,16 +500,16 @@ static int i915_get_bridge_dev(struct drm_i915_private *dev_priv)
>
> /* Allocate space for the MCH regs if needed, return nonzero on error */
> static int
> -intel_alloc_mchbar_resource(struct drm_i915_private *dev_priv)
> +intel_alloc_mchbar_resource(struct drm_i915_private *i915)
> {
> - int reg = INTEL_GEN(dev_priv) >= 4 ? MCHBAR_I965 : MCHBAR_I915;
> + int reg = INTEL_GEN(i915) >= 4 ? MCHBAR_I965 : MCHBAR_I915;
> u32 temp_lo, temp_hi = 0;
> u64 mchbar_addr;
> int ret;
>
> - if (INTEL_GEN(dev_priv) >= 4)
> - pci_read_config_dword(dev_priv->bridge_dev, reg + 4, &temp_hi);
> - pci_read_config_dword(dev_priv->bridge_dev, reg, &temp_lo);
> + if (INTEL_GEN(i915) >= 4)
> + pci_read_config_dword(i915->bridge_dev, reg + 4, &temp_hi);
> + pci_read_config_dword(i915->bridge_dev, reg, &temp_lo);
> mchbar_addr = ((u64)temp_hi << 32) | temp_lo;
>
> /* If ACPI doesn't have it, assume we need to allocate it ourselves */
> @@ -520,47 +520,47 @@ intel_alloc_mchbar_resource(struct drm_i915_private *dev_priv)
> #endif
>
> /* Get some space for it */
> - dev_priv->mch_res.name = "i915 MCHBAR";
> - dev_priv->mch_res.flags = IORESOURCE_MEM;
> - ret = pci_bus_alloc_resource(dev_priv->bridge_dev->bus,
> - &dev_priv->mch_res,
> + i915->mch_res.name = "i915 MCHBAR";
> + i915->mch_res.flags = IORESOURCE_MEM;
> + ret = pci_bus_alloc_resource(i915->bridge_dev->bus,
> + &i915->mch_res,
> MCHBAR_SIZE, MCHBAR_SIZE,
> PCIBIOS_MIN_MEM,
> 0, pcibios_align_resource,
> - dev_priv->bridge_dev);
> + i915->bridge_dev);
> if (ret) {
> DRM_DEBUG_DRIVER("failed bus alloc: %d\n", ret);
> - dev_priv->mch_res.start = 0;
> + i915->mch_res.start = 0;
> return ret;
> }
>
> - if (INTEL_GEN(dev_priv) >= 4)
> - pci_write_config_dword(dev_priv->bridge_dev, reg + 4,
> - upper_32_bits(dev_priv->mch_res.start));
> + if (INTEL_GEN(i915) >= 4)
> + pci_write_config_dword(i915->bridge_dev, reg + 4,
> + upper_32_bits(i915->mch_res.start));
>
> - pci_write_config_dword(dev_priv->bridge_dev, reg,
> - lower_32_bits(dev_priv->mch_res.start));
> + pci_write_config_dword(i915->bridge_dev, reg,
> + lower_32_bits(i915->mch_res.start));
> return 0;
> }
>
> /* Setup MCHBAR if possible, return true if we should disable it again */
> static void
> -intel_setup_mchbar(struct drm_i915_private *dev_priv)
> +intel_setup_mchbar(struct drm_i915_private *i915)
> {
> - int mchbar_reg = INTEL_GEN(dev_priv) >= 4 ? MCHBAR_I965 : MCHBAR_I915;
> + int mchbar_reg = INTEL_GEN(i915) >= 4 ? MCHBAR_I965 : MCHBAR_I915;
> u32 temp;
> bool enabled;
>
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> return;
>
> - dev_priv->mchbar_need_disable = false;
> + i915->mchbar_need_disable = false;
>
> - if (IS_I915G(dev_priv) || IS_I915GM(dev_priv)) {
> - pci_read_config_dword(dev_priv->bridge_dev, DEVEN, &temp);
> + if (IS_I915G(i915) || IS_I915GM(i915)) {
> + pci_read_config_dword(i915->bridge_dev, DEVEN, &temp);
> enabled = !!(temp & DEVEN_MCHBAR_EN);
> } else {
> - pci_read_config_dword(dev_priv->bridge_dev, mchbar_reg, &temp);
> + pci_read_config_dword(i915->bridge_dev, mchbar_reg, &temp);
> enabled = temp & 1;
> }
>
> @@ -568,56 +568,56 @@ intel_setup_mchbar(struct drm_i915_private *dev_priv)
> if (enabled)
> return;
>
> - if (intel_alloc_mchbar_resource(dev_priv))
> + if (intel_alloc_mchbar_resource(i915))
> return;
>
> - dev_priv->mchbar_need_disable = true;
> + i915->mchbar_need_disable = true;
>
> /* Space is allocated or reserved, so enable it. */
> - if (IS_I915G(dev_priv) || IS_I915GM(dev_priv)) {
> - pci_write_config_dword(dev_priv->bridge_dev, DEVEN,
> + if (IS_I915G(i915) || IS_I915GM(i915)) {
> + pci_write_config_dword(i915->bridge_dev, DEVEN,
> temp | DEVEN_MCHBAR_EN);
> } else {
> - pci_read_config_dword(dev_priv->bridge_dev, mchbar_reg, &temp);
> - pci_write_config_dword(dev_priv->bridge_dev, mchbar_reg, temp | 1);
> + pci_read_config_dword(i915->bridge_dev, mchbar_reg, &temp);
> + pci_write_config_dword(i915->bridge_dev, mchbar_reg, temp | 1);
> }
> }
>
> static void
> -intel_teardown_mchbar(struct drm_i915_private *dev_priv)
> +intel_teardown_mchbar(struct drm_i915_private *i915)
> {
> - int mchbar_reg = INTEL_GEN(dev_priv) >= 4 ? MCHBAR_I965 : MCHBAR_I915;
> + int mchbar_reg = INTEL_GEN(i915) >= 4 ? MCHBAR_I965 : MCHBAR_I915;
>
> - if (dev_priv->mchbar_need_disable) {
> - if (IS_I915G(dev_priv) || IS_I915GM(dev_priv)) {
> + if (i915->mchbar_need_disable) {
> + if (IS_I915G(i915) || IS_I915GM(i915)) {
> u32 deven_val;
>
> - pci_read_config_dword(dev_priv->bridge_dev, DEVEN,
> + pci_read_config_dword(i915->bridge_dev, DEVEN,
> &deven_val);
> deven_val &= ~DEVEN_MCHBAR_EN;
> - pci_write_config_dword(dev_priv->bridge_dev, DEVEN,
> + pci_write_config_dword(i915->bridge_dev, DEVEN,
> deven_val);
> } else {
> u32 mchbar_val;
>
> - pci_read_config_dword(dev_priv->bridge_dev, mchbar_reg,
> + pci_read_config_dword(i915->bridge_dev, mchbar_reg,
> &mchbar_val);
> mchbar_val &= ~1;
> - pci_write_config_dword(dev_priv->bridge_dev, mchbar_reg,
> + pci_write_config_dword(i915->bridge_dev, mchbar_reg,
> mchbar_val);
> }
> }
>
> - if (dev_priv->mch_res.start)
> - release_resource(&dev_priv->mch_res);
> + if (i915->mch_res.start)
> + release_resource(&i915->mch_res);
> }
>
> /* true = enable decode, false = disable decoder */
> static unsigned int i915_vga_set_decode(void *cookie, bool state)
> {
> - struct drm_i915_private *dev_priv = cookie;
> + struct drm_i915_private *i915 = cookie;
>
> - intel_modeset_vga_set_state(dev_priv, state);
> + intel_modeset_vga_set_state(i915, state);
> if (state)
> return VGA_RSRC_LEGACY_IO | VGA_RSRC_LEGACY_MEM |
> VGA_RSRC_NORMAL_IO | VGA_RSRC_NORMAL_MEM;
> @@ -668,21 +668,21 @@ static const struct vga_switcheroo_client_ops i915_switcheroo_ops = {
>
> static int i915_load_modeset_init(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct pci_dev *pdev = i915->drm.pdev;
> int ret;
>
> if (i915_inject_load_failure())
> return -ENODEV;
>
> - if (HAS_DISPLAY(dev_priv)) {
> - ret = drm_vblank_init(&dev_priv->drm,
> - INTEL_INFO(dev_priv)->num_pipes);
> + if (HAS_DISPLAY(i915)) {
> + ret = drm_vblank_init(&i915->drm,
> + INTEL_INFO(i915)->num_pipes);
> if (ret)
> goto out;
> }
>
> - intel_bios_init(dev_priv);
> + intel_bios_init(i915);
>
> /* If we have > 1 VGA cards, then we need to arbitrate access
> * to the common VGA resources.
> @@ -691,7 +691,7 @@ static int i915_load_modeset_init(struct drm_device *dev)
> * then we do not take part in VGA arbitration and the
> * vga_client_register() fails with -ENODEV.
> */
> - ret = vga_client_register(pdev, dev_priv, NULL, i915_vga_set_decode);
> + ret = vga_client_register(pdev, i915, NULL, i915_vga_set_decode);
> if (ret && ret != -ENODEV)
> goto out;
>
> @@ -702,17 +702,17 @@ static int i915_load_modeset_init(struct drm_device *dev)
> goto cleanup_vga_client;
>
> /* must happen before intel_power_domains_init_hw() on VLV/CHV */
> - intel_update_rawclk(dev_priv);
> + intel_update_rawclk(i915);
>
> - intel_power_domains_init_hw(dev_priv, false);
> + intel_power_domains_init_hw(i915, false);
>
> - intel_csr_ucode_init(dev_priv);
> + intel_csr_ucode_init(i915);
>
> - ret = intel_irq_install(dev_priv);
> + ret = intel_irq_install(i915);
> if (ret)
> goto cleanup_csr;
>
> - intel_gmbus_setup(dev_priv);
> + intel_gmbus_setup(i915);
>
> /* Important: The output setup functions called by modeset_init need
> * working irqs for e.g. gmbus and dp aux transfers. */
> @@ -720,13 +720,13 @@ static int i915_load_modeset_init(struct drm_device *dev)
> if (ret)
> goto cleanup_irq;
>
> - ret = i915_gem_init(dev_priv);
> + ret = i915_gem_init(i915);
> if (ret)
> goto cleanup_modeset;
>
> - intel_overlay_setup(dev_priv);
> + intel_overlay_setup(i915);
>
> - if (!HAS_DISPLAY(dev_priv))
> + if (!HAS_DISPLAY(i915))
> return 0;
>
> ret = intel_fbdev_init(dev);
> @@ -734,24 +734,24 @@ static int i915_load_modeset_init(struct drm_device *dev)
> goto cleanup_gem;
>
> /* Only enable hotplug handling once the fbdev is fully set up. */
> - intel_hpd_init(dev_priv);
> + intel_hpd_init(i915);
>
> - intel_init_ipc(dev_priv);
> + intel_init_ipc(i915);
>
> return 0;
>
> cleanup_gem:
> - i915_gem_suspend(dev_priv);
> - i915_gem_fini_hw(dev_priv);
> - i915_gem_fini(dev_priv);
> + i915_gem_suspend(i915);
> + i915_gem_fini_hw(i915);
> + i915_gem_fini(i915);
> cleanup_modeset:
> intel_modeset_cleanup(dev);
> cleanup_irq:
> drm_irq_uninstall(dev);
> - intel_gmbus_teardown(dev_priv);
> + intel_gmbus_teardown(i915);
> cleanup_csr:
> - intel_csr_ucode_fini(dev_priv);
> - intel_power_domains_fini_hw(dev_priv);
> + intel_csr_ucode_fini(i915);
> + intel_power_domains_fini_hw(i915);
> vga_switcheroo_unregister_client(pdev);
> cleanup_vga_client:
> vga_client_register(pdev, NULL, NULL, NULL);
> @@ -759,11 +759,11 @@ static int i915_load_modeset_init(struct drm_device *dev)
> return ret;
> }
>
> -static int i915_kick_out_firmware_fb(struct drm_i915_private *dev_priv)
> +static int i915_kick_out_firmware_fb(struct drm_i915_private *i915)
> {
> struct apertures_struct *ap;
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> - struct i915_ggtt *ggtt = &dev_priv->ggtt;
> + struct pci_dev *pdev = i915->drm.pdev;
> + struct i915_ggtt *ggtt = &i915->ggtt;
> bool primary;
> int ret;
>
> @@ -784,22 +784,22 @@ static int i915_kick_out_firmware_fb(struct drm_i915_private *dev_priv)
> return ret;
> }
>
> -static void intel_init_dpio(struct drm_i915_private *dev_priv)
> +static void intel_init_dpio(struct drm_i915_private *i915)
> {
> /*
> * IOSF_PORT_DPIO is used for VLV x2 PHY (DP/HDMI B and C),
> * CHV x1 PHY (DP/HDMI D)
> * IOSF_PORT_DPIO_2 is used for CHV x2 PHY (DP/HDMI B and C)
> */
> - if (IS_CHERRYVIEW(dev_priv)) {
> + if (IS_CHERRYVIEW(i915)) {
> DPIO_PHY_IOSF_PORT(DPIO_PHY0) = IOSF_PORT_DPIO_2;
> DPIO_PHY_IOSF_PORT(DPIO_PHY1) = IOSF_PORT_DPIO;
> - } else if (IS_VALLEYVIEW(dev_priv)) {
> + } else if (IS_VALLEYVIEW(i915)) {
> DPIO_PHY_IOSF_PORT(DPIO_PHY0) = IOSF_PORT_DPIO;
> }
> }
>
> -static int i915_workqueues_init(struct drm_i915_private *dev_priv)
> +static int i915_workqueues_init(struct drm_i915_private *i915)
> {
> /*
> * The i915 workqueue is primarily used for batched retirement of
> @@ -815,18 +815,18 @@ static int i915_workqueues_init(struct drm_i915_private *dev_priv)
> * so there is no point in running more than one instance of the
> * workqueue at any time. Use an ordered one.
> */
> - dev_priv->wq = alloc_ordered_workqueue("i915", 0);
> - if (dev_priv->wq == NULL)
> + i915->wq = alloc_ordered_workqueue("i915", 0);
> + if (i915->wq == NULL)
> goto out_err;
>
> - dev_priv->hotplug.dp_wq = alloc_ordered_workqueue("i915-dp", 0);
> - if (dev_priv->hotplug.dp_wq == NULL)
> + i915->hotplug.dp_wq = alloc_ordered_workqueue("i915-dp", 0);
> + if (i915->hotplug.dp_wq == NULL)
> goto out_free_wq;
>
> return 0;
>
> out_free_wq:
> - destroy_workqueue(dev_priv->wq);
> + destroy_workqueue(i915->wq);
> out_err:
> DRM_ERROR("Failed to allocate workqueues.\n");
>
> @@ -842,10 +842,10 @@ static void i915_engines_cleanup(struct drm_i915_private *i915)
> kfree(engine);
> }
>
> -static void i915_workqueues_cleanup(struct drm_i915_private *dev_priv)
> +static void i915_workqueues_cleanup(struct drm_i915_private *i915)
> {
> - destroy_workqueue(dev_priv->hotplug.dp_wq);
> - destroy_workqueue(dev_priv->wq);
> + destroy_workqueue(i915->hotplug.dp_wq);
> + destroy_workqueue(i915->wq);
> }
>
> /*
> @@ -858,14 +858,14 @@ static void i915_workqueues_cleanup(struct drm_i915_private *dev_priv)
> * (workarounds have a habit of persisting!). Anything older than that
> * should be removed along with the complications they introduce.
> */
> -static void intel_detect_preproduction_hw(struct drm_i915_private *dev_priv)
> +static void intel_detect_preproduction_hw(struct drm_i915_private *i915)
> {
> bool pre = false;
>
> - pre |= IS_HSW_EARLY_SDV(dev_priv);
> - pre |= IS_SKL_REVID(dev_priv, 0, SKL_REVID_F0);
> - pre |= IS_BXT_REVID(dev_priv, 0, BXT_REVID_B_LAST);
> - pre |= IS_KBL_REVID(dev_priv, 0, KBL_REVID_A0);
> + pre |= IS_HSW_EARLY_SDV(i915);
> + pre |= IS_SKL_REVID(i915, 0, SKL_REVID_F0);
> + pre |= IS_BXT_REVID(i915, 0, BXT_REVID_B_LAST);
> + pre |= IS_KBL_REVID(i915, 0, KBL_REVID_A0);
>
> if (pre) {
> DRM_ERROR("This is a pre-production stepping. "
> @@ -876,7 +876,7 @@ static void intel_detect_preproduction_hw(struct drm_i915_private *dev_priv)
>
> /**
> * i915_driver_init_early - setup state not requiring device access
> - * @dev_priv: device private
> + * @i915: device private
> *
> * Initialize everything that is a "SW-only" state, that is state not
> * requiring accessing the device or exposing the driver via kernel internal
> @@ -884,152 +884,152 @@ static void intel_detect_preproduction_hw(struct drm_i915_private *dev_priv)
> * system memory allocation, setting up device specific attributes and
> * function hooks not requiring accessing the device.
> */
> -static int i915_driver_init_early(struct drm_i915_private *dev_priv)
> +static int i915_driver_init_early(struct drm_i915_private *i915)
> {
> int ret = 0;
>
> if (i915_inject_load_failure())
> return -ENODEV;
>
> - intel_device_info_subplatform_init(dev_priv);
> + intel_device_info_subplatform_init(i915);
>
> - intel_uncore_init_early(&dev_priv->uncore);
> + intel_uncore_init_early(&i915->uncore);
>
> - spin_lock_init(&dev_priv->irq_lock);
> - spin_lock_init(&dev_priv->gpu_error.lock);
> - mutex_init(&dev_priv->backlight_lock);
> + spin_lock_init(&i915->irq_lock);
> + spin_lock_init(&i915->gpu_error.lock);
> + mutex_init(&i915->backlight_lock);
>
> - mutex_init(&dev_priv->sb_lock);
> - pm_qos_add_request(&dev_priv->sb_qos,
> + mutex_init(&i915->sb_lock);
> + pm_qos_add_request(&i915->sb_qos,
> PM_QOS_CPU_DMA_LATENCY, PM_QOS_DEFAULT_VALUE);
>
> - mutex_init(&dev_priv->av_mutex);
> - mutex_init(&dev_priv->wm.wm_mutex);
> - mutex_init(&dev_priv->pps_mutex);
> - mutex_init(&dev_priv->hdcp_comp_mutex);
> + mutex_init(&i915->av_mutex);
> + mutex_init(&i915->wm.wm_mutex);
> + mutex_init(&i915->pps_mutex);
> + mutex_init(&i915->hdcp_comp_mutex);
>
> - i915_memcpy_init_early(dev_priv);
> - intel_runtime_pm_init_early(dev_priv);
> + i915_memcpy_init_early(i915);
> + intel_runtime_pm_init_early(i915);
>
> - ret = i915_workqueues_init(dev_priv);
> + ret = i915_workqueues_init(i915);
> if (ret < 0)
> goto err_engines;
>
> - ret = i915_gem_init_early(dev_priv);
> + ret = i915_gem_init_early(i915);
> if (ret < 0)
> goto err_workqueues;
>
> /* This must be called before any calls to HAS_PCH_* */
> - intel_detect_pch(dev_priv);
> + intel_detect_pch(i915);
>
> - intel_wopcm_init_early(&dev_priv->wopcm);
> - intel_uc_init_early(dev_priv);
> - intel_pm_setup(dev_priv);
> - intel_init_dpio(dev_priv);
> - ret = intel_power_domains_init(dev_priv);
> + intel_wopcm_init_early(&i915->wopcm);
> + intel_uc_init_early(i915);
> + intel_pm_setup(i915);
> + intel_init_dpio(i915);
> + ret = intel_power_domains_init(i915);
> if (ret < 0)
> goto err_uc;
> - intel_irq_init(dev_priv);
> - intel_hangcheck_init(dev_priv);
> - intel_init_display_hooks(dev_priv);
> - intel_init_clock_gating_hooks(dev_priv);
> - intel_init_audio_hooks(dev_priv);
> - intel_display_crc_init(dev_priv);
> + intel_irq_init(i915);
> + intel_hangcheck_init(i915);
> + intel_init_display_hooks(i915);
> + intel_init_clock_gating_hooks(i915);
> + intel_init_audio_hooks(i915);
> + intel_display_crc_init(i915);
>
> - intel_detect_preproduction_hw(dev_priv);
> + intel_detect_preproduction_hw(i915);
>
> return 0;
>
> err_uc:
> - intel_uc_cleanup_early(dev_priv);
> - i915_gem_cleanup_early(dev_priv);
> + intel_uc_cleanup_early(i915);
> + i915_gem_cleanup_early(i915);
> err_workqueues:
> - i915_workqueues_cleanup(dev_priv);
> + i915_workqueues_cleanup(i915);
> err_engines:
> - i915_engines_cleanup(dev_priv);
> + i915_engines_cleanup(i915);
> return ret;
> }
>
> /**
> * i915_driver_cleanup_early - cleanup the setup done in i915_driver_init_early()
> - * @dev_priv: device private
> + * @i915: device private
> */
> -static void i915_driver_cleanup_early(struct drm_i915_private *dev_priv)
> +static void i915_driver_cleanup_early(struct drm_i915_private *i915)
> {
> - intel_irq_fini(dev_priv);
> - intel_power_domains_cleanup(dev_priv);
> - intel_uc_cleanup_early(dev_priv);
> - i915_gem_cleanup_early(dev_priv);
> - i915_workqueues_cleanup(dev_priv);
> - i915_engines_cleanup(dev_priv);
> + intel_irq_fini(i915);
> + intel_power_domains_cleanup(i915);
> + intel_uc_cleanup_early(i915);
> + i915_gem_cleanup_early(i915);
> + i915_workqueues_cleanup(i915);
> + i915_engines_cleanup(i915);
>
> - pm_qos_remove_request(&dev_priv->sb_qos);
> - mutex_destroy(&dev_priv->sb_lock);
> + pm_qos_remove_request(&i915->sb_qos);
> + mutex_destroy(&i915->sb_lock);
> }
>
> /**
> * i915_driver_init_mmio - setup device MMIO
> - * @dev_priv: device private
> + * @i915: device private
> *
> * Setup minimal device state necessary for MMIO accesses later in the
> * initialization sequence. The setup here should avoid any other device-wide
> * side effects or exposing the driver via kernel internal or user space
> * interfaces.
> */
> -static int i915_driver_init_mmio(struct drm_i915_private *dev_priv)
> +static int i915_driver_init_mmio(struct drm_i915_private *i915)
> {
> int ret;
>
> if (i915_inject_load_failure())
> return -ENODEV;
>
> - if (i915_get_bridge_dev(dev_priv))
> + if (i915_get_bridge_dev(i915))
> return -EIO;
>
> - ret = intel_uncore_init_mmio(&dev_priv->uncore);
> + ret = intel_uncore_init_mmio(&i915->uncore);
> if (ret < 0)
> goto err_bridge;
>
> /* Try to make sure MCHBAR is enabled before poking at it */
> - intel_setup_mchbar(dev_priv);
> + intel_setup_mchbar(i915);
>
> - intel_device_info_init_mmio(dev_priv);
> + intel_device_info_init_mmio(i915);
>
> - intel_uncore_prune_mmio_domains(&dev_priv->uncore);
> + intel_uncore_prune_mmio_domains(&i915->uncore);
>
> - intel_uc_init_mmio(dev_priv);
> + intel_uc_init_mmio(i915);
>
> - ret = intel_engines_init_mmio(dev_priv);
> + ret = intel_engines_init_mmio(i915);
> if (ret)
> goto err_uncore;
>
> - i915_gem_init_mmio(dev_priv);
> + i915_gem_init_mmio(i915);
>
> return 0;
>
> err_uncore:
> - intel_teardown_mchbar(dev_priv);
> - intel_uncore_fini_mmio(&dev_priv->uncore);
> + intel_teardown_mchbar(i915);
> + intel_uncore_fini_mmio(&i915->uncore);
> err_bridge:
> - pci_dev_put(dev_priv->bridge_dev);
> + pci_dev_put(i915->bridge_dev);
>
> return ret;
> }
>
> /**
> * i915_driver_cleanup_mmio - cleanup the setup done in i915_driver_init_mmio()
> - * @dev_priv: device private
> + * @i915: device private
> */
> -static void i915_driver_cleanup_mmio(struct drm_i915_private *dev_priv)
> +static void i915_driver_cleanup_mmio(struct drm_i915_private *i915)
> {
> - intel_teardown_mchbar(dev_priv);
> - intel_uncore_fini_mmio(&dev_priv->uncore);
> - pci_dev_put(dev_priv->bridge_dev);
> + intel_teardown_mchbar(i915);
> + intel_uncore_fini_mmio(&i915->uncore);
> + pci_dev_put(i915->bridge_dev);
> }
>
> -static void intel_sanitize_options(struct drm_i915_private *dev_priv)
> +static void intel_sanitize_options(struct drm_i915_private *i915)
> {
> - intel_gvt_sanitize_options(dev_priv);
> + intel_gvt_sanitize_options(i915);
> }
>
> #define DRAM_TYPE_STR(type) [INTEL_DRAM_ ## type] = #type
> @@ -1131,11 +1131,11 @@ skl_is_16gb_dimm(const struct dram_dimm_info *dimm)
> }
>
> static void
> -skl_dram_get_dimm_info(struct drm_i915_private *dev_priv,
> +skl_dram_get_dimm_info(struct drm_i915_private *i915,
> struct dram_dimm_info *dimm,
> int channel, char dimm_name, u16 val)
> {
> - if (INTEL_GEN(dev_priv) >= 10) {
> + if (INTEL_GEN(i915) >= 10) {
> dimm->size = cnl_get_dimm_size(val);
> dimm->width = cnl_get_dimm_width(val);
> dimm->ranks = cnl_get_dimm_ranks(val);
> @@ -1151,13 +1151,13 @@ skl_dram_get_dimm_info(struct drm_i915_private *dev_priv,
> }
>
> static int
> -skl_dram_get_channel_info(struct drm_i915_private *dev_priv,
> +skl_dram_get_channel_info(struct drm_i915_private *i915,
> struct dram_channel_info *ch,
> int channel, u32 val)
> {
> - skl_dram_get_dimm_info(dev_priv, &ch->dimm_l,
> + skl_dram_get_dimm_info(i915, &ch->dimm_l,
> channel, 'L', val & 0xffff);
> - skl_dram_get_dimm_info(dev_priv, &ch->dimm_s,
> + skl_dram_get_dimm_info(i915, &ch->dimm_s,
> channel, 'S', val >> 16);
>
> if (ch->dimm_l.size == 0 && ch->dimm_s.size == 0) {
> @@ -1192,20 +1192,20 @@ intel_is_dram_symmetric(const struct dram_channel_info *ch0,
> }
>
> static int
> -skl_dram_get_channels_info(struct drm_i915_private *dev_priv)
> +skl_dram_get_channels_info(struct drm_i915_private *i915)
> {
> - struct dram_info *dram_info = &dev_priv->dram_info;
> + struct dram_info *dram_info = &i915->dram_info;
> struct dram_channel_info ch0 = {}, ch1 = {};
> u32 val;
> int ret;
>
> val = I915_READ(SKL_MAD_DIMM_CH0_0_0_0_MCHBAR_MCMAIN);
> - ret = skl_dram_get_channel_info(dev_priv, &ch0, 0, val);
> + ret = skl_dram_get_channel_info(i915, &ch0, 0, val);
> if (ret == 0)
> dram_info->num_channels++;
>
> val = I915_READ(SKL_MAD_DIMM_CH1_0_0_0_MCHBAR_MCMAIN);
> - ret = skl_dram_get_channel_info(dev_priv, &ch1, 1, val);
> + ret = skl_dram_get_channel_info(i915, &ch1, 1, val);
> if (ret == 0)
> dram_info->num_channels++;
>
> @@ -1239,7 +1239,7 @@ skl_dram_get_channels_info(struct drm_i915_private *dev_priv)
> }
>
> static enum intel_dram_type
> -skl_get_dram_type(struct drm_i915_private *dev_priv)
> +skl_get_dram_type(struct drm_i915_private *i915)
> {
> u32 val;
>
> @@ -1261,16 +1261,16 @@ skl_get_dram_type(struct drm_i915_private *dev_priv)
> }
>
> static int
> -skl_get_dram_info(struct drm_i915_private *dev_priv)
> +skl_get_dram_info(struct drm_i915_private *i915)
> {
> - struct dram_info *dram_info = &dev_priv->dram_info;
> + struct dram_info *dram_info = &i915->dram_info;
> u32 mem_freq_khz, val;
> int ret;
>
> - dram_info->type = skl_get_dram_type(dev_priv);
> + dram_info->type = skl_get_dram_type(i915);
> DRM_DEBUG_KMS("DRAM type: %s\n", intel_dram_type_str(dram_info->type));
>
> - ret = skl_dram_get_channels_info(dev_priv);
> + ret = skl_dram_get_channels_info(i915);
> if (ret)
> return ret;
>
> @@ -1370,9 +1370,9 @@ static void bxt_get_dimm_info(struct dram_dimm_info *dimm,
> }
>
> static int
> -bxt_get_dram_info(struct drm_i915_private *dev_priv)
> +bxt_get_dram_info(struct drm_i915_private *i915)
> {
> - struct dram_info *dram_info = &dev_priv->dram_info;
> + struct dram_info *dram_info = &i915->dram_info;
> u32 dram_channels;
> u32 mem_freq_khz, val;
> u8 num_active_channels;
> @@ -1443,9 +1443,9 @@ bxt_get_dram_info(struct drm_i915_private *dev_priv)
> }
>
> static void
> -intel_get_dram_info(struct drm_i915_private *dev_priv)
> +intel_get_dram_info(struct drm_i915_private *i915)
> {
> - struct dram_info *dram_info = &dev_priv->dram_info;
> + struct dram_info *dram_info = &i915->dram_info;
> int ret;
>
> /*
> @@ -1453,15 +1453,15 @@ intel_get_dram_info(struct drm_i915_private *dev_priv)
> * This is only used for the level 0 watermark latency
> * w/a which does not apply to bxt/glk.
> */
> - dram_info->is_16gb_dimm = !IS_GEN9_LP(dev_priv);
> + dram_info->is_16gb_dimm = !IS_GEN9_LP(i915);
>
> - if (INTEL_GEN(dev_priv) < 9)
> + if (INTEL_GEN(i915) < 9)
> return;
>
> - if (IS_GEN9_LP(dev_priv))
> - ret = bxt_get_dram_info(dev_priv);
> + if (IS_GEN9_LP(i915))
> + ret = bxt_get_dram_info(i915);
> else
> - ret = skl_get_dram_info(dev_priv);
> + ret = skl_get_dram_info(i915);
> if (ret)
> return;
>
> @@ -1473,7 +1473,7 @@ intel_get_dram_info(struct drm_i915_private *dev_priv)
> dram_info->ranks, yesno(dram_info->is_16gb_dimm));
> }
>
> -static u32 gen9_edram_size_mb(struct drm_i915_private *dev_priv, u32 cap)
> +static u32 gen9_edram_size_mb(struct drm_i915_private *i915, u32 cap)
> {
> const unsigned int ways[8] = { 4, 8, 12, 16, 16, 16, 16, 16 };
> const unsigned int sets[4] = { 1, 1, 2, 2 };
> @@ -1483,16 +1483,16 @@ static u32 gen9_edram_size_mb(struct drm_i915_private *dev_priv, u32 cap)
> sets[EDRAM_SETS_IDX(cap)];
> }
>
> -static void edram_detect(struct drm_i915_private *dev_priv)
> +static void edram_detect(struct drm_i915_private *i915)
> {
> u32 edram_cap = 0;
>
> - if (!(IS_HASWELL(dev_priv) ||
> - IS_BROADWELL(dev_priv) ||
> - INTEL_GEN(dev_priv) >= 9))
> + if (!(IS_HASWELL(i915) ||
> + IS_BROADWELL(i915) ||
> + INTEL_GEN(i915) >= 9))
> return;
>
> - edram_cap = __raw_uncore_read32(&dev_priv->uncore, HSW_EDRAM_CAP);
> + edram_cap = __raw_uncore_read32(&i915->uncore, HSW_EDRAM_CAP);
>
> /* NB: We can't write IDICR yet because we don't have gt funcs set up */
>
> @@ -1503,63 +1503,63 @@ static void edram_detect(struct drm_i915_private *dev_priv)
> * The needed capability bits for size calculation are not there with
> * pre gen9 so return 128MB always.
> */
> - if (INTEL_GEN(dev_priv) < 9)
> - dev_priv->edram_size_mb = 128;
> + if (INTEL_GEN(i915) < 9)
> + i915->edram_size_mb = 128;
> else
> - dev_priv->edram_size_mb =
> - gen9_edram_size_mb(dev_priv, edram_cap);
> + i915->edram_size_mb =
> + gen9_edram_size_mb(i915, edram_cap);
>
> - DRM_INFO("Found %uMB of eDRAM\n", dev_priv->edram_size_mb);
> + DRM_INFO("Found %uMB of eDRAM\n", i915->edram_size_mb);
> }
>
> /**
> * i915_driver_init_hw - setup state requiring device access
> - * @dev_priv: device private
> + * @i915: device private
> *
> * Setup state that requires accessing the device, but doesn't require
> * exposing the driver via kernel internal or userspace interfaces.
> */
> -static int i915_driver_init_hw(struct drm_i915_private *dev_priv)
> +static int i915_driver_init_hw(struct drm_i915_private *i915)
> {
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct pci_dev *pdev = i915->drm.pdev;
> int ret;
>
> if (i915_inject_load_failure())
> return -ENODEV;
>
> - intel_device_info_runtime_init(dev_priv);
> + intel_device_info_runtime_init(i915);
>
> - if (HAS_PPGTT(dev_priv)) {
> - if (intel_vgpu_active(dev_priv) &&
> - !intel_vgpu_has_full_ppgtt(dev_priv)) {
> - i915_report_error(dev_priv,
> + if (HAS_PPGTT(i915)) {
> + if (intel_vgpu_active(i915) &&
> + !intel_vgpu_has_full_ppgtt(i915)) {
> + i915_report_error(i915,
> "incompatible vGPU found, support for isolated ppGTT required\n");
> return -ENXIO;
> }
> }
>
> - if (HAS_EXECLISTS(dev_priv)) {
> + if (HAS_EXECLISTS(i915)) {
> /*
> * Older GVT emulation depends upon intercepting CSB mmio,
> * which we no longer use, preferring to use the HWSP cache
> * instead.
> */
> - if (intel_vgpu_active(dev_priv) &&
> - !intel_vgpu_has_hwsp_emulation(dev_priv)) {
> - i915_report_error(dev_priv,
> + if (intel_vgpu_active(i915) &&
> + !intel_vgpu_has_hwsp_emulation(i915)) {
> + i915_report_error(i915,
> "old vGPU host found, support for HWSP emulation required\n");
> return -ENXIO;
> }
> }
>
> - intel_sanitize_options(dev_priv);
> + intel_sanitize_options(i915);
>
> /* needs to be done before ggtt probe */
> - edram_detect(dev_priv);
> + edram_detect(i915);
>
> - i915_perf_init(dev_priv);
> + i915_perf_init(i915);
>
> - ret = i915_ggtt_probe_hw(dev_priv);
> + ret = i915_ggtt_probe_hw(i915);
> if (ret)
> goto err_perf;
>
> @@ -1567,7 +1567,7 @@ static int i915_driver_init_hw(struct drm_i915_private *dev_priv)
> * WARNING: Apparently we must kick fbdev drivers before vgacon,
> * otherwise the vga fbdev driver falls over.
> */
> - ret = i915_kick_out_firmware_fb(dev_priv);
> + ret = i915_kick_out_firmware_fb(i915);
> if (ret) {
> DRM_ERROR("failed to remove conflicting framebuffer drivers\n");
> goto err_ggtt;
> @@ -1579,11 +1579,11 @@ static int i915_driver_init_hw(struct drm_i915_private *dev_priv)
> goto err_ggtt;
> }
>
> - ret = i915_ggtt_init_hw(dev_priv);
> + ret = i915_ggtt_init_hw(i915);
> if (ret)
> goto err_ggtt;
>
> - ret = i915_ggtt_enable_hw(dev_priv);
> + ret = i915_ggtt_enable_hw(i915);
> if (ret) {
> DRM_ERROR("failed to enable GGTT\n");
> goto err_ggtt;
> @@ -1592,7 +1592,7 @@ static int i915_driver_init_hw(struct drm_i915_private *dev_priv)
> pci_set_master(pdev);
>
> /* overlay on gen2 is broken and can't address above 1G */
> - if (IS_GEN(dev_priv, 2)) {
> + if (IS_GEN(i915, 2)) {
> ret = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(30));
> if (ret) {
> DRM_ERROR("failed to set DMA mask\n");
> @@ -1609,7 +1609,7 @@ static int i915_driver_init_hw(struct drm_i915_private *dev_priv)
> * behaviour if any general state is accessed within a page above 4GB,
> * which also needs to be handled carefully.
> */
> - if (IS_I965G(dev_priv) || IS_I965GM(dev_priv)) {
> + if (IS_I965G(i915) || IS_I965GM(i915)) {
> ret = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32));
>
> if (ret) {
> @@ -1619,13 +1619,13 @@ static int i915_driver_init_hw(struct drm_i915_private *dev_priv)
> }
> }
>
> - pm_qos_add_request(&dev_priv->pm_qos, PM_QOS_CPU_DMA_LATENCY,
> + pm_qos_add_request(&i915->pm_qos, PM_QOS_CPU_DMA_LATENCY,
> PM_QOS_DEFAULT_VALUE);
>
> - intel_uncore_sanitize(dev_priv);
> + intel_uncore_sanitize(i915);
>
> - intel_gt_init_workarounds(dev_priv);
> - i915_gem_load_init_fences(dev_priv);
> + intel_gt_init_workarounds(i915);
> + i915_gem_load_init_fences(i915);
>
> /* On the 945G/GM, the chipset reports the MSI capability on the
> * integrated graphics even though the support isn't actually there
> @@ -1646,94 +1646,94 @@ static int i915_driver_init_hw(struct drm_i915_private *dev_priv)
> * device. The kernel then disables that interrupt source and so
> * prevents the other device from working properly.
> */
> - if (INTEL_GEN(dev_priv) >= 5) {
> + if (INTEL_GEN(i915) >= 5) {
> if (pci_enable_msi(pdev) < 0)
> DRM_DEBUG_DRIVER("can't enable MSI");
> }
>
> - ret = intel_gvt_init(dev_priv);
> + ret = intel_gvt_init(i915);
> if (ret)
> goto err_msi;
>
> - intel_opregion_setup(dev_priv);
> + intel_opregion_setup(i915);
> /*
> * Fill the dram structure to get the system raw bandwidth and
> * dram info. This will be used for memory latency calculation.
> */
> - intel_get_dram_info(dev_priv);
> + intel_get_dram_info(i915);
>
> - intel_bw_init_hw(dev_priv);
> + intel_bw_init_hw(i915);
>
> return 0;
>
> err_msi:
> if (pdev->msi_enabled)
> pci_disable_msi(pdev);
> - pm_qos_remove_request(&dev_priv->pm_qos);
> + pm_qos_remove_request(&i915->pm_qos);
> err_ggtt:
> - i915_ggtt_cleanup_hw(dev_priv);
> + i915_ggtt_cleanup_hw(i915);
> err_perf:
> - i915_perf_fini(dev_priv);
> + i915_perf_fini(i915);
> return ret;
> }
>
> /**
> * i915_driver_cleanup_hw - cleanup the setup done in i915_driver_init_hw()
> - * @dev_priv: device private
> + * @i915: device private
> */
> -static void i915_driver_cleanup_hw(struct drm_i915_private *dev_priv)
> +static void i915_driver_cleanup_hw(struct drm_i915_private *i915)
> {
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct pci_dev *pdev = i915->drm.pdev;
>
> - i915_perf_fini(dev_priv);
> + i915_perf_fini(i915);
>
> if (pdev->msi_enabled)
> pci_disable_msi(pdev);
>
> - pm_qos_remove_request(&dev_priv->pm_qos);
> + pm_qos_remove_request(&i915->pm_qos);
> }
>
> /**
> * i915_driver_register - register the driver with the rest of the system
> - * @dev_priv: device private
> + * @i915: device private
> *
> * Perform any steps necessary to make the driver available via kernel
> * internal or userspace interfaces.
> */
> -static void i915_driver_register(struct drm_i915_private *dev_priv)
> +static void i915_driver_register(struct drm_i915_private *i915)
> {
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_device *dev = &i915->drm;
>
> - i915_gem_shrinker_register(dev_priv);
> - i915_pmu_register(dev_priv);
> + i915_gem_shrinker_register(i915);
> + i915_pmu_register(i915);
>
> /*
> * Notify a valid surface after modesetting,
> * when running inside a VM.
> */
> - if (intel_vgpu_active(dev_priv))
> + if (intel_vgpu_active(i915))
> I915_WRITE(vgtif_reg(display_ready), VGT_DRV_DISPLAY_READY);
>
> /* Reveal our presence to userspace */
> if (drm_dev_register(dev, 0) == 0) {
> - i915_debugfs_register(dev_priv);
> - i915_setup_sysfs(dev_priv);
> + i915_debugfs_register(i915);
> + i915_setup_sysfs(i915);
>
> /* Depends on sysfs having been initialized */
> - i915_perf_register(dev_priv);
> + i915_perf_register(i915);
> } else
> DRM_ERROR("Failed to register driver for userspace access!\n");
>
> - if (HAS_DISPLAY(dev_priv)) {
> + if (HAS_DISPLAY(i915)) {
> /* Must be done after probing outputs */
> - intel_opregion_register(dev_priv);
> + intel_opregion_register(i915);
> acpi_video_register();
> }
>
> - if (IS_GEN(dev_priv, 5))
> - intel_gpu_ips_init(dev_priv);
> + if (IS_GEN(i915, 5))
> + intel_gpu_ips_init(i915);
>
> - intel_audio_init(dev_priv);
> + intel_audio_init(i915);
>
> /*
> * Some ports require correctly set-up hpd registers for detection to
> @@ -1748,60 +1748,60 @@ static void i915_driver_register(struct drm_i915_private *dev_priv)
> * We need to coordinate the hotplugs with the asynchronous fbdev
> * configuration, for which we use the fbdev->async_cookie.
> */
> - if (HAS_DISPLAY(dev_priv))
> + if (HAS_DISPLAY(i915))
> drm_kms_helper_poll_init(dev);
>
> - intel_power_domains_enable(dev_priv);
> - intel_runtime_pm_enable(dev_priv);
> + intel_power_domains_enable(i915);
> + intel_runtime_pm_enable(i915);
> }
>
> /**
> * i915_driver_unregister - cleanup the registration done in i915_driver_regiser()
> - * @dev_priv: device private
> + * @i915: device private
> */
> -static void i915_driver_unregister(struct drm_i915_private *dev_priv)
> +static void i915_driver_unregister(struct drm_i915_private *i915)
> {
> - intel_runtime_pm_disable(dev_priv);
> - intel_power_domains_disable(dev_priv);
> + intel_runtime_pm_disable(i915);
> + intel_power_domains_disable(i915);
>
> - intel_fbdev_unregister(dev_priv);
> - intel_audio_deinit(dev_priv);
> + intel_fbdev_unregister(i915);
> + intel_audio_deinit(i915);
>
> /*
> * After flushing the fbdev (incl. a late async config which will
> * have delayed queuing of a hotplug event), then flush the hotplug
> * events.
> */
> - drm_kms_helper_poll_fini(&dev_priv->drm);
> + drm_kms_helper_poll_fini(&i915->drm);
>
> intel_gpu_ips_teardown();
> acpi_video_unregister();
> - intel_opregion_unregister(dev_priv);
> + intel_opregion_unregister(i915);
>
> - i915_perf_unregister(dev_priv);
> - i915_pmu_unregister(dev_priv);
> + i915_perf_unregister(i915);
> + i915_pmu_unregister(i915);
>
> - i915_teardown_sysfs(dev_priv);
> - drm_dev_unplug(&dev_priv->drm);
> + i915_teardown_sysfs(i915);
> + drm_dev_unplug(&i915->drm);
>
> - i915_gem_shrinker_unregister(dev_priv);
> + i915_gem_shrinker_unregister(i915);
> }
>
> -static void i915_welcome_messages(struct drm_i915_private *dev_priv)
> +static void i915_welcome_messages(struct drm_i915_private *i915)
> {
> if (drm_debug & DRM_UT_DRIVER) {
> struct drm_printer p = drm_debug_printer("i915 device info:");
>
> drm_printf(&p, "pciid=0x%04x rev=0x%02x platform=%s (subplatform=0x%x) gen=%i\n",
> - INTEL_DEVID(dev_priv),
> - INTEL_REVID(dev_priv),
> - intel_platform_name(INTEL_INFO(dev_priv)->platform),
> - intel_subplatform(RUNTIME_INFO(dev_priv),
> - INTEL_INFO(dev_priv)->platform),
> - INTEL_GEN(dev_priv));
> -
> - intel_device_info_dump_flags(INTEL_INFO(dev_priv), &p);
> - intel_device_info_dump_runtime(RUNTIME_INFO(dev_priv), &p);
> + INTEL_DEVID(i915),
> + INTEL_REVID(i915),
> + intel_platform_name(INTEL_INFO(i915)->platform),
> + intel_subplatform(RUNTIME_INFO(i915),
> + INTEL_INFO(i915)->platform),
> + INTEL_GEN(i915));
> +
> + intel_device_info_dump_flags(INTEL_INFO(i915), &p);
> + intel_device_info_dump_runtime(RUNTIME_INFO(i915), &p);
> }
>
> if (IS_ENABLED(CONFIG_DRM_I915_DEBUG))
> @@ -1871,126 +1871,126 @@ int i915_driver_load(struct pci_dev *pdev, const struct pci_device_id *ent)
> {
> const struct intel_device_info *match_info =
> (struct intel_device_info *)ent->driver_data;
> - struct drm_i915_private *dev_priv;
> + struct drm_i915_private *i915;
> int ret;
>
> - dev_priv = i915_driver_create(pdev, ent);
> - if (IS_ERR(dev_priv))
> - return PTR_ERR(dev_priv);
> + i915 = i915_driver_create(pdev, ent);
> + if (IS_ERR(i915))
> + return PTR_ERR(i915);
>
> /* Disable nuclear pageflip by default on pre-ILK */
> if (!i915_modparams.nuclear_pageflip && match_info->gen < 5)
> - dev_priv->drm.driver_features &= ~DRIVER_ATOMIC;
> + i915->drm.driver_features &= ~DRIVER_ATOMIC;
>
> ret = pci_enable_device(pdev);
> if (ret)
> goto out_fini;
>
> - ret = i915_driver_init_early(dev_priv);
> + ret = i915_driver_init_early(i915);
> if (ret < 0)
> goto out_pci_disable;
>
> - disable_rpm_wakeref_asserts(dev_priv);
> + disable_rpm_wakeref_asserts(i915);
>
> - ret = i915_driver_init_mmio(dev_priv);
> + ret = i915_driver_init_mmio(i915);
> if (ret < 0)
> goto out_runtime_pm_put;
>
> - ret = i915_driver_init_hw(dev_priv);
> + ret = i915_driver_init_hw(i915);
> if (ret < 0)
> goto out_cleanup_mmio;
>
> - ret = i915_load_modeset_init(&dev_priv->drm);
> + ret = i915_load_modeset_init(&i915->drm);
> if (ret < 0)
> goto out_cleanup_hw;
>
> - i915_driver_register(dev_priv);
> + i915_driver_register(i915);
>
> - enable_rpm_wakeref_asserts(dev_priv);
> + enable_rpm_wakeref_asserts(i915);
>
> - i915_welcome_messages(dev_priv);
> + i915_welcome_messages(i915);
>
> return 0;
>
> out_cleanup_hw:
> - i915_driver_cleanup_hw(dev_priv);
> - i915_ggtt_cleanup_hw(dev_priv);
> + i915_driver_cleanup_hw(i915);
> + i915_ggtt_cleanup_hw(i915);
> out_cleanup_mmio:
> - i915_driver_cleanup_mmio(dev_priv);
> + i915_driver_cleanup_mmio(i915);
> out_runtime_pm_put:
> - enable_rpm_wakeref_asserts(dev_priv);
> - i915_driver_cleanup_early(dev_priv);
> + enable_rpm_wakeref_asserts(i915);
> + i915_driver_cleanup_early(i915);
> out_pci_disable:
> pci_disable_device(pdev);
> out_fini:
> - i915_load_error(dev_priv, "Device initialization failed (%d)\n", ret);
> - i915_driver_destroy(dev_priv);
> + i915_load_error(i915, "Device initialization failed (%d)\n", ret);
> + i915_driver_destroy(i915);
> return ret;
> }
>
> void i915_driver_unload(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct pci_dev *pdev = i915->drm.pdev;
>
> - disable_rpm_wakeref_asserts(dev_priv);
> + disable_rpm_wakeref_asserts(i915);
>
> - i915_driver_unregister(dev_priv);
> + i915_driver_unregister(i915);
>
> /*
> * After unregistering the device to prevent any new users, cancel
> * all in-flight requests so that we can quickly unbind the active
> * resources.
> */
> - i915_gem_set_wedged(dev_priv);
> + i915_gem_set_wedged(i915);
>
> /* Flush any external code that still may be under the RCU lock */
> synchronize_rcu();
>
> - i915_gem_suspend(dev_priv);
> + i915_gem_suspend(i915);
>
> drm_atomic_helper_shutdown(dev);
>
> - intel_gvt_cleanup(dev_priv);
> + intel_gvt_cleanup(i915);
>
> intel_modeset_cleanup(dev);
>
> - intel_bios_cleanup(dev_priv);
> + intel_bios_cleanup(i915);
>
> vga_switcheroo_unregister_client(pdev);
> vga_client_register(pdev, NULL, NULL, NULL);
>
> - intel_csr_ucode_fini(dev_priv);
> + intel_csr_ucode_fini(i915);
>
> /* Free error state after interrupts are fully disabled. */
> - cancel_delayed_work_sync(&dev_priv->gpu_error.hangcheck_work);
> - i915_reset_error_state(dev_priv);
> + cancel_delayed_work_sync(&i915->gpu_error.hangcheck_work);
> + i915_reset_error_state(i915);
>
> - i915_gem_fini_hw(dev_priv);
> + i915_gem_fini_hw(i915);
>
> - intel_power_domains_fini_hw(dev_priv);
> + intel_power_domains_fini_hw(i915);
>
> - i915_driver_cleanup_hw(dev_priv);
> + i915_driver_cleanup_hw(i915);
>
> - enable_rpm_wakeref_asserts(dev_priv);
> + enable_rpm_wakeref_asserts(i915);
> }
>
> static void i915_driver_release(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
>
> - disable_rpm_wakeref_asserts(dev_priv);
> + disable_rpm_wakeref_asserts(i915);
>
> - i915_gem_fini(dev_priv);
> + i915_gem_fini(i915);
>
> - i915_ggtt_cleanup_hw(dev_priv);
> - i915_driver_cleanup_mmio(dev_priv);
> + i915_ggtt_cleanup_hw(i915);
> + i915_driver_cleanup_mmio(i915);
>
> - enable_rpm_wakeref_asserts(dev_priv);
> - intel_runtime_pm_cleanup(dev_priv);
> + enable_rpm_wakeref_asserts(i915);
> + intel_runtime_pm_cleanup(i915);
>
> - i915_driver_cleanup_early(dev_priv);
> - i915_driver_destroy(dev_priv);
> + i915_driver_cleanup_early(i915);
> + i915_driver_destroy(i915);
> }
>
> static int i915_driver_open(struct drm_device *dev, struct drm_file *file)
> @@ -2035,9 +2035,9 @@ static void i915_driver_postclose(struct drm_device *dev, struct drm_file *file)
> kfree(file_priv);
> }
>
> -static void intel_suspend_encoders(struct drm_i915_private *dev_priv)
> +static void intel_suspend_encoders(struct drm_i915_private *i915)
> {
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_device *dev = &i915->drm;
> struct intel_encoder *encoder;
>
> drm_modeset_lock_all(dev);
> @@ -2047,11 +2047,11 @@ static void intel_suspend_encoders(struct drm_i915_private *dev_priv)
> drm_modeset_unlock_all(dev);
> }
>
> -static int vlv_resume_prepare(struct drm_i915_private *dev_priv,
> +static int vlv_resume_prepare(struct drm_i915_private *i915,
> bool rpm_resume);
> -static int vlv_suspend_complete(struct drm_i915_private *dev_priv);
> +static int vlv_suspend_complete(struct drm_i915_private *i915);
>
> -static bool suspend_to_idle(struct drm_i915_private *dev_priv)
> +static bool suspend_to_idle(struct drm_i915_private *i915)
> {
> #if IS_ENABLED(CONFIG_ACPI_SLEEP)
> if (acpi_target_system_state() < ACPI_STATE_S3)
> @@ -2077,15 +2077,15 @@ static int i915_drm_prepare(struct drm_device *dev)
>
> static int i915_drm_suspend(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct pci_dev *pdev = i915->drm.pdev;
> pci_power_t opregion_target_state;
>
> - disable_rpm_wakeref_asserts(dev_priv);
> + disable_rpm_wakeref_asserts(i915);
>
> /* We do a lot of poking in a lot of registers, make sure they work
> * properly. */
> - intel_power_domains_disable(dev_priv);
> + intel_power_domains_disable(i915);
>
> drm_kms_helper_poll_disable(dev);
>
> @@ -2093,40 +2093,40 @@ static int i915_drm_suspend(struct drm_device *dev)
>
> intel_display_suspend(dev);
>
> - intel_dp_mst_suspend(dev_priv);
> + intel_dp_mst_suspend(i915);
>
> - intel_runtime_pm_disable_interrupts(dev_priv);
> - intel_hpd_cancel_work(dev_priv);
> + intel_runtime_pm_disable_interrupts(i915);
> + intel_hpd_cancel_work(i915);
>
> - intel_suspend_encoders(dev_priv);
> + intel_suspend_encoders(i915);
>
> - intel_suspend_hw(dev_priv);
> + intel_suspend_hw(i915);
>
> - i915_gem_suspend_gtt_mappings(dev_priv);
> + i915_gem_suspend_gtt_mappings(i915);
>
> - i915_save_state(dev_priv);
> + i915_save_state(i915);
>
> - opregion_target_state = suspend_to_idle(dev_priv) ? PCI_D1 : PCI_D3cold;
> - intel_opregion_suspend(dev_priv, opregion_target_state);
> + opregion_target_state = suspend_to_idle(i915) ? PCI_D1 : PCI_D3cold;
> + intel_opregion_suspend(i915, opregion_target_state);
>
> intel_fbdev_set_suspend(dev, FBINFO_STATE_SUSPENDED, true);
>
> - dev_priv->suspend_count++;
> + i915->suspend_count++;
>
> - intel_csr_ucode_suspend(dev_priv);
> + intel_csr_ucode_suspend(i915);
>
> - enable_rpm_wakeref_asserts(dev_priv);
> + enable_rpm_wakeref_asserts(i915);
>
> return 0;
> }
>
> static enum i915_drm_suspend_mode
> -get_suspend_mode(struct drm_i915_private *dev_priv, bool hibernate)
> +get_suspend_mode(struct drm_i915_private *i915, bool hibernate)
> {
> if (hibernate)
> return I915_DRM_SUSPEND_HIBERNATE;
>
> - if (suspend_to_idle(dev_priv))
> + if (suspend_to_idle(i915))
> return I915_DRM_SUSPEND_IDLE;
>
> return I915_DRM_SUSPEND_MEM;
> @@ -2134,30 +2134,30 @@ get_suspend_mode(struct drm_i915_private *dev_priv, bool hibernate)
>
> static int i915_drm_suspend_late(struct drm_device *dev, bool hibernation)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct pci_dev *pdev = i915->drm.pdev;
> int ret;
>
> - disable_rpm_wakeref_asserts(dev_priv);
> + disable_rpm_wakeref_asserts(i915);
>
> - i915_gem_suspend_late(dev_priv);
> + i915_gem_suspend_late(i915);
>
> - intel_uncore_suspend(&dev_priv->uncore);
> + intel_uncore_suspend(&i915->uncore);
>
> - intel_power_domains_suspend(dev_priv,
> - get_suspend_mode(dev_priv, hibernation));
> + intel_power_domains_suspend(i915,
> + get_suspend_mode(i915, hibernation));
>
> ret = 0;
> - if (INTEL_GEN(dev_priv) >= 11 || IS_GEN9_LP(dev_priv))
> - bxt_enable_dc9(dev_priv);
> - else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> - hsw_enable_pc8(dev_priv);
> - else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> - ret = vlv_suspend_complete(dev_priv);
> + if (INTEL_GEN(i915) >= 11 || IS_GEN9_LP(i915))
> + bxt_enable_dc9(i915);
> + else if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> + hsw_enable_pc8(i915);
> + else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> + ret = vlv_suspend_complete(i915);
>
> if (ret) {
> DRM_ERROR("Suspend complete failed: %d\n", ret);
> - intel_power_domains_resume(dev_priv);
> + intel_power_domains_resume(i915);
>
> goto out;
> }
> @@ -2175,13 +2175,13 @@ static int i915_drm_suspend_late(struct drm_device *dev, bool hibernation)
> * Fujitsu FSC S7110
> * Acer Aspire 1830T
> */
> - if (!(hibernation && INTEL_GEN(dev_priv) < 6))
> + if (!(hibernation && INTEL_GEN(i915) < 6))
> pci_set_power_state(pdev, PCI_D3hot);
>
> out:
> - enable_rpm_wakeref_asserts(dev_priv);
> - if (!dev_priv->uncore.user_forcewake.count)
> - intel_runtime_pm_cleanup(dev_priv);
> + enable_rpm_wakeref_asserts(i915);
> + if (!i915->uncore.user_forcewake.count)
> + intel_runtime_pm_cleanup(i915);
>
> return ret;
> }
> @@ -2212,24 +2212,24 @@ static int i915_suspend_switcheroo(struct drm_device *dev, pm_message_t state)
>
> static int i915_drm_resume(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> int ret;
>
> - disable_rpm_wakeref_asserts(dev_priv);
> - intel_sanitize_gt_powersave(dev_priv);
> + disable_rpm_wakeref_asserts(i915);
> + intel_sanitize_gt_powersave(i915);
>
> - i915_gem_sanitize(dev_priv);
> + i915_gem_sanitize(i915);
>
> - ret = i915_ggtt_enable_hw(dev_priv);
> + ret = i915_ggtt_enable_hw(i915);
> if (ret)
> DRM_ERROR("failed to re-enable GGTT\n");
>
> - intel_csr_ucode_resume(dev_priv);
> + intel_csr_ucode_resume(i915);
>
> - i915_restore_state(dev_priv);
> - intel_pps_unlock_regs_wa(dev_priv);
> + i915_restore_state(i915);
> + intel_pps_unlock_regs_wa(i915);
>
> - intel_init_pch_refclk(dev_priv);
> + intel_init_pch_refclk(i915);
>
> /*
> * Interrupts have to be enabled before any batches are run. If not the
> @@ -2241,21 +2241,21 @@ static int i915_drm_resume(struct drm_device *dev)
> * Modeset enabling in intel_modeset_init_hw() also needs working
> * interrupts.
> */
> - intel_runtime_pm_enable_interrupts(dev_priv);
> + intel_runtime_pm_enable_interrupts(i915);
>
> drm_mode_config_reset(dev);
>
> - i915_gem_resume(dev_priv);
> + i915_gem_resume(i915);
>
> intel_modeset_init_hw(dev);
> - intel_init_clock_gating(dev_priv);
> + intel_init_clock_gating(i915);
>
> - spin_lock_irq(&dev_priv->irq_lock);
> - if (dev_priv->display.hpd_irq_setup)
> - dev_priv->display.hpd_irq_setup(dev_priv);
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
> + if (i915->display.hpd_irq_setup)
> + i915->display.hpd_irq_setup(i915);
> + spin_unlock_irq(&i915->irq_lock);
>
> - intel_dp_mst_resume(dev_priv);
> + intel_dp_mst_resume(i915);
>
> intel_display_resume(dev);
>
> @@ -2267,23 +2267,23 @@ static int i915_drm_resume(struct drm_device *dev)
> * bother with the tiny race here where we might lose hotplug
> * notifications.
> * */
> - intel_hpd_init(dev_priv);
> + intel_hpd_init(i915);
>
> - intel_opregion_resume(dev_priv);
> + intel_opregion_resume(i915);
>
> intel_fbdev_set_suspend(dev, FBINFO_STATE_RUNNING, false);
>
> - intel_power_domains_enable(dev_priv);
> + intel_power_domains_enable(i915);
>
> - enable_rpm_wakeref_asserts(dev_priv);
> + enable_rpm_wakeref_asserts(i915);
>
> return 0;
> }
>
> static int i915_drm_resume_early(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct pci_dev *pdev = i915->drm.pdev;
> int ret;
>
> /*
> @@ -2330,32 +2330,32 @@ static int i915_drm_resume_early(struct drm_device *dev)
>
> pci_set_master(pdev);
>
> - disable_rpm_wakeref_asserts(dev_priv);
> + disable_rpm_wakeref_asserts(i915);
>
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> - ret = vlv_resume_prepare(dev_priv, false);
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> + ret = vlv_resume_prepare(i915, false);
> if (ret)
> DRM_ERROR("Resume prepare failed: %d, continuing anyway\n",
> ret);
>
> - intel_uncore_resume_early(&dev_priv->uncore);
> + intel_uncore_resume_early(&i915->uncore);
>
> - i915_check_and_clear_faults(dev_priv);
> + i915_check_and_clear_faults(i915);
>
> - if (INTEL_GEN(dev_priv) >= 11 || IS_GEN9_LP(dev_priv)) {
> - gen9_sanitize_dc_state(dev_priv);
> - bxt_disable_dc9(dev_priv);
> - } else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) {
> - hsw_disable_pc8(dev_priv);
> + if (INTEL_GEN(i915) >= 11 || IS_GEN9_LP(i915)) {
> + gen9_sanitize_dc_state(i915);
> + bxt_disable_dc9(i915);
> + } else if (IS_HASWELL(i915) || IS_BROADWELL(i915)) {
> + hsw_disable_pc8(i915);
> }
>
> - intel_uncore_sanitize(dev_priv);
> + intel_uncore_sanitize(i915);
>
> - intel_power_domains_resume(dev_priv);
> + intel_power_domains_resume(i915);
>
> - intel_gt_sanitize(dev_priv, true);
> + intel_gt_sanitize(i915, true);
>
> - enable_rpm_wakeref_asserts(dev_priv);
> + enable_rpm_wakeref_asserts(i915);
>
> return ret;
> }
> @@ -2540,9 +2540,9 @@ static int i915_pm_restore(struct device *kdev)
> * a black-box for the driver. Further investigation is needed to reduce the
> * saved/restored registers even further, by following the same 3 criteria.
> */
> -static void vlv_save_gunit_s0ix_state(struct drm_i915_private *dev_priv)
> +static void vlv_save_gunit_s0ix_state(struct drm_i915_private *i915)
> {
> - struct vlv_s0ix_state *s = &dev_priv->vlv_s0ix_state;
> + struct vlv_s0ix_state *s = &i915->vlv_s0ix_state;
> int i;
>
> /* GAM 0x4000-0x4770 */
> @@ -2621,9 +2621,9 @@ static void vlv_save_gunit_s0ix_state(struct drm_i915_private *dev_priv)
> */
> }
>
> -static void vlv_restore_gunit_s0ix_state(struct drm_i915_private *dev_priv)
> +static void vlv_restore_gunit_s0ix_state(struct drm_i915_private *i915)
> {
> - struct vlv_s0ix_state *s = &dev_priv->vlv_s0ix_state;
> + struct vlv_s0ix_state *s = &i915->vlv_s0ix_state;
> u32 val;
> int i;
>
> @@ -2732,7 +2732,7 @@ static int vlv_wait_for_pw_status(struct drm_i915_private *i915,
> return ret;
> }
>
> -int vlv_force_gfx_clock(struct drm_i915_private *dev_priv, bool force_on)
> +int vlv_force_gfx_clock(struct drm_i915_private *i915, bool force_on)
> {
> u32 val;
> int err;
> @@ -2746,7 +2746,7 @@ int vlv_force_gfx_clock(struct drm_i915_private *dev_priv, bool force_on)
> if (!force_on)
> return 0;
>
> - err = intel_wait_for_register(&dev_priv->uncore,
> + err = intel_wait_for_register(&i915->uncore,
> VLV_GTLC_SURVIVABILITY_REG,
> VLV_GFX_CLK_STATUS_BIT,
> VLV_GFX_CLK_STATUS_BIT,
> @@ -2758,7 +2758,7 @@ int vlv_force_gfx_clock(struct drm_i915_private *dev_priv, bool force_on)
> return err;
> }
>
> -static int vlv_allow_gt_wake(struct drm_i915_private *dev_priv, bool allow)
> +static int vlv_allow_gt_wake(struct drm_i915_private *i915, bool allow)
> {
> u32 mask;
> u32 val;
> @@ -2774,14 +2774,14 @@ static int vlv_allow_gt_wake(struct drm_i915_private *dev_priv, bool allow)
> mask = VLV_GTLC_ALLOWWAKEACK;
> val = allow ? mask : 0;
>
> - err = vlv_wait_for_pw_status(dev_priv, mask, val);
> + err = vlv_wait_for_pw_status(i915, mask, val);
> if (err)
> DRM_ERROR("timeout disabling GT waking\n");
>
> return err;
> }
>
> -static void vlv_wait_for_gt_wells(struct drm_i915_private *dev_priv,
> +static void vlv_wait_for_gt_wells(struct drm_i915_private *i915,
> bool wait_for_on)
> {
> u32 mask;
> @@ -2797,12 +2797,12 @@ static void vlv_wait_for_gt_wells(struct drm_i915_private *dev_priv,
> * This can fail to turn off the rc6 if the GPU is stuck after a failed
> * reset and we are trying to force the machine to sleep.
> */
> - if (vlv_wait_for_pw_status(dev_priv, mask, val))
> + if (vlv_wait_for_pw_status(i915, mask, val))
> DRM_DEBUG_DRIVER("timeout waiting for GT wells to go %s\n",
> onoff(wait_for_on));
> }
>
> -static void vlv_check_no_gt_access(struct drm_i915_private *dev_priv)
> +static void vlv_check_no_gt_access(struct drm_i915_private *i915)
> {
> if (!(I915_READ(VLV_GTLC_PW_STATUS) & VLV_GTLC_ALLOWWAKEERR))
> return;
> @@ -2811,7 +2811,7 @@ static void vlv_check_no_gt_access(struct drm_i915_private *dev_priv)
> I915_WRITE(VLV_GTLC_PW_STATUS, VLV_GTLC_ALLOWWAKEERR);
> }
>
> -static int vlv_suspend_complete(struct drm_i915_private *dev_priv)
> +static int vlv_suspend_complete(struct drm_i915_private *i915)
> {
> u32 mask;
> int err;
> @@ -2820,25 +2820,25 @@ static int vlv_suspend_complete(struct drm_i915_private *dev_priv)
> * Bspec defines the following GT well on flags as debug only, so
> * don't treat them as hard failures.
> */
> - vlv_wait_for_gt_wells(dev_priv, false);
> + vlv_wait_for_gt_wells(i915, false);
>
> mask = VLV_GTLC_RENDER_CTX_EXISTS | VLV_GTLC_MEDIA_CTX_EXISTS;
> WARN_ON((I915_READ(VLV_GTLC_WAKE_CTRL) & mask) != mask);
>
> - vlv_check_no_gt_access(dev_priv);
> + vlv_check_no_gt_access(i915);
>
> - err = vlv_force_gfx_clock(dev_priv, true);
> + err = vlv_force_gfx_clock(i915, true);
> if (err)
> goto err1;
>
> - err = vlv_allow_gt_wake(dev_priv, false);
> + err = vlv_allow_gt_wake(i915, false);
> if (err)
> goto err2;
>
> - if (!IS_CHERRYVIEW(dev_priv))
> - vlv_save_gunit_s0ix_state(dev_priv);
> + if (!IS_CHERRYVIEW(i915))
> + vlv_save_gunit_s0ix_state(i915);
>
> - err = vlv_force_gfx_clock(dev_priv, false);
> + err = vlv_force_gfx_clock(i915, false);
> if (err)
> goto err2;
>
> @@ -2846,14 +2846,14 @@ static int vlv_suspend_complete(struct drm_i915_private *dev_priv)
>
> err2:
> /* For safety always re-enable waking and disable gfx clock forcing */
> - vlv_allow_gt_wake(dev_priv, true);
> + vlv_allow_gt_wake(i915, true);
> err1:
> - vlv_force_gfx_clock(dev_priv, false);
> + vlv_force_gfx_clock(i915, false);
>
> return err;
> }
>
> -static int vlv_resume_prepare(struct drm_i915_private *dev_priv,
> +static int vlv_resume_prepare(struct drm_i915_private *i915,
> bool rpm_resume)
> {
> int err;
> @@ -2864,23 +2864,23 @@ static int vlv_resume_prepare(struct drm_i915_private *dev_priv,
> * can do at this point. Return the first error code (which will also
> * leave RPM permanently disabled).
> */
> - ret = vlv_force_gfx_clock(dev_priv, true);
> + ret = vlv_force_gfx_clock(i915, true);
>
> - if (!IS_CHERRYVIEW(dev_priv))
> - vlv_restore_gunit_s0ix_state(dev_priv);
> + if (!IS_CHERRYVIEW(i915))
> + vlv_restore_gunit_s0ix_state(i915);
>
> - err = vlv_allow_gt_wake(dev_priv, true);
> + err = vlv_allow_gt_wake(i915, true);
> if (!ret)
> ret = err;
>
> - err = vlv_force_gfx_clock(dev_priv, false);
> + err = vlv_force_gfx_clock(i915, false);
> if (!ret)
> ret = err;
>
> - vlv_check_no_gt_access(dev_priv);
> + vlv_check_no_gt_access(i915);
>
> if (rpm_resume)
> - intel_init_clock_gating(dev_priv);
> + intel_init_clock_gating(i915);
>
> return ret;
> }
> @@ -2889,80 +2889,80 @@ static int intel_runtime_suspend(struct device *kdev)
> {
> struct pci_dev *pdev = to_pci_dev(kdev);
> struct drm_device *dev = pci_get_drvdata(pdev);
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> int ret;
>
> - if (WARN_ON_ONCE(!(dev_priv->gt_pm.rc6.enabled && HAS_RC6(dev_priv))))
> + if (WARN_ON_ONCE(!(i915->gt_pm.rc6.enabled && HAS_RC6(i915))))
> return -ENODEV;
>
> - if (WARN_ON_ONCE(!HAS_RUNTIME_PM(dev_priv)))
> + if (WARN_ON_ONCE(!HAS_RUNTIME_PM(i915)))
> return -ENODEV;
>
> DRM_DEBUG_KMS("Suspending device\n");
>
> - disable_rpm_wakeref_asserts(dev_priv);
> + disable_rpm_wakeref_asserts(i915);
>
> /*
> * We are safe here against re-faults, since the fault handler takes
> * an RPM reference.
> */
> - i915_gem_runtime_suspend(dev_priv);
> + i915_gem_runtime_suspend(i915);
>
> - intel_uc_runtime_suspend(dev_priv);
> + intel_uc_runtime_suspend(i915);
>
> - intel_runtime_pm_disable_interrupts(dev_priv);
> + intel_runtime_pm_disable_interrupts(i915);
>
> - intel_uncore_suspend(&dev_priv->uncore);
> + intel_uncore_suspend(&i915->uncore);
>
> ret = 0;
> - if (INTEL_GEN(dev_priv) >= 11) {
> - icl_display_core_uninit(dev_priv);
> - bxt_enable_dc9(dev_priv);
> - } else if (IS_GEN9_LP(dev_priv)) {
> - bxt_display_core_uninit(dev_priv);
> - bxt_enable_dc9(dev_priv);
> - } else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) {
> - hsw_enable_pc8(dev_priv);
> - } else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> - ret = vlv_suspend_complete(dev_priv);
> + if (INTEL_GEN(i915) >= 11) {
> + icl_display_core_uninit(i915);
> + bxt_enable_dc9(i915);
> + } else if (IS_GEN9_LP(i915)) {
> + bxt_display_core_uninit(i915);
> + bxt_enable_dc9(i915);
> + } else if (IS_HASWELL(i915) || IS_BROADWELL(i915)) {
> + hsw_enable_pc8(i915);
> + } else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> + ret = vlv_suspend_complete(i915);
> }
>
> if (ret) {
> DRM_ERROR("Runtime suspend failed, disabling it (%d)\n", ret);
> - intel_uncore_runtime_resume(&dev_priv->uncore);
> + intel_uncore_runtime_resume(&i915->uncore);
>
> - intel_runtime_pm_enable_interrupts(dev_priv);
> + intel_runtime_pm_enable_interrupts(i915);
>
> - intel_uc_resume(dev_priv);
> + intel_uc_resume(i915);
>
> - i915_gem_init_swizzling(dev_priv);
> - i915_gem_restore_fences(dev_priv);
> + i915_gem_init_swizzling(i915);
> + i915_gem_restore_fences(i915);
>
> - enable_rpm_wakeref_asserts(dev_priv);
> + enable_rpm_wakeref_asserts(i915);
>
> return ret;
> }
>
> - enable_rpm_wakeref_asserts(dev_priv);
> - intel_runtime_pm_cleanup(dev_priv);
> + enable_rpm_wakeref_asserts(i915);
> + intel_runtime_pm_cleanup(i915);
>
> - if (intel_uncore_arm_unclaimed_mmio_detection(&dev_priv->uncore))
> + if (intel_uncore_arm_unclaimed_mmio_detection(&i915->uncore))
> DRM_ERROR("Unclaimed access detected prior to suspending\n");
>
> - dev_priv->runtime_pm.suspended = true;
> + i915->runtime_pm.suspended = true;
>
> /*
> * FIXME: We really should find a document that references the arguments
> * used below!
> */
> - if (IS_BROADWELL(dev_priv)) {
> + if (IS_BROADWELL(i915)) {
> /*
> * On Broadwell, if we use PCI_D1 the PCH DDI ports will stop
> * being detected, and the call we do at intel_runtime_resume()
> * won't be able to restore them. Since PCI_D3hot matches the
> * actual specification and appears to be working, use it.
> */
> - intel_opregion_notify_adapter(dev_priv, PCI_D3hot);
> + intel_opregion_notify_adapter(i915, PCI_D3hot);
> } else {
> /*
> * current versions of firmware which depend on this opregion
> @@ -2971,13 +2971,13 @@ static int intel_runtime_suspend(struct device *kdev)
> * to distinguish it from notifications that might be sent via
> * the suspend path.
> */
> - intel_opregion_notify_adapter(dev_priv, PCI_D1);
> + intel_opregion_notify_adapter(i915, PCI_D1);
> }
>
> - assert_forcewakes_inactive(&dev_priv->uncore);
> + assert_forcewakes_inactive(&i915->uncore);
>
> - if (!IS_VALLEYVIEW(dev_priv) && !IS_CHERRYVIEW(dev_priv))
> - intel_hpd_poll_init(dev_priv);
> + if (!IS_VALLEYVIEW(i915) && !IS_CHERRYVIEW(i915))
> + intel_hpd_poll_init(i915);
>
> DRM_DEBUG_KMS("Device suspended\n");
> return 0;
> @@ -2987,69 +2987,69 @@ static int intel_runtime_resume(struct device *kdev)
> {
> struct pci_dev *pdev = to_pci_dev(kdev);
> struct drm_device *dev = pci_get_drvdata(pdev);
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> int ret = 0;
>
> - if (WARN_ON_ONCE(!HAS_RUNTIME_PM(dev_priv)))
> + if (WARN_ON_ONCE(!HAS_RUNTIME_PM(i915)))
> return -ENODEV;
>
> DRM_DEBUG_KMS("Resuming device\n");
>
> - WARN_ON_ONCE(atomic_read(&dev_priv->runtime_pm.wakeref_count));
> - disable_rpm_wakeref_asserts(dev_priv);
> + WARN_ON_ONCE(atomic_read(&i915->runtime_pm.wakeref_count));
> + disable_rpm_wakeref_asserts(i915);
>
> - intel_opregion_notify_adapter(dev_priv, PCI_D0);
> - dev_priv->runtime_pm.suspended = false;
> - if (intel_uncore_unclaimed_mmio(&dev_priv->uncore))
> + intel_opregion_notify_adapter(i915, PCI_D0);
> + i915->runtime_pm.suspended = false;
> + if (intel_uncore_unclaimed_mmio(&i915->uncore))
> DRM_DEBUG_DRIVER("Unclaimed access during suspend, bios?\n");
>
> - if (INTEL_GEN(dev_priv) >= 11) {
> - bxt_disable_dc9(dev_priv);
> - icl_display_core_init(dev_priv, true);
> - if (dev_priv->csr.dmc_payload) {
> - if (dev_priv->csr.allowed_dc_mask &
> + if (INTEL_GEN(i915) >= 11) {
> + bxt_disable_dc9(i915);
> + icl_display_core_init(i915, true);
> + if (i915->csr.dmc_payload) {
> + if (i915->csr.allowed_dc_mask &
> DC_STATE_EN_UPTO_DC6)
> - skl_enable_dc6(dev_priv);
> - else if (dev_priv->csr.allowed_dc_mask &
> + skl_enable_dc6(i915);
> + else if (i915->csr.allowed_dc_mask &
> DC_STATE_EN_UPTO_DC5)
> - gen9_enable_dc5(dev_priv);
> + gen9_enable_dc5(i915);
> }
> - } else if (IS_GEN9_LP(dev_priv)) {
> - bxt_disable_dc9(dev_priv);
> - bxt_display_core_init(dev_priv, true);
> - if (dev_priv->csr.dmc_payload &&
> - (dev_priv->csr.allowed_dc_mask & DC_STATE_EN_UPTO_DC5))
> - gen9_enable_dc5(dev_priv);
> - } else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) {
> - hsw_disable_pc8(dev_priv);
> - } else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> - ret = vlv_resume_prepare(dev_priv, true);
> + } else if (IS_GEN9_LP(i915)) {
> + bxt_disable_dc9(i915);
> + bxt_display_core_init(i915, true);
> + if (i915->csr.dmc_payload &&
> + (i915->csr.allowed_dc_mask & DC_STATE_EN_UPTO_DC5))
> + gen9_enable_dc5(i915);
> + } else if (IS_HASWELL(i915) || IS_BROADWELL(i915)) {
> + hsw_disable_pc8(i915);
> + } else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> + ret = vlv_resume_prepare(i915, true);
> }
>
> - intel_uncore_runtime_resume(&dev_priv->uncore);
> + intel_uncore_runtime_resume(&i915->uncore);
>
> - intel_runtime_pm_enable_interrupts(dev_priv);
> + intel_runtime_pm_enable_interrupts(i915);
>
> - intel_uc_resume(dev_priv);
> + intel_uc_resume(i915);
>
> /*
> * No point of rolling back things in case of an error, as the best
> * we can do is to hope that things will still work (and disable RPM).
> */
> - i915_gem_init_swizzling(dev_priv);
> - i915_gem_restore_fences(dev_priv);
> + i915_gem_init_swizzling(i915);
> + i915_gem_restore_fences(i915);
>
> /*
> * On VLV/CHV display interrupts are part of the display
> * power well, so hpd is reinitialized from there. For
> * everyone else do it here.
> */
> - if (!IS_VALLEYVIEW(dev_priv) && !IS_CHERRYVIEW(dev_priv))
> - intel_hpd_init(dev_priv);
> + if (!IS_VALLEYVIEW(i915) && !IS_CHERRYVIEW(i915))
> + intel_hpd_init(i915);
>
> - intel_enable_ipc(dev_priv);
> + intel_enable_ipc(i915);
>
> - enable_rpm_wakeref_asserts(dev_priv);
> + enable_rpm_wakeref_asserts(i915);
>
> if (ret)
> DRM_ERROR("Runtime resume failed, disabling it (%d)\n", ret);
> diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h
> index eb98d285d290..a1adb79ca2c3 100644
> --- a/drivers/gpu/drm/i915/i915_drv.h
> +++ b/drivers/gpu/drm/i915/i915_drv.h
> @@ -208,7 +208,7 @@ struct i915_mm_struct;
> struct i915_mmu_object;
>
> struct drm_i915_file_private {
> - struct drm_i915_private *dev_priv;
> + struct drm_i915_private *i915;
> struct drm_file *file;
>
> struct {
> @@ -278,12 +278,12 @@ struct dpll;
> struct intel_cdclk_state;
>
> struct drm_i915_display_funcs {
> - void (*get_cdclk)(struct drm_i915_private *dev_priv,
> + void (*get_cdclk)(struct drm_i915_private *i915,
> struct intel_cdclk_state *cdclk_state);
> - void (*set_cdclk)(struct drm_i915_private *dev_priv,
> + void (*set_cdclk)(struct drm_i915_private *i915,
> const struct intel_cdclk_state *cdclk_state,
> enum pipe pipe);
> - int (*get_fifo_size)(struct drm_i915_private *dev_priv,
> + int (*get_fifo_size)(struct drm_i915_private *i915,
> enum i9xx_plane_id i9xx_plane);
> int (*compute_pipe_wm)(struct intel_crtc_state *cstate);
> int (*compute_intermediate_wm)(struct intel_crtc_state *newstate);
> @@ -317,8 +317,8 @@ struct drm_i915_display_funcs {
> const struct drm_connector_state *old_conn_state);
> void (*fdi_link_train)(struct intel_crtc *crtc,
> const struct intel_crtc_state *crtc_state);
> - void (*init_clock_gating)(struct drm_i915_private *dev_priv);
> - void (*hpd_irq_setup)(struct drm_i915_private *dev_priv);
> + void (*init_clock_gating)(struct drm_i915_private *i915);
> + void (*hpd_irq_setup)(struct drm_i915_private *i915);
> /* clock updates for mode set */
> /* cursor updates */
> /* render clock increase/decrease */
> @@ -552,7 +552,7 @@ struct intel_gmbus {
> u32 reg0;
> i915_reg_t gpio_reg;
> struct i2c_algo_bit_data bit_algo;
> - struct drm_i915_private *dev_priv;
> + struct drm_i915_private *i915;
> };
>
> struct i915_suspend_saved_registers {
> @@ -641,7 +641,7 @@ struct intel_rps {
>
> /*
> * work, interrupts_enabled and pm_iir are protected by
> - * dev_priv->irq_lock
> + * i915->irq_lock
> */
> struct work_struct work;
> bool interrupts_enabled;
> @@ -1240,9 +1240,9 @@ struct i915_perf_stream_ops {
> */
> struct i915_perf_stream {
> /**
> - * @dev_priv: i915 drm device
> + * @i915: i915 drm device
> */
> - struct drm_i915_private *dev_priv;
> + struct drm_i915_private *i915;
>
> /**
> * @link: Links the stream into ``&drm_i915_private->streams``
> @@ -1302,20 +1302,20 @@ struct i915_oa_ops {
> * @is_valid_b_counter_reg: Validates register's address for
> * programming boolean counters for a particular platform.
> */
> - bool (*is_valid_b_counter_reg)(struct drm_i915_private *dev_priv,
> + bool (*is_valid_b_counter_reg)(struct drm_i915_private *i915,
> u32 addr);
>
> /**
> * @is_valid_mux_reg: Validates register's address for programming mux
> * for a particular platform.
> */
> - bool (*is_valid_mux_reg)(struct drm_i915_private *dev_priv, u32 addr);
> + bool (*is_valid_mux_reg)(struct drm_i915_private *i915, u32 addr);
>
> /**
> * @is_valid_flex_reg: Validates register's address for programming
> * flex EU filtering for a particular platform.
> */
> - bool (*is_valid_flex_reg)(struct drm_i915_private *dev_priv, u32 addr);
> + bool (*is_valid_flex_reg)(struct drm_i915_private *i915, u32 addr);
>
> /**
> * @enable_metric_set: Selects and applies any MUX configuration to set
> @@ -1329,7 +1329,7 @@ struct i915_oa_ops {
> * @disable_metric_set: Remove system constraints associated with using
> * the OA unit.
> */
> - void (*disable_metric_set)(struct drm_i915_private *dev_priv);
> + void (*disable_metric_set)(struct drm_i915_private *i915);
>
> /**
> * @oa_enable: Enable periodic sampling
> @@ -1357,7 +1357,7 @@ struct i915_oa_ops {
> * handling the OA unit tail pointer race that affects multiple
> * generations.
> */
> - u32 (*oa_hw_tail_read)(struct drm_i915_private *dev_priv);
> + u32 (*oa_hw_tail_read)(struct drm_i915_private *i915);
> };
>
> struct intel_cdclk_state {
> @@ -1750,13 +1750,13 @@ struct drm_i915_private {
>
> /*
> * Lock associated with adding/modifying/removing OA configs
> - * in dev_priv->perf.metrics_idr.
> + * in i915->perf.metrics_idr.
> */
> struct mutex metrics_lock;
>
> /*
> * List of dynamic configurations, you need to hold
> - * dev_priv->perf.metrics_lock to access it.
> + * i915->perf.metrics_lock to access it.
> */
> struct idr metrics_idr;
>
> @@ -1772,7 +1772,7 @@ struct drm_i915_private {
> * The stream currently using the OA unit. If accessed
> * outside a syscall associated to its file
> * descriptor, you need to hold
> - * dev_priv->drm.struct_mutex.
> + * i915->drm.struct_mutex.
> */
> struct i915_perf_stream *exclusive_stream;
>
> @@ -2066,15 +2066,15 @@ enum hdmi_force_audio {
> GENMASK(INTEL_FRONTBUFFER_BITS_PER_PIPE * ((pipe) + 1) - 1, \
> INTEL_FRONTBUFFER_BITS_PER_PIPE * (pipe))
>
> -#define INTEL_INFO(dev_priv) (&(dev_priv)->__info)
> -#define RUNTIME_INFO(dev_priv) (&(dev_priv)->__runtime)
> -#define DRIVER_CAPS(dev_priv) (&(dev_priv)->caps)
> +#define INTEL_INFO(i915) (&(i915)->__info)
> +#define RUNTIME_INFO(i915) (&(i915)->__runtime)
> +#define DRIVER_CAPS(i915) (&(i915)->caps)
>
> -#define INTEL_GEN(dev_priv) (INTEL_INFO(dev_priv)->gen)
> -#define INTEL_DEVID(dev_priv) (RUNTIME_INFO(dev_priv)->device_id)
> +#define INTEL_GEN(i915) (INTEL_INFO(i915)->gen)
> +#define INTEL_DEVID(i915) (RUNTIME_INFO(i915)->device_id)
>
> #define REVID_FOREVER 0xff
> -#define INTEL_REVID(dev_priv) ((dev_priv)->drm.pdev->revision)
> +#define INTEL_REVID(i915) ((i915)->drm.pdev->revision)
>
> #define INTEL_GEN_MASK(s, e) ( \
> BUILD_BUG_ON_ZERO(!__builtin_constant_p(s)) + \
> @@ -2082,12 +2082,12 @@ enum hdmi_force_audio {
> GENMASK((e) - 1, (s) - 1))
>
> /* Returns true if Gen is in inclusive range [Start, End] */
> -#define IS_GEN_RANGE(dev_priv, s, e) \
> - (!!(INTEL_INFO(dev_priv)->gen_mask & INTEL_GEN_MASK((s), (e))))
> +#define IS_GEN_RANGE(i915, s, e) \
> + (!!(INTEL_INFO(i915)->gen_mask & INTEL_GEN_MASK((s), (e))))
>
> -#define IS_GEN(dev_priv, n) \
> +#define IS_GEN(i915, n) \
> (BUILD_BUG_ON_ZERO(!__builtin_constant_p(n)) + \
> - INTEL_INFO(dev_priv)->gen == (n))
> + INTEL_INFO(i915)->gen == (n))
>
> /*
> * Return true if revision is in range [since,until] inclusive.
> @@ -2159,88 +2159,88 @@ IS_SUBPLATFORM(const struct drm_i915_private *i915,
> return ((mask << (msb - pb)) & (mask << (msb - s))) & BIT(msb);
> }
>
> -#define IS_MOBILE(dev_priv) (INTEL_INFO(dev_priv)->is_mobile)
> -
> -#define IS_I830(dev_priv) IS_PLATFORM(dev_priv, INTEL_I830)
> -#define IS_I845G(dev_priv) IS_PLATFORM(dev_priv, INTEL_I845G)
> -#define IS_I85X(dev_priv) IS_PLATFORM(dev_priv, INTEL_I85X)
> -#define IS_I865G(dev_priv) IS_PLATFORM(dev_priv, INTEL_I865G)
> -#define IS_I915G(dev_priv) IS_PLATFORM(dev_priv, INTEL_I915G)
> -#define IS_I915GM(dev_priv) IS_PLATFORM(dev_priv, INTEL_I915GM)
> -#define IS_I945G(dev_priv) IS_PLATFORM(dev_priv, INTEL_I945G)
> -#define IS_I945GM(dev_priv) IS_PLATFORM(dev_priv, INTEL_I945GM)
> -#define IS_I965G(dev_priv) IS_PLATFORM(dev_priv, INTEL_I965G)
> -#define IS_I965GM(dev_priv) IS_PLATFORM(dev_priv, INTEL_I965GM)
> -#define IS_G45(dev_priv) IS_PLATFORM(dev_priv, INTEL_G45)
> -#define IS_GM45(dev_priv) IS_PLATFORM(dev_priv, INTEL_GM45)
> -#define IS_G4X(dev_priv) (IS_G45(dev_priv) || IS_GM45(dev_priv))
> -#define IS_PINEVIEW(dev_priv) IS_PLATFORM(dev_priv, INTEL_PINEVIEW)
> -#define IS_G33(dev_priv) IS_PLATFORM(dev_priv, INTEL_G33)
> -#define IS_IRONLAKE(dev_priv) IS_PLATFORM(dev_priv, INTEL_IRONLAKE)
> -#define IS_IRONLAKE_M(dev_priv) \
> - (IS_PLATFORM(dev_priv, INTEL_IRONLAKE) && IS_MOBILE(dev_priv))
> -#define IS_IVYBRIDGE(dev_priv) IS_PLATFORM(dev_priv, INTEL_IVYBRIDGE)
> -#define IS_IVB_GT1(dev_priv) (IS_IVYBRIDGE(dev_priv) && \
> - INTEL_INFO(dev_priv)->gt == 1)
> -#define IS_VALLEYVIEW(dev_priv) IS_PLATFORM(dev_priv, INTEL_VALLEYVIEW)
> -#define IS_CHERRYVIEW(dev_priv) IS_PLATFORM(dev_priv, INTEL_CHERRYVIEW)
> -#define IS_HASWELL(dev_priv) IS_PLATFORM(dev_priv, INTEL_HASWELL)
> -#define IS_BROADWELL(dev_priv) IS_PLATFORM(dev_priv, INTEL_BROADWELL)
> -#define IS_SKYLAKE(dev_priv) IS_PLATFORM(dev_priv, INTEL_SKYLAKE)
> -#define IS_BROXTON(dev_priv) IS_PLATFORM(dev_priv, INTEL_BROXTON)
> -#define IS_KABYLAKE(dev_priv) IS_PLATFORM(dev_priv, INTEL_KABYLAKE)
> -#define IS_GEMINILAKE(dev_priv) IS_PLATFORM(dev_priv, INTEL_GEMINILAKE)
> -#define IS_COFFEELAKE(dev_priv) IS_PLATFORM(dev_priv, INTEL_COFFEELAKE)
> -#define IS_CANNONLAKE(dev_priv) IS_PLATFORM(dev_priv, INTEL_CANNONLAKE)
> -#define IS_ICELAKE(dev_priv) IS_PLATFORM(dev_priv, INTEL_ICELAKE)
> -#define IS_ELKHARTLAKE(dev_priv) IS_PLATFORM(dev_priv, INTEL_ELKHARTLAKE)
> -#define IS_HSW_EARLY_SDV(dev_priv) (IS_HASWELL(dev_priv) && \
> - (INTEL_DEVID(dev_priv) & 0xFF00) == 0x0C00)
> -#define IS_BDW_ULT(dev_priv) \
> - IS_SUBPLATFORM(dev_priv, INTEL_BROADWELL, INTEL_SUBPLATFORM_ULT)
> -#define IS_BDW_ULX(dev_priv) \
> - IS_SUBPLATFORM(dev_priv, INTEL_BROADWELL, INTEL_SUBPLATFORM_ULX)
> -#define IS_BDW_GT3(dev_priv) (IS_BROADWELL(dev_priv) && \
> - INTEL_INFO(dev_priv)->gt == 3)
> -#define IS_HSW_ULT(dev_priv) \
> - IS_SUBPLATFORM(dev_priv, INTEL_HASWELL, INTEL_SUBPLATFORM_ULT)
> -#define IS_HSW_GT3(dev_priv) (IS_HASWELL(dev_priv) && \
> - INTEL_INFO(dev_priv)->gt == 3)
> -#define IS_HSW_GT1(dev_priv) (IS_HASWELL(dev_priv) && \
> - INTEL_INFO(dev_priv)->gt == 1)
> +#define IS_MOBILE(i915) (INTEL_INFO(i915)->is_mobile)
> +
> +#define IS_I830(i915) IS_PLATFORM(i915, INTEL_I830)
> +#define IS_I845G(i915) IS_PLATFORM(i915, INTEL_I845G)
> +#define IS_I85X(i915) IS_PLATFORM(i915, INTEL_I85X)
> +#define IS_I865G(i915) IS_PLATFORM(i915, INTEL_I865G)
> +#define IS_I915G(i915) IS_PLATFORM(i915, INTEL_I915G)
> +#define IS_I915GM(i915) IS_PLATFORM(i915, INTEL_I915GM)
> +#define IS_I945G(i915) IS_PLATFORM(i915, INTEL_I945G)
> +#define IS_I945GM(i915) IS_PLATFORM(i915, INTEL_I945GM)
> +#define IS_I965G(i915) IS_PLATFORM(i915, INTEL_I965G)
> +#define IS_I965GM(i915) IS_PLATFORM(i915, INTEL_I965GM)
> +#define IS_G45(i915) IS_PLATFORM(i915, INTEL_G45)
> +#define IS_GM45(i915) IS_PLATFORM(i915, INTEL_GM45)
> +#define IS_G4X(i915) (IS_G45(i915) || IS_GM45(i915))
> +#define IS_PINEVIEW(i915) IS_PLATFORM(i915, INTEL_PINEVIEW)
> +#define IS_G33(i915) IS_PLATFORM(i915, INTEL_G33)
> +#define IS_IRONLAKE(i915) IS_PLATFORM(i915, INTEL_IRONLAKE)
> +#define IS_IRONLAKE_M(i915) \
> + (IS_PLATFORM(i915, INTEL_IRONLAKE) && IS_MOBILE(i915))
> +#define IS_IVYBRIDGE(i915) IS_PLATFORM(i915, INTEL_IVYBRIDGE)
> +#define IS_IVB_GT1(i915) (IS_IVYBRIDGE(i915) && \
> + INTEL_INFO(i915)->gt == 1)
> +#define IS_VALLEYVIEW(i915) IS_PLATFORM(i915, INTEL_VALLEYVIEW)
> +#define IS_CHERRYVIEW(i915) IS_PLATFORM(i915, INTEL_CHERRYVIEW)
> +#define IS_HASWELL(i915) IS_PLATFORM(i915, INTEL_HASWELL)
> +#define IS_BROADWELL(i915) IS_PLATFORM(i915, INTEL_BROADWELL)
> +#define IS_SKYLAKE(i915) IS_PLATFORM(i915, INTEL_SKYLAKE)
> +#define IS_BROXTON(i915) IS_PLATFORM(i915, INTEL_BROXTON)
> +#define IS_KABYLAKE(i915) IS_PLATFORM(i915, INTEL_KABYLAKE)
> +#define IS_GEMINILAKE(i915) IS_PLATFORM(i915, INTEL_GEMINILAKE)
> +#define IS_COFFEELAKE(i915) IS_PLATFORM(i915, INTEL_COFFEELAKE)
> +#define IS_CANNONLAKE(i915) IS_PLATFORM(i915, INTEL_CANNONLAKE)
> +#define IS_ICELAKE(i915) IS_PLATFORM(i915, INTEL_ICELAKE)
> +#define IS_ELKHARTLAKE(i915) IS_PLATFORM(i915, INTEL_ELKHARTLAKE)
> +#define IS_HSW_EARLY_SDV(i915) (IS_HASWELL(i915) && \
> + (INTEL_DEVID(i915) & 0xFF00) == 0x0C00)
> +#define IS_BDW_ULT(i915) \
> + IS_SUBPLATFORM(i915, INTEL_BROADWELL, INTEL_SUBPLATFORM_ULT)
> +#define IS_BDW_ULX(i915) \
> + IS_SUBPLATFORM(i915, INTEL_BROADWELL, INTEL_SUBPLATFORM_ULX)
> +#define IS_BDW_GT3(i915) (IS_BROADWELL(i915) && \
> + INTEL_INFO(i915)->gt == 3)
> +#define IS_HSW_ULT(i915) \
> + IS_SUBPLATFORM(i915, INTEL_HASWELL, INTEL_SUBPLATFORM_ULT)
> +#define IS_HSW_GT3(i915) (IS_HASWELL(i915) && \
> + INTEL_INFO(i915)->gt == 3)
> +#define IS_HSW_GT1(i915) (IS_HASWELL(i915) && \
> + INTEL_INFO(i915)->gt == 1)
> /* ULX machines are also considered ULT. */
> -#define IS_HSW_ULX(dev_priv) \
> - IS_SUBPLATFORM(dev_priv, INTEL_HASWELL, INTEL_SUBPLATFORM_ULX)
> -#define IS_SKL_ULT(dev_priv) \
> - IS_SUBPLATFORM(dev_priv, INTEL_SKYLAKE, INTEL_SUBPLATFORM_ULT)
> -#define IS_SKL_ULX(dev_priv) \
> - IS_SUBPLATFORM(dev_priv, INTEL_SKYLAKE, INTEL_SUBPLATFORM_ULX)
> -#define IS_KBL_ULT(dev_priv) \
> - IS_SUBPLATFORM(dev_priv, INTEL_KABYLAKE, INTEL_SUBPLATFORM_ULT)
> -#define IS_KBL_ULX(dev_priv) \
> - IS_SUBPLATFORM(dev_priv, INTEL_KABYLAKE, INTEL_SUBPLATFORM_ULX)
> -#define IS_SKL_GT2(dev_priv) (IS_SKYLAKE(dev_priv) && \
> - INTEL_INFO(dev_priv)->gt == 2)
> -#define IS_SKL_GT3(dev_priv) (IS_SKYLAKE(dev_priv) && \
> - INTEL_INFO(dev_priv)->gt == 3)
> -#define IS_SKL_GT4(dev_priv) (IS_SKYLAKE(dev_priv) && \
> - INTEL_INFO(dev_priv)->gt == 4)
> -#define IS_KBL_GT2(dev_priv) (IS_KABYLAKE(dev_priv) && \
> - INTEL_INFO(dev_priv)->gt == 2)
> -#define IS_KBL_GT3(dev_priv) (IS_KABYLAKE(dev_priv) && \
> - INTEL_INFO(dev_priv)->gt == 3)
> -#define IS_CFL_ULT(dev_priv) \
> - IS_SUBPLATFORM(dev_priv, INTEL_COFFEELAKE, INTEL_SUBPLATFORM_ULT)
> -#define IS_CFL_ULX(dev_priv) \
> - IS_SUBPLATFORM(dev_priv, INTEL_COFFEELAKE, INTEL_SUBPLATFORM_ULX)
> -#define IS_CFL_GT2(dev_priv) (IS_COFFEELAKE(dev_priv) && \
> - INTEL_INFO(dev_priv)->gt == 2)
> -#define IS_CFL_GT3(dev_priv) (IS_COFFEELAKE(dev_priv) && \
> - INTEL_INFO(dev_priv)->gt == 3)
> -#define IS_CNL_WITH_PORT_F(dev_priv) \
> - IS_SUBPLATFORM(dev_priv, INTEL_CANNONLAKE, INTEL_SUBPLATFORM_PORTF)
> -#define IS_ICL_WITH_PORT_F(dev_priv) \
> - IS_SUBPLATFORM(dev_priv, INTEL_ICELAKE, INTEL_SUBPLATFORM_PORTF)
> +#define IS_HSW_ULX(i915) \
> + IS_SUBPLATFORM(i915, INTEL_HASWELL, INTEL_SUBPLATFORM_ULX)
> +#define IS_SKL_ULT(i915) \
> + IS_SUBPLATFORM(i915, INTEL_SKYLAKE, INTEL_SUBPLATFORM_ULT)
> +#define IS_SKL_ULX(i915) \
> + IS_SUBPLATFORM(i915, INTEL_SKYLAKE, INTEL_SUBPLATFORM_ULX)
> +#define IS_KBL_ULT(i915) \
> + IS_SUBPLATFORM(i915, INTEL_KABYLAKE, INTEL_SUBPLATFORM_ULT)
> +#define IS_KBL_ULX(i915) \
> + IS_SUBPLATFORM(i915, INTEL_KABYLAKE, INTEL_SUBPLATFORM_ULX)
> +#define IS_SKL_GT2(i915) (IS_SKYLAKE(i915) && \
> + INTEL_INFO(i915)->gt == 2)
> +#define IS_SKL_GT3(i915) (IS_SKYLAKE(i915) && \
> + INTEL_INFO(i915)->gt == 3)
> +#define IS_SKL_GT4(i915) (IS_SKYLAKE(i915) && \
> + INTEL_INFO(i915)->gt == 4)
> +#define IS_KBL_GT2(i915) (IS_KABYLAKE(i915) && \
> + INTEL_INFO(i915)->gt == 2)
> +#define IS_KBL_GT3(i915) (IS_KABYLAKE(i915) && \
> + INTEL_INFO(i915)->gt == 3)
> +#define IS_CFL_ULT(i915) \
> + IS_SUBPLATFORM(i915, INTEL_COFFEELAKE, INTEL_SUBPLATFORM_ULT)
> +#define IS_CFL_ULX(i915) \
> + IS_SUBPLATFORM(i915, INTEL_COFFEELAKE, INTEL_SUBPLATFORM_ULX)
> +#define IS_CFL_GT2(i915) (IS_COFFEELAKE(i915) && \
> + INTEL_INFO(i915)->gt == 2)
> +#define IS_CFL_GT3(i915) (IS_COFFEELAKE(i915) && \
> + INTEL_INFO(i915)->gt == 3)
> +#define IS_CNL_WITH_PORT_F(i915) \
> + IS_SUBPLATFORM(i915, INTEL_CANNONLAKE, INTEL_SUBPLATFORM_PORTF)
> +#define IS_ICL_WITH_PORT_F(i915) \
> + IS_SUBPLATFORM(i915, INTEL_ICELAKE, INTEL_SUBPLATFORM_PORTF)
>
> #define SKL_REVID_A0 0x0
> #define SKL_REVID_B0 0x1
> @@ -2259,8 +2259,8 @@ IS_SUBPLATFORM(const struct drm_i915_private *i915,
> #define BXT_REVID_B_LAST 0x8
> #define BXT_REVID_C0 0x9
>
> -#define IS_BXT_REVID(dev_priv, since, until) \
> - (IS_BROXTON(dev_priv) && IS_REVID(dev_priv, since, until))
> +#define IS_BXT_REVID(i915, since, until) \
> + (IS_BROXTON(i915) && IS_REVID(i915, since, until))
>
> #define KBL_REVID_A0 0x0
> #define KBL_REVID_B0 0x1
> @@ -2268,14 +2268,14 @@ IS_SUBPLATFORM(const struct drm_i915_private *i915,
> #define KBL_REVID_D0 0x3
> #define KBL_REVID_E0 0x4
>
> -#define IS_KBL_REVID(dev_priv, since, until) \
> - (IS_KABYLAKE(dev_priv) && IS_REVID(dev_priv, since, until))
> +#define IS_KBL_REVID(i915, since, until) \
> + (IS_KABYLAKE(i915) && IS_REVID(i915, since, until))
>
> #define GLK_REVID_A0 0x0
> #define GLK_REVID_A1 0x1
>
> -#define IS_GLK_REVID(dev_priv, since, until) \
> - (IS_GEMINILAKE(dev_priv) && IS_REVID(dev_priv, since, until))
> +#define IS_GLK_REVID(i915, since, until) \
> + (IS_GEMINILAKE(i915) && IS_REVID(i915, since, until))
>
> #define CNL_REVID_A0 0x0
> #define CNL_REVID_B0 0x1
> @@ -2293,122 +2293,122 @@ IS_SUBPLATFORM(const struct drm_i915_private *i915,
> #define IS_ICL_REVID(p, since, until) \
> (IS_ICELAKE(p) && IS_REVID(p, since, until))
>
> -#define IS_LP(dev_priv) (INTEL_INFO(dev_priv)->is_lp)
> -#define IS_GEN9_LP(dev_priv) (IS_GEN(dev_priv, 9) && IS_LP(dev_priv))
> -#define IS_GEN9_BC(dev_priv) (IS_GEN(dev_priv, 9) && !IS_LP(dev_priv))
> +#define IS_LP(i915) (INTEL_INFO(i915)->is_lp)
> +#define IS_GEN9_LP(i915) (IS_GEN(i915, 9) && IS_LP(i915))
> +#define IS_GEN9_BC(i915) (IS_GEN(i915, 9) && !IS_LP(i915))
>
> -#define HAS_ENGINE(dev_priv, id) (INTEL_INFO(dev_priv)->engine_mask & BIT(id))
> +#define HAS_ENGINE(i915, id) (INTEL_INFO(i915)->engine_mask & BIT(id))
>
> -#define ENGINE_INSTANCES_MASK(dev_priv, first, count) ({ \
> +#define ENGINE_INSTANCES_MASK(i915, first, count) ({ \
> unsigned int first__ = (first); \
> unsigned int count__ = (count); \
> - (INTEL_INFO(dev_priv)->engine_mask & \
> + (INTEL_INFO(i915)->engine_mask & \
> GENMASK(first__ + count__ - 1, first__)) >> first__; \
> })
> -#define VDBOX_MASK(dev_priv) \
> - ENGINE_INSTANCES_MASK(dev_priv, VCS0, I915_MAX_VCS)
> -#define VEBOX_MASK(dev_priv) \
> - ENGINE_INSTANCES_MASK(dev_priv, VECS0, I915_MAX_VECS)
> -
> -#define HAS_LLC(dev_priv) (INTEL_INFO(dev_priv)->has_llc)
> -#define HAS_SNOOP(dev_priv) (INTEL_INFO(dev_priv)->has_snoop)
> -#define HAS_EDRAM(dev_priv) ((dev_priv)->edram_size_mb)
> -#define HAS_WT(dev_priv) ((IS_HASWELL(dev_priv) || \
> - IS_BROADWELL(dev_priv)) && HAS_EDRAM(dev_priv))
> -
> -#define HWS_NEEDS_PHYSICAL(dev_priv) (INTEL_INFO(dev_priv)->hws_needs_physical)
> -
> -#define HAS_LOGICAL_RING_CONTEXTS(dev_priv) \
> - (INTEL_INFO(dev_priv)->has_logical_ring_contexts)
> -#define HAS_LOGICAL_RING_ELSQ(dev_priv) \
> - (INTEL_INFO(dev_priv)->has_logical_ring_elsq)
> -#define HAS_LOGICAL_RING_PREEMPTION(dev_priv) \
> - (INTEL_INFO(dev_priv)->has_logical_ring_preemption)
> -
> -#define HAS_EXECLISTS(dev_priv) HAS_LOGICAL_RING_CONTEXTS(dev_priv)
> -
> -#define INTEL_PPGTT(dev_priv) (INTEL_INFO(dev_priv)->ppgtt_type)
> -#define HAS_PPGTT(dev_priv) \
> - (INTEL_PPGTT(dev_priv) != INTEL_PPGTT_NONE)
> -#define HAS_FULL_PPGTT(dev_priv) \
> - (INTEL_PPGTT(dev_priv) >= INTEL_PPGTT_FULL)
> -
> -#define HAS_PAGE_SIZES(dev_priv, sizes) ({ \
> +#define VDBOX_MASK(i915) \
> + ENGINE_INSTANCES_MASK(i915, VCS0, I915_MAX_VCS)
> +#define VEBOX_MASK(i915) \
> + ENGINE_INSTANCES_MASK(i915, VECS0, I915_MAX_VECS)
> +
> +#define HAS_LLC(i915) (INTEL_INFO(i915)->has_llc)
> +#define HAS_SNOOP(i915) (INTEL_INFO(i915)->has_snoop)
> +#define HAS_EDRAM(i915) ((i915)->edram_size_mb)
> +#define HAS_WT(i915) ((IS_HASWELL(i915) || \
> + IS_BROADWELL(i915)) && HAS_EDRAM(i915))
> +
> +#define HWS_NEEDS_PHYSICAL(i915) (INTEL_INFO(i915)->hws_needs_physical)
> +
> +#define HAS_LOGICAL_RING_CONTEXTS(i915) \
> + (INTEL_INFO(i915)->has_logical_ring_contexts)
> +#define HAS_LOGICAL_RING_ELSQ(i915) \
> + (INTEL_INFO(i915)->has_logical_ring_elsq)
> +#define HAS_LOGICAL_RING_PREEMPTION(i915) \
> + (INTEL_INFO(i915)->has_logical_ring_preemption)
> +
> +#define HAS_EXECLISTS(i915) HAS_LOGICAL_RING_CONTEXTS(i915)
> +
> +#define INTEL_PPGTT(i915) (INTEL_INFO(i915)->ppgtt_type)
> +#define HAS_PPGTT(i915) \
> + (INTEL_PPGTT(i915) != INTEL_PPGTT_NONE)
> +#define HAS_FULL_PPGTT(i915) \
> + (INTEL_PPGTT(i915) >= INTEL_PPGTT_FULL)
> +
> +#define HAS_PAGE_SIZES(i915, sizes) ({ \
> GEM_BUG_ON((sizes) == 0); \
> - ((sizes) & ~INTEL_INFO(dev_priv)->page_sizes) == 0; \
> + ((sizes) & ~INTEL_INFO(i915)->page_sizes) == 0; \
> })
>
> -#define HAS_OVERLAY(dev_priv) (INTEL_INFO(dev_priv)->display.has_overlay)
> -#define OVERLAY_NEEDS_PHYSICAL(dev_priv) \
> - (INTEL_INFO(dev_priv)->display.overlay_needs_physical)
> +#define HAS_OVERLAY(i915) (INTEL_INFO(i915)->display.has_overlay)
> +#define OVERLAY_NEEDS_PHYSICAL(i915) \
> + (INTEL_INFO(i915)->display.overlay_needs_physical)
>
> /* Early gen2 have a totally busted CS tlb and require pinned batches. */
> -#define HAS_BROKEN_CS_TLB(dev_priv) (IS_I830(dev_priv) || IS_I845G(dev_priv))
> +#define HAS_BROKEN_CS_TLB(i915) (IS_I830(i915) || IS_I845G(i915))
>
> /* WaRsDisableCoarsePowerGating:skl,cnl */
> -#define NEEDS_WaRsDisableCoarsePowerGating(dev_priv) \
> - (IS_CANNONLAKE(dev_priv) || \
> - IS_SKL_GT3(dev_priv) || IS_SKL_GT4(dev_priv))
> +#define NEEDS_WaRsDisableCoarsePowerGating(i915) \
> + (IS_CANNONLAKE(i915) || \
> + IS_SKL_GT3(i915) || IS_SKL_GT4(i915))
>
> -#define HAS_GMBUS_IRQ(dev_priv) (INTEL_GEN(dev_priv) >= 4)
> -#define HAS_GMBUS_BURST_READ(dev_priv) (INTEL_GEN(dev_priv) >= 10 || \
> - IS_GEMINILAKE(dev_priv) || \
> - IS_KABYLAKE(dev_priv))
> +#define HAS_GMBUS_IRQ(i915) (INTEL_GEN(i915) >= 4)
> +#define HAS_GMBUS_BURST_READ(i915) (INTEL_GEN(i915) >= 10 || \
> + IS_GEMINILAKE(i915) || \
> + IS_KABYLAKE(i915))
>
> /* With the 945 and later, Y tiling got adjusted so that it was 32 128-byte
> * rows, which changed the alignment requirements and fence programming.
> */
> -#define HAS_128_BYTE_Y_TILING(dev_priv) (!IS_GEN(dev_priv, 2) && \
> - !(IS_I915G(dev_priv) || \
> - IS_I915GM(dev_priv)))
> -#define SUPPORTS_TV(dev_priv) (INTEL_INFO(dev_priv)->display.supports_tv)
> -#define I915_HAS_HOTPLUG(dev_priv) (INTEL_INFO(dev_priv)->display.has_hotplug)
> +#define HAS_128_BYTE_Y_TILING(i915) (!IS_GEN(i915, 2) && \
> + !(IS_I915G(i915) || \
> + IS_I915GM(i915)))
> +#define SUPPORTS_TV(i915) (INTEL_INFO(i915)->display.supports_tv)
> +#define I915_HAS_HOTPLUG(i915) (INTEL_INFO(i915)->display.has_hotplug)
>
> -#define HAS_FW_BLC(dev_priv) (INTEL_GEN(dev_priv) > 2)
> -#define HAS_FBC(dev_priv) (INTEL_INFO(dev_priv)->display.has_fbc)
> -#define HAS_CUR_FBC(dev_priv) (!HAS_GMCH(dev_priv) && INTEL_GEN(dev_priv) >= 7)
> +#define HAS_FW_BLC(i915) (INTEL_GEN(i915) > 2)
> +#define HAS_FBC(i915) (INTEL_INFO(i915)->display.has_fbc)
> +#define HAS_CUR_FBC(i915) (!HAS_GMCH(i915) && INTEL_GEN(i915) >= 7)
>
> -#define HAS_IPS(dev_priv) (IS_HSW_ULT(dev_priv) || IS_BROADWELL(dev_priv))
> +#define HAS_IPS(i915) (IS_HSW_ULT(i915) || IS_BROADWELL(i915))
>
> -#define HAS_DP_MST(dev_priv) (INTEL_INFO(dev_priv)->display.has_dp_mst)
> +#define HAS_DP_MST(i915) (INTEL_INFO(i915)->display.has_dp_mst)
>
> -#define HAS_DDI(dev_priv) (INTEL_INFO(dev_priv)->display.has_ddi)
> -#define HAS_FPGA_DBG_UNCLAIMED(dev_priv) (INTEL_INFO(dev_priv)->has_fpga_dbg)
> -#define HAS_PSR(dev_priv) (INTEL_INFO(dev_priv)->display.has_psr)
> -#define HAS_TRANSCODER_EDP(dev_priv) (INTEL_INFO(dev_priv)->trans_offsets[TRANSCODER_EDP] != 0)
> +#define HAS_DDI(i915) (INTEL_INFO(i915)->display.has_ddi)
> +#define HAS_FPGA_DBG_UNCLAIMED(i915) (INTEL_INFO(i915)->has_fpga_dbg)
> +#define HAS_PSR(i915) (INTEL_INFO(i915)->display.has_psr)
> +#define HAS_TRANSCODER_EDP(i915) (INTEL_INFO(i915)->trans_offsets[TRANSCODER_EDP] != 0)
>
> -#define HAS_RC6(dev_priv) (INTEL_INFO(dev_priv)->has_rc6)
> -#define HAS_RC6p(dev_priv) (INTEL_INFO(dev_priv)->has_rc6p)
> -#define HAS_RC6pp(dev_priv) (false) /* HW was never validated */
> +#define HAS_RC6(i915) (INTEL_INFO(i915)->has_rc6)
> +#define HAS_RC6p(i915) (INTEL_INFO(i915)->has_rc6p)
> +#define HAS_RC6pp(i915) (false) /* HW was never validated */
>
> -#define HAS_RPS(dev_priv) (INTEL_INFO(dev_priv)->has_rps)
> +#define HAS_RPS(i915) (INTEL_INFO(i915)->has_rps)
>
> -#define HAS_CSR(dev_priv) (INTEL_INFO(dev_priv)->display.has_csr)
> +#define HAS_CSR(i915) (INTEL_INFO(i915)->display.has_csr)
>
> -#define HAS_RUNTIME_PM(dev_priv) (INTEL_INFO(dev_priv)->has_runtime_pm)
> -#define HAS_64BIT_RELOC(dev_priv) (INTEL_INFO(dev_priv)->has_64bit_reloc)
> +#define HAS_RUNTIME_PM(i915) (INTEL_INFO(i915)->has_runtime_pm)
> +#define HAS_64BIT_RELOC(i915) (INTEL_INFO(i915)->has_64bit_reloc)
>
> -#define HAS_IPC(dev_priv) (INTEL_INFO(dev_priv)->display.has_ipc)
> +#define HAS_IPC(i915) (INTEL_INFO(i915)->display.has_ipc)
>
> /*
> * For now, anything with a GuC requires uCode loading, and then supports
> * command submission once loaded. But these are logically independent
> * properties, so we have separate macros to test them.
> */
> -#define HAS_GUC(dev_priv) (INTEL_INFO(dev_priv)->has_guc)
> -#define HAS_GUC_UCODE(dev_priv) (HAS_GUC(dev_priv))
> -#define HAS_GUC_SCHED(dev_priv) (HAS_GUC(dev_priv))
> +#define HAS_GUC(i915) (INTEL_INFO(i915)->has_guc)
> +#define HAS_GUC_UCODE(i915) (HAS_GUC(i915))
> +#define HAS_GUC_SCHED(i915) (HAS_GUC(i915))
>
> /* For now, anything with a GuC has also HuC */
> -#define HAS_HUC(dev_priv) (HAS_GUC(dev_priv))
> -#define HAS_HUC_UCODE(dev_priv) (HAS_GUC(dev_priv))
> +#define HAS_HUC(i915) (HAS_GUC(i915))
> +#define HAS_HUC_UCODE(i915) (HAS_GUC(i915))
>
> /* Having a GuC is not the same as using a GuC */
> -#define USES_GUC(dev_priv) intel_uc_is_using_guc(dev_priv)
> -#define USES_GUC_SUBMISSION(dev_priv) intel_uc_is_using_guc_submission(dev_priv)
> -#define USES_HUC(dev_priv) intel_uc_is_using_huc(dev_priv)
> +#define USES_GUC(i915) intel_uc_is_using_guc(i915)
> +#define USES_GUC_SUBMISSION(i915) intel_uc_is_using_guc_submission(i915)
> +#define USES_HUC(i915) intel_uc_is_using_huc(i915)
>
> -#define HAS_POOLED_EU(dev_priv) (INTEL_INFO(dev_priv)->has_pooled_eu)
> +#define HAS_POOLED_EU(i915) (INTEL_INFO(i915)->has_pooled_eu)
>
> #define INTEL_PCH_DEVICE_ID_MASK 0xff80
> #define INTEL_PCH_IBX_DEVICE_ID_TYPE 0x3b00
> @@ -2429,36 +2429,36 @@ IS_SUBPLATFORM(const struct drm_i915_private *i915,
> #define INTEL_PCH_P3X_DEVICE_ID_TYPE 0x7000
> #define INTEL_PCH_QEMU_DEVICE_ID_TYPE 0x2900 /* qemu q35 has 2918 */
>
> -#define INTEL_PCH_TYPE(dev_priv) ((dev_priv)->pch_type)
> -#define INTEL_PCH_ID(dev_priv) ((dev_priv)->pch_id)
> -#define HAS_PCH_ICP(dev_priv) (INTEL_PCH_TYPE(dev_priv) == PCH_ICP)
> -#define HAS_PCH_CNP(dev_priv) (INTEL_PCH_TYPE(dev_priv) == PCH_CNP)
> -#define HAS_PCH_SPT(dev_priv) (INTEL_PCH_TYPE(dev_priv) == PCH_SPT)
> -#define HAS_PCH_LPT(dev_priv) (INTEL_PCH_TYPE(dev_priv) == PCH_LPT)
> -#define HAS_PCH_LPT_LP(dev_priv) \
> - (INTEL_PCH_ID(dev_priv) == INTEL_PCH_LPT_LP_DEVICE_ID_TYPE || \
> - INTEL_PCH_ID(dev_priv) == INTEL_PCH_WPT_LP_DEVICE_ID_TYPE)
> -#define HAS_PCH_LPT_H(dev_priv) \
> - (INTEL_PCH_ID(dev_priv) == INTEL_PCH_LPT_DEVICE_ID_TYPE || \
> - INTEL_PCH_ID(dev_priv) == INTEL_PCH_WPT_DEVICE_ID_TYPE)
> -#define HAS_PCH_CPT(dev_priv) (INTEL_PCH_TYPE(dev_priv) == PCH_CPT)
> -#define HAS_PCH_IBX(dev_priv) (INTEL_PCH_TYPE(dev_priv) == PCH_IBX)
> -#define HAS_PCH_NOP(dev_priv) (INTEL_PCH_TYPE(dev_priv) == PCH_NOP)
> -#define HAS_PCH_SPLIT(dev_priv) (INTEL_PCH_TYPE(dev_priv) != PCH_NONE)
> -
> -#define HAS_GMCH(dev_priv) (INTEL_INFO(dev_priv)->display.has_gmch)
> -
> -#define HAS_LSPCON(dev_priv) (INTEL_GEN(dev_priv) >= 9)
> +#define INTEL_PCH_TYPE(i915) ((i915)->pch_type)
> +#define INTEL_PCH_ID(i915) ((i915)->pch_id)
> +#define HAS_PCH_ICP(i915) (INTEL_PCH_TYPE(i915) == PCH_ICP)
> +#define HAS_PCH_CNP(i915) (INTEL_PCH_TYPE(i915) == PCH_CNP)
> +#define HAS_PCH_SPT(i915) (INTEL_PCH_TYPE(i915) == PCH_SPT)
> +#define HAS_PCH_LPT(i915) (INTEL_PCH_TYPE(i915) == PCH_LPT)
> +#define HAS_PCH_LPT_LP(i915) \
> + (INTEL_PCH_ID(i915) == INTEL_PCH_LPT_LP_DEVICE_ID_TYPE || \
> + INTEL_PCH_ID(i915) == INTEL_PCH_WPT_LP_DEVICE_ID_TYPE)
> +#define HAS_PCH_LPT_H(i915) \
> + (INTEL_PCH_ID(i915) == INTEL_PCH_LPT_DEVICE_ID_TYPE || \
> + INTEL_PCH_ID(i915) == INTEL_PCH_WPT_DEVICE_ID_TYPE)
> +#define HAS_PCH_CPT(i915) (INTEL_PCH_TYPE(i915) == PCH_CPT)
> +#define HAS_PCH_IBX(i915) (INTEL_PCH_TYPE(i915) == PCH_IBX)
> +#define HAS_PCH_NOP(i915) (INTEL_PCH_TYPE(i915) == PCH_NOP)
> +#define HAS_PCH_SPLIT(i915) (INTEL_PCH_TYPE(i915) != PCH_NONE)
> +
> +#define HAS_GMCH(i915) (INTEL_INFO(i915)->display.has_gmch)
> +
> +#define HAS_LSPCON(i915) (INTEL_GEN(i915) >= 9)
>
> /* DPF == dynamic parity feature */
> -#define HAS_L3_DPF(dev_priv) (INTEL_INFO(dev_priv)->has_l3_dpf)
> -#define NUM_L3_SLICES(dev_priv) (IS_HSW_GT3(dev_priv) ? \
> - 2 : HAS_L3_DPF(dev_priv))
> +#define HAS_L3_DPF(i915) (INTEL_INFO(i915)->has_l3_dpf)
> +#define NUM_L3_SLICES(i915) (IS_HSW_GT3(i915) ? \
> + 2 : HAS_L3_DPF(i915))
>
> #define GT_FREQUENCY_MULTIPLIER 50
> #define GEN9_FREQ_SCALER 3
>
> -#define HAS_DISPLAY(dev_priv) (INTEL_INFO(dev_priv)->num_pipes > 0)
> +#define HAS_DISPLAY(i915) (INTEL_INFO(i915)->num_pipes > 0)
>
> #include "i915_trace.h"
>
> @@ -2471,24 +2471,24 @@ static inline bool intel_vtd_active(void)
> return false;
> }
>
> -static inline bool intel_scanout_needs_vtd_wa(struct drm_i915_private *dev_priv)
> +static inline bool intel_scanout_needs_vtd_wa(struct drm_i915_private *i915)
> {
> - return INTEL_GEN(dev_priv) >= 6 && intel_vtd_active();
> + return INTEL_GEN(i915) >= 6 && intel_vtd_active();
> }
>
> static inline bool
> -intel_ggtt_update_needs_vtd_wa(struct drm_i915_private *dev_priv)
> +intel_ggtt_update_needs_vtd_wa(struct drm_i915_private *i915)
> {
> - return IS_BROXTON(dev_priv) && intel_vtd_active();
> + return IS_BROXTON(i915) && intel_vtd_active();
> }
>
> /* i915_drv.c */
> void __printf(3, 4)
> -__i915_printk(struct drm_i915_private *dev_priv, const char *level,
> +__i915_printk(struct drm_i915_private *i915, const char *level,
> const char *fmt, ...);
>
> -#define i915_report_error(dev_priv, fmt, ...) \
> - __i915_printk(dev_priv, KERN_ERR, fmt, ##__VA_ARGS__)
> +#define i915_report_error(i915, fmt, ...) \
> + __i915_printk(i915, KERN_ERR, fmt, ##__VA_ARGS__)
>
> #ifdef CONFIG_COMPAT
> extern long i915_compat_ioctl(struct file *filp, unsigned int cmd,
> @@ -2503,12 +2503,12 @@ extern int i915_driver_load(struct pci_dev *pdev,
> extern void i915_driver_unload(struct drm_device *dev);
>
> extern void intel_engine_init_hangcheck(struct intel_engine_cs *engine);
> -extern void intel_hangcheck_init(struct drm_i915_private *dev_priv);
> -int vlv_force_gfx_clock(struct drm_i915_private *dev_priv, bool on);
> +extern void intel_hangcheck_init(struct drm_i915_private *i915);
> +int vlv_force_gfx_clock(struct drm_i915_private *i915, bool on);
>
> -u32 intel_calculate_mcr_s_ss_select(struct drm_i915_private *dev_priv);
> +u32 intel_calculate_mcr_s_ss_select(struct drm_i915_private *i915);
>
> -static inline void i915_queue_hangcheck(struct drm_i915_private *dev_priv)
> +static inline void i915_queue_hangcheck(struct drm_i915_private *i915)
> {
> unsigned long delay;
>
> @@ -2522,28 +2522,28 @@ static inline void i915_queue_hangcheck(struct drm_i915_private *dev_priv)
>
> delay = round_jiffies_up_relative(DRM_I915_HANGCHECK_JIFFIES);
> queue_delayed_work(system_long_wq,
> - &dev_priv->gpu_error.hangcheck_work, delay);
> + &i915->gpu_error.hangcheck_work, delay);
> }
>
> -static inline bool intel_gvt_active(struct drm_i915_private *dev_priv)
> +static inline bool intel_gvt_active(struct drm_i915_private *i915)
> {
> - return dev_priv->gvt;
> + return i915->gvt;
> }
>
> -static inline bool intel_vgpu_active(struct drm_i915_private *dev_priv)
> +static inline bool intel_vgpu_active(struct drm_i915_private *i915)
> {
> - return dev_priv->vgpu.active;
> + return i915->vgpu.active;
> }
>
> /* i915_gem.c */
> -int i915_gem_init_userptr(struct drm_i915_private *dev_priv);
> -void i915_gem_cleanup_userptr(struct drm_i915_private *dev_priv);
> +int i915_gem_init_userptr(struct drm_i915_private *i915);
> +void i915_gem_cleanup_userptr(struct drm_i915_private *i915);
> void i915_gem_sanitize(struct drm_i915_private *i915);
> -int i915_gem_init_early(struct drm_i915_private *dev_priv);
> -void i915_gem_cleanup_early(struct drm_i915_private *dev_priv);
> -void i915_gem_load_init_fences(struct drm_i915_private *dev_priv);
> -int i915_gem_freeze(struct drm_i915_private *dev_priv);
> -int i915_gem_freeze_late(struct drm_i915_private *dev_priv);
> +int i915_gem_init_early(struct drm_i915_private *i915);
> +void i915_gem_cleanup_early(struct drm_i915_private *i915);
> +void i915_gem_load_init_fences(struct drm_i915_private *i915);
> +int i915_gem_freeze(struct drm_i915_private *i915);
> +int i915_gem_freeze_late(struct drm_i915_private *i915);
>
> static inline void i915_gem_drain_freed_objects(struct drm_i915_private *i915)
> {
> @@ -2591,7 +2591,7 @@ i915_gem_object_ggtt_pin(struct drm_i915_gem_object *obj,
>
> int i915_gem_object_unbind(struct drm_i915_gem_object *obj);
>
> -void i915_gem_runtime_suspend(struct drm_i915_private *dev_priv);
> +void i915_gem_runtime_suspend(struct drm_i915_private *i915);
>
> static inline int __must_check
> i915_mutex_lock_interruptible(struct drm_device *dev)
> @@ -2633,20 +2633,20 @@ static inline u32 i915_reset_engine_count(struct i915_gpu_error *error,
> return READ_ONCE(error->reset_engine_count[engine->id]);
> }
>
> -void i915_gem_set_wedged(struct drm_i915_private *dev_priv);
> -bool i915_gem_unset_wedged(struct drm_i915_private *dev_priv);
> +void i915_gem_set_wedged(struct drm_i915_private *i915);
> +bool i915_gem_unset_wedged(struct drm_i915_private *i915);
>
> void i915_gem_init_mmio(struct drm_i915_private *i915);
> -int __must_check i915_gem_init(struct drm_i915_private *dev_priv);
> -int __must_check i915_gem_init_hw(struct drm_i915_private *dev_priv);
> -void i915_gem_init_swizzling(struct drm_i915_private *dev_priv);
> -void i915_gem_fini_hw(struct drm_i915_private *dev_priv);
> -void i915_gem_fini(struct drm_i915_private *dev_priv);
> -int i915_gem_wait_for_idle(struct drm_i915_private *dev_priv,
> +int __must_check i915_gem_init(struct drm_i915_private *i915);
> +int __must_check i915_gem_init_hw(struct drm_i915_private *i915);
> +void i915_gem_init_swizzling(struct drm_i915_private *i915);
> +void i915_gem_fini_hw(struct drm_i915_private *i915);
> +void i915_gem_fini(struct drm_i915_private *i915);
> +int i915_gem_wait_for_idle(struct drm_i915_private *i915,
> unsigned int flags, long timeout);
> -void i915_gem_suspend(struct drm_i915_private *dev_priv);
> -void i915_gem_suspend_late(struct drm_i915_private *dev_priv);
> -void i915_gem_resume(struct drm_i915_private *dev_priv);
> +void i915_gem_suspend(struct drm_i915_private *i915);
> +void i915_gem_suspend_late(struct drm_i915_private *i915);
> +void i915_gem_resume(struct drm_i915_private *i915);
> vm_fault_t i915_gem_fault(struct vm_fault *vmf);
>
> int i915_gem_open(struct drm_i915_private *i915, struct drm_file *file);
> @@ -2663,12 +2663,12 @@ struct dma_buf *i915_gem_prime_export(struct drm_device *dev,
>
> /* i915_gem_fence_reg.c */
> struct drm_i915_fence_reg *
> -i915_reserve_fence(struct drm_i915_private *dev_priv);
> +i915_reserve_fence(struct drm_i915_private *i915);
> void i915_unreserve_fence(struct drm_i915_fence_reg *fence);
>
> -void i915_gem_restore_fences(struct drm_i915_private *dev_priv);
> +void i915_gem_restore_fences(struct drm_i915_private *i915);
>
> -void i915_gem_detect_bit_6_swizzle(struct drm_i915_private *dev_priv);
> +void i915_gem_detect_bit_6_swizzle(struct drm_i915_private *i915);
> void i915_gem_object_do_bit_17_swizzle(struct drm_i915_gem_object *obj,
> struct sg_table *pages);
> void i915_gem_object_save_bit_17_swizzle(struct drm_i915_gem_object *obj,
> @@ -2715,40 +2715,40 @@ int __must_check i915_gem_evict_for_node(struct i915_address_space *vm,
> unsigned int flags);
> int i915_gem_evict_vm(struct i915_address_space *vm);
>
> -void i915_gem_flush_ggtt_writes(struct drm_i915_private *dev_priv);
> +void i915_gem_flush_ggtt_writes(struct drm_i915_private *i915);
>
> /* belongs in i915_gem_gtt.h */
> -static inline void i915_gem_chipset_flush(struct drm_i915_private *dev_priv)
> +static inline void i915_gem_chipset_flush(struct drm_i915_private *i915)
> {
> wmb();
> - if (INTEL_GEN(dev_priv) < 6)
> + if (INTEL_GEN(i915) < 6)
> intel_gtt_chipset_flush();
> }
>
> /* i915_gem_stolen.c */
> -int i915_gem_stolen_insert_node(struct drm_i915_private *dev_priv,
> +int i915_gem_stolen_insert_node(struct drm_i915_private *i915,
> struct drm_mm_node *node, u64 size,
> unsigned alignment);
> -int i915_gem_stolen_insert_node_in_range(struct drm_i915_private *dev_priv,
> +int i915_gem_stolen_insert_node_in_range(struct drm_i915_private *i915,
> struct drm_mm_node *node, u64 size,
> unsigned alignment, u64 start,
> u64 end);
> -void i915_gem_stolen_remove_node(struct drm_i915_private *dev_priv,
> +void i915_gem_stolen_remove_node(struct drm_i915_private *i915,
> struct drm_mm_node *node);
> -int i915_gem_init_stolen(struct drm_i915_private *dev_priv);
> -void i915_gem_cleanup_stolen(struct drm_i915_private *dev_priv);
> +int i915_gem_init_stolen(struct drm_i915_private *i915);
> +void i915_gem_cleanup_stolen(struct drm_i915_private *i915);
> struct drm_i915_gem_object *
> -i915_gem_object_create_stolen(struct drm_i915_private *dev_priv,
> +i915_gem_object_create_stolen(struct drm_i915_private *i915,
> resource_size_t size);
> struct drm_i915_gem_object *
> -i915_gem_object_create_stolen_for_preallocated(struct drm_i915_private *dev_priv,
> +i915_gem_object_create_stolen_for_preallocated(struct drm_i915_private *i915,
> resource_size_t stolen_offset,
> resource_size_t gtt_offset,
> resource_size_t size);
>
> /* i915_gem_internal.c */
> struct drm_i915_gem_object *
> -i915_gem_object_create_internal(struct drm_i915_private *dev_priv,
> +i915_gem_object_create_internal(struct drm_i915_private *i915,
> phys_addr_t size);
>
> /* i915_gem_shrinker.c */
> @@ -2771,21 +2771,21 @@ void i915_gem_shrinker_taints_mutex(struct drm_i915_private *i915,
> /* i915_gem_tiling.c */
> static inline bool i915_gem_object_needs_bit17_swizzle(struct drm_i915_gem_object *obj)
> {
> - struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
> + struct drm_i915_private *i915 = to_i915(obj->base.dev);
>
> - return dev_priv->mm.bit_6_swizzle_x == I915_BIT_6_SWIZZLE_9_10_17 &&
> + return i915->mm.bit_6_swizzle_x == I915_BIT_6_SWIZZLE_9_10_17 &&
> i915_gem_object_is_tiled(obj);
> }
>
> -u32 i915_gem_fence_size(struct drm_i915_private *dev_priv, u32 size,
> +u32 i915_gem_fence_size(struct drm_i915_private *i915, u32 size,
> unsigned int tiling, unsigned int stride);
> -u32 i915_gem_fence_alignment(struct drm_i915_private *dev_priv, u32 size,
> +u32 i915_gem_fence_alignment(struct drm_i915_private *i915, u32 size,
> unsigned int tiling, unsigned int stride);
>
> const char *i915_cache_level_str(struct drm_i915_private *i915, int type);
>
> /* i915_cmd_parser.c */
> -int i915_cmd_parser_get_version(struct drm_i915_private *dev_priv);
> +int i915_cmd_parser_get_version(struct drm_i915_private *i915);
> void intel_engine_init_cmd_parser(struct intel_engine_cs *engine);
> void intel_engine_cleanup_cmd_parser(struct intel_engine_cs *engine);
> int intel_engine_cmd_parser(struct intel_engine_cs *engine,
> @@ -2796,52 +2796,52 @@ int intel_engine_cmd_parser(struct intel_engine_cs *engine,
> bool is_master);
>
> /* i915_perf.c */
> -extern void i915_perf_init(struct drm_i915_private *dev_priv);
> -extern void i915_perf_fini(struct drm_i915_private *dev_priv);
> -extern void i915_perf_register(struct drm_i915_private *dev_priv);
> -extern void i915_perf_unregister(struct drm_i915_private *dev_priv);
> +extern void i915_perf_init(struct drm_i915_private *i915);
> +extern void i915_perf_fini(struct drm_i915_private *i915);
> +extern void i915_perf_register(struct drm_i915_private *i915);
> +extern void i915_perf_unregister(struct drm_i915_private *i915);
>
> /* i915_suspend.c */
> -extern int i915_save_state(struct drm_i915_private *dev_priv);
> -extern int i915_restore_state(struct drm_i915_private *dev_priv);
> +extern int i915_save_state(struct drm_i915_private *i915);
> +extern int i915_restore_state(struct drm_i915_private *i915);
>
> /* i915_sysfs.c */
> -void i915_setup_sysfs(struct drm_i915_private *dev_priv);
> -void i915_teardown_sysfs(struct drm_i915_private *dev_priv);
> +void i915_setup_sysfs(struct drm_i915_private *i915);
> +void i915_teardown_sysfs(struct drm_i915_private *i915);
>
> /* intel_device_info.c */
> static inline struct intel_device_info *
> -mkwrite_device_info(struct drm_i915_private *dev_priv)
> +mkwrite_device_info(struct drm_i915_private *i915)
> {
> - return (struct intel_device_info *)INTEL_INFO(dev_priv);
> + return (struct intel_device_info *)INTEL_INFO(i915);
> }
>
> /* modesetting */
> extern void intel_modeset_init_hw(struct drm_device *dev);
> extern int intel_modeset_init(struct drm_device *dev);
> extern void intel_modeset_cleanup(struct drm_device *dev);
> -extern int intel_modeset_vga_set_state(struct drm_i915_private *dev_priv,
> +extern int intel_modeset_vga_set_state(struct drm_i915_private *i915,
> bool state);
> extern void intel_display_resume(struct drm_device *dev);
> -extern void i915_redisable_vga(struct drm_i915_private *dev_priv);
> -extern void i915_redisable_vga_power_on(struct drm_i915_private *dev_priv);
> -extern void intel_init_pch_refclk(struct drm_i915_private *dev_priv);
> +extern void i915_redisable_vga(struct drm_i915_private *i915);
> +extern void i915_redisable_vga_power_on(struct drm_i915_private *i915);
> +extern void intel_init_pch_refclk(struct drm_i915_private *i915);
>
> int i915_reg_read_ioctl(struct drm_device *dev, void *data,
> struct drm_file *file);
>
> extern struct intel_display_error_state *
> -intel_display_capture_error_state(struct drm_i915_private *dev_priv);
> +intel_display_capture_error_state(struct drm_i915_private *i915);
> extern void intel_display_print_error_state(struct drm_i915_error_state_buf *e,
> struct intel_display_error_state *error);
>
> #define __I915_REG_OP(op__, dev_priv__, ...) \
> intel_uncore_##op__(&(dev_priv__)->uncore, __VA_ARGS__)
>
> -#define I915_READ(reg__) __I915_REG_OP(read, dev_priv, (reg__))
> -#define I915_WRITE(reg__, val__) __I915_REG_OP(write, dev_priv, (reg__), (val__))
> +#define I915_READ(reg__) __I915_REG_OP(read, i915, (reg__))
> +#define I915_WRITE(reg__, val__) __I915_REG_OP(write, i915, (reg__), (val__))
>
> -#define POSTING_READ(reg__) __I915_REG_OP(posting_read, dev_priv, (reg__))
> +#define POSTING_READ(reg__) __I915_REG_OP(posting_read, i915, (reg__))
>
> /* These are untraced mmio-accessors that are only valid to be used inside
> * critical sections, such as inside IRQ handlers, where forcewake is explicitly
> @@ -2851,13 +2851,13 @@ extern void intel_display_print_error_state(struct drm_i915_error_state_buf *e,
> *
> * As an example, these accessors can possibly be used between:
> *
> - * spin_lock_irq(&dev_priv->uncore.lock);
> + * spin_lock_irq(&i915->uncore.lock);
> * intel_uncore_forcewake_get__locked();
> *
> * and
> *
> * intel_uncore_forcewake_put__locked();
> - * spin_unlock_irq(&dev_priv->uncore.lock);
> + * spin_unlock_irq(&i915->uncore.lock);
> *
> *
> * Note: some registers may not need forcewake held, so
> @@ -2866,18 +2866,18 @@ extern void intel_display_print_error_state(struct drm_i915_error_state_buf *e,
> *
> * Certain architectures will die if the same cacheline is concurrently accessed
> * by different clients (e.g. on Ivybridge). Access to registers should
> - * therefore generally be serialised, by either the dev_priv->uncore.lock or
> + * therefore generally be serialised, by either the i915->uncore.lock or
> * a more localised lock guarding all access to that bank of registers.
> */
> -#define I915_READ_FW(reg__) __I915_REG_OP(read_fw, dev_priv, (reg__))
> -#define I915_WRITE_FW(reg__, val__) __I915_REG_OP(write_fw, dev_priv, (reg__), (val__))
> +#define I915_READ_FW(reg__) __I915_REG_OP(read_fw, i915, (reg__))
> +#define I915_WRITE_FW(reg__, val__) __I915_REG_OP(write_fw, i915, (reg__), (val__))
>
> /* "Broadcast RGB" property */
> #define INTEL_BROADCAST_RGB_AUTO 0
> #define INTEL_BROADCAST_RGB_FULL 1
> #define INTEL_BROADCAST_RGB_LIMITED 2
>
> -void i915_memcpy_init_early(struct drm_i915_private *dev_priv);
> +void i915_memcpy_init_early(struct drm_i915_private *i915);
> bool i915_memcpy_from_wc(void *dst, const void *src, unsigned long len);
>
> /* The movntdqa instructions used for memcpy-from-wc require 16-byte alignment,
> diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c
> index 4017ecf561f6..1774c7604076 100644
> --- a/drivers/gpu/drm/i915/i915_gem.c
> +++ b/drivers/gpu/drm/i915/i915_gem.c
> @@ -149,7 +149,7 @@ i915_gem_phys_pwrite(struct drm_i915_gem_object *obj,
>
> static int
> i915_gem_create(struct drm_file *file,
> - struct drm_i915_private *dev_priv,
> + struct drm_i915_private *i915,
> u64 *size_p,
> u32 *handle_p)
> {
> @@ -163,7 +163,7 @@ i915_gem_create(struct drm_file *file,
> return -EINVAL;
>
> /* Allocate the new object */
> - obj = i915_gem_object_create_shmem(dev_priv, size);
> + obj = i915_gem_object_create_shmem(i915, size);
> if (IS_ERR(obj))
> return PTR_ERR(obj);
>
> @@ -223,16 +223,16 @@ int
> i915_gem_create_ioctl(struct drm_device *dev, void *data,
> struct drm_file *file)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct drm_i915_gem_create *args = data;
>
> - i915_gem_flush_free_objects(dev_priv);
> + i915_gem_flush_free_objects(i915);
>
> - return i915_gem_create(file, dev_priv,
> + return i915_gem_create(file, i915,
> &args->size, &args->handle);
> }
>
> -void i915_gem_flush_ggtt_writes(struct drm_i915_private *dev_priv)
> +void i915_gem_flush_ggtt_writes(struct drm_i915_private *i915)
> {
> intel_wakeref_t wakeref;
>
> @@ -257,13 +257,13 @@ void i915_gem_flush_ggtt_writes(struct drm_i915_private *dev_priv)
>
> wmb();
>
> - if (INTEL_INFO(dev_priv)->has_coherent_ggtt)
> + if (INTEL_INFO(i915)->has_coherent_ggtt)
> return;
>
> - i915_gem_chipset_flush(dev_priv);
> + i915_gem_chipset_flush(i915);
>
> - with_intel_runtime_pm(dev_priv, wakeref) {
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + with_intel_runtime_pm(i915, wakeref) {
> + struct intel_uncore *uncore = &i915->uncore;
>
> spin_lock_irq(&uncore->lock);
> intel_uncore_posting_read_fw(uncore,
> @@ -884,7 +884,7 @@ i915_gem_sw_finish_ioctl(struct drm_device *dev, void *data,
> return 0;
> }
>
> -void i915_gem_runtime_suspend(struct drm_i915_private *dev_priv)
> +void i915_gem_runtime_suspend(struct drm_i915_private *i915)
> {
> struct drm_i915_gem_object *obj, *on;
> int i;
> @@ -897,15 +897,15 @@ void i915_gem_runtime_suspend(struct drm_i915_private *dev_priv)
> */
>
> list_for_each_entry_safe(obj, on,
> - &dev_priv->mm.userfault_list, userfault_link)
> + &i915->mm.userfault_list, userfault_link)
> __i915_gem_object_release_mmap(obj);
>
> /* The fence will be lost when the device powers down. If any were
> * in use by hardware (i.e. they are pinned), we should not be powering
> * down! All other fences will be reacquired by the user upon waking.
> */
> - for (i = 0; i < dev_priv->num_fence_regs; i++) {
> - struct drm_i915_fence_reg *reg = &dev_priv->fence_regs[i];
> + for (i = 0; i < i915->num_fence_regs; i++) {
> + struct drm_i915_fence_reg *reg = &i915->fence_regs[i];
>
> /* Ideally we want to assert that the fence register is not
> * live at this point (i.e. that no piece of code will be
> @@ -1020,8 +1020,8 @@ i915_gem_object_ggtt_pin(struct drm_i915_gem_object *obj,
> u64 alignment,
> u64 flags)
> {
> - struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
> - struct i915_address_space *vm = &dev_priv->ggtt.vm;
> + struct drm_i915_private *i915 = to_i915(obj->base.dev);
> + struct i915_address_space *vm = &i915->ggtt.vm;
> struct i915_vma *vma;
> int ret;
>
> @@ -1036,7 +1036,7 @@ i915_gem_object_ggtt_pin(struct drm_i915_gem_object *obj,
> * the object in and out of the Global GTT and
> * waste a lot of cycles under the mutex.
> */
> - if (obj->base.size > dev_priv->ggtt.mappable_end)
> + if (obj->base.size > i915->ggtt.mappable_end)
> return ERR_PTR(-E2BIG);
>
> /* If NONBLOCK is set the caller is optimistically
> @@ -1055,7 +1055,7 @@ i915_gem_object_ggtt_pin(struct drm_i915_gem_object *obj,
> * we could try to minimise harm to others.
> */
> if (flags & PIN_NONBLOCK &&
> - obj->base.size > dev_priv->ggtt.mappable_end / 2)
> + obj->base.size > i915->ggtt.mappable_end / 2)
> return ERR_PTR(-ENOSPC);
> }
>
> @@ -1069,7 +1069,7 @@ i915_gem_object_ggtt_pin(struct drm_i915_gem_object *obj,
> return ERR_PTR(-ENOSPC);
>
> if (flags & PIN_MAPPABLE &&
> - vma->fence_size > dev_priv->ggtt.mappable_end / 2)
> + vma->fence_size > i915->ggtt.mappable_end / 2)
> return ERR_PTR(-ENOSPC);
> }
>
> @@ -1202,30 +1202,30 @@ void i915_gem_sanitize(struct drm_i915_private *i915)
> mutex_unlock(&i915->drm.struct_mutex);
> }
>
> -void i915_gem_init_swizzling(struct drm_i915_private *dev_priv)
> +void i915_gem_init_swizzling(struct drm_i915_private *i915)
> {
> - if (INTEL_GEN(dev_priv) < 5 ||
> - dev_priv->mm.bit_6_swizzle_x == I915_BIT_6_SWIZZLE_NONE)
> + if (INTEL_GEN(i915) < 5 ||
> + i915->mm.bit_6_swizzle_x == I915_BIT_6_SWIZZLE_NONE)
> return;
>
> I915_WRITE(DISP_ARB_CTL, I915_READ(DISP_ARB_CTL) |
> DISP_TILE_SURFACE_SWIZZLING);
>
> - if (IS_GEN(dev_priv, 5))
> + if (IS_GEN(i915, 5))
> return;
>
> I915_WRITE(TILECTL, I915_READ(TILECTL) | TILECTL_SWZCTL);
> - if (IS_GEN(dev_priv, 6))
> + if (IS_GEN(i915, 6))
> I915_WRITE(ARB_MODE, _MASKED_BIT_ENABLE(ARB_MODE_SWIZZLE_SNB));
> - else if (IS_GEN(dev_priv, 7))
> + else if (IS_GEN(i915, 7))
> I915_WRITE(ARB_MODE, _MASKED_BIT_ENABLE(ARB_MODE_SWIZZLE_IVB));
> - else if (IS_GEN(dev_priv, 8))
> + else if (IS_GEN(i915, 8))
> I915_WRITE(GAMTARBMODE, _MASKED_BIT_ENABLE(ARB_MODE_SWIZZLE_BDW));
> else
> BUG();
> }
>
> -static void init_unused_ring(struct drm_i915_private *dev_priv, u32 base)
> +static void init_unused_ring(struct drm_i915_private *i915, u32 base)
> {
> I915_WRITE(RING_CTL(base), 0);
> I915_WRITE(RING_HEAD(base), 0);
> @@ -1233,45 +1233,45 @@ static void init_unused_ring(struct drm_i915_private *dev_priv, u32 base)
> I915_WRITE(RING_START(base), 0);
> }
>
> -static void init_unused_rings(struct drm_i915_private *dev_priv)
> +static void init_unused_rings(struct drm_i915_private *i915)
> {
> - if (IS_I830(dev_priv)) {
> - init_unused_ring(dev_priv, PRB1_BASE);
> - init_unused_ring(dev_priv, SRB0_BASE);
> - init_unused_ring(dev_priv, SRB1_BASE);
> - init_unused_ring(dev_priv, SRB2_BASE);
> - init_unused_ring(dev_priv, SRB3_BASE);
> - } else if (IS_GEN(dev_priv, 2)) {
> - init_unused_ring(dev_priv, SRB0_BASE);
> - init_unused_ring(dev_priv, SRB1_BASE);
> - } else if (IS_GEN(dev_priv, 3)) {
> - init_unused_ring(dev_priv, PRB1_BASE);
> - init_unused_ring(dev_priv, PRB2_BASE);
> + if (IS_I830(i915)) {
> + init_unused_ring(i915, PRB1_BASE);
> + init_unused_ring(i915, SRB0_BASE);
> + init_unused_ring(i915, SRB1_BASE);
> + init_unused_ring(i915, SRB2_BASE);
> + init_unused_ring(i915, SRB3_BASE);
> + } else if (IS_GEN(i915, 2)) {
> + init_unused_ring(i915, SRB0_BASE);
> + init_unused_ring(i915, SRB1_BASE);
> + } else if (IS_GEN(i915, 3)) {
> + init_unused_ring(i915, PRB1_BASE);
> + init_unused_ring(i915, PRB2_BASE);
> }
> }
>
> -int i915_gem_init_hw(struct drm_i915_private *dev_priv)
> +int i915_gem_init_hw(struct drm_i915_private *i915)
> {
> int ret;
>
> - dev_priv->gt.last_init_time = ktime_get();
> + i915->gt.last_init_time = ktime_get();
>
> /* Double layer security blanket, see i915_gem_init() */
> - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
>
> - if (HAS_EDRAM(dev_priv) && INTEL_GEN(dev_priv) < 9)
> + if (HAS_EDRAM(i915) && INTEL_GEN(i915) < 9)
> I915_WRITE(HSW_IDICR, I915_READ(HSW_IDICR) | IDIHASHMSK(0xf));
>
> - if (IS_HASWELL(dev_priv))
> - I915_WRITE(MI_PREDICATE_RESULT_2, IS_HSW_GT3(dev_priv) ?
> + if (IS_HASWELL(i915))
> + I915_WRITE(MI_PREDICATE_RESULT_2, IS_HSW_GT3(i915) ?
> LOWER_SLICE_ENABLED : LOWER_SLICE_DISABLED);
>
> /* Apply the GT workarounds... */
> - intel_gt_apply_workarounds(dev_priv);
> + intel_gt_apply_workarounds(i915);
> /* ...and determine whether they are sticking. */
> - intel_gt_verify_workarounds(dev_priv, "init");
> + intel_gt_verify_workarounds(i915, "init");
>
> - i915_gem_init_swizzling(dev_priv);
> + i915_gem_init_swizzling(i915);
>
> /*
> * At least 830 can leave some of the unused rings
> @@ -1279,48 +1279,48 @@ int i915_gem_init_hw(struct drm_i915_private *dev_priv)
> * will prevent c3 entry. Makes sure all unused rings
> * are totally idle.
> */
> - init_unused_rings(dev_priv);
> + init_unused_rings(i915);
>
> - BUG_ON(!dev_priv->kernel_context);
> - ret = i915_terminally_wedged(dev_priv);
> + BUG_ON(!i915->kernel_context);
> + ret = i915_terminally_wedged(i915);
> if (ret)
> goto out;
>
> - ret = i915_ppgtt_init_hw(dev_priv);
> + ret = i915_ppgtt_init_hw(i915);
> if (ret) {
> DRM_ERROR("Enabling PPGTT failed (%d)\n", ret);
> goto out;
> }
>
> - ret = intel_wopcm_init_hw(&dev_priv->wopcm);
> + ret = intel_wopcm_init_hw(&i915->wopcm);
> if (ret) {
> DRM_ERROR("Enabling WOPCM failed (%d)\n", ret);
> goto out;
> }
>
> /* We can't enable contexts until all firmware is loaded */
> - ret = intel_uc_init_hw(dev_priv);
> + ret = intel_uc_init_hw(i915);
> if (ret) {
> DRM_ERROR("Enabling uc failed (%d)\n", ret);
> goto out;
> }
>
> - intel_mocs_init_l3cc_table(dev_priv);
> + intel_mocs_init_l3cc_table(i915);
>
> /* Only when the HW is re-initialised, can we replay the requests */
> - ret = intel_engines_resume(dev_priv);
> + ret = intel_engines_resume(i915);
> if (ret)
> goto cleanup_uc;
>
> - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
>
> - intel_engines_set_scheduler_caps(dev_priv);
> + intel_engines_set_scheduler_caps(i915);
> return 0;
>
> cleanup_uc:
> - intel_uc_fini_hw(dev_priv);
> + intel_uc_fini_hw(i915);
> out:
> - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
>
> return ret;
> }
> @@ -1505,28 +1505,28 @@ static int intel_engines_verify_workarounds(struct drm_i915_private *i915)
> return err;
> }
>
> -int i915_gem_init(struct drm_i915_private *dev_priv)
> +int i915_gem_init(struct drm_i915_private *i915)
> {
> int ret;
>
> /* We need to fallback to 4K pages if host doesn't support huge gtt. */
> - if (intel_vgpu_active(dev_priv) && !intel_vgpu_has_huge_gtt(dev_priv))
> - mkwrite_device_info(dev_priv)->page_sizes =
> + if (intel_vgpu_active(i915) && !intel_vgpu_has_huge_gtt(i915))
> + mkwrite_device_info(i915)->page_sizes =
> I915_GTT_PAGE_SIZE_4K;
>
> - dev_priv->mm.unordered_timeline = dma_fence_context_alloc(1);
> + i915->mm.unordered_timeline = dma_fence_context_alloc(1);
>
> - i915_timelines_init(dev_priv);
> + i915_timelines_init(i915);
>
> - ret = i915_gem_init_userptr(dev_priv);
> + ret = i915_gem_init_userptr(i915);
> if (ret)
> return ret;
>
> - ret = intel_uc_init_misc(dev_priv);
> + ret = intel_uc_init_misc(i915);
> if (ret)
> return ret;
>
> - ret = intel_wopcm_init(&dev_priv->wopcm);
> + ret = intel_wopcm_init(&i915->wopcm);
> if (ret)
> goto err_uc_misc;
>
> @@ -1536,47 +1536,47 @@ int i915_gem_init(struct drm_i915_private *dev_priv)
> * we hold the forcewake during initialisation these problems
> * just magically go away.
> */
> - mutex_lock(&dev_priv->drm.struct_mutex);
> - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> + mutex_lock(&i915->drm.struct_mutex);
> + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
>
> - ret = i915_gem_init_ggtt(dev_priv);
> + ret = i915_gem_init_ggtt(i915);
> if (ret) {
> GEM_BUG_ON(ret == -EIO);
> goto err_unlock;
> }
>
> - ret = i915_gem_init_scratch(dev_priv,
> - IS_GEN(dev_priv, 2) ? SZ_256K : PAGE_SIZE);
> + ret = i915_gem_init_scratch(i915,
> + IS_GEN(i915, 2) ? SZ_256K : PAGE_SIZE);
> if (ret) {
> GEM_BUG_ON(ret == -EIO);
> goto err_ggtt;
> }
>
> - ret = intel_engines_setup(dev_priv);
> + ret = intel_engines_setup(i915);
> if (ret) {
> GEM_BUG_ON(ret == -EIO);
> goto err_unlock;
> }
>
> - ret = i915_gem_contexts_init(dev_priv);
> + ret = i915_gem_contexts_init(i915);
> if (ret) {
> GEM_BUG_ON(ret == -EIO);
> goto err_scratch;
> }
>
> - ret = intel_engines_init(dev_priv);
> + ret = intel_engines_init(i915);
> if (ret) {
> GEM_BUG_ON(ret == -EIO);
> goto err_context;
> }
>
> - intel_init_gt_powersave(dev_priv);
> + intel_init_gt_powersave(i915);
>
> - ret = intel_uc_init(dev_priv);
> + ret = intel_uc_init(i915);
> if (ret)
> goto err_pm;
>
> - ret = i915_gem_init_hw(dev_priv);
> + ret = i915_gem_init_hw(i915);
> if (ret)
> goto err_uc_init;
>
> @@ -1589,13 +1589,13 @@ int i915_gem_init(struct drm_i915_private *dev_priv)
> *
> * FIXME: break up the workarounds and apply them at the right time!
> */
> - intel_init_clock_gating(dev_priv);
> + intel_init_clock_gating(i915);
>
> - ret = intel_engines_verify_workarounds(dev_priv);
> + ret = intel_engines_verify_workarounds(i915);
> if (ret)
> goto err_init_hw;
>
> - ret = __intel_engines_record_defaults(dev_priv);
> + ret = __intel_engines_record_defaults(i915);
> if (ret)
> goto err_init_hw;
>
> @@ -1609,8 +1609,8 @@ int i915_gem_init(struct drm_i915_private *dev_priv)
> goto err_init_hw;
> }
>
> - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> + mutex_unlock(&i915->drm.struct_mutex);
>
> return 0;
>
> @@ -1621,107 +1621,107 @@ int i915_gem_init(struct drm_i915_private *dev_priv)
> * driver doesn't explode during runtime.
> */
> err_init_hw:
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + mutex_unlock(&i915->drm.struct_mutex);
>
> - i915_gem_set_wedged(dev_priv);
> - i915_gem_suspend(dev_priv);
> - i915_gem_suspend_late(dev_priv);
> + i915_gem_set_wedged(i915);
> + i915_gem_suspend(i915);
> + i915_gem_suspend_late(i915);
>
> - i915_gem_drain_workqueue(dev_priv);
> + i915_gem_drain_workqueue(i915);
>
> - mutex_lock(&dev_priv->drm.struct_mutex);
> - intel_uc_fini_hw(dev_priv);
> + mutex_lock(&i915->drm.struct_mutex);
> + intel_uc_fini_hw(i915);
> err_uc_init:
> - intel_uc_fini(dev_priv);
> + intel_uc_fini(i915);
> err_pm:
> if (ret != -EIO) {
> - intel_cleanup_gt_powersave(dev_priv);
> - intel_engines_cleanup(dev_priv);
> + intel_cleanup_gt_powersave(i915);
> + intel_engines_cleanup(i915);
> }
> err_context:
> if (ret != -EIO)
> - i915_gem_contexts_fini(dev_priv);
> + i915_gem_contexts_fini(i915);
> err_scratch:
> - i915_gem_fini_scratch(dev_priv);
> + i915_gem_fini_scratch(i915);
> err_ggtt:
> err_unlock:
> - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> + mutex_unlock(&i915->drm.struct_mutex);
>
> err_uc_misc:
> - intel_uc_fini_misc(dev_priv);
> + intel_uc_fini_misc(i915);
>
> if (ret != -EIO) {
> - i915_gem_cleanup_userptr(dev_priv);
> - i915_timelines_fini(dev_priv);
> + i915_gem_cleanup_userptr(i915);
> + i915_timelines_fini(i915);
> }
>
> if (ret == -EIO) {
> - mutex_lock(&dev_priv->drm.struct_mutex);
> + mutex_lock(&i915->drm.struct_mutex);
>
> /*
> * Allow engine initialisation to fail by marking the GPU as
> * wedged. But we only want to do this where the GPU is angry,
> * for all other failure, such as an allocation failure, bail.
> */
> - if (!i915_reset_failed(dev_priv)) {
> - i915_load_error(dev_priv,
> + if (!i915_reset_failed(i915)) {
> + i915_load_error(i915,
> "Failed to initialize GPU, declaring it wedged!\n");
> - i915_gem_set_wedged(dev_priv);
> + i915_gem_set_wedged(i915);
> }
>
> /* Minimal basic recovery for KMS */
> - ret = i915_ggtt_enable_hw(dev_priv);
> - i915_gem_restore_gtt_mappings(dev_priv);
> - i915_gem_restore_fences(dev_priv);
> - intel_init_clock_gating(dev_priv);
> + ret = i915_ggtt_enable_hw(i915);
> + i915_gem_restore_gtt_mappings(i915);
> + i915_gem_restore_fences(i915);
> + intel_init_clock_gating(i915);
>
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + mutex_unlock(&i915->drm.struct_mutex);
> }
>
> - i915_gem_drain_freed_objects(dev_priv);
> + i915_gem_drain_freed_objects(i915);
> return ret;
> }
>
> -void i915_gem_fini_hw(struct drm_i915_private *dev_priv)
> +void i915_gem_fini_hw(struct drm_i915_private *i915)
> {
> - GEM_BUG_ON(dev_priv->gt.awake);
> + GEM_BUG_ON(i915->gt.awake);
>
> - intel_wakeref_auto_fini(&dev_priv->mm.userfault_wakeref);
> + intel_wakeref_auto_fini(&i915->mm.userfault_wakeref);
>
> - i915_gem_suspend_late(dev_priv);
> - intel_disable_gt_powersave(dev_priv);
> + i915_gem_suspend_late(i915);
> + intel_disable_gt_powersave(i915);
>
> /* Flush any outstanding unpin_work. */
> - i915_gem_drain_workqueue(dev_priv);
> + i915_gem_drain_workqueue(i915);
>
> - mutex_lock(&dev_priv->drm.struct_mutex);
> - intel_uc_fini_hw(dev_priv);
> - intel_uc_fini(dev_priv);
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + mutex_lock(&i915->drm.struct_mutex);
> + intel_uc_fini_hw(i915);
> + intel_uc_fini(i915);
> + mutex_unlock(&i915->drm.struct_mutex);
>
> - i915_gem_drain_freed_objects(dev_priv);
> + i915_gem_drain_freed_objects(i915);
> }
>
> -void i915_gem_fini(struct drm_i915_private *dev_priv)
> +void i915_gem_fini(struct drm_i915_private *i915)
> {
> - mutex_lock(&dev_priv->drm.struct_mutex);
> - intel_engines_cleanup(dev_priv);
> - i915_gem_contexts_fini(dev_priv);
> - i915_gem_fini_scratch(dev_priv);
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + mutex_lock(&i915->drm.struct_mutex);
> + intel_engines_cleanup(i915);
> + i915_gem_contexts_fini(i915);
> + i915_gem_fini_scratch(i915);
> + mutex_unlock(&i915->drm.struct_mutex);
>
> - intel_wa_list_free(&dev_priv->gt_wa_list);
> + intel_wa_list_free(&i915->gt_wa_list);
>
> - intel_cleanup_gt_powersave(dev_priv);
> + intel_cleanup_gt_powersave(i915);
>
> - intel_uc_fini_misc(dev_priv);
> - i915_gem_cleanup_userptr(dev_priv);
> - i915_timelines_fini(dev_priv);
> + intel_uc_fini_misc(i915);
> + i915_gem_cleanup_userptr(i915);
> + i915_timelines_fini(i915);
>
> - i915_gem_drain_freed_objects(dev_priv);
> + i915_gem_drain_freed_objects(i915);
>
> - WARN_ON(!list_empty(&dev_priv->contexts.list));
> + WARN_ON(!list_empty(&i915->contexts.list));
> }
>
> void i915_gem_init_mmio(struct drm_i915_private *i915)
> @@ -1730,35 +1730,35 @@ void i915_gem_init_mmio(struct drm_i915_private *i915)
> }
>
> void
> -i915_gem_load_init_fences(struct drm_i915_private *dev_priv)
> +i915_gem_load_init_fences(struct drm_i915_private *i915)
> {
> int i;
>
> - if (INTEL_GEN(dev_priv) >= 7 && !IS_VALLEYVIEW(dev_priv) &&
> - !IS_CHERRYVIEW(dev_priv))
> - dev_priv->num_fence_regs = 32;
> - else if (INTEL_GEN(dev_priv) >= 4 ||
> - IS_I945G(dev_priv) || IS_I945GM(dev_priv) ||
> - IS_G33(dev_priv) || IS_PINEVIEW(dev_priv))
> - dev_priv->num_fence_regs = 16;
> + if (INTEL_GEN(i915) >= 7 && !IS_VALLEYVIEW(i915) &&
> + !IS_CHERRYVIEW(i915))
> + i915->num_fence_regs = 32;
> + else if (INTEL_GEN(i915) >= 4 ||
> + IS_I945G(i915) || IS_I945GM(i915) ||
> + IS_G33(i915) || IS_PINEVIEW(i915))
> + i915->num_fence_regs = 16;
> else
> - dev_priv->num_fence_regs = 8;
> + i915->num_fence_regs = 8;
>
> - if (intel_vgpu_active(dev_priv))
> - dev_priv->num_fence_regs =
> + if (intel_vgpu_active(i915))
> + i915->num_fence_regs =
> I915_READ(vgtif_reg(avail_rs.fence_num));
>
> /* Initialize fence registers to zero */
> - for (i = 0; i < dev_priv->num_fence_regs; i++) {
> - struct drm_i915_fence_reg *fence = &dev_priv->fence_regs[i];
> + for (i = 0; i < i915->num_fence_regs; i++) {
> + struct drm_i915_fence_reg *fence = &i915->fence_regs[i];
>
> - fence->i915 = dev_priv;
> + fence->i915 = i915;
> fence->id = i;
> - list_add_tail(&fence->link, &dev_priv->mm.fence_list);
> + list_add_tail(&fence->link, &i915->mm.fence_list);
> }
> - i915_gem_restore_fences(dev_priv);
> + i915_gem_restore_fences(i915);
>
> - i915_gem_detect_bit_6_swizzle(dev_priv);
> + i915_gem_detect_bit_6_swizzle(i915);
> }
>
> static void i915_gem_init__mm(struct drm_i915_private *i915)
> @@ -1778,56 +1778,56 @@ static void i915_gem_init__mm(struct drm_i915_private *i915)
> i915_gem_init__objects(i915);
> }
>
> -int i915_gem_init_early(struct drm_i915_private *dev_priv)
> +int i915_gem_init_early(struct drm_i915_private *i915)
> {
> static struct lock_class_key reset_key;
> int err;
>
> - intel_gt_pm_init(dev_priv);
> + intel_gt_pm_init(i915);
>
> - INIT_LIST_HEAD(&dev_priv->gt.active_rings);
> - INIT_LIST_HEAD(&dev_priv->gt.closed_vma);
> - spin_lock_init(&dev_priv->gt.closed_lock);
> - lockdep_init_map(&dev_priv->gt.reset_lockmap,
> + INIT_LIST_HEAD(&i915->gt.active_rings);
> + INIT_LIST_HEAD(&i915->gt.closed_vma);
> + spin_lock_init(&i915->gt.closed_lock);
> + lockdep_init_map(&i915->gt.reset_lockmap,
> "i915.reset", &reset_key, 0);
>
> - i915_gem_init__mm(dev_priv);
> - i915_gem_init__pm(dev_priv);
> + i915_gem_init__mm(i915);
> + i915_gem_init__pm(i915);
>
> - init_waitqueue_head(&dev_priv->gpu_error.wait_queue);
> - init_waitqueue_head(&dev_priv->gpu_error.reset_queue);
> - mutex_init(&dev_priv->gpu_error.wedge_mutex);
> - init_srcu_struct(&dev_priv->gpu_error.reset_backoff_srcu);
> + init_waitqueue_head(&i915->gpu_error.wait_queue);
> + init_waitqueue_head(&i915->gpu_error.reset_queue);
> + mutex_init(&i915->gpu_error.wedge_mutex);
> + init_srcu_struct(&i915->gpu_error.reset_backoff_srcu);
>
> - atomic_set(&dev_priv->mm.bsd_engine_dispatch_index, 0);
> + atomic_set(&i915->mm.bsd_engine_dispatch_index, 0);
>
> - spin_lock_init(&dev_priv->fb_tracking.lock);
> + spin_lock_init(&i915->fb_tracking.lock);
>
> - err = i915_gemfs_init(dev_priv);
> + err = i915_gemfs_init(i915);
> if (err)
> DRM_NOTE("Unable to create a private tmpfs mount, hugepage support will be disabled(%d).\n", err);
>
> return 0;
> }
>
> -void i915_gem_cleanup_early(struct drm_i915_private *dev_priv)
> +void i915_gem_cleanup_early(struct drm_i915_private *i915)
> {
> - i915_gem_drain_freed_objects(dev_priv);
> - GEM_BUG_ON(!llist_empty(&dev_priv->mm.free_list));
> - GEM_BUG_ON(atomic_read(&dev_priv->mm.free_count));
> - WARN_ON(dev_priv->mm.shrink_count);
> + i915_gem_drain_freed_objects(i915);
> + GEM_BUG_ON(!llist_empty(&i915->mm.free_list));
> + GEM_BUG_ON(atomic_read(&i915->mm.free_count));
> + WARN_ON(i915->mm.shrink_count);
>
> - cleanup_srcu_struct(&dev_priv->gpu_error.reset_backoff_srcu);
> + cleanup_srcu_struct(&i915->gpu_error.reset_backoff_srcu);
>
> - i915_gemfs_fini(dev_priv);
> + i915_gemfs_fini(i915);
> }
>
> -int i915_gem_freeze(struct drm_i915_private *dev_priv)
> +int i915_gem_freeze(struct drm_i915_private *i915)
> {
> /* Discard all purgeable objects, let userspace recover those as
> * required after resuming.
> */
> - i915_gem_shrink_all(dev_priv);
> + i915_gem_shrink_all(i915);
>
> return 0;
> }
> @@ -1895,7 +1895,7 @@ int i915_gem_open(struct drm_i915_private *i915, struct drm_file *file)
> return -ENOMEM;
>
> file->driver_priv = file_priv;
> - file_priv->dev_priv = i915;
> + file_priv->i915 = i915;
> file_priv->file = file;
>
> spin_lock_init(&file_priv->mm.lock);
> diff --git a/drivers/gpu/drm/i915/i915_gem_evict.c b/drivers/gpu/drm/i915/i915_gem_evict.c
> index a5783c4cb98b..4244d9c54d66 100644
> --- a/drivers/gpu/drm/i915/i915_gem_evict.c
> +++ b/drivers/gpu/drm/i915/i915_gem_evict.c
> @@ -99,7 +99,7 @@ i915_gem_evict_something(struct i915_address_space *vm,
> u64 start, u64 end,
> unsigned flags)
> {
> - struct drm_i915_private *dev_priv = vm->i915;
> + struct drm_i915_private *i915 = vm->i915;
> struct drm_mm_scan scan;
> struct list_head eviction_list;
> struct i915_vma *vma, *next;
> @@ -138,7 +138,7 @@ i915_gem_evict_something(struct i915_address_space *vm,
> * retiring.
> */
> if (!(flags & PIN_NONBLOCK))
> - i915_retire_requests(dev_priv);
> + i915_retire_requests(i915);
>
> search_again:
> active = NULL;
> @@ -211,7 +211,7 @@ i915_gem_evict_something(struct i915_address_space *vm,
> if (I915_SELFTEST_ONLY(igt_evict_ctl.fail_if_busy))
> return -EBUSY;
>
> - ret = ggtt_flush(dev_priv);
> + ret = ggtt_flush(i915);
> if (ret)
> return ret;
>
> diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.c b/drivers/gpu/drm/i915/i915_gem_gtt.c
> index e70675bfb51d..42d88dae9c5d 100644
> --- a/drivers/gpu/drm/i915/i915_gem_gtt.c
> +++ b/drivers/gpu/drm/i915/i915_gem_gtt.c
> @@ -1294,7 +1294,7 @@ static int gen8_init_scratch(struct i915_address_space *vm)
> static int gen8_ppgtt_notify_vgt(struct i915_ppgtt *ppgtt, bool create)
> {
> struct i915_address_space *vm = &ppgtt->vm;
> - struct drm_i915_private *dev_priv = vm->i915;
> + struct drm_i915_private *i915 = vm->i915;
> enum vgt_g2v_type msg;
> int i;
>
> @@ -1692,7 +1692,7 @@ static inline void gen6_write_pde(const struct gen6_ppgtt *ppgtt,
> ppgtt->pd_addr + pde);
> }
>
> -static void gen7_ppgtt_enable(struct drm_i915_private *dev_priv)
> +static void gen7_ppgtt_enable(struct drm_i915_private *i915)
> {
> struct intel_engine_cs *engine;
> u32 ecochk, ecobits;
> @@ -1702,7 +1702,7 @@ static void gen7_ppgtt_enable(struct drm_i915_private *dev_priv)
> I915_WRITE(GAC_ECO_BITS, ecobits | ECOBITS_PPGTT_CACHE64B);
>
> ecochk = I915_READ(GAM_ECOCHK);
> - if (IS_HASWELL(dev_priv)) {
> + if (IS_HASWELL(i915)) {
> ecochk |= ECOCHK_PPGTT_WB_HSW;
> } else {
> ecochk |= ECOCHK_PPGTT_LLC_IVB;
> @@ -1710,7 +1710,7 @@ static void gen7_ppgtt_enable(struct drm_i915_private *dev_priv)
> }
> I915_WRITE(GAM_ECOCHK, ecochk);
>
> - for_each_engine(engine, dev_priv, id) {
> + for_each_engine(engine, i915, id) {
> /* GFX_MODE is per-ring on gen7+ */
> ENGINE_WRITE(engine,
> RING_MODE_GEN7,
> @@ -1718,7 +1718,7 @@ static void gen7_ppgtt_enable(struct drm_i915_private *dev_priv)
> }
> }
>
> -static void gen6_ppgtt_enable(struct drm_i915_private *dev_priv)
> +static void gen6_ppgtt_enable(struct drm_i915_private *i915)
> {
> u32 ecochk, gab_ctl, ecobits;
>
> @@ -1732,7 +1732,7 @@ static void gen6_ppgtt_enable(struct drm_i915_private *dev_priv)
> ecochk = I915_READ(GAM_ECOCHK);
> I915_WRITE(GAM_ECOCHK, ecochk | ECOCHK_SNB_BIT | ECOCHK_PPGTT_CACHE64B);
>
> - if (HAS_PPGTT(dev_priv)) /* may be disabled for VT-d */
> + if (HAS_PPGTT(i915)) /* may be disabled for VT-d */
> I915_WRITE(GFX_MODE, _MASKED_BIT_ENABLE(GFX_PPGTT_ENABLE));
> }
>
> @@ -2185,20 +2185,20 @@ static struct i915_ppgtt *gen6_ppgtt_create(struct drm_i915_private *i915)
> return ERR_PTR(err);
> }
>
> -static void gtt_write_workarounds(struct drm_i915_private *dev_priv)
> +static void gtt_write_workarounds(struct drm_i915_private *i915)
> {
> /* This function is for gtt related workarounds. This function is
> * called on driver load and after a GPU reset, so you can place
> * workarounds here even if they get overwritten by GPU reset.
> */
> /* WaIncreaseDefaultTLBEntries:chv,bdw,skl,bxt,kbl,glk,cfl,cnl,icl */
> - if (IS_BROADWELL(dev_priv))
> + if (IS_BROADWELL(i915))
> I915_WRITE(GEN8_L3_LRA_1_GPGPU, GEN8_L3_LRA_1_GPGPU_DEFAULT_VALUE_BDW);
> - else if (IS_CHERRYVIEW(dev_priv))
> + else if (IS_CHERRYVIEW(i915))
> I915_WRITE(GEN8_L3_LRA_1_GPGPU, GEN8_L3_LRA_1_GPGPU_DEFAULT_VALUE_CHV);
> - else if (IS_GEN9_LP(dev_priv))
> + else if (IS_GEN9_LP(i915))
> I915_WRITE(GEN8_L3_LRA_1_GPGPU, GEN9_L3_LRA_1_GPGPU_DEFAULT_VALUE_BXT);
> - else if (INTEL_GEN(dev_priv) >= 9)
> + else if (INTEL_GEN(i915) >= 9)
> I915_WRITE(GEN8_L3_LRA_1_GPGPU, GEN9_L3_LRA_1_GPGPU_DEFAULT_VALUE_SKL);
>
> /*
> @@ -2212,21 +2212,21 @@ static void gtt_write_workarounds(struct drm_i915_private *dev_priv)
> * 32K pages, but we don't currently have any support for it in our
> * driver.
> */
> - if (HAS_PAGE_SIZES(dev_priv, I915_GTT_PAGE_SIZE_64K) &&
> - INTEL_GEN(dev_priv) <= 10)
> + if (HAS_PAGE_SIZES(i915, I915_GTT_PAGE_SIZE_64K) &&
> + INTEL_GEN(i915) <= 10)
> I915_WRITE(GEN8_GAMW_ECO_DEV_RW_IA,
> I915_READ(GEN8_GAMW_ECO_DEV_RW_IA) |
> GAMW_ECO_ENABLE_64K_IPS_FIELD);
> }
>
> -int i915_ppgtt_init_hw(struct drm_i915_private *dev_priv)
> +int i915_ppgtt_init_hw(struct drm_i915_private *i915)
> {
> - gtt_write_workarounds(dev_priv);
> + gtt_write_workarounds(i915);
>
> - if (IS_GEN(dev_priv, 6))
> - gen6_ppgtt_enable(dev_priv);
> - else if (IS_GEN(dev_priv, 7))
> - gen7_ppgtt_enable(dev_priv);
> + if (IS_GEN(i915, 6))
> + gen6_ppgtt_enable(i915);
> + else if (IS_GEN(i915, 7))
> + gen7_ppgtt_enable(i915);
>
> return 0;
> }
> @@ -2293,29 +2293,29 @@ void i915_vm_release(struct kref *kref)
> /* Certain Gen5 chipsets require require idling the GPU before
> * unmapping anything from the GTT when VT-d is enabled.
> */
> -static bool needs_idle_maps(struct drm_i915_private *dev_priv)
> +static bool needs_idle_maps(struct drm_i915_private *i915)
> {
> /* Query intel_iommu to see if we need the workaround. Presumably that
> * was loaded first.
> */
> - return IS_GEN(dev_priv, 5) && IS_MOBILE(dev_priv) && intel_vtd_active();
> + return IS_GEN(i915, 5) && IS_MOBILE(i915) && intel_vtd_active();
> }
>
> -void i915_gem_suspend_gtt_mappings(struct drm_i915_private *dev_priv)
> +void i915_gem_suspend_gtt_mappings(struct drm_i915_private *i915)
> {
> - struct i915_ggtt *ggtt = &dev_priv->ggtt;
> + struct i915_ggtt *ggtt = &i915->ggtt;
>
> /* Don't bother messing with faults pre GEN6 as we have little
> * documentation supporting that it's a good idea.
> */
> - if (INTEL_GEN(dev_priv) < 6)
> + if (INTEL_GEN(i915) < 6)
> return;
>
> - i915_check_and_clear_faults(dev_priv);
> + i915_check_and_clear_faults(i915);
>
> ggtt->vm.clear_range(&ggtt->vm, 0, ggtt->vm.total);
>
> - i915_ggtt_invalidate(dev_priv);
> + i915_ggtt_invalidate(i915);
> }
>
> int i915_gem_gtt_prepare_pages(struct drm_i915_gem_object *obj,
> @@ -2411,7 +2411,7 @@ static void gen6_ggtt_insert_page(struct i915_address_space *vm,
> * Binds an object into the global gtt with the specified cache level. The object
> * will be accessible to the GPU via commands whose operands reference offsets
> * within the global GTT as well as accessible by the GPU through the GMADR
> - * mapped BAR (dev_priv->mm.gtt->gtt).
> + * mapped BAR (i915->mm.gtt->gtt).
> */
> static void gen6_ggtt_insert_entries(struct i915_address_space *vm,
> struct i915_vma *vma,
> @@ -2461,7 +2461,7 @@ static void gen8_ggtt_clear_range(struct i915_address_space *vm,
>
> static void bxt_vtd_ggtt_wa(struct i915_address_space *vm)
> {
> - struct drm_i915_private *dev_priv = vm->i915;
> + struct drm_i915_private *i915 = vm->i915;
>
> /*
> * Make sure the internal GAM fifo has been cleared of all GTT
> @@ -2705,12 +2705,12 @@ static void aliasing_gtt_unbind_vma(struct i915_vma *vma)
> void i915_gem_gtt_finish_pages(struct drm_i915_gem_object *obj,
> struct sg_table *pages)
> {
> - struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
> - struct device *kdev = &dev_priv->drm.pdev->dev;
> - struct i915_ggtt *ggtt = &dev_priv->ggtt;
> + struct drm_i915_private *i915 = to_i915(obj->base.dev);
> + struct device *kdev = &i915->drm.pdev->dev;
> + struct i915_ggtt *ggtt = &i915->ggtt;
>
> if (unlikely(ggtt->do_idle_maps)) {
> - if (i915_gem_wait_for_idle(dev_priv, 0, MAX_SCHEDULE_TIMEOUT)) {
> + if (i915_gem_wait_for_idle(i915, 0, MAX_SCHEDULE_TIMEOUT)) {
> DRM_ERROR("Failed to wait for idle; VT'd may hang.\n");
> /* Wait a bit, in hopes it avoids the hang */
> udelay(10);
> @@ -2834,7 +2834,7 @@ static void ggtt_release_guc_top(struct i915_ggtt *ggtt)
> drm_mm_remove_node(&ggtt->uc_fw);
> }
>
> -int i915_gem_init_ggtt(struct drm_i915_private *dev_priv)
> +int i915_gem_init_ggtt(struct drm_i915_private *i915)
> {
> /* Let GEM Manage all of the aperture.
> *
> @@ -2845,7 +2845,7 @@ int i915_gem_init_ggtt(struct drm_i915_private *dev_priv)
> * aperture. One page should be enough to keep any prefetching inside
> * of the aperture.
> */
> - struct i915_ggtt *ggtt = &dev_priv->ggtt;
> + struct i915_ggtt *ggtt = &i915->ggtt;
> unsigned long hole_start, hole_end;
> struct drm_mm_node *entry;
> int ret;
> @@ -2857,9 +2857,9 @@ int i915_gem_init_ggtt(struct drm_i915_private *dev_priv)
> * why.
> */
> ggtt->pin_bias = max_t(u32, I915_GTT_PAGE_SIZE,
> - intel_wopcm_guc_size(&dev_priv->wopcm));
> + intel_wopcm_guc_size(&i915->wopcm));
>
> - ret = intel_vgt_balloon(dev_priv);
> + ret = intel_vgt_balloon(i915);
> if (ret)
> return ret;
>
> @@ -2891,8 +2891,8 @@ int i915_gem_init_ggtt(struct drm_i915_private *dev_priv)
> /* And finally clear the reserved guard page */
> ggtt->vm.clear_range(&ggtt->vm, ggtt->vm.total - PAGE_SIZE, PAGE_SIZE);
>
> - if (INTEL_PPGTT(dev_priv) == INTEL_PPGTT_ALIASING) {
> - ret = init_aliasing_ppgtt(dev_priv);
> + if (INTEL_PPGTT(i915) == INTEL_PPGTT_ALIASING) {
> + ret = init_aliasing_ppgtt(i915);
> if (ret)
> goto err_appgtt;
> }
> @@ -2908,18 +2908,18 @@ int i915_gem_init_ggtt(struct drm_i915_private *dev_priv)
>
> /**
> * i915_ggtt_cleanup_hw - Clean up GGTT hardware initialization
> - * @dev_priv: i915 device
> + * @i915: i915 device
> */
> -void i915_ggtt_cleanup_hw(struct drm_i915_private *dev_priv)
> +void i915_ggtt_cleanup_hw(struct drm_i915_private *i915)
> {
> - struct i915_ggtt *ggtt = &dev_priv->ggtt;
> + struct i915_ggtt *ggtt = &i915->ggtt;
> struct i915_vma *vma, *vn;
> struct pagevec *pvec;
>
> ggtt->vm.closed = true;
>
> - mutex_lock(&dev_priv->drm.struct_mutex);
> - fini_aliasing_ppgtt(dev_priv);
> + mutex_lock(&i915->drm.struct_mutex);
> + fini_aliasing_ppgtt(i915);
>
> list_for_each_entry_safe(vma, vn, &ggtt->vm.bound_list, vm_link)
> WARN_ON(i915_vma_unbind(vma));
> @@ -2930,24 +2930,24 @@ void i915_ggtt_cleanup_hw(struct drm_i915_private *dev_priv)
> ggtt_release_guc_top(ggtt);
>
> if (drm_mm_initialized(&ggtt->vm.mm)) {
> - intel_vgt_deballoon(dev_priv);
> + intel_vgt_deballoon(i915);
> i915_address_space_fini(&ggtt->vm);
> }
>
> ggtt->vm.cleanup(&ggtt->vm);
>
> - pvec = &dev_priv->mm.wc_stash.pvec;
> + pvec = &i915->mm.wc_stash.pvec;
> if (pvec->nr) {
> set_pages_array_wb(pvec->pages, pvec->nr);
> __pagevec_release(pvec);
> }
>
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + mutex_unlock(&i915->drm.struct_mutex);
>
> arch_phys_wc_del(ggtt->mtrr);
> io_mapping_fini(&ggtt->iomap);
>
> - i915_gem_cleanup_stolen(dev_priv);
> + i915_gem_cleanup_stolen(i915);
> }
>
> static unsigned int gen6_get_total_gtt_size(u16 snb_gmch_ctl)
> @@ -2986,8 +2986,8 @@ static unsigned int chv_get_total_gtt_size(u16 gmch_ctrl)
>
> static int ggtt_probe_common(struct i915_ggtt *ggtt, u64 size)
> {
> - struct drm_i915_private *dev_priv = ggtt->vm.i915;
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct drm_i915_private *i915 = ggtt->vm.i915;
> + struct pci_dev *pdev = i915->drm.pdev;
> phys_addr_t phys_addr;
> int ret;
>
> @@ -3001,7 +3001,7 @@ static int ggtt_probe_common(struct i915_ggtt *ggtt, u64 size)
> * resort to an uncached mapping. The WC issue is easily caught by the
> * readback check when writing GTT PTE entries.
> */
> - if (IS_GEN9_LP(dev_priv) || INTEL_GEN(dev_priv) >= 10)
> + if (IS_GEN9_LP(i915) || INTEL_GEN(i915) >= 10)
> ggtt->gsm = ioremap_nocache(phys_addr, size);
> else
> ggtt->gsm = ioremap_wc(phys_addr, size);
> @@ -3135,9 +3135,9 @@ void intel_ppat_put(const struct intel_ppat_entry *entry)
> kref_put(&ppat->entries[index].ref, release_ppat);
> }
>
> -static void cnl_private_pat_update_hw(struct drm_i915_private *dev_priv)
> +static void cnl_private_pat_update_hw(struct drm_i915_private *i915)
> {
> - struct intel_ppat *ppat = &dev_priv->ppat;
> + struct intel_ppat *ppat = &i915->ppat;
> int i;
>
> for_each_set_bit(i, ppat->dirty, ppat->max_entries) {
> @@ -3146,9 +3146,9 @@ static void cnl_private_pat_update_hw(struct drm_i915_private *dev_priv)
> }
> }
>
> -static void bdw_private_pat_update_hw(struct drm_i915_private *dev_priv)
> +static void bdw_private_pat_update_hw(struct drm_i915_private *i915)
> {
> - struct intel_ppat *ppat = &dev_priv->ppat;
> + struct intel_ppat *ppat = &i915->ppat;
> u64 pat = 0;
> int i;
>
> @@ -3293,16 +3293,16 @@ static void gen6_gmch_remove(struct i915_address_space *vm)
> cleanup_scratch_page(vm);
> }
>
> -static void setup_private_pat(struct drm_i915_private *dev_priv)
> +static void setup_private_pat(struct drm_i915_private *i915)
> {
> - struct intel_ppat *ppat = &dev_priv->ppat;
> + struct intel_ppat *ppat = &i915->ppat;
> int i;
>
> - ppat->i915 = dev_priv;
> + ppat->i915 = i915;
>
> - if (INTEL_GEN(dev_priv) >= 10)
> + if (INTEL_GEN(i915) >= 10)
> cnl_setup_private_ppat(ppat);
> - else if (IS_CHERRYVIEW(dev_priv) || IS_GEN9_LP(dev_priv))
> + else if (IS_CHERRYVIEW(i915) || IS_GEN9_LP(i915))
> chv_setup_private_ppat(ppat);
> else
> bdw_setup_private_ppat(ppat);
> @@ -3315,13 +3315,13 @@ static void setup_private_pat(struct drm_i915_private *dev_priv)
> set_bit(i, ppat->dirty);
> }
>
> - ppat->update_hw(dev_priv);
> + ppat->update_hw(i915);
> }
>
> static int gen8_gmch_probe(struct i915_ggtt *ggtt)
> {
> - struct drm_i915_private *dev_priv = ggtt->vm.i915;
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct drm_i915_private *i915 = ggtt->vm.i915;
> + struct pci_dev *pdev = i915->drm.pdev;
> unsigned int size;
> u16 snb_gmch_ctl;
> int err;
> @@ -3339,7 +3339,7 @@ static int gen8_gmch_probe(struct i915_ggtt *ggtt)
> DRM_ERROR("Can't set DMA mask/consistent mask (%d)\n", err);
>
> pci_read_config_word(pdev, SNB_GMCH_CTRL, &snb_gmch_ctl);
> - if (IS_CHERRYVIEW(dev_priv))
> + if (IS_CHERRYVIEW(i915))
> size = chv_get_total_gtt_size(snb_gmch_ctl);
> else
> size = gen8_get_total_gtt_size(snb_gmch_ctl);
> @@ -3348,23 +3348,23 @@ static int gen8_gmch_probe(struct i915_ggtt *ggtt)
> ggtt->vm.cleanup = gen6_gmch_remove;
> ggtt->vm.insert_page = gen8_ggtt_insert_page;
> ggtt->vm.clear_range = nop_clear_range;
> - if (intel_scanout_needs_vtd_wa(dev_priv))
> + if (intel_scanout_needs_vtd_wa(i915))
> ggtt->vm.clear_range = gen8_ggtt_clear_range;
>
> ggtt->vm.insert_entries = gen8_ggtt_insert_entries;
>
> /* Serialize GTT updates with aperture access on BXT if VT-d is on. */
> - if (intel_ggtt_update_needs_vtd_wa(dev_priv) ||
> - IS_CHERRYVIEW(dev_priv) /* fails with concurrent use/update */) {
> + if (intel_ggtt_update_needs_vtd_wa(i915) ||
> + IS_CHERRYVIEW(i915) /* fails with concurrent use/update */) {
> ggtt->vm.insert_entries = bxt_vtd_ggtt_insert_entries__BKL;
> ggtt->vm.insert_page = bxt_vtd_ggtt_insert_page__BKL;
> if (ggtt->vm.clear_range != nop_clear_range)
> ggtt->vm.clear_range = bxt_vtd_ggtt_clear_range__BKL;
>
> /* Prevent recursively calling stop_machine() and deadlocks. */
> - dev_info(dev_priv->drm.dev,
> + dev_info(i915->drm.dev,
> "Disabling error capture for VT-d workaround\n");
> - i915_disable_error_state(dev_priv, -ENODEV);
> + i915_disable_error_state(i915, -ENODEV);
> }
>
> ggtt->invalidate = gen6_ggtt_invalidate;
> @@ -3376,15 +3376,15 @@ static int gen8_gmch_probe(struct i915_ggtt *ggtt)
>
> ggtt->vm.pte_encode = gen8_pte_encode;
>
> - setup_private_pat(dev_priv);
> + setup_private_pat(i915);
>
> return ggtt_probe_common(ggtt, size);
> }
>
> static int gen6_gmch_probe(struct i915_ggtt *ggtt)
> {
> - struct drm_i915_private *dev_priv = ggtt->vm.i915;
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct drm_i915_private *i915 = ggtt->vm.i915;
> + struct pci_dev *pdev = i915->drm.pdev;
> unsigned int size;
> u16 snb_gmch_ctl;
> int err;
> @@ -3413,7 +3413,7 @@ static int gen6_gmch_probe(struct i915_ggtt *ggtt)
> ggtt->vm.total = (size / sizeof(gen6_pte_t)) * I915_GTT_PAGE_SIZE;
>
> ggtt->vm.clear_range = nop_clear_range;
> - if (!HAS_FULL_PPGTT(dev_priv) || intel_scanout_needs_vtd_wa(dev_priv))
> + if (!HAS_FULL_PPGTT(i915) || intel_scanout_needs_vtd_wa(i915))
> ggtt->vm.clear_range = gen6_ggtt_clear_range;
> ggtt->vm.insert_page = gen6_ggtt_insert_page;
> ggtt->vm.insert_entries = gen6_ggtt_insert_entries;
> @@ -3421,13 +3421,13 @@ static int gen6_gmch_probe(struct i915_ggtt *ggtt)
>
> ggtt->invalidate = gen6_ggtt_invalidate;
>
> - if (HAS_EDRAM(dev_priv))
> + if (HAS_EDRAM(i915))
> ggtt->vm.pte_encode = iris_pte_encode;
> - else if (IS_HASWELL(dev_priv))
> + else if (IS_HASWELL(i915))
> ggtt->vm.pte_encode = hsw_pte_encode;
> - else if (IS_VALLEYVIEW(dev_priv))
> + else if (IS_VALLEYVIEW(i915))
> ggtt->vm.pte_encode = byt_pte_encode;
> - else if (INTEL_GEN(dev_priv) >= 7)
> + else if (INTEL_GEN(i915) >= 7)
> ggtt->vm.pte_encode = ivb_pte_encode;
> else
> ggtt->vm.pte_encode = snb_pte_encode;
> @@ -3447,11 +3447,11 @@ static void i915_gmch_remove(struct i915_address_space *vm)
>
> static int i915_gmch_probe(struct i915_ggtt *ggtt)
> {
> - struct drm_i915_private *dev_priv = ggtt->vm.i915;
> + struct drm_i915_private *i915 = ggtt->vm.i915;
> phys_addr_t gmadr_base;
> int ret;
>
> - ret = intel_gmch_probe(dev_priv->bridge_dev, dev_priv->drm.pdev, NULL);
> + ret = intel_gmch_probe(i915->bridge_dev, i915->drm.pdev, NULL);
> if (!ret) {
> DRM_ERROR("failed to set up gmch\n");
> return -EIO;
> @@ -3463,7 +3463,7 @@ static int i915_gmch_probe(struct i915_ggtt *ggtt)
> (struct resource) DEFINE_RES_MEM(gmadr_base,
> ggtt->mappable_end);
>
> - ggtt->do_idle_maps = needs_idle_maps(dev_priv);
> + ggtt->do_idle_maps = needs_idle_maps(i915);
> ggtt->vm.insert_page = i915_ggtt_insert_page;
> ggtt->vm.insert_entries = i915_ggtt_insert_entries;
> ggtt->vm.clear_range = i915_ggtt_clear_range;
> @@ -3484,19 +3484,19 @@ static int i915_gmch_probe(struct i915_ggtt *ggtt)
>
> /**
> * i915_ggtt_probe_hw - Probe GGTT hardware location
> - * @dev_priv: i915 device
> + * @i915: i915 device
> */
> -int i915_ggtt_probe_hw(struct drm_i915_private *dev_priv)
> +int i915_ggtt_probe_hw(struct drm_i915_private *i915)
> {
> - struct i915_ggtt *ggtt = &dev_priv->ggtt;
> + struct i915_ggtt *ggtt = &i915->ggtt;
> int ret;
>
> - ggtt->vm.i915 = dev_priv;
> - ggtt->vm.dma = &dev_priv->drm.pdev->dev;
> + ggtt->vm.i915 = i915;
> + ggtt->vm.dma = &i915->drm.pdev->dev;
>
> - if (INTEL_GEN(dev_priv) <= 5)
> + if (INTEL_GEN(i915) <= 5)
> ret = i915_gmch_probe(ggtt);
> - else if (INTEL_GEN(dev_priv) < 8)
> + else if (INTEL_GEN(i915) < 8)
> ret = gen6_gmch_probe(ggtt);
> else
> ret = gen8_gmch_probe(ggtt);
> @@ -3532,35 +3532,35 @@ int i915_ggtt_probe_hw(struct drm_i915_private *dev_priv)
>
> /**
> * i915_ggtt_init_hw - Initialize GGTT hardware
> - * @dev_priv: i915 device
> + * @i915: i915 device
> */
> -int i915_ggtt_init_hw(struct drm_i915_private *dev_priv)
> +int i915_ggtt_init_hw(struct drm_i915_private *i915)
> {
> - struct i915_ggtt *ggtt = &dev_priv->ggtt;
> + struct i915_ggtt *ggtt = &i915->ggtt;
> int ret;
>
> - stash_init(&dev_priv->mm.wc_stash);
> + stash_init(&i915->mm.wc_stash);
>
> /* Note that we use page colouring to enforce a guard page at the
> * end of the address space. This is required as the CS may prefetch
> * beyond the end of the batch buffer, across the page boundary,
> * and beyond the end of the GTT if we do not provide a guard.
> */
> - mutex_lock(&dev_priv->drm.struct_mutex);
> + mutex_lock(&i915->drm.struct_mutex);
> i915_address_space_init(&ggtt->vm, VM_CLASS_GGTT);
>
> ggtt->vm.is_ggtt = true;
>
> /* Only VLV supports read-only GGTT mappings */
> - ggtt->vm.has_read_only = IS_VALLEYVIEW(dev_priv);
> + ggtt->vm.has_read_only = IS_VALLEYVIEW(i915);
>
> - if (!HAS_LLC(dev_priv) && !HAS_PPGTT(dev_priv))
> + if (!HAS_LLC(i915) && !HAS_PPGTT(i915))
> ggtt->vm.mm.color_adjust = i915_gtt_color_adjust;
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + mutex_unlock(&i915->drm.struct_mutex);
>
> - if (!io_mapping_init_wc(&dev_priv->ggtt.iomap,
> - dev_priv->ggtt.gmadr.start,
> - dev_priv->ggtt.mappable_end)) {
> + if (!io_mapping_init_wc(&i915->ggtt.iomap,
> + i915->ggtt.gmadr.start,
> + i915->ggtt.mappable_end)) {
> ret = -EIO;
> goto out_gtt_cleanup;
> }
> @@ -3571,7 +3571,7 @@ int i915_ggtt_init_hw(struct drm_i915_private *dev_priv)
> * Initialise stolen early so that we may reserve preallocated
> * objects for the BIOS to KMS transition.
> */
> - ret = i915_gem_init_stolen(dev_priv);
> + ret = i915_gem_init_stolen(i915);
> if (ret)
> goto out_gtt_cleanup;
>
> @@ -3582,9 +3582,9 @@ int i915_ggtt_init_hw(struct drm_i915_private *dev_priv)
> return ret;
> }
>
> -int i915_ggtt_enable_hw(struct drm_i915_private *dev_priv)
> +int i915_ggtt_enable_hw(struct drm_i915_private *i915)
> {
> - if (INTEL_GEN(dev_priv) < 6 && !intel_enable_gtt())
> + if (INTEL_GEN(i915) < 6 && !intel_enable_gtt())
> return -EIO;
>
> return 0;
> @@ -3613,12 +3613,12 @@ void i915_ggtt_disable_guc(struct drm_i915_private *i915)
> i915_ggtt_invalidate(i915);
> }
>
> -void i915_gem_restore_gtt_mappings(struct drm_i915_private *dev_priv)
> +void i915_gem_restore_gtt_mappings(struct drm_i915_private *i915)
> {
> - struct i915_ggtt *ggtt = &dev_priv->ggtt;
> + struct i915_ggtt *ggtt = &i915->ggtt;
> struct i915_vma *vma, *vn;
>
> - i915_check_and_clear_faults(dev_priv);
> + i915_check_and_clear_faults(i915);
>
> mutex_lock(&ggtt->vm.mutex);
>
> @@ -3652,15 +3652,15 @@ void i915_gem_restore_gtt_mappings(struct drm_i915_private *dev_priv)
> }
>
> ggtt->vm.closed = false;
> - i915_ggtt_invalidate(dev_priv);
> + i915_ggtt_invalidate(i915);
>
> mutex_unlock(&ggtt->vm.mutex);
>
> - if (INTEL_GEN(dev_priv) >= 8) {
> - struct intel_ppat *ppat = &dev_priv->ppat;
> + if (INTEL_GEN(i915) >= 8) {
> + struct intel_ppat *ppat = &i915->ppat;
>
> bitmap_set(ppat->dirty, 0, ppat->max_entries);
> - dev_priv->ppat.update_hw(dev_priv);
> + i915->ppat.update_hw(i915);
> return;
> }
> }
> diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.h b/drivers/gpu/drm/i915/i915_gem_gtt.h
> index 89437d0a721c..b444552ccc63 100644
> --- a/drivers/gpu/drm/i915/i915_gem_gtt.h
> +++ b/drivers/gpu/drm/i915/i915_gem_gtt.h
> @@ -400,7 +400,7 @@ struct i915_ggtt {
>
> /** "Graphics Stolen Memory" holds the global PTEs */
> void __iomem *gsm;
> - void (*invalidate)(struct drm_i915_private *dev_priv);
> + void (*invalidate)(struct drm_i915_private *i915);
>
> bool do_idle_maps;
>
> @@ -627,17 +627,17 @@ const struct intel_ppat_entry *
> intel_ppat_get(struct drm_i915_private *i915, u8 value);
> void intel_ppat_put(const struct intel_ppat_entry *entry);
>
> -int i915_ggtt_probe_hw(struct drm_i915_private *dev_priv);
> -int i915_ggtt_init_hw(struct drm_i915_private *dev_priv);
> -int i915_ggtt_enable_hw(struct drm_i915_private *dev_priv);
> +int i915_ggtt_probe_hw(struct drm_i915_private *i915);
> +int i915_ggtt_init_hw(struct drm_i915_private *i915);
> +int i915_ggtt_enable_hw(struct drm_i915_private *i915);
> void i915_ggtt_enable_guc(struct drm_i915_private *i915);
> void i915_ggtt_disable_guc(struct drm_i915_private *i915);
> -int i915_gem_init_ggtt(struct drm_i915_private *dev_priv);
> -void i915_ggtt_cleanup_hw(struct drm_i915_private *dev_priv);
> +int i915_gem_init_ggtt(struct drm_i915_private *i915);
> +void i915_ggtt_cleanup_hw(struct drm_i915_private *i915);
>
> -int i915_ppgtt_init_hw(struct drm_i915_private *dev_priv);
> +int i915_ppgtt_init_hw(struct drm_i915_private *i915);
>
> -struct i915_ppgtt *i915_ppgtt_create(struct drm_i915_private *dev_priv);
> +struct i915_ppgtt *i915_ppgtt_create(struct drm_i915_private *i915);
>
> static inline struct i915_address_space *
> i915_vm_get(struct i915_address_space *vm)
> @@ -657,8 +657,8 @@ int gen6_ppgtt_pin(struct i915_ppgtt *base);
> void gen6_ppgtt_unpin(struct i915_ppgtt *base);
> void gen6_ppgtt_unpin_all(struct i915_ppgtt *base);
>
> -void i915_gem_suspend_gtt_mappings(struct drm_i915_private *dev_priv);
> -void i915_gem_restore_gtt_mappings(struct drm_i915_private *dev_priv);
> +void i915_gem_suspend_gtt_mappings(struct drm_i915_private *i915);
> +void i915_gem_restore_gtt_mappings(struct drm_i915_private *i915);
>
> int __must_check i915_gem_gtt_prepare_pages(struct drm_i915_gem_object *obj,
> struct sg_table *pages);
> diff --git a/drivers/gpu/drm/i915/i915_gpu_error.c b/drivers/gpu/drm/i915/i915_gpu_error.c
> index d7f164c9f2eb..335da59031b0 100644
> --- a/drivers/gpu/drm/i915/i915_gpu_error.c
> +++ b/drivers/gpu/drm/i915/i915_gpu_error.c
> @@ -1122,22 +1122,22 @@ static u32 i915_error_generate_code(struct i915_gpu_state *error,
>
> static void gem_record_fences(struct i915_gpu_state *error)
> {
> - struct drm_i915_private *dev_priv = error->i915;
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct drm_i915_private *i915 = error->i915;
> + struct intel_uncore *uncore = &i915->uncore;
> int i;
>
> - if (INTEL_GEN(dev_priv) >= 6) {
> - for (i = 0; i < dev_priv->num_fence_regs; i++)
> + if (INTEL_GEN(i915) >= 6) {
> + for (i = 0; i < i915->num_fence_regs; i++)
> error->fence[i] =
> intel_uncore_read64(uncore,
> FENCE_REG_GEN6_LO(i));
> - } else if (INTEL_GEN(dev_priv) >= 4) {
> - for (i = 0; i < dev_priv->num_fence_regs; i++)
> + } else if (INTEL_GEN(i915) >= 4) {
> + for (i = 0; i < i915->num_fence_regs; i++)
> error->fence[i] =
> intel_uncore_read64(uncore,
> FENCE_REG_965_LO(i));
> } else {
> - for (i = 0; i < dev_priv->num_fence_regs; i++)
> + for (i = 0; i < i915->num_fence_regs; i++)
> error->fence[i] =
> intel_uncore_read(uncore, FENCE_REG(i));
> }
> @@ -1148,23 +1148,23 @@ static void error_record_engine_registers(struct i915_gpu_state *error,
> struct intel_engine_cs *engine,
> struct drm_i915_error_engine *ee)
> {
> - struct drm_i915_private *dev_priv = engine->i915;
> + struct drm_i915_private *i915 = engine->i915;
>
> - if (INTEL_GEN(dev_priv) >= 6) {
> + if (INTEL_GEN(i915) >= 6) {
> ee->rc_psmi = ENGINE_READ(engine, RING_PSMI_CTL);
> - if (INTEL_GEN(dev_priv) >= 8)
> + if (INTEL_GEN(i915) >= 8)
> ee->fault_reg = I915_READ(GEN8_RING_FAULT_REG);
> else
> ee->fault_reg = GEN6_RING_FAULT_REG_READ(engine);
> }
>
> - if (INTEL_GEN(dev_priv) >= 4) {
> + if (INTEL_GEN(i915) >= 4) {
> ee->faddr = ENGINE_READ(engine, RING_DMA_FADD);
> ee->ipeir = ENGINE_READ(engine, RING_IPEIR);
> ee->ipehr = ENGINE_READ(engine, RING_IPEHR);
> ee->instps = ENGINE_READ(engine, RING_INSTPS);
> ee->bbaddr = ENGINE_READ(engine, RING_BBADDR);
> - if (INTEL_GEN(dev_priv) >= 8) {
> + if (INTEL_GEN(i915) >= 8) {
> ee->faddr |= (u64)ENGINE_READ(engine, RING_DMA_FADD_UDW) << 32;
> ee->bbaddr |= (u64)ENGINE_READ(engine, RING_BBADDR_UDW) << 32;
> }
> @@ -1183,13 +1183,13 @@ static void error_record_engine_registers(struct i915_gpu_state *error,
> ee->head = ENGINE_READ(engine, RING_HEAD);
> ee->tail = ENGINE_READ(engine, RING_TAIL);
> ee->ctl = ENGINE_READ(engine, RING_CTL);
> - if (INTEL_GEN(dev_priv) > 2)
> + if (INTEL_GEN(i915) > 2)
> ee->mode = ENGINE_READ(engine, RING_MI_MODE);
>
> - if (!HWS_NEEDS_PHYSICAL(dev_priv)) {
> + if (!HWS_NEEDS_PHYSICAL(i915)) {
> i915_reg_t mmio;
>
> - if (IS_GEN(dev_priv, 7)) {
> + if (IS_GEN(i915, 7)) {
> switch (engine->id) {
> default:
> MISSING_CASE(engine->id);
> @@ -1219,21 +1219,21 @@ static void error_record_engine_registers(struct i915_gpu_state *error,
> ee->idle = intel_engine_is_idle(engine);
> if (!ee->idle)
> ee->hangcheck_timestamp = engine->hangcheck.action_timestamp;
> - ee->reset_count = i915_reset_engine_count(&dev_priv->gpu_error,
> + ee->reset_count = i915_reset_engine_count(&i915->gpu_error,
> engine);
>
> - if (HAS_PPGTT(dev_priv)) {
> + if (HAS_PPGTT(i915)) {
> int i;
>
> ee->vm_info.gfx_mode = ENGINE_READ(engine, RING_MODE_GEN7);
>
> - if (IS_GEN(dev_priv, 6)) {
> + if (IS_GEN(i915, 6)) {
> ee->vm_info.pp_dir_base =
> ENGINE_READ(engine, RING_PP_DIR_BASE_READ);
> - } else if (IS_GEN(dev_priv, 7)) {
> + } else if (IS_GEN(i915, 7)) {
> ee->vm_info.pp_dir_base =
> ENGINE_READ(engine, RING_PP_DIR_BASE);
> - } else if (INTEL_GEN(dev_priv) >= 8) {
> + } else if (INTEL_GEN(i915) >= 8) {
> u32 base = engine->mmio_base;
>
> for (i = 0; i < 4; i++) {
> @@ -1388,7 +1388,7 @@ static void request_record_user_bo(struct i915_request *request,
> }
>
> static struct drm_i915_error_object *
> -capture_object(struct drm_i915_private *dev_priv,
> +capture_object(struct drm_i915_private *i915,
> struct drm_i915_gem_object *obj)
> {
> if (obj && i915_gem_object_has_pages(obj)) {
> @@ -1399,7 +1399,7 @@ capture_object(struct drm_i915_private *dev_priv,
> .obj = obj,
> };
>
> - return i915_error_object_create(dev_priv, &fake);
> + return i915_error_object_create(i915, &fake);
> } else {
> return NULL;
> }
> diff --git a/drivers/gpu/drm/i915/i915_gpu_error.h b/drivers/gpu/drm/i915/i915_gpu_error.h
> index 2ecd0c6a1c94..980f95fb5578 100644
> --- a/drivers/gpu/drm/i915/i915_gpu_error.h
> +++ b/drivers/gpu/drm/i915/i915_gpu_error.h
> @@ -234,7 +234,7 @@ struct i915_gpu_error {
>
> /**
> * Waitqueue to signal when the reset has completed. Used by clients
> - * that wait for dev_priv->mm.wedged to settle.
> + * that wait for i915->mm.wedged to settle.
> */
> wait_queue_head_t reset_queue;
>
> @@ -259,7 +259,7 @@ __printf(2, 3)
> void i915_error_printf(struct drm_i915_error_state_buf *e, const char *f, ...);
>
> struct i915_gpu_state *i915_capture_gpu_state(struct drm_i915_private *i915);
> -void i915_capture_error_state(struct drm_i915_private *dev_priv,
> +void i915_capture_error_state(struct drm_i915_private *i915,
> intel_engine_mask_t engine_mask,
> const char *error_msg);
>
> @@ -286,7 +286,7 @@ void i915_disable_error_state(struct drm_i915_private *i915, int err);
>
> #else
>
> -static inline void i915_capture_error_state(struct drm_i915_private *dev_priv,
> +static inline void i915_capture_error_state(struct drm_i915_private *i915,
> u32 engine_mask,
> const char *error_msg)
> {
> diff --git a/drivers/gpu/drm/i915/i915_irq.c b/drivers/gpu/drm/i915/i915_irq.c
> index cd9edddd6718..81ea65d85b9f 100644
> --- a/drivers/gpu/drm/i915/i915_irq.c
> +++ b/drivers/gpu/drm/i915/i915_irq.c
> @@ -256,18 +256,18 @@ static void gen2_irq_init(struct intel_uncore *uncore,
> #define GEN2_IRQ_INIT(uncore, imr_val, ier_val) \
> gen2_irq_init((uncore), imr_val, ier_val)
>
> -static void gen6_rps_irq_handler(struct drm_i915_private *dev_priv, u32 pm_iir);
> -static void gen9_guc_irq_handler(struct drm_i915_private *dev_priv, u32 pm_iir);
> +static void gen6_rps_irq_handler(struct drm_i915_private *i915, u32 pm_iir);
> +static void gen9_guc_irq_handler(struct drm_i915_private *i915, u32 pm_iir);
>
> /* For display hotplug interrupt */
> static inline void
> -i915_hotplug_interrupt_update_locked(struct drm_i915_private *dev_priv,
> +i915_hotplug_interrupt_update_locked(struct drm_i915_private *i915,
> u32 mask,
> u32 bits)
> {
> u32 val;
>
> - lockdep_assert_held(&dev_priv->irq_lock);
> + lockdep_assert_held(&i915->irq_lock);
> WARN_ON(bits & ~mask);
>
> val = I915_READ(PORT_HOTPLUG_EN);
> @@ -278,7 +278,7 @@ i915_hotplug_interrupt_update_locked(struct drm_i915_private *dev_priv,
>
> /**
> * i915_hotplug_interrupt_update - update hotplug interrupt enable
> - * @dev_priv: driver private
> + * @i915: driver private
> * @mask: bits to update
> * @bits: bits to enable
> * NOTE: the HPD enable bits are modified both inside and outside
> @@ -288,13 +288,13 @@ i915_hotplug_interrupt_update_locked(struct drm_i915_private *dev_priv,
> * held already, this function acquires the lock itself. A non-locking
> * version is also available.
> */
> -void i915_hotplug_interrupt_update(struct drm_i915_private *dev_priv,
> +void i915_hotplug_interrupt_update(struct drm_i915_private *i915,
> u32 mask,
> u32 bits)
> {
> - spin_lock_irq(&dev_priv->irq_lock);
> - i915_hotplug_interrupt_update_locked(dev_priv, mask, bits);
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
> + i915_hotplug_interrupt_update_locked(i915, mask, bits);
> + spin_unlock_irq(&i915->irq_lock);
> }
>
> static u32
> @@ -334,84 +334,84 @@ static bool gen11_reset_one_iir(struct drm_i915_private * const i915,
>
> /**
> * ilk_update_display_irq - update DEIMR
> - * @dev_priv: driver private
> + * @i915: driver private
> * @interrupt_mask: mask of interrupt bits to update
> * @enabled_irq_mask: mask of interrupt bits to enable
> */
> -void ilk_update_display_irq(struct drm_i915_private *dev_priv,
> +void ilk_update_display_irq(struct drm_i915_private *i915,
> u32 interrupt_mask,
> u32 enabled_irq_mask)
> {
> u32 new_val;
>
> - lockdep_assert_held(&dev_priv->irq_lock);
> + lockdep_assert_held(&i915->irq_lock);
>
> WARN_ON(enabled_irq_mask & ~interrupt_mask);
>
> - if (WARN_ON(!intel_irqs_enabled(dev_priv)))
> + if (WARN_ON(!intel_irqs_enabled(i915)))
> return;
>
> - new_val = dev_priv->irq_mask;
> + new_val = i915->irq_mask;
> new_val &= ~interrupt_mask;
> new_val |= (~enabled_irq_mask & interrupt_mask);
>
> - if (new_val != dev_priv->irq_mask) {
> - dev_priv->irq_mask = new_val;
> - I915_WRITE(DEIMR, dev_priv->irq_mask);
> + if (new_val != i915->irq_mask) {
> + i915->irq_mask = new_val;
> + I915_WRITE(DEIMR, i915->irq_mask);
> POSTING_READ(DEIMR);
> }
> }
>
> /**
> * ilk_update_gt_irq - update GTIMR
> - * @dev_priv: driver private
> + * @i915: driver private
> * @interrupt_mask: mask of interrupt bits to update
> * @enabled_irq_mask: mask of interrupt bits to enable
> */
> -static void ilk_update_gt_irq(struct drm_i915_private *dev_priv,
> +static void ilk_update_gt_irq(struct drm_i915_private *i915,
> u32 interrupt_mask,
> u32 enabled_irq_mask)
> {
> - lockdep_assert_held(&dev_priv->irq_lock);
> + lockdep_assert_held(&i915->irq_lock);
>
> WARN_ON(enabled_irq_mask & ~interrupt_mask);
>
> - if (WARN_ON(!intel_irqs_enabled(dev_priv)))
> + if (WARN_ON(!intel_irqs_enabled(i915)))
> return;
>
> - dev_priv->gt_irq_mask &= ~interrupt_mask;
> - dev_priv->gt_irq_mask |= (~enabled_irq_mask & interrupt_mask);
> - I915_WRITE(GTIMR, dev_priv->gt_irq_mask);
> + i915->gt_irq_mask &= ~interrupt_mask;
> + i915->gt_irq_mask |= (~enabled_irq_mask & interrupt_mask);
> + I915_WRITE(GTIMR, i915->gt_irq_mask);
> }
>
> -void gen5_enable_gt_irq(struct drm_i915_private *dev_priv, u32 mask)
> +void gen5_enable_gt_irq(struct drm_i915_private *i915, u32 mask)
> {
> - ilk_update_gt_irq(dev_priv, mask, mask);
> - intel_uncore_posting_read_fw(&dev_priv->uncore, GTIMR);
> + ilk_update_gt_irq(i915, mask, mask);
> + intel_uncore_posting_read_fw(&i915->uncore, GTIMR);
> }
>
> -void gen5_disable_gt_irq(struct drm_i915_private *dev_priv, u32 mask)
> +void gen5_disable_gt_irq(struct drm_i915_private *i915, u32 mask)
> {
> - ilk_update_gt_irq(dev_priv, mask, 0);
> + ilk_update_gt_irq(i915, mask, 0);
> }
>
> -static i915_reg_t gen6_pm_iir(struct drm_i915_private *dev_priv)
> +static i915_reg_t gen6_pm_iir(struct drm_i915_private *i915)
> {
> - WARN_ON_ONCE(INTEL_GEN(dev_priv) >= 11);
> + WARN_ON_ONCE(INTEL_GEN(i915) >= 11);
>
> - return INTEL_GEN(dev_priv) >= 8 ? GEN8_GT_IIR(2) : GEN6_PMIIR;
> + return INTEL_GEN(i915) >= 8 ? GEN8_GT_IIR(2) : GEN6_PMIIR;
> }
>
> -static void write_pm_imr(struct drm_i915_private *dev_priv)
> +static void write_pm_imr(struct drm_i915_private *i915)
> {
> i915_reg_t reg;
> - u32 mask = dev_priv->pm_imr;
> + u32 mask = i915->pm_imr;
>
> - if (INTEL_GEN(dev_priv) >= 11) {
> + if (INTEL_GEN(i915) >= 11) {
> reg = GEN11_GPM_WGBOXPERF_INTR_MASK;
> /* pm is in upper half */
> mask = mask << 16;
> - } else if (INTEL_GEN(dev_priv) >= 8) {
> + } else if (INTEL_GEN(i915) >= 8) {
> reg = GEN8_GT_IMR(2);
> } else {
> reg = GEN6_PMIMR;
> @@ -421,16 +421,16 @@ static void write_pm_imr(struct drm_i915_private *dev_priv)
> POSTING_READ(reg);
> }
>
> -static void write_pm_ier(struct drm_i915_private *dev_priv)
> +static void write_pm_ier(struct drm_i915_private *i915)
> {
> i915_reg_t reg;
> - u32 mask = dev_priv->pm_ier;
> + u32 mask = i915->pm_ier;
>
> - if (INTEL_GEN(dev_priv) >= 11) {
> + if (INTEL_GEN(i915) >= 11) {
> reg = GEN11_GPM_WGBOXPERF_INTR_ENABLE;
> /* pm is in upper half */
> mask = mask << 16;
> - } else if (INTEL_GEN(dev_priv) >= 8) {
> + } else if (INTEL_GEN(i915) >= 8) {
> reg = GEN8_GT_IER(2);
> } else {
> reg = GEN6_PMIER;
> @@ -441,11 +441,11 @@ static void write_pm_ier(struct drm_i915_private *dev_priv)
>
> /**
> * snb_update_pm_irq - update GEN6_PMIMR
> - * @dev_priv: driver private
> + * @i915: driver private
> * @interrupt_mask: mask of interrupt bits to update
> * @enabled_irq_mask: mask of interrupt bits to enable
> */
> -static void snb_update_pm_irq(struct drm_i915_private *dev_priv,
> +static void snb_update_pm_irq(struct drm_i915_private *i915,
> u32 interrupt_mask,
> u32 enabled_irq_mask)
> {
> @@ -453,127 +453,127 @@ static void snb_update_pm_irq(struct drm_i915_private *dev_priv,
>
> WARN_ON(enabled_irq_mask & ~interrupt_mask);
>
> - lockdep_assert_held(&dev_priv->irq_lock);
> + lockdep_assert_held(&i915->irq_lock);
>
> - new_val = dev_priv->pm_imr;
> + new_val = i915->pm_imr;
> new_val &= ~interrupt_mask;
> new_val |= (~enabled_irq_mask & interrupt_mask);
>
> - if (new_val != dev_priv->pm_imr) {
> - dev_priv->pm_imr = new_val;
> - write_pm_imr(dev_priv);
> + if (new_val != i915->pm_imr) {
> + i915->pm_imr = new_val;
> + write_pm_imr(i915);
> }
> }
>
> -void gen6_unmask_pm_irq(struct drm_i915_private *dev_priv, u32 mask)
> +void gen6_unmask_pm_irq(struct drm_i915_private *i915, u32 mask)
> {
> - if (WARN_ON(!intel_irqs_enabled(dev_priv)))
> + if (WARN_ON(!intel_irqs_enabled(i915)))
> return;
>
> - snb_update_pm_irq(dev_priv, mask, mask);
> + snb_update_pm_irq(i915, mask, mask);
> }
>
> -static void __gen6_mask_pm_irq(struct drm_i915_private *dev_priv, u32 mask)
> +static void __gen6_mask_pm_irq(struct drm_i915_private *i915, u32 mask)
> {
> - snb_update_pm_irq(dev_priv, mask, 0);
> + snb_update_pm_irq(i915, mask, 0);
> }
>
> -void gen6_mask_pm_irq(struct drm_i915_private *dev_priv, u32 mask)
> +void gen6_mask_pm_irq(struct drm_i915_private *i915, u32 mask)
> {
> - if (WARN_ON(!intel_irqs_enabled(dev_priv)))
> + if (WARN_ON(!intel_irqs_enabled(i915)))
> return;
>
> - __gen6_mask_pm_irq(dev_priv, mask);
> + __gen6_mask_pm_irq(i915, mask);
> }
>
> -static void gen6_reset_pm_iir(struct drm_i915_private *dev_priv, u32 reset_mask)
> +static void gen6_reset_pm_iir(struct drm_i915_private *i915, u32 reset_mask)
> {
> - i915_reg_t reg = gen6_pm_iir(dev_priv);
> + i915_reg_t reg = gen6_pm_iir(i915);
>
> - lockdep_assert_held(&dev_priv->irq_lock);
> + lockdep_assert_held(&i915->irq_lock);
>
> I915_WRITE(reg, reset_mask);
> I915_WRITE(reg, reset_mask);
> POSTING_READ(reg);
> }
>
> -static void gen6_enable_pm_irq(struct drm_i915_private *dev_priv, u32 enable_mask)
> +static void gen6_enable_pm_irq(struct drm_i915_private *i915, u32 enable_mask)
> {
> - lockdep_assert_held(&dev_priv->irq_lock);
> + lockdep_assert_held(&i915->irq_lock);
>
> - dev_priv->pm_ier |= enable_mask;
> - write_pm_ier(dev_priv);
> - gen6_unmask_pm_irq(dev_priv, enable_mask);
> + i915->pm_ier |= enable_mask;
> + write_pm_ier(i915);
> + gen6_unmask_pm_irq(i915, enable_mask);
> /* unmask_pm_irq provides an implicit barrier (POSTING_READ) */
> }
>
> -static void gen6_disable_pm_irq(struct drm_i915_private *dev_priv, u32 disable_mask)
> +static void gen6_disable_pm_irq(struct drm_i915_private *i915, u32 disable_mask)
> {
> - lockdep_assert_held(&dev_priv->irq_lock);
> + lockdep_assert_held(&i915->irq_lock);
>
> - dev_priv->pm_ier &= ~disable_mask;
> - __gen6_mask_pm_irq(dev_priv, disable_mask);
> - write_pm_ier(dev_priv);
> + i915->pm_ier &= ~disable_mask;
> + __gen6_mask_pm_irq(i915, disable_mask);
> + write_pm_ier(i915);
> /* though a barrier is missing here, but don't really need a one */
> }
>
> -void gen11_reset_rps_interrupts(struct drm_i915_private *dev_priv)
> +void gen11_reset_rps_interrupts(struct drm_i915_private *i915)
> {
> - spin_lock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
>
> - while (gen11_reset_one_iir(dev_priv, 0, GEN11_GTPM))
> + while (gen11_reset_one_iir(i915, 0, GEN11_GTPM))
> ;
>
> - dev_priv->gt_pm.rps.pm_iir = 0;
> + i915->gt_pm.rps.pm_iir = 0;
>
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_unlock_irq(&i915->irq_lock);
> }
>
> -void gen6_reset_rps_interrupts(struct drm_i915_private *dev_priv)
> +void gen6_reset_rps_interrupts(struct drm_i915_private *i915)
> {
> - spin_lock_irq(&dev_priv->irq_lock);
> - gen6_reset_pm_iir(dev_priv, GEN6_PM_RPS_EVENTS);
> - dev_priv->gt_pm.rps.pm_iir = 0;
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
> + gen6_reset_pm_iir(i915, GEN6_PM_RPS_EVENTS);
> + i915->gt_pm.rps.pm_iir = 0;
> + spin_unlock_irq(&i915->irq_lock);
> }
>
> -void gen6_enable_rps_interrupts(struct drm_i915_private *dev_priv)
> +void gen6_enable_rps_interrupts(struct drm_i915_private *i915)
> {
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct intel_rps *rps = &i915->gt_pm.rps;
>
> if (READ_ONCE(rps->interrupts_enabled))
> return;
>
> - spin_lock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
> WARN_ON_ONCE(rps->pm_iir);
>
> - if (INTEL_GEN(dev_priv) >= 11)
> - WARN_ON_ONCE(gen11_reset_one_iir(dev_priv, 0, GEN11_GTPM));
> + if (INTEL_GEN(i915) >= 11)
> + WARN_ON_ONCE(gen11_reset_one_iir(i915, 0, GEN11_GTPM));
> else
> - WARN_ON_ONCE(I915_READ(gen6_pm_iir(dev_priv)) & dev_priv->pm_rps_events);
> + WARN_ON_ONCE(I915_READ(gen6_pm_iir(i915)) & i915->pm_rps_events);
>
> rps->interrupts_enabled = true;
> - gen6_enable_pm_irq(dev_priv, dev_priv->pm_rps_events);
> + gen6_enable_pm_irq(i915, i915->pm_rps_events);
>
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_unlock_irq(&i915->irq_lock);
> }
>
> -void gen6_disable_rps_interrupts(struct drm_i915_private *dev_priv)
> +void gen6_disable_rps_interrupts(struct drm_i915_private *i915)
> {
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct intel_rps *rps = &i915->gt_pm.rps;
>
> if (!READ_ONCE(rps->interrupts_enabled))
> return;
>
> - spin_lock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
> rps->interrupts_enabled = false;
>
> - I915_WRITE(GEN6_PMINTRMSK, gen6_sanitize_rps_pm_mask(dev_priv, ~0u));
> + I915_WRITE(GEN6_PMINTRMSK, gen6_sanitize_rps_pm_mask(i915, ~0u));
>
> - gen6_disable_pm_irq(dev_priv, GEN6_PM_RPS_EVENTS);
> + gen6_disable_pm_irq(i915, GEN6_PM_RPS_EVENTS);
>
> - spin_unlock_irq(&dev_priv->irq_lock);
> - synchronize_irq(dev_priv->drm.irq);
> + spin_unlock_irq(&i915->irq_lock);
> + synchronize_irq(i915->drm.irq);
>
> /* Now that we will not be generating any more work, flush any
> * outstanding tasks. As we are called on the RPS idle path,
> @@ -581,48 +581,48 @@ void gen6_disable_rps_interrupts(struct drm_i915_private *dev_priv)
> * state of the worker can be discarded.
> */
> cancel_work_sync(&rps->work);
> - if (INTEL_GEN(dev_priv) >= 11)
> - gen11_reset_rps_interrupts(dev_priv);
> + if (INTEL_GEN(i915) >= 11)
> + gen11_reset_rps_interrupts(i915);
> else
> - gen6_reset_rps_interrupts(dev_priv);
> + gen6_reset_rps_interrupts(i915);
> }
>
> -void gen9_reset_guc_interrupts(struct drm_i915_private *dev_priv)
> +void gen9_reset_guc_interrupts(struct drm_i915_private *i915)
> {
> - assert_rpm_wakelock_held(dev_priv);
> + assert_rpm_wakelock_held(i915);
>
> - spin_lock_irq(&dev_priv->irq_lock);
> - gen6_reset_pm_iir(dev_priv, dev_priv->pm_guc_events);
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
> + gen6_reset_pm_iir(i915, i915->pm_guc_events);
> + spin_unlock_irq(&i915->irq_lock);
> }
>
> -void gen9_enable_guc_interrupts(struct drm_i915_private *dev_priv)
> +void gen9_enable_guc_interrupts(struct drm_i915_private *i915)
> {
> - assert_rpm_wakelock_held(dev_priv);
> + assert_rpm_wakelock_held(i915);
>
> - spin_lock_irq(&dev_priv->irq_lock);
> - if (!dev_priv->guc.interrupts.enabled) {
> - WARN_ON_ONCE(I915_READ(gen6_pm_iir(dev_priv)) &
> - dev_priv->pm_guc_events);
> - dev_priv->guc.interrupts.enabled = true;
> - gen6_enable_pm_irq(dev_priv, dev_priv->pm_guc_events);
> + spin_lock_irq(&i915->irq_lock);
> + if (!i915->guc.interrupts.enabled) {
> + WARN_ON_ONCE(I915_READ(gen6_pm_iir(i915)) &
> + i915->pm_guc_events);
> + i915->guc.interrupts.enabled = true;
> + gen6_enable_pm_irq(i915, i915->pm_guc_events);
> }
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_unlock_irq(&i915->irq_lock);
> }
>
> -void gen9_disable_guc_interrupts(struct drm_i915_private *dev_priv)
> +void gen9_disable_guc_interrupts(struct drm_i915_private *i915)
> {
> - assert_rpm_wakelock_held(dev_priv);
> + assert_rpm_wakelock_held(i915);
>
> - spin_lock_irq(&dev_priv->irq_lock);
> - dev_priv->guc.interrupts.enabled = false;
> + spin_lock_irq(&i915->irq_lock);
> + i915->guc.interrupts.enabled = false;
>
> - gen6_disable_pm_irq(dev_priv, dev_priv->pm_guc_events);
> + gen6_disable_pm_irq(i915, i915->pm_guc_events);
>
> - spin_unlock_irq(&dev_priv->irq_lock);
> - synchronize_irq(dev_priv->drm.irq);
> + spin_unlock_irq(&i915->irq_lock);
> + synchronize_irq(i915->drm.irq);
>
> - gen9_reset_guc_interrupts(dev_priv);
> + gen9_reset_guc_interrupts(i915);
> }
>
> void gen11_reset_guc_interrupts(struct drm_i915_private *i915)
> @@ -632,53 +632,53 @@ void gen11_reset_guc_interrupts(struct drm_i915_private *i915)
> spin_unlock_irq(&i915->irq_lock);
> }
>
> -void gen11_enable_guc_interrupts(struct drm_i915_private *dev_priv)
> +void gen11_enable_guc_interrupts(struct drm_i915_private *i915)
> {
> - spin_lock_irq(&dev_priv->irq_lock);
> - if (!dev_priv->guc.interrupts.enabled) {
> + spin_lock_irq(&i915->irq_lock);
> + if (!i915->guc.interrupts.enabled) {
> u32 events = REG_FIELD_PREP(ENGINE1_MASK,
> GEN11_GUC_INTR_GUC2HOST);
>
> - WARN_ON_ONCE(gen11_reset_one_iir(dev_priv, 0, GEN11_GUC));
> + WARN_ON_ONCE(gen11_reset_one_iir(i915, 0, GEN11_GUC));
> I915_WRITE(GEN11_GUC_SG_INTR_ENABLE, events);
> I915_WRITE(GEN11_GUC_SG_INTR_MASK, ~events);
> - dev_priv->guc.interrupts.enabled = true;
> + i915->guc.interrupts.enabled = true;
> }
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_unlock_irq(&i915->irq_lock);
> }
>
> -void gen11_disable_guc_interrupts(struct drm_i915_private *dev_priv)
> +void gen11_disable_guc_interrupts(struct drm_i915_private *i915)
> {
> - spin_lock_irq(&dev_priv->irq_lock);
> - dev_priv->guc.interrupts.enabled = false;
> + spin_lock_irq(&i915->irq_lock);
> + i915->guc.interrupts.enabled = false;
>
> I915_WRITE(GEN11_GUC_SG_INTR_MASK, ~0);
> I915_WRITE(GEN11_GUC_SG_INTR_ENABLE, 0);
>
> - spin_unlock_irq(&dev_priv->irq_lock);
> - synchronize_irq(dev_priv->drm.irq);
> + spin_unlock_irq(&i915->irq_lock);
> + synchronize_irq(i915->drm.irq);
>
> - gen11_reset_guc_interrupts(dev_priv);
> + gen11_reset_guc_interrupts(i915);
> }
>
> /**
> * bdw_update_port_irq - update DE port interrupt
> - * @dev_priv: driver private
> + * @i915: driver private
> * @interrupt_mask: mask of interrupt bits to update
> * @enabled_irq_mask: mask of interrupt bits to enable
> */
> -static void bdw_update_port_irq(struct drm_i915_private *dev_priv,
> +static void bdw_update_port_irq(struct drm_i915_private *i915,
> u32 interrupt_mask,
> u32 enabled_irq_mask)
> {
> u32 new_val;
> u32 old_val;
>
> - lockdep_assert_held(&dev_priv->irq_lock);
> + lockdep_assert_held(&i915->irq_lock);
>
> WARN_ON(enabled_irq_mask & ~interrupt_mask);
>
> - if (WARN_ON(!intel_irqs_enabled(dev_priv)))
> + if (WARN_ON(!intel_irqs_enabled(i915)))
> return;
>
> old_val = I915_READ(GEN8_DE_PORT_IMR);
> @@ -695,43 +695,43 @@ static void bdw_update_port_irq(struct drm_i915_private *dev_priv,
>
> /**
> * bdw_update_pipe_irq - update DE pipe interrupt
> - * @dev_priv: driver private
> + * @i915: driver private
> * @pipe: pipe whose interrupt to update
> * @interrupt_mask: mask of interrupt bits to update
> * @enabled_irq_mask: mask of interrupt bits to enable
> */
> -void bdw_update_pipe_irq(struct drm_i915_private *dev_priv,
> +void bdw_update_pipe_irq(struct drm_i915_private *i915,
> enum pipe pipe,
> u32 interrupt_mask,
> u32 enabled_irq_mask)
> {
> u32 new_val;
>
> - lockdep_assert_held(&dev_priv->irq_lock);
> + lockdep_assert_held(&i915->irq_lock);
>
> WARN_ON(enabled_irq_mask & ~interrupt_mask);
>
> - if (WARN_ON(!intel_irqs_enabled(dev_priv)))
> + if (WARN_ON(!intel_irqs_enabled(i915)))
> return;
>
> - new_val = dev_priv->de_irq_mask[pipe];
> + new_val = i915->de_irq_mask[pipe];
> new_val &= ~interrupt_mask;
> new_val |= (~enabled_irq_mask & interrupt_mask);
>
> - if (new_val != dev_priv->de_irq_mask[pipe]) {
> - dev_priv->de_irq_mask[pipe] = new_val;
> - I915_WRITE(GEN8_DE_PIPE_IMR(pipe), dev_priv->de_irq_mask[pipe]);
> + if (new_val != i915->de_irq_mask[pipe]) {
> + i915->de_irq_mask[pipe] = new_val;
> + I915_WRITE(GEN8_DE_PIPE_IMR(pipe), i915->de_irq_mask[pipe]);
> POSTING_READ(GEN8_DE_PIPE_IMR(pipe));
> }
> }
>
> /**
> * ibx_display_interrupt_update - update SDEIMR
> - * @dev_priv: driver private
> + * @i915: driver private
> * @interrupt_mask: mask of interrupt bits to update
> * @enabled_irq_mask: mask of interrupt bits to enable
> */
> -void ibx_display_interrupt_update(struct drm_i915_private *dev_priv,
> +void ibx_display_interrupt_update(struct drm_i915_private *i915,
> u32 interrupt_mask,
> u32 enabled_irq_mask)
> {
> @@ -741,24 +741,24 @@ void ibx_display_interrupt_update(struct drm_i915_private *dev_priv,
>
> WARN_ON(enabled_irq_mask & ~interrupt_mask);
>
> - lockdep_assert_held(&dev_priv->irq_lock);
> + lockdep_assert_held(&i915->irq_lock);
>
> - if (WARN_ON(!intel_irqs_enabled(dev_priv)))
> + if (WARN_ON(!intel_irqs_enabled(i915)))
> return;
>
> I915_WRITE(SDEIMR, sdeimr);
> POSTING_READ(SDEIMR);
> }
>
> -u32 i915_pipestat_enable_mask(struct drm_i915_private *dev_priv,
> +u32 i915_pipestat_enable_mask(struct drm_i915_private *i915,
> enum pipe pipe)
> {
> - u32 status_mask = dev_priv->pipestat_irq_mask[pipe];
> + u32 status_mask = i915->pipestat_irq_mask[pipe];
> u32 enable_mask = status_mask << 16;
>
> - lockdep_assert_held(&dev_priv->irq_lock);
> + lockdep_assert_held(&i915->irq_lock);
>
> - if (INTEL_GEN(dev_priv) < 5)
> + if (INTEL_GEN(i915) < 5)
> goto out;
>
> /*
> @@ -791,7 +791,7 @@ u32 i915_pipestat_enable_mask(struct drm_i915_private *dev_priv,
> return enable_mask;
> }
>
> -void i915_enable_pipestat(struct drm_i915_private *dev_priv,
> +void i915_enable_pipestat(struct drm_i915_private *i915,
> enum pipe pipe, u32 status_mask)
> {
> i915_reg_t reg = PIPESTAT(pipe);
> @@ -801,20 +801,20 @@ void i915_enable_pipestat(struct drm_i915_private *dev_priv,
> "pipe %c: status_mask=0x%x\n",
> pipe_name(pipe), status_mask);
>
> - lockdep_assert_held(&dev_priv->irq_lock);
> - WARN_ON(!intel_irqs_enabled(dev_priv));
> + lockdep_assert_held(&i915->irq_lock);
> + WARN_ON(!intel_irqs_enabled(i915));
>
> - if ((dev_priv->pipestat_irq_mask[pipe] & status_mask) == status_mask)
> + if ((i915->pipestat_irq_mask[pipe] & status_mask) == status_mask)
> return;
>
> - dev_priv->pipestat_irq_mask[pipe] |= status_mask;
> - enable_mask = i915_pipestat_enable_mask(dev_priv, pipe);
> + i915->pipestat_irq_mask[pipe] |= status_mask;
> + enable_mask = i915_pipestat_enable_mask(i915, pipe);
>
> I915_WRITE(reg, enable_mask | status_mask);
> POSTING_READ(reg);
> }
>
> -void i915_disable_pipestat(struct drm_i915_private *dev_priv,
> +void i915_disable_pipestat(struct drm_i915_private *i915,
> enum pipe pipe, u32 status_mask)
> {
> i915_reg_t reg = PIPESTAT(pipe);
> @@ -824,44 +824,44 @@ void i915_disable_pipestat(struct drm_i915_private *dev_priv,
> "pipe %c: status_mask=0x%x\n",
> pipe_name(pipe), status_mask);
>
> - lockdep_assert_held(&dev_priv->irq_lock);
> - WARN_ON(!intel_irqs_enabled(dev_priv));
> + lockdep_assert_held(&i915->irq_lock);
> + WARN_ON(!intel_irqs_enabled(i915));
>
> - if ((dev_priv->pipestat_irq_mask[pipe] & status_mask) == 0)
> + if ((i915->pipestat_irq_mask[pipe] & status_mask) == 0)
> return;
>
> - dev_priv->pipestat_irq_mask[pipe] &= ~status_mask;
> - enable_mask = i915_pipestat_enable_mask(dev_priv, pipe);
> + i915->pipestat_irq_mask[pipe] &= ~status_mask;
> + enable_mask = i915_pipestat_enable_mask(i915, pipe);
>
> I915_WRITE(reg, enable_mask | status_mask);
> POSTING_READ(reg);
> }
>
> -static bool i915_has_asle(struct drm_i915_private *dev_priv)
> +static bool i915_has_asle(struct drm_i915_private *i915)
> {
> - if (!dev_priv->opregion.asle)
> + if (!i915->opregion.asle)
> return false;
>
> - return IS_PINEVIEW(dev_priv) || IS_MOBILE(dev_priv);
> + return IS_PINEVIEW(i915) || IS_MOBILE(i915);
> }
>
> /**
> * i915_enable_asle_pipestat - enable ASLE pipestat for OpRegion
> - * @dev_priv: i915 device private
> + * @i915: i915 device private
> */
> -static void i915_enable_asle_pipestat(struct drm_i915_private *dev_priv)
> +static void i915_enable_asle_pipestat(struct drm_i915_private *i915)
> {
> - if (!i915_has_asle(dev_priv))
> + if (!i915_has_asle(i915))
> return;
>
> - spin_lock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
>
> - i915_enable_pipestat(dev_priv, PIPE_B, PIPE_LEGACY_BLC_EVENT_STATUS);
> - if (INTEL_GEN(dev_priv) >= 4)
> - i915_enable_pipestat(dev_priv, PIPE_A,
> + i915_enable_pipestat(i915, PIPE_B, PIPE_LEGACY_BLC_EVENT_STATUS);
> + if (INTEL_GEN(i915) >= 4)
> + i915_enable_pipestat(i915, PIPE_A,
> PIPE_LEGACY_BLC_EVENT_STATUS);
>
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_unlock_irq(&i915->irq_lock);
> }
>
> /*
> @@ -919,7 +919,7 @@ static void i915_enable_asle_pipestat(struct drm_i915_private *dev_priv)
> */
> static u32 i915_get_vblank_counter(struct drm_device *dev, unsigned int pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct drm_vblank_crtc *vblank = &dev->vblank[pipe];
> const struct drm_display_mode *mode = &vblank->hwmode;
> i915_reg_t high_frame, low_frame;
> @@ -955,7 +955,7 @@ static u32 i915_get_vblank_counter(struct drm_device *dev, unsigned int pipe)
> high_frame = PIPEFRAME(pipe);
> low_frame = PIPEFRAMEPIXEL(pipe);
>
> - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> + spin_lock_irqsave(&i915->uncore.lock, irqflags);
>
> /*
> * High & low register fields aren't synchronized, so make sure
> @@ -968,7 +968,7 @@ static u32 i915_get_vblank_counter(struct drm_device *dev, unsigned int pipe)
> high2 = I915_READ_FW(high_frame) & PIPE_FRAME_HIGH_MASK;
> } while (high1 != high2);
>
> - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> + spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
>
> high1 >>= PIPE_FRAME_HIGH_SHIFT;
> pixel = low & PIPE_PIXEL_MASK;
> @@ -984,7 +984,7 @@ static u32 i915_get_vblank_counter(struct drm_device *dev, unsigned int pipe)
>
> static u32 g4x_get_vblank_counter(struct drm_device *dev, unsigned int pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
>
> return I915_READ(PIPE_FRMCOUNT_G4X(pipe));
> }
> @@ -999,7 +999,7 @@ static u32 g4x_get_vblank_counter(struct drm_device *dev, unsigned int pipe)
> */
> static u32 __intel_get_crtc_scanline_from_timestamp(struct intel_crtc *crtc)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> struct drm_vblank_crtc *vblank =
> &crtc->base.dev->vblank[drm_crtc_index(&crtc->base)];
> const struct drm_display_mode *mode = &vblank->hwmode;
> @@ -1044,7 +1044,7 @@ static u32 __intel_get_crtc_scanline_from_timestamp(struct intel_crtc *crtc)
> static int __intel_get_crtc_scanline(struct intel_crtc *crtc)
> {
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> const struct drm_display_mode *mode;
> struct drm_vblank_crtc *vblank;
> enum pipe pipe = crtc->pipe;
> @@ -1063,7 +1063,7 @@ static int __intel_get_crtc_scanline(struct intel_crtc *crtc)
> if (mode->flags & DRM_MODE_FLAG_INTERLACE)
> vtotal /= 2;
>
> - if (IS_GEN(dev_priv, 2))
> + if (IS_GEN(i915, 2))
> position = I915_READ_FW(PIPEDSL(pipe)) & DSL_LINEMASK_GEN2;
> else
> position = I915_READ_FW(PIPEDSL(pipe)) & DSL_LINEMASK_GEN3;
> @@ -1080,7 +1080,7 @@ static int __intel_get_crtc_scanline(struct intel_crtc *crtc)
> * problem. We may need to extend this to include other platforms,
> * but so far testing only shows the problem on HSW.
> */
> - if (HAS_DDI(dev_priv) && !position) {
> + if (HAS_DDI(i915) && !position) {
> int i, temp;
>
> for (i = 0; i < 100; i++) {
> @@ -1105,14 +1105,14 @@ static bool i915_get_crtc_scanoutpos(struct drm_device *dev, unsigned int pipe,
> ktime_t *stime, ktime_t *etime,
> const struct drm_display_mode *mode)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct intel_crtc *intel_crtc = intel_get_crtc_for_pipe(dev_priv,
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct intel_crtc *intel_crtc = intel_get_crtc_for_pipe(i915,
> pipe);
> int position;
> int vbl_start, vbl_end, hsync_start, htotal, vtotal;
> unsigned long irqflags;
> - bool use_scanline_counter = INTEL_GEN(dev_priv) >= 5 ||
> - IS_G4X(dev_priv) || IS_GEN(dev_priv, 2) ||
> + bool use_scanline_counter = INTEL_GEN(i915) >= 5 ||
> + IS_G4X(i915) || IS_GEN(i915, 2) ||
> mode->private_flags & I915_MODE_FLAG_USE_SCANLINE_COUNTER;
>
> if (WARN_ON(!mode->crtc_clock)) {
> @@ -1138,7 +1138,7 @@ static bool i915_get_crtc_scanoutpos(struct drm_device *dev, unsigned int pipe,
> * register reads, potentially with preemption disabled, so the
> * following code must not block on uncore.lock.
> */
> - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> + spin_lock_irqsave(&i915->uncore.lock, irqflags);
>
> /* preempt_disable_rt() should go right here in PREEMPT_RT patchset. */
>
> @@ -1193,7 +1193,7 @@ static bool i915_get_crtc_scanoutpos(struct drm_device *dev, unsigned int pipe,
>
> /* preempt_enable_rt() should go right here in PREEMPT_RT patchset. */
>
> - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> + spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
>
> /*
> * While in vblank, position will be negative
> @@ -1219,20 +1219,20 @@ static bool i915_get_crtc_scanoutpos(struct drm_device *dev, unsigned int pipe,
>
> int intel_get_crtc_scanline(struct intel_crtc *crtc)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> unsigned long irqflags;
> int position;
>
> - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> + spin_lock_irqsave(&i915->uncore.lock, irqflags);
> position = __intel_get_crtc_scanline(crtc);
> - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> + spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
>
> return position;
> }
>
> -static void ironlake_rps_change_irq_handler(struct drm_i915_private *dev_priv)
> +static void ironlake_rps_change_irq_handler(struct drm_i915_private *i915)
> {
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct intel_uncore *uncore = &i915->uncore;
> u32 busy_up, busy_down, max_avg, min_avg;
> u8 new_delay;
>
> @@ -1242,7 +1242,7 @@ static void ironlake_rps_change_irq_handler(struct drm_i915_private *dev_priv)
> MEMINTRSTS,
> intel_uncore_read(uncore, MEMINTRSTS));
>
> - new_delay = dev_priv->ips.cur_delay;
> + new_delay = i915->ips.cur_delay;
>
> intel_uncore_write16(uncore, MEMINTRSTS, MEMINT_EVAL_CHG);
> busy_up = intel_uncore_read(uncore, RCPREVBSYTUPAVG);
> @@ -1252,26 +1252,26 @@ static void ironlake_rps_change_irq_handler(struct drm_i915_private *dev_priv)
>
> /* Handle RCS change request from hw */
> if (busy_up > max_avg) {
> - if (dev_priv->ips.cur_delay != dev_priv->ips.max_delay)
> - new_delay = dev_priv->ips.cur_delay - 1;
> - if (new_delay < dev_priv->ips.max_delay)
> - new_delay = dev_priv->ips.max_delay;
> + if (i915->ips.cur_delay != i915->ips.max_delay)
> + new_delay = i915->ips.cur_delay - 1;
> + if (new_delay < i915->ips.max_delay)
> + new_delay = i915->ips.max_delay;
> } else if (busy_down < min_avg) {
> - if (dev_priv->ips.cur_delay != dev_priv->ips.min_delay)
> - new_delay = dev_priv->ips.cur_delay + 1;
> - if (new_delay > dev_priv->ips.min_delay)
> - new_delay = dev_priv->ips.min_delay;
> + if (i915->ips.cur_delay != i915->ips.min_delay)
> + new_delay = i915->ips.cur_delay + 1;
> + if (new_delay > i915->ips.min_delay)
> + new_delay = i915->ips.min_delay;
> }
>
> - if (ironlake_set_drps(dev_priv, new_delay))
> - dev_priv->ips.cur_delay = new_delay;
> + if (ironlake_set_drps(i915, new_delay))
> + i915->ips.cur_delay = new_delay;
>
> spin_unlock(&mchdev_lock);
>
> return;
> }
>
> -static void vlv_c0_read(struct drm_i915_private *dev_priv,
> +static void vlv_c0_read(struct drm_i915_private *i915,
> struct intel_rps_ei *ei)
> {
> ei->ktime = ktime_get_raw();
> @@ -1279,14 +1279,14 @@ static void vlv_c0_read(struct drm_i915_private *dev_priv,
> ei->media_c0 = I915_READ(VLV_MEDIA_C0_COUNT);
> }
>
> -void gen6_rps_reset_ei(struct drm_i915_private *dev_priv)
> +void gen6_rps_reset_ei(struct drm_i915_private *i915)
> {
> - memset(&dev_priv->gt_pm.rps.ei, 0, sizeof(dev_priv->gt_pm.rps.ei));
> + memset(&i915->gt_pm.rps.ei, 0, sizeof(i915->gt_pm.rps.ei));
> }
>
> -static u32 vlv_wa_c0_ei(struct drm_i915_private *dev_priv, u32 pm_iir)
> +static u32 vlv_wa_c0_ei(struct drm_i915_private *i915, u32 pm_iir)
> {
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct intel_rps *rps = &i915->gt_pm.rps;
> const struct intel_rps_ei *prev = &rps->ei;
> struct intel_rps_ei now;
> u32 events = 0;
> @@ -1294,7 +1294,7 @@ static u32 vlv_wa_c0_ei(struct drm_i915_private *dev_priv, u32 pm_iir)
> if ((pm_iir & GEN6_PM_RP_UP_EI_EXPIRED) == 0)
> return 0;
>
> - vlv_c0_read(dev_priv, &now);
> + vlv_c0_read(i915, &now);
>
> if (prev->ktime) {
> u64 time, c0;
> @@ -1302,7 +1302,7 @@ static u32 vlv_wa_c0_ei(struct drm_i915_private *dev_priv, u32 pm_iir)
>
> time = ktime_us_delta(now.ktime, prev->ktime);
>
> - time *= dev_priv->czclk_freq;
> + time *= i915->czclk_freq;
>
> /* Workload can be split between render + media,
> * e.g. SwapBuffers being blitted in X after being rendered in
> @@ -1326,28 +1326,28 @@ static u32 vlv_wa_c0_ei(struct drm_i915_private *dev_priv, u32 pm_iir)
>
> static void gen6_pm_rps_work(struct work_struct *work)
> {
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> container_of(work, struct drm_i915_private, gt_pm.rps.work);
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct intel_rps *rps = &i915->gt_pm.rps;
> bool client_boost = false;
> int new_delay, adj, min, max;
> u32 pm_iir = 0;
>
> - spin_lock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
> if (rps->interrupts_enabled) {
> pm_iir = fetch_and_zero(&rps->pm_iir);
> client_boost = atomic_read(&rps->num_waiters);
> }
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_unlock_irq(&i915->irq_lock);
>
> /* Make sure we didn't queue anything we're not going to process. */
> - WARN_ON(pm_iir & ~dev_priv->pm_rps_events);
> - if ((pm_iir & dev_priv->pm_rps_events) == 0 && !client_boost)
> + WARN_ON(pm_iir & ~i915->pm_rps_events);
> + if ((pm_iir & i915->pm_rps_events) == 0 && !client_boost)
> goto out;
>
> mutex_lock(&rps->lock);
>
> - pm_iir |= vlv_wa_c0_ei(dev_priv, pm_iir);
> + pm_iir |= vlv_wa_c0_ei(i915, pm_iir);
>
> adj = rps->last_adj;
> new_delay = rps->cur_freq;
> @@ -1362,7 +1362,7 @@ static void gen6_pm_rps_work(struct work_struct *work)
> if (adj > 0)
> adj *= 2;
> else /* CHV needs even encode values */
> - adj = IS_CHERRYVIEW(dev_priv) ? 2 : 1;
> + adj = IS_CHERRYVIEW(i915) ? 2 : 1;
>
> if (new_delay >= rps->max_freq_softlimit)
> adj = 0;
> @@ -1378,7 +1378,7 @@ static void gen6_pm_rps_work(struct work_struct *work)
> if (adj < 0)
> adj *= 2;
> else /* CHV needs even encode values */
> - adj = IS_CHERRYVIEW(dev_priv) ? -2 : -1;
> + adj = IS_CHERRYVIEW(i915) ? -2 : -1;
>
> if (new_delay <= rps->min_freq_softlimit)
> adj = 0;
> @@ -1406,7 +1406,7 @@ static void gen6_pm_rps_work(struct work_struct *work)
> new_delay += adj;
> new_delay = clamp_t(int, new_delay, min, max);
>
> - if (intel_set_rps(dev_priv, new_delay)) {
> + if (intel_set_rps(i915, new_delay)) {
> DRM_DEBUG_DRIVER("Failed to set new GPU frequency\n");
> rps->last_adj = 0;
> }
> @@ -1415,10 +1415,10 @@ static void gen6_pm_rps_work(struct work_struct *work)
>
> out:
> /* Make sure not to corrupt PMIMR state used by ringbuffer on GEN6 */
> - spin_lock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
> if (rps->interrupts_enabled)
> - gen6_unmask_pm_irq(dev_priv, dev_priv->pm_rps_events);
> - spin_unlock_irq(&dev_priv->irq_lock);
> + gen6_unmask_pm_irq(i915, i915->pm_rps_events);
> + spin_unlock_irq(&i915->irq_lock);
> }
>
>
> @@ -1433,8 +1433,8 @@ static void gen6_pm_rps_work(struct work_struct *work)
> */
> static void ivybridge_parity_work(struct work_struct *work)
> {
> - struct drm_i915_private *dev_priv =
> - container_of(work, typeof(*dev_priv), l3_parity.error_work);
> + struct drm_i915_private *i915 =
> + container_of(work, typeof(*i915), l3_parity.error_work);
> u32 error_status, row, bank, subbank;
> char *parity_event[6];
> u32 misccpctl;
> @@ -1444,24 +1444,24 @@ static void ivybridge_parity_work(struct work_struct *work)
> * In order to prevent a get/put style interface, acquire struct mutex
> * any time we access those registers.
> */
> - mutex_lock(&dev_priv->drm.struct_mutex);
> + mutex_lock(&i915->drm.struct_mutex);
>
> /* If we've screwed up tracking, just let the interrupt fire again */
> - if (WARN_ON(!dev_priv->l3_parity.which_slice))
> + if (WARN_ON(!i915->l3_parity.which_slice))
> goto out;
>
> misccpctl = I915_READ(GEN7_MISCCPCTL);
> I915_WRITE(GEN7_MISCCPCTL, misccpctl & ~GEN7_DOP_CLOCK_GATE_ENABLE);
> POSTING_READ(GEN7_MISCCPCTL);
>
> - while ((slice = ffs(dev_priv->l3_parity.which_slice)) != 0) {
> + while ((slice = ffs(i915->l3_parity.which_slice)) != 0) {
> i915_reg_t reg;
>
> slice--;
> - if (WARN_ON_ONCE(slice >= NUM_L3_SLICES(dev_priv)))
> + if (WARN_ON_ONCE(slice >= NUM_L3_SLICES(i915)))
> break;
>
> - dev_priv->l3_parity.which_slice &= ~(1<<slice);
> + i915->l3_parity.which_slice &= ~(1<<slice);
>
> reg = GEN7_L3CDERRST1(slice);
>
> @@ -1480,7 +1480,7 @@ static void ivybridge_parity_work(struct work_struct *work)
> parity_event[4] = kasprintf(GFP_KERNEL, "SLICE=%d", slice);
> parity_event[5] = NULL;
>
> - kobject_uevent_env(&dev_priv->drm.primary->kdev->kobj,
> + kobject_uevent_env(&i915->drm.primary->kdev->kobj,
> KOBJ_CHANGE, parity_event);
>
> DRM_DEBUG("Parity error: Slice = %d, Row = %d, Bank = %d, Sub bank = %d.\n",
> @@ -1495,60 +1495,60 @@ static void ivybridge_parity_work(struct work_struct *work)
> I915_WRITE(GEN7_MISCCPCTL, misccpctl);
>
> out:
> - WARN_ON(dev_priv->l3_parity.which_slice);
> - spin_lock_irq(&dev_priv->irq_lock);
> - gen5_enable_gt_irq(dev_priv, GT_PARITY_ERROR(dev_priv));
> - spin_unlock_irq(&dev_priv->irq_lock);
> + WARN_ON(i915->l3_parity.which_slice);
> + spin_lock_irq(&i915->irq_lock);
> + gen5_enable_gt_irq(i915, GT_PARITY_ERROR(i915));
> + spin_unlock_irq(&i915->irq_lock);
>
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + mutex_unlock(&i915->drm.struct_mutex);
> }
>
> -static void ivybridge_parity_error_irq_handler(struct drm_i915_private *dev_priv,
> +static void ivybridge_parity_error_irq_handler(struct drm_i915_private *i915,
> u32 iir)
> {
> - if (!HAS_L3_DPF(dev_priv))
> + if (!HAS_L3_DPF(i915))
> return;
>
> - spin_lock(&dev_priv->irq_lock);
> - gen5_disable_gt_irq(dev_priv, GT_PARITY_ERROR(dev_priv));
> - spin_unlock(&dev_priv->irq_lock);
> + spin_lock(&i915->irq_lock);
> + gen5_disable_gt_irq(i915, GT_PARITY_ERROR(i915));
> + spin_unlock(&i915->irq_lock);
>
> - iir &= GT_PARITY_ERROR(dev_priv);
> + iir &= GT_PARITY_ERROR(i915);
> if (iir & GT_RENDER_L3_PARITY_ERROR_INTERRUPT_S1)
> - dev_priv->l3_parity.which_slice |= 1 << 1;
> + i915->l3_parity.which_slice |= 1 << 1;
>
> if (iir & GT_RENDER_L3_PARITY_ERROR_INTERRUPT)
> - dev_priv->l3_parity.which_slice |= 1 << 0;
> + i915->l3_parity.which_slice |= 1 << 0;
>
> - queue_work(dev_priv->wq, &dev_priv->l3_parity.error_work);
> + queue_work(i915->wq, &i915->l3_parity.error_work);
> }
>
> -static void ilk_gt_irq_handler(struct drm_i915_private *dev_priv,
> +static void ilk_gt_irq_handler(struct drm_i915_private *i915,
> u32 gt_iir)
> {
> if (gt_iir & GT_RENDER_USER_INTERRUPT)
> - intel_engine_breadcrumbs_irq(dev_priv->engine[RCS0]);
> + intel_engine_breadcrumbs_irq(i915->engine[RCS0]);
> if (gt_iir & ILK_BSD_USER_INTERRUPT)
> - intel_engine_breadcrumbs_irq(dev_priv->engine[VCS0]);
> + intel_engine_breadcrumbs_irq(i915->engine[VCS0]);
> }
>
> -static void snb_gt_irq_handler(struct drm_i915_private *dev_priv,
> +static void snb_gt_irq_handler(struct drm_i915_private *i915,
> u32 gt_iir)
> {
> if (gt_iir & GT_RENDER_USER_INTERRUPT)
> - intel_engine_breadcrumbs_irq(dev_priv->engine[RCS0]);
> + intel_engine_breadcrumbs_irq(i915->engine[RCS0]);
> if (gt_iir & GT_BSD_USER_INTERRUPT)
> - intel_engine_breadcrumbs_irq(dev_priv->engine[VCS0]);
> + intel_engine_breadcrumbs_irq(i915->engine[VCS0]);
> if (gt_iir & GT_BLT_USER_INTERRUPT)
> - intel_engine_breadcrumbs_irq(dev_priv->engine[BCS0]);
> + intel_engine_breadcrumbs_irq(i915->engine[BCS0]);
>
> if (gt_iir & (GT_BLT_CS_ERROR_INTERRUPT |
> GT_BSD_CS_ERROR_INTERRUPT |
> GT_RENDER_CS_MASTER_ERROR_INTERRUPT))
> DRM_DEBUG("Command parser error, gt_iir 0x%08x\n", gt_iir);
>
> - if (gt_iir & GT_PARITY_ERROR(dev_priv))
> - ivybridge_parity_error_irq_handler(dev_priv, gt_iir);
> + if (gt_iir & GT_PARITY_ERROR(i915))
> + ivybridge_parity_error_irq_handler(i915, gt_iir);
> }
>
> static void
> @@ -1763,7 +1763,7 @@ static bool i9xx_port_hotplug_long_detect(enum hpd_pin pin, u32 val)
> *
> * Note that the caller is expected to zero out the masks initially.
> */
> -static void intel_get_hpd_pins(struct drm_i915_private *dev_priv,
> +static void intel_get_hpd_pins(struct drm_i915_private *i915,
> u32 *pin_mask, u32 *long_mask,
> u32 hotplug_trigger, u32 dig_hotplug_reg,
> const u32 hpd[HPD_NUM_PINS],
> @@ -1786,25 +1786,25 @@ static void intel_get_hpd_pins(struct drm_i915_private *dev_priv,
>
> }
>
> -static void gmbus_irq_handler(struct drm_i915_private *dev_priv)
> +static void gmbus_irq_handler(struct drm_i915_private *i915)
> {
> - wake_up_all(&dev_priv->gmbus_wait_queue);
> + wake_up_all(&i915->gmbus_wait_queue);
> }
>
> -static void dp_aux_irq_handler(struct drm_i915_private *dev_priv)
> +static void dp_aux_irq_handler(struct drm_i915_private *i915)
> {
> - wake_up_all(&dev_priv->gmbus_wait_queue);
> + wake_up_all(&i915->gmbus_wait_queue);
> }
>
> #if defined(CONFIG_DEBUG_FS)
> -static void display_pipe_crc_irq_handler(struct drm_i915_private *dev_priv,
> +static void display_pipe_crc_irq_handler(struct drm_i915_private *i915,
> enum pipe pipe,
> u32 crc0, u32 crc1,
> u32 crc2, u32 crc3,
> u32 crc4)
> {
> - struct intel_pipe_crc *pipe_crc = &dev_priv->pipe_crc[pipe];
> - struct intel_crtc *crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> + struct intel_pipe_crc *pipe_crc = &i915->pipe_crc[pipe];
> + struct intel_crtc *crtc = intel_get_crtc_for_pipe(i915, pipe);
> u32 crcs[5] = { crc0, crc1, crc2, crc3, crc4 };
>
> trace_intel_pipe_crc(crtc, crcs);
> @@ -1819,7 +1819,7 @@ static void display_pipe_crc_irq_handler(struct drm_i915_private *dev_priv,
> * don't trust that one either.
> */
> if (pipe_crc->skipped <= 0 ||
> - (INTEL_GEN(dev_priv) >= 8 && pipe_crc->skipped == 1)) {
> + (INTEL_GEN(i915) >= 8 && pipe_crc->skipped == 1)) {
> pipe_crc->skipped++;
> spin_unlock(&pipe_crc->lock);
> return;
> @@ -1832,7 +1832,7 @@ static void display_pipe_crc_irq_handler(struct drm_i915_private *dev_priv,
> }
> #else
> static inline void
> -display_pipe_crc_irq_handler(struct drm_i915_private *dev_priv,
> +display_pipe_crc_irq_handler(struct drm_i915_private *i915,
> enum pipe pipe,
> u32 crc0, u32 crc1,
> u32 crc2, u32 crc3,
> @@ -1840,18 +1840,18 @@ display_pipe_crc_irq_handler(struct drm_i915_private *dev_priv,
> #endif
>
>
> -static void hsw_pipe_crc_irq_handler(struct drm_i915_private *dev_priv,
> +static void hsw_pipe_crc_irq_handler(struct drm_i915_private *i915,
> enum pipe pipe)
> {
> - display_pipe_crc_irq_handler(dev_priv, pipe,
> + display_pipe_crc_irq_handler(i915, pipe,
> I915_READ(PIPE_CRC_RES_1_IVB(pipe)),
> 0, 0, 0, 0);
> }
>
> -static void ivb_pipe_crc_irq_handler(struct drm_i915_private *dev_priv,
> +static void ivb_pipe_crc_irq_handler(struct drm_i915_private *i915,
> enum pipe pipe)
> {
> - display_pipe_crc_irq_handler(dev_priv, pipe,
> + display_pipe_crc_irq_handler(i915, pipe,
> I915_READ(PIPE_CRC_RES_1_IVB(pipe)),
> I915_READ(PIPE_CRC_RES_2_IVB(pipe)),
> I915_READ(PIPE_CRC_RES_3_IVB(pipe)),
> @@ -1859,22 +1859,22 @@ static void ivb_pipe_crc_irq_handler(struct drm_i915_private *dev_priv,
> I915_READ(PIPE_CRC_RES_5_IVB(pipe)));
> }
>
> -static void i9xx_pipe_crc_irq_handler(struct drm_i915_private *dev_priv,
> +static void i9xx_pipe_crc_irq_handler(struct drm_i915_private *i915,
> enum pipe pipe)
> {
> u32 res1, res2;
>
> - if (INTEL_GEN(dev_priv) >= 3)
> + if (INTEL_GEN(i915) >= 3)
> res1 = I915_READ(PIPE_CRC_RES_RES1_I915(pipe));
> else
> res1 = 0;
>
> - if (INTEL_GEN(dev_priv) >= 5 || IS_G4X(dev_priv))
> + if (INTEL_GEN(i915) >= 5 || IS_G4X(i915))
> res2 = I915_READ(PIPE_CRC_RES_RES2_G4X(pipe));
> else
> res2 = 0;
>
> - display_pipe_crc_irq_handler(dev_priv, pipe,
> + display_pipe_crc_irq_handler(i915, pipe,
> I915_READ(PIPE_CRC_RES_RED(pipe)),
> I915_READ(PIPE_CRC_RES_GREEN(pipe)),
> I915_READ(PIPE_CRC_RES_BLUE(pipe)),
> @@ -1903,34 +1903,34 @@ static void gen11_rps_irq_handler(struct drm_i915_private *i915, u32 pm_iir)
> schedule_work(&rps->work);
> }
>
> -static void gen6_rps_irq_handler(struct drm_i915_private *dev_priv, u32 pm_iir)
> +static void gen6_rps_irq_handler(struct drm_i915_private *i915, u32 pm_iir)
> {
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct intel_rps *rps = &i915->gt_pm.rps;
>
> - if (pm_iir & dev_priv->pm_rps_events) {
> - spin_lock(&dev_priv->irq_lock);
> - gen6_mask_pm_irq(dev_priv, pm_iir & dev_priv->pm_rps_events);
> + if (pm_iir & i915->pm_rps_events) {
> + spin_lock(&i915->irq_lock);
> + gen6_mask_pm_irq(i915, pm_iir & i915->pm_rps_events);
> if (rps->interrupts_enabled) {
> - rps->pm_iir |= pm_iir & dev_priv->pm_rps_events;
> + rps->pm_iir |= pm_iir & i915->pm_rps_events;
> schedule_work(&rps->work);
> }
> - spin_unlock(&dev_priv->irq_lock);
> + spin_unlock(&i915->irq_lock);
> }
>
> - if (INTEL_GEN(dev_priv) >= 8)
> + if (INTEL_GEN(i915) >= 8)
> return;
>
> if (pm_iir & PM_VEBOX_USER_INTERRUPT)
> - intel_engine_breadcrumbs_irq(dev_priv->engine[VECS0]);
> + intel_engine_breadcrumbs_irq(i915->engine[VECS0]);
>
> if (pm_iir & PM_VEBOX_CS_ERROR_INTERRUPT)
> DRM_DEBUG("Command parser error, pm_iir 0x%08x\n", pm_iir);
> }
>
> -static void gen9_guc_irq_handler(struct drm_i915_private *dev_priv, u32 gt_iir)
> +static void gen9_guc_irq_handler(struct drm_i915_private *i915, u32 gt_iir)
> {
> if (gt_iir & GEN9_GUC_TO_HOST_INT_EVENT)
> - intel_guc_to_host_event_handler(&dev_priv->guc);
> + intel_guc_to_host_event_handler(&i915->guc);
> }
>
> static void gen11_guc_irq_handler(struct drm_i915_private *i915, u16 iir)
> @@ -1939,32 +1939,32 @@ static void gen11_guc_irq_handler(struct drm_i915_private *i915, u16 iir)
> intel_guc_to_host_event_handler(&i915->guc);
> }
>
> -static void i9xx_pipestat_irq_reset(struct drm_i915_private *dev_priv)
> +static void i9xx_pipestat_irq_reset(struct drm_i915_private *i915)
> {
> enum pipe pipe;
>
> - for_each_pipe(dev_priv, pipe) {
> + for_each_pipe(i915, pipe) {
> I915_WRITE(PIPESTAT(pipe),
> PIPESTAT_INT_STATUS_MASK |
> PIPE_FIFO_UNDERRUN_STATUS);
>
> - dev_priv->pipestat_irq_mask[pipe] = 0;
> + i915->pipestat_irq_mask[pipe] = 0;
> }
> }
>
> -static void i9xx_pipestat_irq_ack(struct drm_i915_private *dev_priv,
> +static void i9xx_pipestat_irq_ack(struct drm_i915_private *i915,
> u32 iir, u32 pipe_stats[I915_MAX_PIPES])
> {
> int pipe;
>
> - spin_lock(&dev_priv->irq_lock);
> + spin_lock(&i915->irq_lock);
>
> - if (!dev_priv->display_irqs_enabled) {
> - spin_unlock(&dev_priv->irq_lock);
> + if (!i915->display_irqs_enabled) {
> + spin_unlock(&i915->irq_lock);
> return;
> }
>
> - for_each_pipe(dev_priv, pipe) {
> + for_each_pipe(i915, pipe) {
> i915_reg_t reg;
> u32 status_mask, enable_mask, iir_bit = 0;
>
> @@ -1991,14 +1991,14 @@ static void i9xx_pipestat_irq_ack(struct drm_i915_private *dev_priv,
> break;
> }
> if (iir & iir_bit)
> - status_mask |= dev_priv->pipestat_irq_mask[pipe];
> + status_mask |= i915->pipestat_irq_mask[pipe];
>
> if (!status_mask)
> continue;
>
> reg = PIPESTAT(pipe);
> pipe_stats[pipe] = I915_READ(reg) & status_mask;
> - enable_mask = i915_pipestat_enable_mask(dev_priv, pipe);
> + enable_mask = i915_pipestat_enable_mask(i915, pipe);
>
> /*
> * Clear the PIPE*STAT regs before the IIR
> @@ -2014,104 +2014,104 @@ static void i9xx_pipestat_irq_ack(struct drm_i915_private *dev_priv,
> I915_WRITE(reg, enable_mask);
> }
> }
> - spin_unlock(&dev_priv->irq_lock);
> + spin_unlock(&i915->irq_lock);
> }
>
> -static void i8xx_pipestat_irq_handler(struct drm_i915_private *dev_priv,
> +static void i8xx_pipestat_irq_handler(struct drm_i915_private *i915,
> u16 iir, u32 pipe_stats[I915_MAX_PIPES])
> {
> enum pipe pipe;
>
> - for_each_pipe(dev_priv, pipe) {
> + for_each_pipe(i915, pipe) {
> if (pipe_stats[pipe] & PIPE_VBLANK_INTERRUPT_STATUS)
> - drm_handle_vblank(&dev_priv->drm, pipe);
> + drm_handle_vblank(&i915->drm, pipe);
>
> if (pipe_stats[pipe] & PIPE_CRC_DONE_INTERRUPT_STATUS)
> - i9xx_pipe_crc_irq_handler(dev_priv, pipe);
> + i9xx_pipe_crc_irq_handler(i915, pipe);
>
> if (pipe_stats[pipe] & PIPE_FIFO_UNDERRUN_STATUS)
> - intel_cpu_fifo_underrun_irq_handler(dev_priv, pipe);
> + intel_cpu_fifo_underrun_irq_handler(i915, pipe);
> }
> }
>
> -static void i915_pipestat_irq_handler(struct drm_i915_private *dev_priv,
> +static void i915_pipestat_irq_handler(struct drm_i915_private *i915,
> u32 iir, u32 pipe_stats[I915_MAX_PIPES])
> {
> bool blc_event = false;
> enum pipe pipe;
>
> - for_each_pipe(dev_priv, pipe) {
> + for_each_pipe(i915, pipe) {
> if (pipe_stats[pipe] & PIPE_VBLANK_INTERRUPT_STATUS)
> - drm_handle_vblank(&dev_priv->drm, pipe);
> + drm_handle_vblank(&i915->drm, pipe);
>
> if (pipe_stats[pipe] & PIPE_LEGACY_BLC_EVENT_STATUS)
> blc_event = true;
>
> if (pipe_stats[pipe] & PIPE_CRC_DONE_INTERRUPT_STATUS)
> - i9xx_pipe_crc_irq_handler(dev_priv, pipe);
> + i9xx_pipe_crc_irq_handler(i915, pipe);
>
> if (pipe_stats[pipe] & PIPE_FIFO_UNDERRUN_STATUS)
> - intel_cpu_fifo_underrun_irq_handler(dev_priv, pipe);
> + intel_cpu_fifo_underrun_irq_handler(i915, pipe);
> }
>
> if (blc_event || (iir & I915_ASLE_INTERRUPT))
> - intel_opregion_asle_intr(dev_priv);
> + intel_opregion_asle_intr(i915);
> }
>
> -static void i965_pipestat_irq_handler(struct drm_i915_private *dev_priv,
> +static void i965_pipestat_irq_handler(struct drm_i915_private *i915,
> u32 iir, u32 pipe_stats[I915_MAX_PIPES])
> {
> bool blc_event = false;
> enum pipe pipe;
>
> - for_each_pipe(dev_priv, pipe) {
> + for_each_pipe(i915, pipe) {
> if (pipe_stats[pipe] & PIPE_START_VBLANK_INTERRUPT_STATUS)
> - drm_handle_vblank(&dev_priv->drm, pipe);
> + drm_handle_vblank(&i915->drm, pipe);
>
> if (pipe_stats[pipe] & PIPE_LEGACY_BLC_EVENT_STATUS)
> blc_event = true;
>
> if (pipe_stats[pipe] & PIPE_CRC_DONE_INTERRUPT_STATUS)
> - i9xx_pipe_crc_irq_handler(dev_priv, pipe);
> + i9xx_pipe_crc_irq_handler(i915, pipe);
>
> if (pipe_stats[pipe] & PIPE_FIFO_UNDERRUN_STATUS)
> - intel_cpu_fifo_underrun_irq_handler(dev_priv, pipe);
> + intel_cpu_fifo_underrun_irq_handler(i915, pipe);
> }
>
> if (blc_event || (iir & I915_ASLE_INTERRUPT))
> - intel_opregion_asle_intr(dev_priv);
> + intel_opregion_asle_intr(i915);
>
> if (pipe_stats[0] & PIPE_GMBUS_INTERRUPT_STATUS)
> - gmbus_irq_handler(dev_priv);
> + gmbus_irq_handler(i915);
> }
>
> -static void valleyview_pipestat_irq_handler(struct drm_i915_private *dev_priv,
> +static void valleyview_pipestat_irq_handler(struct drm_i915_private *i915,
> u32 pipe_stats[I915_MAX_PIPES])
> {
> enum pipe pipe;
>
> - for_each_pipe(dev_priv, pipe) {
> + for_each_pipe(i915, pipe) {
> if (pipe_stats[pipe] & PIPE_START_VBLANK_INTERRUPT_STATUS)
> - drm_handle_vblank(&dev_priv->drm, pipe);
> + drm_handle_vblank(&i915->drm, pipe);
>
> if (pipe_stats[pipe] & PIPE_CRC_DONE_INTERRUPT_STATUS)
> - i9xx_pipe_crc_irq_handler(dev_priv, pipe);
> + i9xx_pipe_crc_irq_handler(i915, pipe);
>
> if (pipe_stats[pipe] & PIPE_FIFO_UNDERRUN_STATUS)
> - intel_cpu_fifo_underrun_irq_handler(dev_priv, pipe);
> + intel_cpu_fifo_underrun_irq_handler(i915, pipe);
> }
>
> if (pipe_stats[0] & PIPE_GMBUS_INTERRUPT_STATUS)
> - gmbus_irq_handler(dev_priv);
> + gmbus_irq_handler(i915);
> }
>
> -static u32 i9xx_hpd_irq_ack(struct drm_i915_private *dev_priv)
> +static u32 i9xx_hpd_irq_ack(struct drm_i915_private *i915)
> {
> u32 hotplug_status = 0, hotplug_status_mask;
> int i;
>
> - if (IS_G4X(dev_priv) ||
> - IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> + if (IS_G4X(i915) ||
> + IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> hotplug_status_mask = HOTPLUG_INT_STATUS_G4X |
> DP_AUX_CHANNEL_MASK_INT_STATUS_G4X;
> else
> @@ -2143,35 +2143,35 @@ static u32 i9xx_hpd_irq_ack(struct drm_i915_private *dev_priv)
> return hotplug_status;
> }
>
> -static void i9xx_hpd_irq_handler(struct drm_i915_private *dev_priv,
> +static void i9xx_hpd_irq_handler(struct drm_i915_private *i915,
> u32 hotplug_status)
> {
> u32 pin_mask = 0, long_mask = 0;
>
> - if (IS_G4X(dev_priv) || IS_VALLEYVIEW(dev_priv) ||
> - IS_CHERRYVIEW(dev_priv)) {
> + if (IS_G4X(i915) || IS_VALLEYVIEW(i915) ||
> + IS_CHERRYVIEW(i915)) {
> u32 hotplug_trigger = hotplug_status & HOTPLUG_INT_STATUS_G4X;
>
> if (hotplug_trigger) {
> - intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask,
> + intel_get_hpd_pins(i915, &pin_mask, &long_mask,
> hotplug_trigger, hotplug_trigger,
> hpd_status_g4x,
> i9xx_port_hotplug_long_detect);
>
> - intel_hpd_irq_handler(dev_priv, pin_mask, long_mask);
> + intel_hpd_irq_handler(i915, pin_mask, long_mask);
> }
>
> if (hotplug_status & DP_AUX_CHANNEL_MASK_INT_STATUS_G4X)
> - dp_aux_irq_handler(dev_priv);
> + dp_aux_irq_handler(i915);
> } else {
> u32 hotplug_trigger = hotplug_status & HOTPLUG_INT_STATUS_I915;
>
> if (hotplug_trigger) {
> - intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask,
> + intel_get_hpd_pins(i915, &pin_mask, &long_mask,
> hotplug_trigger, hotplug_trigger,
> hpd_status_i915,
> i9xx_port_hotplug_long_detect);
> - intel_hpd_irq_handler(dev_priv, pin_mask, long_mask);
> + intel_hpd_irq_handler(i915, pin_mask, long_mask);
> }
> }
> }
> @@ -2179,14 +2179,14 @@ static void i9xx_hpd_irq_handler(struct drm_i915_private *dev_priv,
> static irqreturn_t valleyview_irq_handler(int irq, void *arg)
> {
> struct drm_device *dev = arg;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> irqreturn_t ret = IRQ_NONE;
>
> - if (!intel_irqs_enabled(dev_priv))
> + if (!intel_irqs_enabled(i915))
> return IRQ_NONE;
>
> /* IRQs are synced during runtime_suspend, we don't require a wakeref */
> - disable_rpm_wakeref_asserts(dev_priv);
> + disable_rpm_wakeref_asserts(i915);
>
> do {
> u32 iir, gt_iir, pm_iir;
> @@ -2226,15 +2226,15 @@ static irqreturn_t valleyview_irq_handler(int irq, void *arg)
> I915_WRITE(GEN6_PMIIR, pm_iir);
>
> if (iir & I915_DISPLAY_PORT_INTERRUPT)
> - hotplug_status = i9xx_hpd_irq_ack(dev_priv);
> + hotplug_status = i9xx_hpd_irq_ack(i915);
>
> /* Call regardless, as some status bits might not be
> * signalled in iir */
> - i9xx_pipestat_irq_ack(dev_priv, iir, pipe_stats);
> + i9xx_pipestat_irq_ack(i915, iir, pipe_stats);
>
> if (iir & (I915_LPE_PIPE_A_INTERRUPT |
> I915_LPE_PIPE_B_INTERRUPT))
> - intel_lpe_audio_irq_handler(dev_priv);
> + intel_lpe_audio_irq_handler(i915);
>
> /*
> * VLV_IIR is single buffered, and reflects the level
> @@ -2247,17 +2247,17 @@ static irqreturn_t valleyview_irq_handler(int irq, void *arg)
> I915_WRITE(VLV_MASTER_IER, MASTER_INTERRUPT_ENABLE);
>
> if (gt_iir)
> - snb_gt_irq_handler(dev_priv, gt_iir);
> + snb_gt_irq_handler(i915, gt_iir);
> if (pm_iir)
> - gen6_rps_irq_handler(dev_priv, pm_iir);
> + gen6_rps_irq_handler(i915, pm_iir);
>
> if (hotplug_status)
> - i9xx_hpd_irq_handler(dev_priv, hotplug_status);
> + i9xx_hpd_irq_handler(i915, hotplug_status);
>
> - valleyview_pipestat_irq_handler(dev_priv, pipe_stats);
> + valleyview_pipestat_irq_handler(i915, pipe_stats);
> } while (0);
>
> - enable_rpm_wakeref_asserts(dev_priv);
> + enable_rpm_wakeref_asserts(i915);
>
> return ret;
> }
> @@ -2265,14 +2265,14 @@ static irqreturn_t valleyview_irq_handler(int irq, void *arg)
> static irqreturn_t cherryview_irq_handler(int irq, void *arg)
> {
> struct drm_device *dev = arg;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> irqreturn_t ret = IRQ_NONE;
>
> - if (!intel_irqs_enabled(dev_priv))
> + if (!intel_irqs_enabled(i915))
> return IRQ_NONE;
>
> /* IRQs are synced during runtime_suspend, we don't require a wakeref */
> - disable_rpm_wakeref_asserts(dev_priv);
> + disable_rpm_wakeref_asserts(i915);
>
> do {
> u32 master_ctl, iir;
> @@ -2306,19 +2306,19 @@ static irqreturn_t cherryview_irq_handler(int irq, void *arg)
> ier = I915_READ(VLV_IER);
> I915_WRITE(VLV_IER, 0);
>
> - gen8_gt_irq_ack(dev_priv, master_ctl, gt_iir);
> + gen8_gt_irq_ack(i915, master_ctl, gt_iir);
>
> if (iir & I915_DISPLAY_PORT_INTERRUPT)
> - hotplug_status = i9xx_hpd_irq_ack(dev_priv);
> + hotplug_status = i9xx_hpd_irq_ack(i915);
>
> /* Call regardless, as some status bits might not be
> * signalled in iir */
> - i9xx_pipestat_irq_ack(dev_priv, iir, pipe_stats);
> + i9xx_pipestat_irq_ack(i915, iir, pipe_stats);
>
> if (iir & (I915_LPE_PIPE_A_INTERRUPT |
> I915_LPE_PIPE_B_INTERRUPT |
> I915_LPE_PIPE_C_INTERRUPT))
> - intel_lpe_audio_irq_handler(dev_priv);
> + intel_lpe_audio_irq_handler(i915);
>
> /*
> * VLV_IIR is single buffered, and reflects the level
> @@ -2330,20 +2330,20 @@ static irqreturn_t cherryview_irq_handler(int irq, void *arg)
> I915_WRITE(VLV_IER, ier);
> I915_WRITE(GEN8_MASTER_IRQ, GEN8_MASTER_IRQ_CONTROL);
>
> - gen8_gt_irq_handler(dev_priv, master_ctl, gt_iir);
> + gen8_gt_irq_handler(i915, master_ctl, gt_iir);
>
> if (hotplug_status)
> - i9xx_hpd_irq_handler(dev_priv, hotplug_status);
> + i9xx_hpd_irq_handler(i915, hotplug_status);
>
> - valleyview_pipestat_irq_handler(dev_priv, pipe_stats);
> + valleyview_pipestat_irq_handler(i915, pipe_stats);
> } while (0);
>
> - enable_rpm_wakeref_asserts(dev_priv);
> + enable_rpm_wakeref_asserts(i915);
>
> return ret;
> }
>
> -static void ibx_hpd_irq_handler(struct drm_i915_private *dev_priv,
> +static void ibx_hpd_irq_handler(struct drm_i915_private *i915,
> u32 hotplug_trigger,
> const u32 hpd[HPD_NUM_PINS])
> {
> @@ -2368,19 +2368,19 @@ static void ibx_hpd_irq_handler(struct drm_i915_private *dev_priv,
> if (!hotplug_trigger)
> return;
>
> - intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask, hotplug_trigger,
> + intel_get_hpd_pins(i915, &pin_mask, &long_mask, hotplug_trigger,
> dig_hotplug_reg, hpd,
> pch_port_hotplug_long_detect);
>
> - intel_hpd_irq_handler(dev_priv, pin_mask, long_mask);
> + intel_hpd_irq_handler(i915, pin_mask, long_mask);
> }
>
> -static void ibx_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
> +static void ibx_irq_handler(struct drm_i915_private *i915, u32 pch_iir)
> {
> int pipe;
> u32 hotplug_trigger = pch_iir & SDE_HOTPLUG_MASK;
>
> - ibx_hpd_irq_handler(dev_priv, hotplug_trigger, hpd_ibx);
> + ibx_hpd_irq_handler(i915, hotplug_trigger, hpd_ibx);
>
> if (pch_iir & SDE_AUDIO_POWER_MASK) {
> int port = ffs((pch_iir & SDE_AUDIO_POWER_MASK) >>
> @@ -2390,10 +2390,10 @@ static void ibx_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
> }
>
> if (pch_iir & SDE_AUX_MASK)
> - dp_aux_irq_handler(dev_priv);
> + dp_aux_irq_handler(i915);
>
> if (pch_iir & SDE_GMBUS)
> - gmbus_irq_handler(dev_priv);
> + gmbus_irq_handler(i915);
>
> if (pch_iir & SDE_AUDIO_HDCP_MASK)
> DRM_DEBUG_DRIVER("PCH HDCP audio interrupt\n");
> @@ -2405,7 +2405,7 @@ static void ibx_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
> DRM_ERROR("PCH poison interrupt\n");
>
> if (pch_iir & SDE_FDI_MASK)
> - for_each_pipe(dev_priv, pipe)
> + for_each_pipe(i915, pipe)
> DRM_DEBUG_DRIVER(" pipe %c FDI IIR: 0x%08x\n",
> pipe_name(pipe),
> I915_READ(FDI_RX_IIR(pipe)));
> @@ -2417,13 +2417,13 @@ static void ibx_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
> DRM_DEBUG_DRIVER("PCH transcoder CRC error interrupt\n");
>
> if (pch_iir & SDE_TRANSA_FIFO_UNDER)
> - intel_pch_fifo_underrun_irq_handler(dev_priv, PIPE_A);
> + intel_pch_fifo_underrun_irq_handler(i915, PIPE_A);
>
> if (pch_iir & SDE_TRANSB_FIFO_UNDER)
> - intel_pch_fifo_underrun_irq_handler(dev_priv, PIPE_B);
> + intel_pch_fifo_underrun_irq_handler(i915, PIPE_B);
> }
>
> -static void ivb_err_int_handler(struct drm_i915_private *dev_priv)
> +static void ivb_err_int_handler(struct drm_i915_private *i915)
> {
> u32 err_int = I915_READ(GEN7_ERR_INT);
> enum pipe pipe;
> @@ -2431,22 +2431,22 @@ static void ivb_err_int_handler(struct drm_i915_private *dev_priv)
> if (err_int & ERR_INT_POISON)
> DRM_ERROR("Poison interrupt\n");
>
> - for_each_pipe(dev_priv, pipe) {
> + for_each_pipe(i915, pipe) {
> if (err_int & ERR_INT_FIFO_UNDERRUN(pipe))
> - intel_cpu_fifo_underrun_irq_handler(dev_priv, pipe);
> + intel_cpu_fifo_underrun_irq_handler(i915, pipe);
>
> if (err_int & ERR_INT_PIPE_CRC_DONE(pipe)) {
> - if (IS_IVYBRIDGE(dev_priv))
> - ivb_pipe_crc_irq_handler(dev_priv, pipe);
> + if (IS_IVYBRIDGE(i915))
> + ivb_pipe_crc_irq_handler(i915, pipe);
> else
> - hsw_pipe_crc_irq_handler(dev_priv, pipe);
> + hsw_pipe_crc_irq_handler(i915, pipe);
> }
> }
>
> I915_WRITE(GEN7_ERR_INT, err_int);
> }
>
> -static void cpt_serr_int_handler(struct drm_i915_private *dev_priv)
> +static void cpt_serr_int_handler(struct drm_i915_private *i915)
> {
> u32 serr_int = I915_READ(SERR_INT);
> enum pipe pipe;
> @@ -2454,19 +2454,19 @@ static void cpt_serr_int_handler(struct drm_i915_private *dev_priv)
> if (serr_int & SERR_INT_POISON)
> DRM_ERROR("PCH poison interrupt\n");
>
> - for_each_pipe(dev_priv, pipe)
> + for_each_pipe(i915, pipe)
> if (serr_int & SERR_INT_TRANS_FIFO_UNDERRUN(pipe))
> - intel_pch_fifo_underrun_irq_handler(dev_priv, pipe);
> + intel_pch_fifo_underrun_irq_handler(i915, pipe);
>
> I915_WRITE(SERR_INT, serr_int);
> }
>
> -static void cpt_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
> +static void cpt_irq_handler(struct drm_i915_private *i915, u32 pch_iir)
> {
> int pipe;
> u32 hotplug_trigger = pch_iir & SDE_HOTPLUG_MASK_CPT;
>
> - ibx_hpd_irq_handler(dev_priv, hotplug_trigger, hpd_cpt);
> + ibx_hpd_irq_handler(i915, hotplug_trigger, hpd_cpt);
>
> if (pch_iir & SDE_AUDIO_POWER_MASK_CPT) {
> int port = ffs((pch_iir & SDE_AUDIO_POWER_MASK_CPT) >>
> @@ -2476,10 +2476,10 @@ static void cpt_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
> }
>
> if (pch_iir & SDE_AUX_MASK_CPT)
> - dp_aux_irq_handler(dev_priv);
> + dp_aux_irq_handler(i915);
>
> if (pch_iir & SDE_GMBUS_CPT)
> - gmbus_irq_handler(dev_priv);
> + gmbus_irq_handler(i915);
>
> if (pch_iir & SDE_AUDIO_CP_REQ_CPT)
> DRM_DEBUG_DRIVER("Audio CP request interrupt\n");
> @@ -2488,16 +2488,16 @@ static void cpt_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
> DRM_DEBUG_DRIVER("Audio CP change interrupt\n");
>
> if (pch_iir & SDE_FDI_MASK_CPT)
> - for_each_pipe(dev_priv, pipe)
> + for_each_pipe(i915, pipe)
> DRM_DEBUG_DRIVER(" pipe %c FDI IIR: 0x%08x\n",
> pipe_name(pipe),
> I915_READ(FDI_RX_IIR(pipe)));
>
> if (pch_iir & SDE_ERROR_CPT)
> - cpt_serr_int_handler(dev_priv);
> + cpt_serr_int_handler(i915);
> }
>
> -static void icp_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
> +static void icp_irq_handler(struct drm_i915_private *i915, u32 pch_iir)
> {
> u32 ddi_hotplug_trigger = pch_iir & SDE_DDI_MASK_ICP;
> u32 tc_hotplug_trigger = pch_iir & SDE_TC_MASK_ICP;
> @@ -2509,7 +2509,7 @@ static void icp_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
> dig_hotplug_reg = I915_READ(SHOTPLUG_CTL_DDI);
> I915_WRITE(SHOTPLUG_CTL_DDI, dig_hotplug_reg);
>
> - intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask,
> + intel_get_hpd_pins(i915, &pin_mask, &long_mask,
> ddi_hotplug_trigger,
> dig_hotplug_reg, hpd_icp,
> icp_ddi_port_hotplug_long_detect);
> @@ -2521,20 +2521,20 @@ static void icp_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
> dig_hotplug_reg = I915_READ(SHOTPLUG_CTL_TC);
> I915_WRITE(SHOTPLUG_CTL_TC, dig_hotplug_reg);
>
> - intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask,
> + intel_get_hpd_pins(i915, &pin_mask, &long_mask,
> tc_hotplug_trigger,
> dig_hotplug_reg, hpd_icp,
> icp_tc_port_hotplug_long_detect);
> }
>
> if (pin_mask)
> - intel_hpd_irq_handler(dev_priv, pin_mask, long_mask);
> + intel_hpd_irq_handler(i915, pin_mask, long_mask);
>
> if (pch_iir & SDE_GMBUS_ICP)
> - gmbus_irq_handler(dev_priv);
> + gmbus_irq_handler(i915);
> }
>
> -static void spt_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
> +static void spt_irq_handler(struct drm_i915_private *i915, u32 pch_iir)
> {
> u32 hotplug_trigger = pch_iir & SDE_HOTPLUG_MASK_SPT &
> ~SDE_PORTE_HOTPLUG_SPT;
> @@ -2547,7 +2547,7 @@ static void spt_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
> dig_hotplug_reg = I915_READ(PCH_PORT_HOTPLUG);
> I915_WRITE(PCH_PORT_HOTPLUG, dig_hotplug_reg);
>
> - intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask,
> + intel_get_hpd_pins(i915, &pin_mask, &long_mask,
> hotplug_trigger, dig_hotplug_reg, hpd_spt,
> spt_port_hotplug_long_detect);
> }
> @@ -2558,19 +2558,19 @@ static void spt_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
> dig_hotplug_reg = I915_READ(PCH_PORT_HOTPLUG2);
> I915_WRITE(PCH_PORT_HOTPLUG2, dig_hotplug_reg);
>
> - intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask,
> + intel_get_hpd_pins(i915, &pin_mask, &long_mask,
> hotplug2_trigger, dig_hotplug_reg, hpd_spt,
> spt_port_hotplug2_long_detect);
> }
>
> if (pin_mask)
> - intel_hpd_irq_handler(dev_priv, pin_mask, long_mask);
> + intel_hpd_irq_handler(i915, pin_mask, long_mask);
>
> if (pch_iir & SDE_GMBUS_CPT)
> - gmbus_irq_handler(dev_priv);
> + gmbus_irq_handler(i915);
> }
>
> -static void ilk_hpd_irq_handler(struct drm_i915_private *dev_priv,
> +static void ilk_hpd_irq_handler(struct drm_i915_private *i915,
> u32 hotplug_trigger,
> const u32 hpd[HPD_NUM_PINS])
> {
> @@ -2579,94 +2579,94 @@ static void ilk_hpd_irq_handler(struct drm_i915_private *dev_priv,
> dig_hotplug_reg = I915_READ(DIGITAL_PORT_HOTPLUG_CNTRL);
> I915_WRITE(DIGITAL_PORT_HOTPLUG_CNTRL, dig_hotplug_reg);
>
> - intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask, hotplug_trigger,
> + intel_get_hpd_pins(i915, &pin_mask, &long_mask, hotplug_trigger,
> dig_hotplug_reg, hpd,
> ilk_port_hotplug_long_detect);
>
> - intel_hpd_irq_handler(dev_priv, pin_mask, long_mask);
> + intel_hpd_irq_handler(i915, pin_mask, long_mask);
> }
>
> -static void ilk_display_irq_handler(struct drm_i915_private *dev_priv,
> +static void ilk_display_irq_handler(struct drm_i915_private *i915,
> u32 de_iir)
> {
> enum pipe pipe;
> u32 hotplug_trigger = de_iir & DE_DP_A_HOTPLUG;
>
> if (hotplug_trigger)
> - ilk_hpd_irq_handler(dev_priv, hotplug_trigger, hpd_ilk);
> + ilk_hpd_irq_handler(i915, hotplug_trigger, hpd_ilk);
>
> if (de_iir & DE_AUX_CHANNEL_A)
> - dp_aux_irq_handler(dev_priv);
> + dp_aux_irq_handler(i915);
>
> if (de_iir & DE_GSE)
> - intel_opregion_asle_intr(dev_priv);
> + intel_opregion_asle_intr(i915);
>
> if (de_iir & DE_POISON)
> DRM_ERROR("Poison interrupt\n");
>
> - for_each_pipe(dev_priv, pipe) {
> + for_each_pipe(i915, pipe) {
> if (de_iir & DE_PIPE_VBLANK(pipe))
> - drm_handle_vblank(&dev_priv->drm, pipe);
> + drm_handle_vblank(&i915->drm, pipe);
>
> if (de_iir & DE_PIPE_FIFO_UNDERRUN(pipe))
> - intel_cpu_fifo_underrun_irq_handler(dev_priv, pipe);
> + intel_cpu_fifo_underrun_irq_handler(i915, pipe);
>
> if (de_iir & DE_PIPE_CRC_DONE(pipe))
> - i9xx_pipe_crc_irq_handler(dev_priv, pipe);
> + i9xx_pipe_crc_irq_handler(i915, pipe);
> }
>
> /* check event from PCH */
> if (de_iir & DE_PCH_EVENT) {
> u32 pch_iir = I915_READ(SDEIIR);
>
> - if (HAS_PCH_CPT(dev_priv))
> - cpt_irq_handler(dev_priv, pch_iir);
> + if (HAS_PCH_CPT(i915))
> + cpt_irq_handler(i915, pch_iir);
> else
> - ibx_irq_handler(dev_priv, pch_iir);
> + ibx_irq_handler(i915, pch_iir);
>
> /* should clear PCH hotplug event before clear CPU irq */
> I915_WRITE(SDEIIR, pch_iir);
> }
>
> - if (IS_GEN(dev_priv, 5) && de_iir & DE_PCU_EVENT)
> - ironlake_rps_change_irq_handler(dev_priv);
> + if (IS_GEN(i915, 5) && de_iir & DE_PCU_EVENT)
> + ironlake_rps_change_irq_handler(i915);
> }
>
> -static void ivb_display_irq_handler(struct drm_i915_private *dev_priv,
> +static void ivb_display_irq_handler(struct drm_i915_private *i915,
> u32 de_iir)
> {
> enum pipe pipe;
> u32 hotplug_trigger = de_iir & DE_DP_A_HOTPLUG_IVB;
>
> if (hotplug_trigger)
> - ilk_hpd_irq_handler(dev_priv, hotplug_trigger, hpd_ivb);
> + ilk_hpd_irq_handler(i915, hotplug_trigger, hpd_ivb);
>
> if (de_iir & DE_ERR_INT_IVB)
> - ivb_err_int_handler(dev_priv);
> + ivb_err_int_handler(i915);
>
> if (de_iir & DE_EDP_PSR_INT_HSW) {
> u32 psr_iir = I915_READ(EDP_PSR_IIR);
>
> - intel_psr_irq_handler(dev_priv, psr_iir);
> + intel_psr_irq_handler(i915, psr_iir);
> I915_WRITE(EDP_PSR_IIR, psr_iir);
> }
>
> if (de_iir & DE_AUX_CHANNEL_A_IVB)
> - dp_aux_irq_handler(dev_priv);
> + dp_aux_irq_handler(i915);
>
> if (de_iir & DE_GSE_IVB)
> - intel_opregion_asle_intr(dev_priv);
> + intel_opregion_asle_intr(i915);
>
> - for_each_pipe(dev_priv, pipe) {
> + for_each_pipe(i915, pipe) {
> if (de_iir & (DE_PIPE_VBLANK_IVB(pipe)))
> - drm_handle_vblank(&dev_priv->drm, pipe);
> + drm_handle_vblank(&i915->drm, pipe);
> }
>
> /* check event from PCH */
> - if (!HAS_PCH_NOP(dev_priv) && (de_iir & DE_PCH_EVENT_IVB)) {
> + if (!HAS_PCH_NOP(i915) && (de_iir & DE_PCH_EVENT_IVB)) {
> u32 pch_iir = I915_READ(SDEIIR);
>
> - cpt_irq_handler(dev_priv, pch_iir);
> + cpt_irq_handler(i915, pch_iir);
>
> /* clear PCH hotplug event before clear CPU irq */
> I915_WRITE(SDEIIR, pch_iir);
> @@ -2684,15 +2684,15 @@ static void ivb_display_irq_handler(struct drm_i915_private *dev_priv,
> static irqreturn_t ironlake_irq_handler(int irq, void *arg)
> {
> struct drm_device *dev = arg;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> u32 de_iir, gt_iir, de_ier, sde_ier = 0;
> irqreturn_t ret = IRQ_NONE;
>
> - if (!intel_irqs_enabled(dev_priv))
> + if (!intel_irqs_enabled(i915))
> return IRQ_NONE;
>
> /* IRQs are synced during runtime_suspend, we don't require a wakeref */
> - disable_rpm_wakeref_asserts(dev_priv);
> + disable_rpm_wakeref_asserts(i915);
>
> /* disable master interrupt before clearing iir */
> de_ier = I915_READ(DEIER);
> @@ -2703,7 +2703,7 @@ static irqreturn_t ironlake_irq_handler(int irq, void *arg)
> * able to process them after we restore SDEIER (as soon as we restore
> * it, we'll get an interrupt if SDEIIR still has something to process
> * due to its back queue). */
> - if (!HAS_PCH_NOP(dev_priv)) {
> + if (!HAS_PCH_NOP(i915)) {
> sde_ier = I915_READ(SDEIER);
> I915_WRITE(SDEIER, 0);
> }
> @@ -2714,42 +2714,42 @@ static irqreturn_t ironlake_irq_handler(int irq, void *arg)
> if (gt_iir) {
> I915_WRITE(GTIIR, gt_iir);
> ret = IRQ_HANDLED;
> - if (INTEL_GEN(dev_priv) >= 6)
> - snb_gt_irq_handler(dev_priv, gt_iir);
> + if (INTEL_GEN(i915) >= 6)
> + snb_gt_irq_handler(i915, gt_iir);
> else
> - ilk_gt_irq_handler(dev_priv, gt_iir);
> + ilk_gt_irq_handler(i915, gt_iir);
> }
>
> de_iir = I915_READ(DEIIR);
> if (de_iir) {
> I915_WRITE(DEIIR, de_iir);
> ret = IRQ_HANDLED;
> - if (INTEL_GEN(dev_priv) >= 7)
> - ivb_display_irq_handler(dev_priv, de_iir);
> + if (INTEL_GEN(i915) >= 7)
> + ivb_display_irq_handler(i915, de_iir);
> else
> - ilk_display_irq_handler(dev_priv, de_iir);
> + ilk_display_irq_handler(i915, de_iir);
> }
>
> - if (INTEL_GEN(dev_priv) >= 6) {
> + if (INTEL_GEN(i915) >= 6) {
> u32 pm_iir = I915_READ(GEN6_PMIIR);
> if (pm_iir) {
> I915_WRITE(GEN6_PMIIR, pm_iir);
> ret = IRQ_HANDLED;
> - gen6_rps_irq_handler(dev_priv, pm_iir);
> + gen6_rps_irq_handler(i915, pm_iir);
> }
> }
>
> I915_WRITE(DEIER, de_ier);
> - if (!HAS_PCH_NOP(dev_priv))
> + if (!HAS_PCH_NOP(i915))
> I915_WRITE(SDEIER, sde_ier);
>
> /* IRQs are synced during runtime_suspend, we don't require a wakeref */
> - enable_rpm_wakeref_asserts(dev_priv);
> + enable_rpm_wakeref_asserts(i915);
>
> return ret;
> }
>
> -static void bxt_hpd_irq_handler(struct drm_i915_private *dev_priv,
> +static void bxt_hpd_irq_handler(struct drm_i915_private *i915,
> u32 hotplug_trigger,
> const u32 hpd[HPD_NUM_PINS])
> {
> @@ -2758,14 +2758,14 @@ static void bxt_hpd_irq_handler(struct drm_i915_private *dev_priv,
> dig_hotplug_reg = I915_READ(PCH_PORT_HOTPLUG);
> I915_WRITE(PCH_PORT_HOTPLUG, dig_hotplug_reg);
>
> - intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask, hotplug_trigger,
> + intel_get_hpd_pins(i915, &pin_mask, &long_mask, hotplug_trigger,
> dig_hotplug_reg, hpd,
> bxt_port_hotplug_long_detect);
>
> - intel_hpd_irq_handler(dev_priv, pin_mask, long_mask);
> + intel_hpd_irq_handler(i915, pin_mask, long_mask);
> }
>
> -static void gen11_hpd_irq_handler(struct drm_i915_private *dev_priv, u32 iir)
> +static void gen11_hpd_irq_handler(struct drm_i915_private *i915, u32 iir)
> {
> u32 pin_mask = 0, long_mask = 0;
> u32 trigger_tc = iir & GEN11_DE_TC_HOTPLUG_MASK;
> @@ -2777,7 +2777,7 @@ static void gen11_hpd_irq_handler(struct drm_i915_private *dev_priv, u32 iir)
> dig_hotplug_reg = I915_READ(GEN11_TC_HOTPLUG_CTL);
> I915_WRITE(GEN11_TC_HOTPLUG_CTL, dig_hotplug_reg);
>
> - intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask, trigger_tc,
> + intel_get_hpd_pins(i915, &pin_mask, &long_mask, trigger_tc,
> dig_hotplug_reg, hpd_gen11,
> gen11_port_hotplug_long_detect);
> }
> @@ -2788,30 +2788,30 @@ static void gen11_hpd_irq_handler(struct drm_i915_private *dev_priv, u32 iir)
> dig_hotplug_reg = I915_READ(GEN11_TBT_HOTPLUG_CTL);
> I915_WRITE(GEN11_TBT_HOTPLUG_CTL, dig_hotplug_reg);
>
> - intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask, trigger_tbt,
> + intel_get_hpd_pins(i915, &pin_mask, &long_mask, trigger_tbt,
> dig_hotplug_reg, hpd_gen11,
> gen11_port_hotplug_long_detect);
> }
>
> if (pin_mask)
> - intel_hpd_irq_handler(dev_priv, pin_mask, long_mask);
> + intel_hpd_irq_handler(i915, pin_mask, long_mask);
> else
> DRM_ERROR("Unexpected DE HPD interrupt 0x%08x\n", iir);
> }
>
> -static u32 gen8_de_port_aux_mask(struct drm_i915_private *dev_priv)
> +static u32 gen8_de_port_aux_mask(struct drm_i915_private *i915)
> {
> u32 mask = GEN8_AUX_CHANNEL_A;
>
> - if (INTEL_GEN(dev_priv) >= 9)
> + if (INTEL_GEN(i915) >= 9)
> mask |= GEN9_AUX_CHANNEL_B |
> GEN9_AUX_CHANNEL_C |
> GEN9_AUX_CHANNEL_D;
>
> - if (IS_CNL_WITH_PORT_F(dev_priv))
> + if (IS_CNL_WITH_PORT_F(i915))
> mask |= CNL_AUX_CHANNEL_F;
>
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> mask |= ICL_AUX_CHANNEL_E |
> CNL_AUX_CHANNEL_F;
>
> @@ -2819,7 +2819,7 @@ static u32 gen8_de_port_aux_mask(struct drm_i915_private *dev_priv)
> }
>
> static irqreturn_t
> -gen8_de_irq_handler(struct drm_i915_private *dev_priv, u32 master_ctl)
> +gen8_de_irq_handler(struct drm_i915_private *i915, u32 master_ctl)
> {
> irqreturn_t ret = IRQ_NONE;
> u32 iir;
> @@ -2834,14 +2834,14 @@ gen8_de_irq_handler(struct drm_i915_private *dev_priv, u32 master_ctl)
> ret = IRQ_HANDLED;
>
> if (iir & GEN8_DE_MISC_GSE) {
> - intel_opregion_asle_intr(dev_priv);
> + intel_opregion_asle_intr(i915);
> found = true;
> }
>
> if (iir & GEN8_DE_EDP_PSR) {
> u32 psr_iir = I915_READ(EDP_PSR_IIR);
>
> - intel_psr_irq_handler(dev_priv, psr_iir);
> + intel_psr_irq_handler(i915, psr_iir);
> I915_WRITE(EDP_PSR_IIR, psr_iir);
> found = true;
> }
> @@ -2853,12 +2853,12 @@ gen8_de_irq_handler(struct drm_i915_private *dev_priv, u32 master_ctl)
> DRM_ERROR("The master control interrupt lied (DE MISC)!\n");
> }
>
> - if (INTEL_GEN(dev_priv) >= 11 && (master_ctl & GEN11_DE_HPD_IRQ)) {
> + if (INTEL_GEN(i915) >= 11 && (master_ctl & GEN11_DE_HPD_IRQ)) {
> iir = I915_READ(GEN11_DE_HPD_IIR);
> if (iir) {
> I915_WRITE(GEN11_DE_HPD_IIR, iir);
> ret = IRQ_HANDLED;
> - gen11_hpd_irq_handler(dev_priv, iir);
> + gen11_hpd_irq_handler(i915, iir);
> } else {
> DRM_ERROR("The master control interrupt lied, (DE HPD)!\n");
> }
> @@ -2873,29 +2873,29 @@ gen8_de_irq_handler(struct drm_i915_private *dev_priv, u32 master_ctl)
> I915_WRITE(GEN8_DE_PORT_IIR, iir);
> ret = IRQ_HANDLED;
>
> - if (iir & gen8_de_port_aux_mask(dev_priv)) {
> - dp_aux_irq_handler(dev_priv);
> + if (iir & gen8_de_port_aux_mask(i915)) {
> + dp_aux_irq_handler(i915);
> found = true;
> }
>
> - if (IS_GEN9_LP(dev_priv)) {
> + if (IS_GEN9_LP(i915)) {
> tmp_mask = iir & BXT_DE_PORT_HOTPLUG_MASK;
> if (tmp_mask) {
> - bxt_hpd_irq_handler(dev_priv, tmp_mask,
> + bxt_hpd_irq_handler(i915, tmp_mask,
> hpd_bxt);
> found = true;
> }
> - } else if (IS_BROADWELL(dev_priv)) {
> + } else if (IS_BROADWELL(i915)) {
> tmp_mask = iir & GEN8_PORT_DP_A_HOTPLUG;
> if (tmp_mask) {
> - ilk_hpd_irq_handler(dev_priv,
> + ilk_hpd_irq_handler(i915,
> tmp_mask, hpd_bdw);
> found = true;
> }
> }
>
> - if (IS_GEN9_LP(dev_priv) && (iir & BXT_DE_PORT_GMBUS)) {
> - gmbus_irq_handler(dev_priv);
> + if (IS_GEN9_LP(i915) && (iir & BXT_DE_PORT_GMBUS)) {
> + gmbus_irq_handler(i915);
> found = true;
> }
>
> @@ -2906,7 +2906,7 @@ gen8_de_irq_handler(struct drm_i915_private *dev_priv, u32 master_ctl)
> DRM_ERROR("The master control interrupt lied (DE PORT)!\n");
> }
>
> - for_each_pipe(dev_priv, pipe) {
> + for_each_pipe(i915, pipe) {
> u32 fault_errors;
>
> if (!(master_ctl & GEN8_DE_PIPE_IRQ(pipe)))
> @@ -2922,16 +2922,16 @@ gen8_de_irq_handler(struct drm_i915_private *dev_priv, u32 master_ctl)
> I915_WRITE(GEN8_DE_PIPE_IIR(pipe), iir);
>
> if (iir & GEN8_PIPE_VBLANK)
> - drm_handle_vblank(&dev_priv->drm, pipe);
> + drm_handle_vblank(&i915->drm, pipe);
>
> if (iir & GEN8_PIPE_CDCLK_CRC_DONE)
> - hsw_pipe_crc_irq_handler(dev_priv, pipe);
> + hsw_pipe_crc_irq_handler(i915, pipe);
>
> if (iir & GEN8_PIPE_FIFO_UNDERRUN)
> - intel_cpu_fifo_underrun_irq_handler(dev_priv, pipe);
> + intel_cpu_fifo_underrun_irq_handler(i915, pipe);
>
> fault_errors = iir;
> - if (INTEL_GEN(dev_priv) >= 9)
> + if (INTEL_GEN(i915) >= 9)
> fault_errors &= GEN9_DE_PIPE_IRQ_FAULT_ERRORS;
> else
> fault_errors &= GEN8_DE_PIPE_IRQ_FAULT_ERRORS;
> @@ -2942,7 +2942,7 @@ gen8_de_irq_handler(struct drm_i915_private *dev_priv, u32 master_ctl)
> fault_errors);
> }
>
> - if (HAS_PCH_SPLIT(dev_priv) && !HAS_PCH_NOP(dev_priv) &&
> + if (HAS_PCH_SPLIT(i915) && !HAS_PCH_NOP(i915) &&
> master_ctl & GEN8_DE_PCH_IRQ) {
> /*
> * FIXME(BDW): Assume for now that the new interrupt handling
> @@ -2954,12 +2954,12 @@ gen8_de_irq_handler(struct drm_i915_private *dev_priv, u32 master_ctl)
> I915_WRITE(SDEIIR, iir);
> ret = IRQ_HANDLED;
>
> - if (INTEL_PCH_TYPE(dev_priv) >= PCH_ICP)
> - icp_irq_handler(dev_priv, iir);
> - else if (INTEL_PCH_TYPE(dev_priv) >= PCH_SPT)
> - spt_irq_handler(dev_priv, iir);
> + if (INTEL_PCH_TYPE(i915) >= PCH_ICP)
> + icp_irq_handler(i915, iir);
> + else if (INTEL_PCH_TYPE(i915) >= PCH_SPT)
> + spt_irq_handler(i915, iir);
> else
> - cpt_irq_handler(dev_priv, iir);
> + cpt_irq_handler(i915, iir);
> } else {
> /*
> * Like on previous PCH there seems to be something
> @@ -2992,12 +2992,12 @@ static inline void gen8_master_intr_enable(void __iomem * const regs)
>
> static irqreturn_t gen8_irq_handler(int irq, void *arg)
> {
> - struct drm_i915_private *dev_priv = to_i915(arg);
> - void __iomem * const regs = dev_priv->uncore.regs;
> + struct drm_i915_private *i915 = to_i915(arg);
> + void __iomem * const regs = i915->uncore.regs;
> u32 master_ctl;
> u32 gt_iir[4];
>
> - if (!intel_irqs_enabled(dev_priv))
> + if (!intel_irqs_enabled(i915))
> return IRQ_NONE;
>
> master_ctl = gen8_master_intr_disable(regs);
> @@ -3007,18 +3007,18 @@ static irqreturn_t gen8_irq_handler(int irq, void *arg)
> }
>
> /* Find, clear, then process each source of interrupt */
> - gen8_gt_irq_ack(dev_priv, master_ctl, gt_iir);
> + gen8_gt_irq_ack(i915, master_ctl, gt_iir);
>
> /* IRQs are synced during runtime_suspend, we don't require a wakeref */
> if (master_ctl & ~GEN8_GT_IRQS) {
> - disable_rpm_wakeref_asserts(dev_priv);
> - gen8_de_irq_handler(dev_priv, master_ctl);
> - enable_rpm_wakeref_asserts(dev_priv);
> + disable_rpm_wakeref_asserts(i915);
> + gen8_de_irq_handler(i915, master_ctl);
> + enable_rpm_wakeref_asserts(i915);
> }
>
> gen8_master_intr_enable(regs);
>
> - gen8_gt_irq_handler(dev_priv, master_ctl, gt_iir);
> + gen8_gt_irq_handler(i915, master_ctl, gt_iir);
>
> return IRQ_HANDLED;
> }
> @@ -3149,9 +3149,9 @@ gen11_gt_irq_handler(struct drm_i915_private * const i915,
> }
>
> static u32
> -gen11_gu_misc_irq_ack(struct drm_i915_private *dev_priv, const u32 master_ctl)
> +gen11_gu_misc_irq_ack(struct drm_i915_private *i915, const u32 master_ctl)
> {
> - void __iomem * const regs = dev_priv->uncore.regs;
> + void __iomem * const regs = i915->uncore.regs;
> u32 iir;
>
> if (!(master_ctl & GEN11_GU_MISC_IRQ))
> @@ -3165,10 +3165,10 @@ gen11_gu_misc_irq_ack(struct drm_i915_private *dev_priv, const u32 master_ctl)
> }
>
> static void
> -gen11_gu_misc_irq_handler(struct drm_i915_private *dev_priv, const u32 iir)
> +gen11_gu_misc_irq_handler(struct drm_i915_private *i915, const u32 iir)
> {
> if (iir & GEN11_GU_MISC_GSE)
> - intel_opregion_asle_intr(dev_priv);
> + intel_opregion_asle_intr(i915);
> }
>
> static inline u32 gen11_master_intr_disable(void __iomem * const regs)
> @@ -3235,54 +3235,54 @@ static irqreturn_t gen11_irq_handler(int irq, void *arg)
> */
> static int i8xx_enable_vblank(struct drm_device *dev, unsigned int pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> unsigned long irqflags;
>
> - spin_lock_irqsave(&dev_priv->irq_lock, irqflags);
> - i915_enable_pipestat(dev_priv, pipe, PIPE_VBLANK_INTERRUPT_STATUS);
> - spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags);
> + spin_lock_irqsave(&i915->irq_lock, irqflags);
> + i915_enable_pipestat(i915, pipe, PIPE_VBLANK_INTERRUPT_STATUS);
> + spin_unlock_irqrestore(&i915->irq_lock, irqflags);
>
> return 0;
> }
>
> static int i945gm_enable_vblank(struct drm_device *dev, unsigned int pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
>
> - if (dev_priv->i945gm_vblank.enabled++ == 0)
> - schedule_work(&dev_priv->i945gm_vblank.work);
> + if (i915->i945gm_vblank.enabled++ == 0)
> + schedule_work(&i915->i945gm_vblank.work);
>
> return i8xx_enable_vblank(dev, pipe);
> }
>
> static int i965_enable_vblank(struct drm_device *dev, unsigned int pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> unsigned long irqflags;
>
> - spin_lock_irqsave(&dev_priv->irq_lock, irqflags);
> - i915_enable_pipestat(dev_priv, pipe,
> + spin_lock_irqsave(&i915->irq_lock, irqflags);
> + i915_enable_pipestat(i915, pipe,
> PIPE_START_VBLANK_INTERRUPT_STATUS);
> - spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags);
> + spin_unlock_irqrestore(&i915->irq_lock, irqflags);
>
> return 0;
> }
>
> static int ironlake_enable_vblank(struct drm_device *dev, unsigned int pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> unsigned long irqflags;
> - u32 bit = INTEL_GEN(dev_priv) >= 7 ?
> + u32 bit = INTEL_GEN(i915) >= 7 ?
> DE_PIPE_VBLANK_IVB(pipe) : DE_PIPE_VBLANK(pipe);
>
> - spin_lock_irqsave(&dev_priv->irq_lock, irqflags);
> - ilk_enable_display_irq(dev_priv, bit);
> - spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags);
> + spin_lock_irqsave(&i915->irq_lock, irqflags);
> + ilk_enable_display_irq(i915, bit);
> + spin_unlock_irqrestore(&i915->irq_lock, irqflags);
>
> /* Even though there is no DMC, frame counter can get stuck when
> * PSR is active as no frames are generated.
> */
> - if (HAS_PSR(dev_priv))
> + if (HAS_PSR(i915))
> drm_vblank_restore(dev, pipe);
>
> return 0;
> @@ -3290,17 +3290,17 @@ static int ironlake_enable_vblank(struct drm_device *dev, unsigned int pipe)
>
> static int gen8_enable_vblank(struct drm_device *dev, unsigned int pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> unsigned long irqflags;
>
> - spin_lock_irqsave(&dev_priv->irq_lock, irqflags);
> - bdw_enable_pipe_irq(dev_priv, pipe, GEN8_PIPE_VBLANK);
> - spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags);
> + spin_lock_irqsave(&i915->irq_lock, irqflags);
> + bdw_enable_pipe_irq(i915, pipe, GEN8_PIPE_VBLANK);
> + spin_unlock_irqrestore(&i915->irq_lock, irqflags);
>
> /* Even if there is no DMC, frame counter can get stuck when
> * PSR is active as no frames are generated, so check only for PSR.
> */
> - if (HAS_PSR(dev_priv))
> + if (HAS_PSR(i915))
> drm_vblank_restore(dev, pipe);
>
> return 0;
> @@ -3311,60 +3311,60 @@ static int gen8_enable_vblank(struct drm_device *dev, unsigned int pipe)
> */
> static void i8xx_disable_vblank(struct drm_device *dev, unsigned int pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> unsigned long irqflags;
>
> - spin_lock_irqsave(&dev_priv->irq_lock, irqflags);
> - i915_disable_pipestat(dev_priv, pipe, PIPE_VBLANK_INTERRUPT_STATUS);
> - spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags);
> + spin_lock_irqsave(&i915->irq_lock, irqflags);
> + i915_disable_pipestat(i915, pipe, PIPE_VBLANK_INTERRUPT_STATUS);
> + spin_unlock_irqrestore(&i915->irq_lock, irqflags);
> }
>
> static void i945gm_disable_vblank(struct drm_device *dev, unsigned int pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
>
> i8xx_disable_vblank(dev, pipe);
>
> - if (--dev_priv->i945gm_vblank.enabled == 0)
> - schedule_work(&dev_priv->i945gm_vblank.work);
> + if (--i915->i945gm_vblank.enabled == 0)
> + schedule_work(&i915->i945gm_vblank.work);
> }
>
> static void i965_disable_vblank(struct drm_device *dev, unsigned int pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> unsigned long irqflags;
>
> - spin_lock_irqsave(&dev_priv->irq_lock, irqflags);
> - i915_disable_pipestat(dev_priv, pipe,
> + spin_lock_irqsave(&i915->irq_lock, irqflags);
> + i915_disable_pipestat(i915, pipe,
> PIPE_START_VBLANK_INTERRUPT_STATUS);
> - spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags);
> + spin_unlock_irqrestore(&i915->irq_lock, irqflags);
> }
>
> static void ironlake_disable_vblank(struct drm_device *dev, unsigned int pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> unsigned long irqflags;
> - u32 bit = INTEL_GEN(dev_priv) >= 7 ?
> + u32 bit = INTEL_GEN(i915) >= 7 ?
> DE_PIPE_VBLANK_IVB(pipe) : DE_PIPE_VBLANK(pipe);
>
> - spin_lock_irqsave(&dev_priv->irq_lock, irqflags);
> - ilk_disable_display_irq(dev_priv, bit);
> - spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags);
> + spin_lock_irqsave(&i915->irq_lock, irqflags);
> + ilk_disable_display_irq(i915, bit);
> + spin_unlock_irqrestore(&i915->irq_lock, irqflags);
> }
>
> static void gen8_disable_vblank(struct drm_device *dev, unsigned int pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> unsigned long irqflags;
>
> - spin_lock_irqsave(&dev_priv->irq_lock, irqflags);
> - bdw_disable_pipe_irq(dev_priv, pipe, GEN8_PIPE_VBLANK);
> - spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags);
> + spin_lock_irqsave(&i915->irq_lock, irqflags);
> + bdw_disable_pipe_irq(i915, pipe, GEN8_PIPE_VBLANK);
> + spin_unlock_irqrestore(&i915->irq_lock, irqflags);
> }
>
> static void i945gm_vblank_work_func(struct work_struct *work)
> {
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> container_of(work, struct drm_i915_private, i945gm_vblank.work);
>
> /*
> @@ -3372,9 +3372,9 @@ static void i945gm_vblank_work_func(struct work_struct *work)
> * hence we want to prevent C3 usage while vblank interrupts
> * are enabled.
> */
> - pm_qos_update_request(&dev_priv->i945gm_vblank.pm_qos,
> - READ_ONCE(dev_priv->i945gm_vblank.enabled) ?
> - dev_priv->i945gm_vblank.c3_disable_latency :
> + pm_qos_update_request(&i915->i945gm_vblank.pm_qos,
> + READ_ONCE(i915->i945gm_vblank.enabled) ?
> + i915->i945gm_vblank.c3_disable_latency :
> PM_QOS_DEFAULT_VALUE);
> }
>
> @@ -3398,34 +3398,34 @@ static int cstate_disable_latency(const char *name)
> return 0;
> }
>
> -static void i945gm_vblank_work_init(struct drm_i915_private *dev_priv)
> +static void i945gm_vblank_work_init(struct drm_i915_private *i915)
> {
> - INIT_WORK(&dev_priv->i945gm_vblank.work,
> + INIT_WORK(&i915->i945gm_vblank.work,
> i945gm_vblank_work_func);
>
> - dev_priv->i945gm_vblank.c3_disable_latency =
> + i915->i945gm_vblank.c3_disable_latency =
> cstate_disable_latency("C3");
> - pm_qos_add_request(&dev_priv->i945gm_vblank.pm_qos,
> + pm_qos_add_request(&i915->i945gm_vblank.pm_qos,
> PM_QOS_CPU_DMA_LATENCY,
> PM_QOS_DEFAULT_VALUE);
> }
>
> -static void i945gm_vblank_work_fini(struct drm_i915_private *dev_priv)
> +static void i945gm_vblank_work_fini(struct drm_i915_private *i915)
> {
> - cancel_work_sync(&dev_priv->i945gm_vblank.work);
> - pm_qos_remove_request(&dev_priv->i945gm_vblank.pm_qos);
> + cancel_work_sync(&i915->i945gm_vblank.work);
> + pm_qos_remove_request(&i915->i945gm_vblank.pm_qos);
> }
>
> -static void ibx_irq_reset(struct drm_i915_private *dev_priv)
> +static void ibx_irq_reset(struct drm_i915_private *i915)
> {
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct intel_uncore *uncore = &i915->uncore;
>
> - if (HAS_PCH_NOP(dev_priv))
> + if (HAS_PCH_NOP(i915))
> return;
>
> GEN3_IRQ_RESET(uncore, SDE);
>
> - if (HAS_PCH_CPT(dev_priv) || HAS_PCH_LPT(dev_priv))
> + if (HAS_PCH_CPT(i915) || HAS_PCH_LPT(i915))
> I915_WRITE(SERR_INT, 0xffffffff);
> }
>
> @@ -3439,9 +3439,9 @@ static void ibx_irq_reset(struct drm_i915_private *dev_priv)
> */
> static void ibx_irq_pre_postinstall(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
>
> - if (HAS_PCH_NOP(dev_priv))
> + if (HAS_PCH_NOP(i915))
> return;
>
> WARN_ON(I915_READ(SDEIER) != 0);
> @@ -3449,36 +3449,36 @@ static void ibx_irq_pre_postinstall(struct drm_device *dev)
> POSTING_READ(SDEIER);
> }
>
> -static void gen5_gt_irq_reset(struct drm_i915_private *dev_priv)
> +static void gen5_gt_irq_reset(struct drm_i915_private *i915)
> {
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct intel_uncore *uncore = &i915->uncore;
>
> GEN3_IRQ_RESET(uncore, GT);
> - if (INTEL_GEN(dev_priv) >= 6)
> + if (INTEL_GEN(i915) >= 6)
> GEN3_IRQ_RESET(uncore, GEN6_PM);
> }
>
> -static void vlv_display_irq_reset(struct drm_i915_private *dev_priv)
> +static void vlv_display_irq_reset(struct drm_i915_private *i915)
> {
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct intel_uncore *uncore = &i915->uncore;
>
> - if (IS_CHERRYVIEW(dev_priv))
> + if (IS_CHERRYVIEW(i915))
> I915_WRITE(DPINVGTT, DPINVGTT_STATUS_MASK_CHV);
> else
> I915_WRITE(DPINVGTT, DPINVGTT_STATUS_MASK);
>
> - i915_hotplug_interrupt_update_locked(dev_priv, 0xffffffff, 0);
> + i915_hotplug_interrupt_update_locked(i915, 0xffffffff, 0);
> I915_WRITE(PORT_HOTPLUG_STAT, I915_READ(PORT_HOTPLUG_STAT));
>
> - i9xx_pipestat_irq_reset(dev_priv);
> + i9xx_pipestat_irq_reset(i915);
>
> GEN3_IRQ_RESET(uncore, VLV_);
> - dev_priv->irq_mask = ~0u;
> + i915->irq_mask = ~0u;
> }
>
> -static void vlv_display_irq_postinstall(struct drm_i915_private *dev_priv)
> +static void vlv_display_irq_postinstall(struct drm_i915_private *i915)
> {
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct intel_uncore *uncore = &i915->uncore;
>
> u32 pipestat_mask;
> u32 enable_mask;
> @@ -3486,9 +3486,9 @@ static void vlv_display_irq_postinstall(struct drm_i915_private *dev_priv)
>
> pipestat_mask = PIPE_CRC_DONE_INTERRUPT_STATUS;
>
> - i915_enable_pipestat(dev_priv, PIPE_A, PIPE_GMBUS_INTERRUPT_STATUS);
> - for_each_pipe(dev_priv, pipe)
> - i915_enable_pipestat(dev_priv, pipe, pipestat_mask);
> + i915_enable_pipestat(i915, PIPE_A, PIPE_GMBUS_INTERRUPT_STATUS);
> + for_each_pipe(i915, pipe)
> + i915_enable_pipestat(i915, pipe, pipestat_mask);
>
> enable_mask = I915_DISPLAY_PORT_INTERRUPT |
> I915_DISPLAY_PIPE_A_EVENT_INTERRUPT |
> @@ -3496,56 +3496,56 @@ static void vlv_display_irq_postinstall(struct drm_i915_private *dev_priv)
> I915_LPE_PIPE_A_INTERRUPT |
> I915_LPE_PIPE_B_INTERRUPT;
>
> - if (IS_CHERRYVIEW(dev_priv))
> + if (IS_CHERRYVIEW(i915))
> enable_mask |= I915_DISPLAY_PIPE_C_EVENT_INTERRUPT |
> I915_LPE_PIPE_C_INTERRUPT;
>
> - WARN_ON(dev_priv->irq_mask != ~0u);
> + WARN_ON(i915->irq_mask != ~0u);
>
> - dev_priv->irq_mask = ~enable_mask;
> + i915->irq_mask = ~enable_mask;
>
> - GEN3_IRQ_INIT(uncore, VLV_, dev_priv->irq_mask, enable_mask);
> + GEN3_IRQ_INIT(uncore, VLV_, i915->irq_mask, enable_mask);
> }
>
> /* drm_dma.h hooks
> */
> static void ironlake_irq_reset(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct intel_uncore *uncore = &i915->uncore;
>
> GEN3_IRQ_RESET(uncore, DE);
> - if (IS_GEN(dev_priv, 7))
> + if (IS_GEN(i915, 7))
> I915_WRITE(GEN7_ERR_INT, 0xffffffff);
>
> - if (IS_HASWELL(dev_priv)) {
> + if (IS_HASWELL(i915)) {
> I915_WRITE(EDP_PSR_IMR, 0xffffffff);
> I915_WRITE(EDP_PSR_IIR, 0xffffffff);
> }
>
> - gen5_gt_irq_reset(dev_priv);
> + gen5_gt_irq_reset(i915);
>
> - ibx_irq_reset(dev_priv);
> + ibx_irq_reset(i915);
> }
>
> static void valleyview_irq_reset(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
>
> I915_WRITE(VLV_MASTER_IER, 0);
> POSTING_READ(VLV_MASTER_IER);
>
> - gen5_gt_irq_reset(dev_priv);
> + gen5_gt_irq_reset(i915);
>
> - spin_lock_irq(&dev_priv->irq_lock);
> - if (dev_priv->display_irqs_enabled)
> - vlv_display_irq_reset(dev_priv);
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
> + if (i915->display_irqs_enabled)
> + vlv_display_irq_reset(i915);
> + spin_unlock_irq(&i915->irq_lock);
> }
>
> -static void gen8_gt_irq_reset(struct drm_i915_private *dev_priv)
> +static void gen8_gt_irq_reset(struct drm_i915_private *i915)
> {
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct intel_uncore *uncore = &i915->uncore;
>
> GEN8_IRQ_RESET_NDX(uncore, GT, 0);
> GEN8_IRQ_RESET_NDX(uncore, GT, 1);
> @@ -3555,19 +3555,19 @@ static void gen8_gt_irq_reset(struct drm_i915_private *dev_priv)
>
> static void gen8_irq_reset(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct intel_uncore *uncore = &i915->uncore;
> int pipe;
>
> - gen8_master_intr_disable(dev_priv->uncore.regs);
> + gen8_master_intr_disable(i915->uncore.regs);
>
> - gen8_gt_irq_reset(dev_priv);
> + gen8_gt_irq_reset(i915);
>
> I915_WRITE(EDP_PSR_IMR, 0xffffffff);
> I915_WRITE(EDP_PSR_IIR, 0xffffffff);
>
> - for_each_pipe(dev_priv, pipe)
> - if (intel_display_power_is_enabled(dev_priv,
> + for_each_pipe(i915, pipe)
> + if (intel_display_power_is_enabled(i915,
> POWER_DOMAIN_PIPE(pipe)))
> GEN8_IRQ_RESET_NDX(uncore, DE_PIPE, pipe);
>
> @@ -3575,11 +3575,11 @@ static void gen8_irq_reset(struct drm_device *dev)
> GEN3_IRQ_RESET(uncore, GEN8_DE_MISC_);
> GEN3_IRQ_RESET(uncore, GEN8_PCU_);
>
> - if (HAS_PCH_SPLIT(dev_priv))
> - ibx_irq_reset(dev_priv);
> + if (HAS_PCH_SPLIT(i915))
> + ibx_irq_reset(i915);
> }
>
> -static void gen11_gt_irq_reset(struct drm_i915_private *dev_priv)
> +static void gen11_gt_irq_reset(struct drm_i915_private *i915)
> {
> /* Disable RCS, BCS, VCS and VECS class engines. */
> I915_WRITE(GEN11_RENDER_COPY_INTR_ENABLE, 0);
> @@ -3600,21 +3600,21 @@ static void gen11_gt_irq_reset(struct drm_i915_private *dev_priv)
>
> static void gen11_irq_reset(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = dev->dev_private;
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct drm_i915_private *i915 = dev->dev_private;
> + struct intel_uncore *uncore = &i915->uncore;
> int pipe;
>
> - gen11_master_intr_disable(dev_priv->uncore.regs);
> + gen11_master_intr_disable(i915->uncore.regs);
>
> - gen11_gt_irq_reset(dev_priv);
> + gen11_gt_irq_reset(i915);
>
> I915_WRITE(GEN11_DISPLAY_INT_CTL, 0);
>
> I915_WRITE(EDP_PSR_IMR, 0xffffffff);
> I915_WRITE(EDP_PSR_IIR, 0xffffffff);
>
> - for_each_pipe(dev_priv, pipe)
> - if (intel_display_power_is_enabled(dev_priv,
> + for_each_pipe(i915, pipe)
> + if (intel_display_power_is_enabled(i915,
> POWER_DOMAIN_PIPE(pipe)))
> GEN8_IRQ_RESET_NDX(uncore, DE_PIPE, pipe);
>
> @@ -3624,87 +3624,87 @@ static void gen11_irq_reset(struct drm_device *dev)
> GEN3_IRQ_RESET(uncore, GEN11_GU_MISC_);
> GEN3_IRQ_RESET(uncore, GEN8_PCU_);
>
> - if (INTEL_PCH_TYPE(dev_priv) >= PCH_ICP)
> + if (INTEL_PCH_TYPE(i915) >= PCH_ICP)
> GEN3_IRQ_RESET(uncore, SDE);
> }
>
> -void gen8_irq_power_well_post_enable(struct drm_i915_private *dev_priv,
> +void gen8_irq_power_well_post_enable(struct drm_i915_private *i915,
> u8 pipe_mask)
> {
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct intel_uncore *uncore = &i915->uncore;
>
> u32 extra_ier = GEN8_PIPE_VBLANK | GEN8_PIPE_FIFO_UNDERRUN;
> enum pipe pipe;
>
> - spin_lock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
>
> - if (!intel_irqs_enabled(dev_priv)) {
> - spin_unlock_irq(&dev_priv->irq_lock);
> + if (!intel_irqs_enabled(i915)) {
> + spin_unlock_irq(&i915->irq_lock);
> return;
> }
>
> - for_each_pipe_masked(dev_priv, pipe, pipe_mask)
> + for_each_pipe_masked(i915, pipe, pipe_mask)
> GEN8_IRQ_INIT_NDX(uncore, DE_PIPE, pipe,
> - dev_priv->de_irq_mask[pipe],
> - ~dev_priv->de_irq_mask[pipe] | extra_ier);
> + i915->de_irq_mask[pipe],
> + ~i915->de_irq_mask[pipe] | extra_ier);
>
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_unlock_irq(&i915->irq_lock);
> }
>
> -void gen8_irq_power_well_pre_disable(struct drm_i915_private *dev_priv,
> +void gen8_irq_power_well_pre_disable(struct drm_i915_private *i915,
> u8 pipe_mask)
> {
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct intel_uncore *uncore = &i915->uncore;
> enum pipe pipe;
>
> - spin_lock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
>
> - if (!intel_irqs_enabled(dev_priv)) {
> - spin_unlock_irq(&dev_priv->irq_lock);
> + if (!intel_irqs_enabled(i915)) {
> + spin_unlock_irq(&i915->irq_lock);
> return;
> }
>
> - for_each_pipe_masked(dev_priv, pipe, pipe_mask)
> + for_each_pipe_masked(i915, pipe, pipe_mask)
> GEN8_IRQ_RESET_NDX(uncore, DE_PIPE, pipe);
>
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_unlock_irq(&i915->irq_lock);
>
> /* make sure we're done processing display irqs */
> - synchronize_irq(dev_priv->drm.irq);
> + synchronize_irq(i915->drm.irq);
> }
>
> static void cherryview_irq_reset(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct intel_uncore *uncore = &i915->uncore;
>
> I915_WRITE(GEN8_MASTER_IRQ, 0);
> POSTING_READ(GEN8_MASTER_IRQ);
>
> - gen8_gt_irq_reset(dev_priv);
> + gen8_gt_irq_reset(i915);
>
> GEN3_IRQ_RESET(uncore, GEN8_PCU_);
>
> - spin_lock_irq(&dev_priv->irq_lock);
> - if (dev_priv->display_irqs_enabled)
> - vlv_display_irq_reset(dev_priv);
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
> + if (i915->display_irqs_enabled)
> + vlv_display_irq_reset(i915);
> + spin_unlock_irq(&i915->irq_lock);
> }
>
> -static u32 intel_hpd_enabled_irqs(struct drm_i915_private *dev_priv,
> +static u32 intel_hpd_enabled_irqs(struct drm_i915_private *i915,
> const u32 hpd[HPD_NUM_PINS])
> {
> struct intel_encoder *encoder;
> u32 enabled_irqs = 0;
>
> - for_each_intel_encoder(&dev_priv->drm, encoder)
> - if (dev_priv->hotplug.stats[encoder->hpd_pin].state == HPD_ENABLED)
> + for_each_intel_encoder(&i915->drm, encoder)
> + if (i915->hotplug.stats[encoder->hpd_pin].state == HPD_ENABLED)
> enabled_irqs |= hpd[encoder->hpd_pin];
>
> return enabled_irqs;
> }
>
> -static void ibx_hpd_detection_setup(struct drm_i915_private *dev_priv)
> +static void ibx_hpd_detection_setup(struct drm_i915_private *i915)
> {
> u32 hotplug;
>
> @@ -3724,29 +3724,29 @@ static void ibx_hpd_detection_setup(struct drm_i915_private *dev_priv)
> * When CPU and PCH are on the same package, port A
> * HPD must be enabled in both north and south.
> */
> - if (HAS_PCH_LPT_LP(dev_priv))
> + if (HAS_PCH_LPT_LP(i915))
> hotplug |= PORTA_HOTPLUG_ENABLE;
> I915_WRITE(PCH_PORT_HOTPLUG, hotplug);
> }
>
> -static void ibx_hpd_irq_setup(struct drm_i915_private *dev_priv)
> +static void ibx_hpd_irq_setup(struct drm_i915_private *i915)
> {
> u32 hotplug_irqs, enabled_irqs;
>
> - if (HAS_PCH_IBX(dev_priv)) {
> + if (HAS_PCH_IBX(i915)) {
> hotplug_irqs = SDE_HOTPLUG_MASK;
> - enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_ibx);
> + enabled_irqs = intel_hpd_enabled_irqs(i915, hpd_ibx);
> } else {
> hotplug_irqs = SDE_HOTPLUG_MASK_CPT;
> - enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_cpt);
> + enabled_irqs = intel_hpd_enabled_irqs(i915, hpd_cpt);
> }
>
> - ibx_display_interrupt_update(dev_priv, hotplug_irqs, enabled_irqs);
> + ibx_display_interrupt_update(i915, hotplug_irqs, enabled_irqs);
>
> - ibx_hpd_detection_setup(dev_priv);
> + ibx_hpd_detection_setup(i915);
> }
>
> -static void icp_hpd_detection_setup(struct drm_i915_private *dev_priv)
> +static void icp_hpd_detection_setup(struct drm_i915_private *i915)
> {
> u32 hotplug;
>
> @@ -3763,19 +3763,19 @@ static void icp_hpd_detection_setup(struct drm_i915_private *dev_priv)
> I915_WRITE(SHOTPLUG_CTL_TC, hotplug);
> }
>
> -static void icp_hpd_irq_setup(struct drm_i915_private *dev_priv)
> +static void icp_hpd_irq_setup(struct drm_i915_private *i915)
> {
> u32 hotplug_irqs, enabled_irqs;
>
> hotplug_irqs = SDE_DDI_MASK_ICP | SDE_TC_MASK_ICP;
> - enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_icp);
> + enabled_irqs = intel_hpd_enabled_irqs(i915, hpd_icp);
>
> - ibx_display_interrupt_update(dev_priv, hotplug_irqs, enabled_irqs);
> + ibx_display_interrupt_update(i915, hotplug_irqs, enabled_irqs);
>
> - icp_hpd_detection_setup(dev_priv);
> + icp_hpd_detection_setup(i915);
> }
>
> -static void gen11_hpd_detection_setup(struct drm_i915_private *dev_priv)
> +static void gen11_hpd_detection_setup(struct drm_i915_private *i915)
> {
> u32 hotplug;
>
> @@ -3794,12 +3794,12 @@ static void gen11_hpd_detection_setup(struct drm_i915_private *dev_priv)
> I915_WRITE(GEN11_TBT_HOTPLUG_CTL, hotplug);
> }
>
> -static void gen11_hpd_irq_setup(struct drm_i915_private *dev_priv)
> +static void gen11_hpd_irq_setup(struct drm_i915_private *i915)
> {
> u32 hotplug_irqs, enabled_irqs;
> u32 val;
>
> - enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_gen11);
> + enabled_irqs = intel_hpd_enabled_irqs(i915, hpd_gen11);
> hotplug_irqs = GEN11_DE_TC_HOTPLUG_MASK | GEN11_DE_TBT_HOTPLUG_MASK;
>
> val = I915_READ(GEN11_DE_HPD_IMR);
> @@ -3807,18 +3807,18 @@ static void gen11_hpd_irq_setup(struct drm_i915_private *dev_priv)
> I915_WRITE(GEN11_DE_HPD_IMR, val);
> POSTING_READ(GEN11_DE_HPD_IMR);
>
> - gen11_hpd_detection_setup(dev_priv);
> + gen11_hpd_detection_setup(i915);
>
> - if (INTEL_PCH_TYPE(dev_priv) >= PCH_ICP)
> - icp_hpd_irq_setup(dev_priv);
> + if (INTEL_PCH_TYPE(i915) >= PCH_ICP)
> + icp_hpd_irq_setup(i915);
> }
>
> -static void spt_hpd_detection_setup(struct drm_i915_private *dev_priv)
> +static void spt_hpd_detection_setup(struct drm_i915_private *i915)
> {
> u32 val, hotplug;
>
> /* Display WA #1179 WaHardHangonHotPlug: cnp */
> - if (HAS_PCH_CNP(dev_priv)) {
> + if (HAS_PCH_CNP(i915)) {
> val = I915_READ(SOUTH_CHICKEN1);
> val &= ~CHASSIS_CLK_REQ_DURATION_MASK;
> val |= CHASSIS_CLK_REQ_DURATION(0xf);
> @@ -3838,19 +3838,19 @@ static void spt_hpd_detection_setup(struct drm_i915_private *dev_priv)
> I915_WRITE(PCH_PORT_HOTPLUG2, hotplug);
> }
>
> -static void spt_hpd_irq_setup(struct drm_i915_private *dev_priv)
> +static void spt_hpd_irq_setup(struct drm_i915_private *i915)
> {
> u32 hotplug_irqs, enabled_irqs;
>
> hotplug_irqs = SDE_HOTPLUG_MASK_SPT;
> - enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_spt);
> + enabled_irqs = intel_hpd_enabled_irqs(i915, hpd_spt);
>
> - ibx_display_interrupt_update(dev_priv, hotplug_irqs, enabled_irqs);
> + ibx_display_interrupt_update(i915, hotplug_irqs, enabled_irqs);
>
> - spt_hpd_detection_setup(dev_priv);
> + spt_hpd_detection_setup(i915);
> }
>
> -static void ilk_hpd_detection_setup(struct drm_i915_private *dev_priv)
> +static void ilk_hpd_detection_setup(struct drm_i915_private *i915)
> {
> u32 hotplug;
>
> @@ -3866,33 +3866,33 @@ static void ilk_hpd_detection_setup(struct drm_i915_private *dev_priv)
> I915_WRITE(DIGITAL_PORT_HOTPLUG_CNTRL, hotplug);
> }
>
> -static void ilk_hpd_irq_setup(struct drm_i915_private *dev_priv)
> +static void ilk_hpd_irq_setup(struct drm_i915_private *i915)
> {
> u32 hotplug_irqs, enabled_irqs;
>
> - if (INTEL_GEN(dev_priv) >= 8) {
> + if (INTEL_GEN(i915) >= 8) {
> hotplug_irqs = GEN8_PORT_DP_A_HOTPLUG;
> - enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_bdw);
> + enabled_irqs = intel_hpd_enabled_irqs(i915, hpd_bdw);
>
> - bdw_update_port_irq(dev_priv, hotplug_irqs, enabled_irqs);
> - } else if (INTEL_GEN(dev_priv) >= 7) {
> + bdw_update_port_irq(i915, hotplug_irqs, enabled_irqs);
> + } else if (INTEL_GEN(i915) >= 7) {
> hotplug_irqs = DE_DP_A_HOTPLUG_IVB;
> - enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_ivb);
> + enabled_irqs = intel_hpd_enabled_irqs(i915, hpd_ivb);
>
> - ilk_update_display_irq(dev_priv, hotplug_irqs, enabled_irqs);
> + ilk_update_display_irq(i915, hotplug_irqs, enabled_irqs);
> } else {
> hotplug_irqs = DE_DP_A_HOTPLUG;
> - enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_ilk);
> + enabled_irqs = intel_hpd_enabled_irqs(i915, hpd_ilk);
>
> - ilk_update_display_irq(dev_priv, hotplug_irqs, enabled_irqs);
> + ilk_update_display_irq(i915, hotplug_irqs, enabled_irqs);
> }
>
> - ilk_hpd_detection_setup(dev_priv);
> + ilk_hpd_detection_setup(i915);
>
> - ibx_hpd_irq_setup(dev_priv);
> + ibx_hpd_irq_setup(i915);
> }
>
> -static void __bxt_hpd_detection_setup(struct drm_i915_private *dev_priv,
> +static void __bxt_hpd_detection_setup(struct drm_i915_private *i915,
> u32 enabled_irqs)
> {
> u32 hotplug;
> @@ -3911,106 +3911,106 @@ static void __bxt_hpd_detection_setup(struct drm_i915_private *dev_priv,
> * for HPD detection logic, update it based on VBT fields.
> */
> if ((enabled_irqs & BXT_DE_PORT_HP_DDIA) &&
> - intel_bios_is_port_hpd_inverted(dev_priv, PORT_A))
> + intel_bios_is_port_hpd_inverted(i915, PORT_A))
> hotplug |= BXT_DDIA_HPD_INVERT;
> if ((enabled_irqs & BXT_DE_PORT_HP_DDIB) &&
> - intel_bios_is_port_hpd_inverted(dev_priv, PORT_B))
> + intel_bios_is_port_hpd_inverted(i915, PORT_B))
> hotplug |= BXT_DDIB_HPD_INVERT;
> if ((enabled_irqs & BXT_DE_PORT_HP_DDIC) &&
> - intel_bios_is_port_hpd_inverted(dev_priv, PORT_C))
> + intel_bios_is_port_hpd_inverted(i915, PORT_C))
> hotplug |= BXT_DDIC_HPD_INVERT;
>
> I915_WRITE(PCH_PORT_HOTPLUG, hotplug);
> }
>
> -static void bxt_hpd_detection_setup(struct drm_i915_private *dev_priv)
> +static void bxt_hpd_detection_setup(struct drm_i915_private *i915)
> {
> - __bxt_hpd_detection_setup(dev_priv, BXT_DE_PORT_HOTPLUG_MASK);
> + __bxt_hpd_detection_setup(i915, BXT_DE_PORT_HOTPLUG_MASK);
> }
>
> -static void bxt_hpd_irq_setup(struct drm_i915_private *dev_priv)
> +static void bxt_hpd_irq_setup(struct drm_i915_private *i915)
> {
> u32 hotplug_irqs, enabled_irqs;
>
> - enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_bxt);
> + enabled_irqs = intel_hpd_enabled_irqs(i915, hpd_bxt);
> hotplug_irqs = BXT_DE_PORT_HOTPLUG_MASK;
>
> - bdw_update_port_irq(dev_priv, hotplug_irqs, enabled_irqs);
> + bdw_update_port_irq(i915, hotplug_irqs, enabled_irqs);
>
> - __bxt_hpd_detection_setup(dev_priv, enabled_irqs);
> + __bxt_hpd_detection_setup(i915, enabled_irqs);
> }
>
> static void ibx_irq_postinstall(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> u32 mask;
>
> - if (HAS_PCH_NOP(dev_priv))
> + if (HAS_PCH_NOP(i915))
> return;
>
> - if (HAS_PCH_IBX(dev_priv))
> + if (HAS_PCH_IBX(i915))
> mask = SDE_GMBUS | SDE_AUX_MASK | SDE_POISON;
> - else if (HAS_PCH_CPT(dev_priv) || HAS_PCH_LPT(dev_priv))
> + else if (HAS_PCH_CPT(i915) || HAS_PCH_LPT(i915))
> mask = SDE_GMBUS_CPT | SDE_AUX_MASK_CPT;
> else
> mask = SDE_GMBUS_CPT;
>
> - gen3_assert_iir_is_zero(&dev_priv->uncore, SDEIIR);
> + gen3_assert_iir_is_zero(&i915->uncore, SDEIIR);
> I915_WRITE(SDEIMR, ~mask);
>
> - if (HAS_PCH_IBX(dev_priv) || HAS_PCH_CPT(dev_priv) ||
> - HAS_PCH_LPT(dev_priv))
> - ibx_hpd_detection_setup(dev_priv);
> + if (HAS_PCH_IBX(i915) || HAS_PCH_CPT(i915) ||
> + HAS_PCH_LPT(i915))
> + ibx_hpd_detection_setup(i915);
> else
> - spt_hpd_detection_setup(dev_priv);
> + spt_hpd_detection_setup(i915);
> }
>
> static void gen5_gt_irq_postinstall(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct intel_uncore *uncore = &i915->uncore;
> u32 pm_irqs, gt_irqs;
>
> pm_irqs = gt_irqs = 0;
>
> - dev_priv->gt_irq_mask = ~0;
> - if (HAS_L3_DPF(dev_priv)) {
> + i915->gt_irq_mask = ~0;
> + if (HAS_L3_DPF(i915)) {
> /* L3 parity interrupt is always unmasked. */
> - dev_priv->gt_irq_mask = ~GT_PARITY_ERROR(dev_priv);
> - gt_irqs |= GT_PARITY_ERROR(dev_priv);
> + i915->gt_irq_mask = ~GT_PARITY_ERROR(i915);
> + gt_irqs |= GT_PARITY_ERROR(i915);
> }
>
> gt_irqs |= GT_RENDER_USER_INTERRUPT;
> - if (IS_GEN(dev_priv, 5)) {
> + if (IS_GEN(i915, 5)) {
> gt_irqs |= ILK_BSD_USER_INTERRUPT;
> } else {
> gt_irqs |= GT_BLT_USER_INTERRUPT | GT_BSD_USER_INTERRUPT;
> }
>
> - GEN3_IRQ_INIT(uncore, GT, dev_priv->gt_irq_mask, gt_irqs);
> + GEN3_IRQ_INIT(uncore, GT, i915->gt_irq_mask, gt_irqs);
>
> - if (INTEL_GEN(dev_priv) >= 6) {
> + if (INTEL_GEN(i915) >= 6) {
> /*
> * RPS interrupts will get enabled/disabled on demand when RPS
> * itself is enabled/disabled.
> */
> - if (HAS_ENGINE(dev_priv, VECS0)) {
> + if (HAS_ENGINE(i915, VECS0)) {
> pm_irqs |= PM_VEBOX_USER_INTERRUPT;
> - dev_priv->pm_ier |= PM_VEBOX_USER_INTERRUPT;
> + i915->pm_ier |= PM_VEBOX_USER_INTERRUPT;
> }
>
> - dev_priv->pm_imr = 0xffffffff;
> - GEN3_IRQ_INIT(uncore, GEN6_PM, dev_priv->pm_imr, pm_irqs);
> + i915->pm_imr = 0xffffffff;
> + GEN3_IRQ_INIT(uncore, GEN6_PM, i915->pm_imr, pm_irqs);
> }
> }
>
> static int ironlake_irq_postinstall(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct intel_uncore *uncore = &i915->uncore;
> u32 display_mask, extra_mask;
>
> - if (INTEL_GEN(dev_priv) >= 7) {
> + if (INTEL_GEN(i915) >= 7) {
> display_mask = (DE_MASTER_IRQ_CONTROL | DE_GSE_IVB |
> DE_PCH_EVENT_IVB | DE_AUX_CHANNEL_A_IVB);
> extra_mask = (DE_PIPEC_VBLANK_IVB | DE_PIPEB_VBLANK_IVB |
> @@ -4025,78 +4025,78 @@ static int ironlake_irq_postinstall(struct drm_device *dev)
> DE_DP_A_HOTPLUG);
> }
>
> - if (IS_HASWELL(dev_priv)) {
> + if (IS_HASWELL(i915)) {
> gen3_assert_iir_is_zero(uncore, EDP_PSR_IIR);
> - intel_psr_irq_control(dev_priv, dev_priv->psr.debug);
> + intel_psr_irq_control(i915, i915->psr.debug);
> display_mask |= DE_EDP_PSR_INT_HSW;
> }
>
> - dev_priv->irq_mask = ~display_mask;
> + i915->irq_mask = ~display_mask;
>
> ibx_irq_pre_postinstall(dev);
>
> - GEN3_IRQ_INIT(uncore, DE, dev_priv->irq_mask,
> + GEN3_IRQ_INIT(uncore, DE, i915->irq_mask,
> display_mask | extra_mask);
>
> gen5_gt_irq_postinstall(dev);
>
> - ilk_hpd_detection_setup(dev_priv);
> + ilk_hpd_detection_setup(i915);
>
> ibx_irq_postinstall(dev);
>
> - if (IS_IRONLAKE_M(dev_priv)) {
> + if (IS_IRONLAKE_M(i915)) {
> /* Enable PCU event interrupts
> *
> * spinlocking not required here for correctness since interrupt
> * setup is guaranteed to run in single-threaded context. But we
> * need it to make the assert_spin_locked happy. */
> - spin_lock_irq(&dev_priv->irq_lock);
> - ilk_enable_display_irq(dev_priv, DE_PCU_EVENT);
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
> + ilk_enable_display_irq(i915, DE_PCU_EVENT);
> + spin_unlock_irq(&i915->irq_lock);
> }
>
> return 0;
> }
>
> -void valleyview_enable_display_irqs(struct drm_i915_private *dev_priv)
> +void valleyview_enable_display_irqs(struct drm_i915_private *i915)
> {
> - lockdep_assert_held(&dev_priv->irq_lock);
> + lockdep_assert_held(&i915->irq_lock);
>
> - if (dev_priv->display_irqs_enabled)
> + if (i915->display_irqs_enabled)
> return;
>
> - dev_priv->display_irqs_enabled = true;
> + i915->display_irqs_enabled = true;
>
> - if (intel_irqs_enabled(dev_priv)) {
> - vlv_display_irq_reset(dev_priv);
> - vlv_display_irq_postinstall(dev_priv);
> + if (intel_irqs_enabled(i915)) {
> + vlv_display_irq_reset(i915);
> + vlv_display_irq_postinstall(i915);
> }
> }
>
> -void valleyview_disable_display_irqs(struct drm_i915_private *dev_priv)
> +void valleyview_disable_display_irqs(struct drm_i915_private *i915)
> {
> - lockdep_assert_held(&dev_priv->irq_lock);
> + lockdep_assert_held(&i915->irq_lock);
>
> - if (!dev_priv->display_irqs_enabled)
> + if (!i915->display_irqs_enabled)
> return;
>
> - dev_priv->display_irqs_enabled = false;
> + i915->display_irqs_enabled = false;
>
> - if (intel_irqs_enabled(dev_priv))
> - vlv_display_irq_reset(dev_priv);
> + if (intel_irqs_enabled(i915))
> + vlv_display_irq_reset(i915);
> }
>
>
> static int valleyview_irq_postinstall(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
>
> gen5_gt_irq_postinstall(dev);
>
> - spin_lock_irq(&dev_priv->irq_lock);
> - if (dev_priv->display_irqs_enabled)
> - vlv_display_irq_postinstall(dev_priv);
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
> + if (i915->display_irqs_enabled)
> + vlv_display_irq_postinstall(i915);
> + spin_unlock_irq(&i915->irq_lock);
>
> I915_WRITE(VLV_MASTER_IER, MASTER_INTERRUPT_ENABLE);
> POSTING_READ(VLV_MASTER_IER);
> @@ -4104,9 +4104,9 @@ static int valleyview_irq_postinstall(struct drm_device *dev)
> return 0;
> }
>
> -static void gen8_gt_irq_postinstall(struct drm_i915_private *dev_priv)
> +static void gen8_gt_irq_postinstall(struct drm_i915_private *i915)
> {
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct intel_uncore *uncore = &i915->uncore;
>
> /* These are interrupts we'll toggle with the ring mask register */
> u32 gt_interrupts[] = {
> @@ -4126,21 +4126,21 @@ static void gen8_gt_irq_postinstall(struct drm_i915_private *dev_priv)
> GT_CONTEXT_SWITCH_INTERRUPT << GEN8_VECS_IRQ_SHIFT)
> };
>
> - dev_priv->pm_ier = 0x0;
> - dev_priv->pm_imr = ~dev_priv->pm_ier;
> + i915->pm_ier = 0x0;
> + i915->pm_imr = ~i915->pm_ier;
> GEN8_IRQ_INIT_NDX(uncore, GT, 0, ~gt_interrupts[0], gt_interrupts[0]);
> GEN8_IRQ_INIT_NDX(uncore, GT, 1, ~gt_interrupts[1], gt_interrupts[1]);
> /*
> * RPS interrupts will get enabled/disabled on demand when RPS itself
> * is enabled/disabled. Same wil be the case for GuC interrupts.
> */
> - GEN8_IRQ_INIT_NDX(uncore, GT, 2, dev_priv->pm_imr, dev_priv->pm_ier);
> + GEN8_IRQ_INIT_NDX(uncore, GT, 2, i915->pm_imr, i915->pm_ier);
> GEN8_IRQ_INIT_NDX(uncore, GT, 3, ~gt_interrupts[3], gt_interrupts[3]);
> }
>
> -static void gen8_de_irq_postinstall(struct drm_i915_private *dev_priv)
> +static void gen8_de_irq_postinstall(struct drm_i915_private *i915)
> {
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct intel_uncore *uncore = &i915->uncore;
>
> u32 de_pipe_masked = GEN8_PIPE_CDCLK_CRC_DONE;
> u32 de_pipe_enables;
> @@ -4149,84 +4149,84 @@ static void gen8_de_irq_postinstall(struct drm_i915_private *dev_priv)
> u32 de_misc_masked = GEN8_DE_EDP_PSR;
> enum pipe pipe;
>
> - if (INTEL_GEN(dev_priv) <= 10)
> + if (INTEL_GEN(i915) <= 10)
> de_misc_masked |= GEN8_DE_MISC_GSE;
>
> - if (INTEL_GEN(dev_priv) >= 9) {
> + if (INTEL_GEN(i915) >= 9) {
> de_pipe_masked |= GEN9_DE_PIPE_IRQ_FAULT_ERRORS;
> de_port_masked |= GEN9_AUX_CHANNEL_B | GEN9_AUX_CHANNEL_C |
> GEN9_AUX_CHANNEL_D;
> - if (IS_GEN9_LP(dev_priv))
> + if (IS_GEN9_LP(i915))
> de_port_masked |= BXT_DE_PORT_GMBUS;
> } else {
> de_pipe_masked |= GEN8_DE_PIPE_IRQ_FAULT_ERRORS;
> }
>
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> de_port_masked |= ICL_AUX_CHANNEL_E;
>
> - if (IS_CNL_WITH_PORT_F(dev_priv) || INTEL_GEN(dev_priv) >= 11)
> + if (IS_CNL_WITH_PORT_F(i915) || INTEL_GEN(i915) >= 11)
> de_port_masked |= CNL_AUX_CHANNEL_F;
>
> de_pipe_enables = de_pipe_masked | GEN8_PIPE_VBLANK |
> GEN8_PIPE_FIFO_UNDERRUN;
>
> de_port_enables = de_port_masked;
> - if (IS_GEN9_LP(dev_priv))
> + if (IS_GEN9_LP(i915))
> de_port_enables |= BXT_DE_PORT_HOTPLUG_MASK;
> - else if (IS_BROADWELL(dev_priv))
> + else if (IS_BROADWELL(i915))
> de_port_enables |= GEN8_PORT_DP_A_HOTPLUG;
>
> gen3_assert_iir_is_zero(uncore, EDP_PSR_IIR);
> - intel_psr_irq_control(dev_priv, dev_priv->psr.debug);
> + intel_psr_irq_control(i915, i915->psr.debug);
>
> - for_each_pipe(dev_priv, pipe) {
> - dev_priv->de_irq_mask[pipe] = ~de_pipe_masked;
> + for_each_pipe(i915, pipe) {
> + i915->de_irq_mask[pipe] = ~de_pipe_masked;
>
> - if (intel_display_power_is_enabled(dev_priv,
> + if (intel_display_power_is_enabled(i915,
> POWER_DOMAIN_PIPE(pipe)))
> GEN8_IRQ_INIT_NDX(uncore, DE_PIPE, pipe,
> - dev_priv->de_irq_mask[pipe],
> + i915->de_irq_mask[pipe],
> de_pipe_enables);
> }
>
> GEN3_IRQ_INIT(uncore, GEN8_DE_PORT_, ~de_port_masked, de_port_enables);
> GEN3_IRQ_INIT(uncore, GEN8_DE_MISC_, ~de_misc_masked, de_misc_masked);
>
> - if (INTEL_GEN(dev_priv) >= 11) {
> + if (INTEL_GEN(i915) >= 11) {
> u32 de_hpd_masked = 0;
> u32 de_hpd_enables = GEN11_DE_TC_HOTPLUG_MASK |
> GEN11_DE_TBT_HOTPLUG_MASK;
>
> GEN3_IRQ_INIT(uncore, GEN11_DE_HPD_, ~de_hpd_masked,
> de_hpd_enables);
> - gen11_hpd_detection_setup(dev_priv);
> - } else if (IS_GEN9_LP(dev_priv)) {
> - bxt_hpd_detection_setup(dev_priv);
> - } else if (IS_BROADWELL(dev_priv)) {
> - ilk_hpd_detection_setup(dev_priv);
> + gen11_hpd_detection_setup(i915);
> + } else if (IS_GEN9_LP(i915)) {
> + bxt_hpd_detection_setup(i915);
> + } else if (IS_BROADWELL(i915)) {
> + ilk_hpd_detection_setup(i915);
> }
> }
>
> static int gen8_irq_postinstall(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
>
> - if (HAS_PCH_SPLIT(dev_priv))
> + if (HAS_PCH_SPLIT(i915))
> ibx_irq_pre_postinstall(dev);
>
> - gen8_gt_irq_postinstall(dev_priv);
> - gen8_de_irq_postinstall(dev_priv);
> + gen8_gt_irq_postinstall(i915);
> + gen8_de_irq_postinstall(i915);
>
> - if (HAS_PCH_SPLIT(dev_priv))
> + if (HAS_PCH_SPLIT(i915))
> ibx_irq_postinstall(dev);
>
> - gen8_master_intr_enable(dev_priv->uncore.regs);
> + gen8_master_intr_enable(i915->uncore.regs);
>
> return 0;
> }
>
> -static void gen11_gt_irq_postinstall(struct drm_i915_private *dev_priv)
> +static void gen11_gt_irq_postinstall(struct drm_i915_private *i915)
> {
> const u32 irqs = GT_RENDER_USER_INTERRUPT | GT_CONTEXT_SWITCH_INTERRUPT;
>
> @@ -4247,8 +4247,8 @@ static void gen11_gt_irq_postinstall(struct drm_i915_private *dev_priv)
> * RPS interrupts will get enabled/disabled on demand when RPS itself
> * is enabled/disabled.
> */
> - dev_priv->pm_ier = 0x0;
> - dev_priv->pm_imr = ~dev_priv->pm_ier;
> + i915->pm_ier = 0x0;
> + i915->pm_imr = ~i915->pm_ier;
> I915_WRITE(GEN11_GPM_WGBOXPERF_INTR_ENABLE, 0);
> I915_WRITE(GEN11_GPM_WGBOXPERF_INTR_MASK, ~0);
>
> @@ -4259,36 +4259,36 @@ static void gen11_gt_irq_postinstall(struct drm_i915_private *dev_priv)
>
> static void icp_irq_postinstall(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> u32 mask = SDE_GMBUS_ICP;
>
> WARN_ON(I915_READ(SDEIER) != 0);
> I915_WRITE(SDEIER, 0xffffffff);
> POSTING_READ(SDEIER);
>
> - gen3_assert_iir_is_zero(&dev_priv->uncore, SDEIIR);
> + gen3_assert_iir_is_zero(&i915->uncore, SDEIIR);
> I915_WRITE(SDEIMR, ~mask);
>
> - icp_hpd_detection_setup(dev_priv);
> + icp_hpd_detection_setup(i915);
> }
>
> static int gen11_irq_postinstall(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = dev->dev_private;
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct drm_i915_private *i915 = dev->dev_private;
> + struct intel_uncore *uncore = &i915->uncore;
> u32 gu_misc_masked = GEN11_GU_MISC_GSE;
>
> - if (INTEL_PCH_TYPE(dev_priv) >= PCH_ICP)
> + if (INTEL_PCH_TYPE(i915) >= PCH_ICP)
> icp_irq_postinstall(dev);
>
> - gen11_gt_irq_postinstall(dev_priv);
> - gen8_de_irq_postinstall(dev_priv);
> + gen11_gt_irq_postinstall(i915);
> + gen8_de_irq_postinstall(i915);
>
> GEN3_IRQ_INIT(uncore, GEN11_GU_MISC_, ~gu_misc_masked, gu_misc_masked);
>
> I915_WRITE(GEN11_DISPLAY_INT_CTL, GEN11_DISPLAY_IRQ_ENABLE);
>
> - gen11_master_intr_enable(dev_priv->uncore.regs);
> + gen11_master_intr_enable(i915->uncore.regs);
> POSTING_READ(GEN11_GFX_MSTR_IRQ);
>
> return 0;
> @@ -4296,14 +4296,14 @@ static int gen11_irq_postinstall(struct drm_device *dev)
>
> static int cherryview_irq_postinstall(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
>
> - gen8_gt_irq_postinstall(dev_priv);
> + gen8_gt_irq_postinstall(i915);
>
> - spin_lock_irq(&dev_priv->irq_lock);
> - if (dev_priv->display_irqs_enabled)
> - vlv_display_irq_postinstall(dev_priv);
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
> + if (i915->display_irqs_enabled)
> + vlv_display_irq_postinstall(i915);
> + spin_unlock_irq(&i915->irq_lock);
>
> I915_WRITE(GEN8_MASTER_IRQ, GEN8_MASTER_IRQ_CONTROL);
> POSTING_READ(GEN8_MASTER_IRQ);
> @@ -4313,18 +4313,18 @@ static int cherryview_irq_postinstall(struct drm_device *dev)
>
> static void i8xx_irq_reset(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct intel_uncore *uncore = &i915->uncore;
>
> - i9xx_pipestat_irq_reset(dev_priv);
> + i9xx_pipestat_irq_reset(i915);
>
> GEN2_IRQ_RESET(uncore);
> }
>
> static int i8xx_irq_postinstall(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct intel_uncore *uncore = &i915->uncore;
> u16 enable_mask;
>
> intel_uncore_write16(uncore,
> @@ -4333,7 +4333,7 @@ static int i8xx_irq_postinstall(struct drm_device *dev)
> I915_ERROR_MEMORY_REFRESH));
>
> /* Unmask the interrupts that we always want on. */
> - dev_priv->irq_mask =
> + i915->irq_mask =
> ~(I915_DISPLAY_PIPE_A_EVENT_INTERRUPT |
> I915_DISPLAY_PIPE_B_EVENT_INTERRUPT |
> I915_MASTER_ERROR_INTERRUPT);
> @@ -4344,14 +4344,14 @@ static int i8xx_irq_postinstall(struct drm_device *dev)
> I915_MASTER_ERROR_INTERRUPT |
> I915_USER_INTERRUPT;
>
> - GEN2_IRQ_INIT(uncore, dev_priv->irq_mask, enable_mask);
> + GEN2_IRQ_INIT(uncore, i915->irq_mask, enable_mask);
>
> /* Interrupt setup is already guaranteed to be single-threaded, this is
> * just to make the assert_spin_locked check happy. */
> - spin_lock_irq(&dev_priv->irq_lock);
> - i915_enable_pipestat(dev_priv, PIPE_A, PIPE_CRC_DONE_INTERRUPT_STATUS);
> - i915_enable_pipestat(dev_priv, PIPE_B, PIPE_CRC_DONE_INTERRUPT_STATUS);
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
> + i915_enable_pipestat(i915, PIPE_A, PIPE_CRC_DONE_INTERRUPT_STATUS);
> + i915_enable_pipestat(i915, PIPE_B, PIPE_CRC_DONE_INTERRUPT_STATUS);
> + spin_unlock_irq(&i915->irq_lock);
>
> return 0;
> }
> @@ -4386,7 +4386,7 @@ static void i8xx_error_irq_ack(struct drm_i915_private *i915,
> intel_uncore_write16(uncore, EMR, emr | *eir_stuck);
> }
>
> -static void i8xx_error_irq_handler(struct drm_i915_private *dev_priv,
> +static void i8xx_error_irq_handler(struct drm_i915_private *i915,
> u16 eir, u16 eir_stuck)
> {
> DRM_DEBUG("Master Error: EIR 0x%04x\n", eir);
> @@ -4395,7 +4395,7 @@ static void i8xx_error_irq_handler(struct drm_i915_private *dev_priv,
> DRM_DEBUG_DRIVER("EIR stuck: 0x%04x, masked\n", eir_stuck);
> }
>
> -static void i9xx_error_irq_ack(struct drm_i915_private *dev_priv,
> +static void i9xx_error_irq_ack(struct drm_i915_private *i915,
> u32 *eir, u32 *eir_stuck)
> {
> u32 emr;
> @@ -4423,7 +4423,7 @@ static void i9xx_error_irq_ack(struct drm_i915_private *dev_priv,
> I915_WRITE(EMR, emr | *eir_stuck);
> }
>
> -static void i9xx_error_irq_handler(struct drm_i915_private *dev_priv,
> +static void i9xx_error_irq_handler(struct drm_i915_private *i915,
> u32 eir, u32 eir_stuck)
> {
> DRM_DEBUG("Master Error, EIR 0x%08x\n", eir);
> @@ -4435,21 +4435,21 @@ static void i9xx_error_irq_handler(struct drm_i915_private *dev_priv,
> static irqreturn_t i8xx_irq_handler(int irq, void *arg)
> {
> struct drm_device *dev = arg;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> irqreturn_t ret = IRQ_NONE;
>
> - if (!intel_irqs_enabled(dev_priv))
> + if (!intel_irqs_enabled(i915))
> return IRQ_NONE;
>
> /* IRQs are synced during runtime_suspend, we don't require a wakeref */
> - disable_rpm_wakeref_asserts(dev_priv);
> + disable_rpm_wakeref_asserts(i915);
>
> do {
> u32 pipe_stats[I915_MAX_PIPES] = {};
> u16 eir = 0, eir_stuck = 0;
> u16 iir;
>
> - iir = intel_uncore_read16(&dev_priv->uncore, GEN2_IIR);
> + iir = intel_uncore_read16(&i915->uncore, GEN2_IIR);
> if (iir == 0)
> break;
>
> @@ -4457,53 +4457,53 @@ static irqreturn_t i8xx_irq_handler(int irq, void *arg)
>
> /* Call regardless, as some status bits might not be
> * signalled in iir */
> - i9xx_pipestat_irq_ack(dev_priv, iir, pipe_stats);
> + i9xx_pipestat_irq_ack(i915, iir, pipe_stats);
>
> if (iir & I915_MASTER_ERROR_INTERRUPT)
> - i8xx_error_irq_ack(dev_priv, &eir, &eir_stuck);
> + i8xx_error_irq_ack(i915, &eir, &eir_stuck);
>
> - intel_uncore_write16(&dev_priv->uncore, GEN2_IIR, iir);
> + intel_uncore_write16(&i915->uncore, GEN2_IIR, iir);
>
> if (iir & I915_USER_INTERRUPT)
> - intel_engine_breadcrumbs_irq(dev_priv->engine[RCS0]);
> + intel_engine_breadcrumbs_irq(i915->engine[RCS0]);
>
> if (iir & I915_MASTER_ERROR_INTERRUPT)
> - i8xx_error_irq_handler(dev_priv, eir, eir_stuck);
> + i8xx_error_irq_handler(i915, eir, eir_stuck);
>
> - i8xx_pipestat_irq_handler(dev_priv, iir, pipe_stats);
> + i8xx_pipestat_irq_handler(i915, iir, pipe_stats);
> } while (0);
>
> - enable_rpm_wakeref_asserts(dev_priv);
> + enable_rpm_wakeref_asserts(i915);
>
> return ret;
> }
>
> static void i915_irq_reset(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct intel_uncore *uncore = &i915->uncore;
>
> - if (I915_HAS_HOTPLUG(dev_priv)) {
> - i915_hotplug_interrupt_update(dev_priv, 0xffffffff, 0);
> + if (I915_HAS_HOTPLUG(i915)) {
> + i915_hotplug_interrupt_update(i915, 0xffffffff, 0);
> I915_WRITE(PORT_HOTPLUG_STAT, I915_READ(PORT_HOTPLUG_STAT));
> }
>
> - i9xx_pipestat_irq_reset(dev_priv);
> + i9xx_pipestat_irq_reset(i915);
>
> GEN3_IRQ_RESET(uncore, GEN2_);
> }
>
> static int i915_irq_postinstall(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct intel_uncore *uncore = &i915->uncore;
> u32 enable_mask;
>
> I915_WRITE(EMR, ~(I915_ERROR_PAGE_TABLE |
> I915_ERROR_MEMORY_REFRESH));
>
> /* Unmask the interrupts that we always want on. */
> - dev_priv->irq_mask =
> + i915->irq_mask =
> ~(I915_ASLE_INTERRUPT |
> I915_DISPLAY_PIPE_A_EVENT_INTERRUPT |
> I915_DISPLAY_PIPE_B_EVENT_INTERRUPT |
> @@ -4516,23 +4516,23 @@ static int i915_irq_postinstall(struct drm_device *dev)
> I915_MASTER_ERROR_INTERRUPT |
> I915_USER_INTERRUPT;
>
> - if (I915_HAS_HOTPLUG(dev_priv)) {
> + if (I915_HAS_HOTPLUG(i915)) {
> /* Enable in IER... */
> enable_mask |= I915_DISPLAY_PORT_INTERRUPT;
> /* and unmask in IMR */
> - dev_priv->irq_mask &= ~I915_DISPLAY_PORT_INTERRUPT;
> + i915->irq_mask &= ~I915_DISPLAY_PORT_INTERRUPT;
> }
>
> - GEN3_IRQ_INIT(uncore, GEN2_, dev_priv->irq_mask, enable_mask);
> + GEN3_IRQ_INIT(uncore, GEN2_, i915->irq_mask, enable_mask);
>
> /* Interrupt setup is already guaranteed to be single-threaded, this is
> * just to make the assert_spin_locked check happy. */
> - spin_lock_irq(&dev_priv->irq_lock);
> - i915_enable_pipestat(dev_priv, PIPE_A, PIPE_CRC_DONE_INTERRUPT_STATUS);
> - i915_enable_pipestat(dev_priv, PIPE_B, PIPE_CRC_DONE_INTERRUPT_STATUS);
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
> + i915_enable_pipestat(i915, PIPE_A, PIPE_CRC_DONE_INTERRUPT_STATUS);
> + i915_enable_pipestat(i915, PIPE_B, PIPE_CRC_DONE_INTERRUPT_STATUS);
> + spin_unlock_irq(&i915->irq_lock);
>
> - i915_enable_asle_pipestat(dev_priv);
> + i915_enable_asle_pipestat(i915);
>
> return 0;
> }
> @@ -4540,14 +4540,14 @@ static int i915_irq_postinstall(struct drm_device *dev)
> static irqreturn_t i915_irq_handler(int irq, void *arg)
> {
> struct drm_device *dev = arg;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> irqreturn_t ret = IRQ_NONE;
>
> - if (!intel_irqs_enabled(dev_priv))
> + if (!intel_irqs_enabled(i915))
> return IRQ_NONE;
>
> /* IRQs are synced during runtime_suspend, we don't require a wakeref */
> - disable_rpm_wakeref_asserts(dev_priv);
> + disable_rpm_wakeref_asserts(i915);
>
> do {
> u32 pipe_stats[I915_MAX_PIPES] = {};
> @@ -4561,53 +4561,53 @@ static irqreturn_t i915_irq_handler(int irq, void *arg)
>
> ret = IRQ_HANDLED;
>
> - if (I915_HAS_HOTPLUG(dev_priv) &&
> + if (I915_HAS_HOTPLUG(i915) &&
> iir & I915_DISPLAY_PORT_INTERRUPT)
> - hotplug_status = i9xx_hpd_irq_ack(dev_priv);
> + hotplug_status = i9xx_hpd_irq_ack(i915);
>
> /* Call regardless, as some status bits might not be
> * signalled in iir */
> - i9xx_pipestat_irq_ack(dev_priv, iir, pipe_stats);
> + i9xx_pipestat_irq_ack(i915, iir, pipe_stats);
>
> if (iir & I915_MASTER_ERROR_INTERRUPT)
> - i9xx_error_irq_ack(dev_priv, &eir, &eir_stuck);
> + i9xx_error_irq_ack(i915, &eir, &eir_stuck);
>
> I915_WRITE(GEN2_IIR, iir);
>
> if (iir & I915_USER_INTERRUPT)
> - intel_engine_breadcrumbs_irq(dev_priv->engine[RCS0]);
> + intel_engine_breadcrumbs_irq(i915->engine[RCS0]);
>
> if (iir & I915_MASTER_ERROR_INTERRUPT)
> - i9xx_error_irq_handler(dev_priv, eir, eir_stuck);
> + i9xx_error_irq_handler(i915, eir, eir_stuck);
>
> if (hotplug_status)
> - i9xx_hpd_irq_handler(dev_priv, hotplug_status);
> + i9xx_hpd_irq_handler(i915, hotplug_status);
>
> - i915_pipestat_irq_handler(dev_priv, iir, pipe_stats);
> + i915_pipestat_irq_handler(i915, iir, pipe_stats);
> } while (0);
>
> - enable_rpm_wakeref_asserts(dev_priv);
> + enable_rpm_wakeref_asserts(i915);
>
> return ret;
> }
>
> static void i965_irq_reset(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct intel_uncore *uncore = &i915->uncore;
>
> - i915_hotplug_interrupt_update(dev_priv, 0xffffffff, 0);
> + i915_hotplug_interrupt_update(i915, 0xffffffff, 0);
> I915_WRITE(PORT_HOTPLUG_STAT, I915_READ(PORT_HOTPLUG_STAT));
>
> - i9xx_pipestat_irq_reset(dev_priv);
> + i9xx_pipestat_irq_reset(i915);
>
> GEN3_IRQ_RESET(uncore, GEN2_);
> }
>
> static int i965_irq_postinstall(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct intel_uncore *uncore = &i915->uncore;
> u32 enable_mask;
> u32 error_mask;
>
> @@ -4615,7 +4615,7 @@ static int i965_irq_postinstall(struct drm_device *dev)
> * Enable some error detection, note the instruction error mask
> * bit is reserved, so we leave it masked.
> */
> - if (IS_G4X(dev_priv)) {
> + if (IS_G4X(i915)) {
> error_mask = ~(GM45_ERROR_PAGE_TABLE |
> GM45_ERROR_MEM_PRIV |
> GM45_ERROR_CP_PRIV |
> @@ -4627,7 +4627,7 @@ static int i965_irq_postinstall(struct drm_device *dev)
> I915_WRITE(EMR, error_mask);
>
> /* Unmask the interrupts that we always want on. */
> - dev_priv->irq_mask =
> + i915->irq_mask =
> ~(I915_ASLE_INTERRUPT |
> I915_DISPLAY_PORT_INTERRUPT |
> I915_DISPLAY_PIPE_A_EVENT_INTERRUPT |
> @@ -4642,43 +4642,43 @@ static int i965_irq_postinstall(struct drm_device *dev)
> I915_MASTER_ERROR_INTERRUPT |
> I915_USER_INTERRUPT;
>
> - if (IS_G4X(dev_priv))
> + if (IS_G4X(i915))
> enable_mask |= I915_BSD_USER_INTERRUPT;
>
> - GEN3_IRQ_INIT(uncore, GEN2_, dev_priv->irq_mask, enable_mask);
> + GEN3_IRQ_INIT(uncore, GEN2_, i915->irq_mask, enable_mask);
>
> /* Interrupt setup is already guaranteed to be single-threaded, this is
> * just to make the assert_spin_locked check happy. */
> - spin_lock_irq(&dev_priv->irq_lock);
> - i915_enable_pipestat(dev_priv, PIPE_A, PIPE_GMBUS_INTERRUPT_STATUS);
> - i915_enable_pipestat(dev_priv, PIPE_A, PIPE_CRC_DONE_INTERRUPT_STATUS);
> - i915_enable_pipestat(dev_priv, PIPE_B, PIPE_CRC_DONE_INTERRUPT_STATUS);
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
> + i915_enable_pipestat(i915, PIPE_A, PIPE_GMBUS_INTERRUPT_STATUS);
> + i915_enable_pipestat(i915, PIPE_A, PIPE_CRC_DONE_INTERRUPT_STATUS);
> + i915_enable_pipestat(i915, PIPE_B, PIPE_CRC_DONE_INTERRUPT_STATUS);
> + spin_unlock_irq(&i915->irq_lock);
>
> - i915_enable_asle_pipestat(dev_priv);
> + i915_enable_asle_pipestat(i915);
>
> return 0;
> }
>
> -static void i915_hpd_irq_setup(struct drm_i915_private *dev_priv)
> +static void i915_hpd_irq_setup(struct drm_i915_private *i915)
> {
> u32 hotplug_en;
>
> - lockdep_assert_held(&dev_priv->irq_lock);
> + lockdep_assert_held(&i915->irq_lock);
>
> /* Note HDMI and DP share hotplug bits */
> /* enable bits are the same for all generations */
> - hotplug_en = intel_hpd_enabled_irqs(dev_priv, hpd_mask_i915);
> + hotplug_en = intel_hpd_enabled_irqs(i915, hpd_mask_i915);
> /* Programming the CRT detection parameters tends
> to generate a spurious hotplug event about three
> seconds later. So just do it once.
> */
> - if (IS_G4X(dev_priv))
> + if (IS_G4X(i915))
> hotplug_en |= CRT_HOTPLUG_ACTIVATION_PERIOD_64;
> hotplug_en |= CRT_HOTPLUG_VOLTAGE_COMPARE_50;
>
> /* Ignore TV since it's buggy */
> - i915_hotplug_interrupt_update_locked(dev_priv,
> + i915_hotplug_interrupt_update_locked(i915,
> HOTPLUG_INT_EN_MASK |
> CRT_HOTPLUG_VOLTAGE_COMPARE_MASK |
> CRT_HOTPLUG_ACTIVATION_PERIOD_64,
> @@ -4688,14 +4688,14 @@ static void i915_hpd_irq_setup(struct drm_i915_private *dev_priv)
> static irqreturn_t i965_irq_handler(int irq, void *arg)
> {
> struct drm_device *dev = arg;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> irqreturn_t ret = IRQ_NONE;
>
> - if (!intel_irqs_enabled(dev_priv))
> + if (!intel_irqs_enabled(i915))
> return IRQ_NONE;
>
> /* IRQs are synced during runtime_suspend, we don't require a wakeref */
> - disable_rpm_wakeref_asserts(dev_priv);
> + disable_rpm_wakeref_asserts(i915);
>
> do {
> u32 pipe_stats[I915_MAX_PIPES] = {};
> @@ -4710,76 +4710,76 @@ static irqreturn_t i965_irq_handler(int irq, void *arg)
> ret = IRQ_HANDLED;
>
> if (iir & I915_DISPLAY_PORT_INTERRUPT)
> - hotplug_status = i9xx_hpd_irq_ack(dev_priv);
> + hotplug_status = i9xx_hpd_irq_ack(i915);
>
> /* Call regardless, as some status bits might not be
> * signalled in iir */
> - i9xx_pipestat_irq_ack(dev_priv, iir, pipe_stats);
> + i9xx_pipestat_irq_ack(i915, iir, pipe_stats);
>
> if (iir & I915_MASTER_ERROR_INTERRUPT)
> - i9xx_error_irq_ack(dev_priv, &eir, &eir_stuck);
> + i9xx_error_irq_ack(i915, &eir, &eir_stuck);
>
> I915_WRITE(GEN2_IIR, iir);
>
> if (iir & I915_USER_INTERRUPT)
> - intel_engine_breadcrumbs_irq(dev_priv->engine[RCS0]);
> + intel_engine_breadcrumbs_irq(i915->engine[RCS0]);
>
> if (iir & I915_BSD_USER_INTERRUPT)
> - intel_engine_breadcrumbs_irq(dev_priv->engine[VCS0]);
> + intel_engine_breadcrumbs_irq(i915->engine[VCS0]);
>
> if (iir & I915_MASTER_ERROR_INTERRUPT)
> - i9xx_error_irq_handler(dev_priv, eir, eir_stuck);
> + i9xx_error_irq_handler(i915, eir, eir_stuck);
>
> if (hotplug_status)
> - i9xx_hpd_irq_handler(dev_priv, hotplug_status);
> + i9xx_hpd_irq_handler(i915, hotplug_status);
>
> - i965_pipestat_irq_handler(dev_priv, iir, pipe_stats);
> + i965_pipestat_irq_handler(i915, iir, pipe_stats);
> } while (0);
>
> - enable_rpm_wakeref_asserts(dev_priv);
> + enable_rpm_wakeref_asserts(i915);
>
> return ret;
> }
>
> /**
> * intel_irq_init - initializes irq support
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> *
> * This function initializes all the irq support including work items, timers
> * and all the vtables. It does not setup the interrupt itself though.
> */
> -void intel_irq_init(struct drm_i915_private *dev_priv)
> +void intel_irq_init(struct drm_i915_private *i915)
> {
> - struct drm_device *dev = &dev_priv->drm;
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct drm_device *dev = &i915->drm;
> + struct intel_rps *rps = &i915->gt_pm.rps;
> int i;
>
> - if (IS_I945GM(dev_priv))
> - i945gm_vblank_work_init(dev_priv);
> + if (IS_I945GM(i915))
> + i945gm_vblank_work_init(i915);
>
> - intel_hpd_init_work(dev_priv);
> + intel_hpd_init_work(i915);
>
> INIT_WORK(&rps->work, gen6_pm_rps_work);
>
> - INIT_WORK(&dev_priv->l3_parity.error_work, ivybridge_parity_work);
> + INIT_WORK(&i915->l3_parity.error_work, ivybridge_parity_work);
> for (i = 0; i < MAX_L3_SLICES; ++i)
> - dev_priv->l3_parity.remap_info[i] = NULL;
> + i915->l3_parity.remap_info[i] = NULL;
>
> - if (HAS_GUC_SCHED(dev_priv) && INTEL_GEN(dev_priv) < 11)
> - dev_priv->pm_guc_events = GEN9_GUC_TO_HOST_INT_EVENT;
> + if (HAS_GUC_SCHED(i915) && INTEL_GEN(i915) < 11)
> + i915->pm_guc_events = GEN9_GUC_TO_HOST_INT_EVENT;
>
> /* Let's track the enabled rps events */
> - if (IS_VALLEYVIEW(dev_priv))
> + if (IS_VALLEYVIEW(i915))
> /* WaGsvRC0ResidencyMethod:vlv */
> - dev_priv->pm_rps_events = GEN6_PM_RP_UP_EI_EXPIRED;
> + i915->pm_rps_events = GEN6_PM_RP_UP_EI_EXPIRED;
> else
> - dev_priv->pm_rps_events = (GEN6_PM_RP_UP_THRESHOLD |
> + i915->pm_rps_events = (GEN6_PM_RP_UP_THRESHOLD |
> GEN6_PM_RP_DOWN_THRESHOLD |
> GEN6_PM_RP_DOWN_TIMEOUT);
>
> /* We share the register with other engine */
> - if (INTEL_GEN(dev_priv) > 9)
> - GEM_WARN_ON(dev_priv->pm_rps_events & 0xffff0000);
> + if (INTEL_GEN(i915) > 9)
> + GEM_WARN_ON(i915->pm_rps_events & 0xffff0000);
>
> rps->pm_intrmsk_mbz = 0;
>
> @@ -4789,15 +4789,15 @@ void intel_irq_init(struct drm_i915_private *dev_priv)
> *
> * TODO: verify if this can be reproduced on VLV,CHV.
> */
> - if (INTEL_GEN(dev_priv) <= 7)
> + if (INTEL_GEN(i915) <= 7)
> rps->pm_intrmsk_mbz |= GEN6_PM_RP_UP_EI_EXPIRED;
>
> - if (INTEL_GEN(dev_priv) >= 8)
> + if (INTEL_GEN(i915) >= 8)
> rps->pm_intrmsk_mbz |= GEN8_PMINTR_DISABLE_REDIRECT_TO_GUC;
>
> - if (INTEL_GEN(dev_priv) >= 5 || IS_G4X(dev_priv))
> + if (INTEL_GEN(i915) >= 5 || IS_G4X(i915))
> dev->driver->get_vblank_counter = g4x_get_vblank_counter;
> - else if (INTEL_GEN(dev_priv) >= 3)
> + else if (INTEL_GEN(i915) >= 3)
> dev->driver->get_vblank_counter = i915_get_vblank_counter;
>
> dev->vblank_disable_immediate = true;
> @@ -4808,83 +4808,83 @@ void intel_irq_init(struct drm_i915_private *dev_priv)
> * outside of the power domain. We defer setting up the display irqs
> * in this case to the runtime pm.
> */
> - dev_priv->display_irqs_enabled = true;
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> - dev_priv->display_irqs_enabled = false;
> + i915->display_irqs_enabled = true;
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> + i915->display_irqs_enabled = false;
>
> - dev_priv->hotplug.hpd_storm_threshold = HPD_STORM_DEFAULT_THRESHOLD;
> + i915->hotplug.hpd_storm_threshold = HPD_STORM_DEFAULT_THRESHOLD;
> /* If we have MST support, we want to avoid doing short HPD IRQ storm
> * detection, as short HPD storms will occur as a natural part of
> * sideband messaging with MST.
> * On older platforms however, IRQ storms can occur with both long and
> * short pulses, as seen on some G4x systems.
> */
> - dev_priv->hotplug.hpd_short_storm_enabled = !HAS_DP_MST(dev_priv);
> + i915->hotplug.hpd_short_storm_enabled = !HAS_DP_MST(i915);
>
> dev->driver->get_vblank_timestamp = drm_calc_vbltimestamp_from_scanoutpos;
> dev->driver->get_scanout_position = i915_get_crtc_scanoutpos;
>
> - if (IS_CHERRYVIEW(dev_priv)) {
> + if (IS_CHERRYVIEW(i915)) {
> dev->driver->irq_handler = cherryview_irq_handler;
> dev->driver->irq_preinstall = cherryview_irq_reset;
> dev->driver->irq_postinstall = cherryview_irq_postinstall;
> dev->driver->irq_uninstall = cherryview_irq_reset;
> dev->driver->enable_vblank = i965_enable_vblank;
> dev->driver->disable_vblank = i965_disable_vblank;
> - dev_priv->display.hpd_irq_setup = i915_hpd_irq_setup;
> - } else if (IS_VALLEYVIEW(dev_priv)) {
> + i915->display.hpd_irq_setup = i915_hpd_irq_setup;
> + } else if (IS_VALLEYVIEW(i915)) {
> dev->driver->irq_handler = valleyview_irq_handler;
> dev->driver->irq_preinstall = valleyview_irq_reset;
> dev->driver->irq_postinstall = valleyview_irq_postinstall;
> dev->driver->irq_uninstall = valleyview_irq_reset;
> dev->driver->enable_vblank = i965_enable_vblank;
> dev->driver->disable_vblank = i965_disable_vblank;
> - dev_priv->display.hpd_irq_setup = i915_hpd_irq_setup;
> - } else if (INTEL_GEN(dev_priv) >= 11) {
> + i915->display.hpd_irq_setup = i915_hpd_irq_setup;
> + } else if (INTEL_GEN(i915) >= 11) {
> dev->driver->irq_handler = gen11_irq_handler;
> dev->driver->irq_preinstall = gen11_irq_reset;
> dev->driver->irq_postinstall = gen11_irq_postinstall;
> dev->driver->irq_uninstall = gen11_irq_reset;
> dev->driver->enable_vblank = gen8_enable_vblank;
> dev->driver->disable_vblank = gen8_disable_vblank;
> - dev_priv->display.hpd_irq_setup = gen11_hpd_irq_setup;
> - } else if (INTEL_GEN(dev_priv) >= 8) {
> + i915->display.hpd_irq_setup = gen11_hpd_irq_setup;
> + } else if (INTEL_GEN(i915) >= 8) {
> dev->driver->irq_handler = gen8_irq_handler;
> dev->driver->irq_preinstall = gen8_irq_reset;
> dev->driver->irq_postinstall = gen8_irq_postinstall;
> dev->driver->irq_uninstall = gen8_irq_reset;
> dev->driver->enable_vblank = gen8_enable_vblank;
> dev->driver->disable_vblank = gen8_disable_vblank;
> - if (IS_GEN9_LP(dev_priv))
> - dev_priv->display.hpd_irq_setup = bxt_hpd_irq_setup;
> - else if (INTEL_PCH_TYPE(dev_priv) >= PCH_SPT)
> - dev_priv->display.hpd_irq_setup = spt_hpd_irq_setup;
> + if (IS_GEN9_LP(i915))
> + i915->display.hpd_irq_setup = bxt_hpd_irq_setup;
> + else if (INTEL_PCH_TYPE(i915) >= PCH_SPT)
> + i915->display.hpd_irq_setup = spt_hpd_irq_setup;
> else
> - dev_priv->display.hpd_irq_setup = ilk_hpd_irq_setup;
> - } else if (HAS_PCH_SPLIT(dev_priv)) {
> + i915->display.hpd_irq_setup = ilk_hpd_irq_setup;
> + } else if (HAS_PCH_SPLIT(i915)) {
> dev->driver->irq_handler = ironlake_irq_handler;
> dev->driver->irq_preinstall = ironlake_irq_reset;
> dev->driver->irq_postinstall = ironlake_irq_postinstall;
> dev->driver->irq_uninstall = ironlake_irq_reset;
> dev->driver->enable_vblank = ironlake_enable_vblank;
> dev->driver->disable_vblank = ironlake_disable_vblank;
> - dev_priv->display.hpd_irq_setup = ilk_hpd_irq_setup;
> + i915->display.hpd_irq_setup = ilk_hpd_irq_setup;
> } else {
> - if (IS_GEN(dev_priv, 2)) {
> + if (IS_GEN(i915, 2)) {
> dev->driver->irq_preinstall = i8xx_irq_reset;
> dev->driver->irq_postinstall = i8xx_irq_postinstall;
> dev->driver->irq_handler = i8xx_irq_handler;
> dev->driver->irq_uninstall = i8xx_irq_reset;
> dev->driver->enable_vblank = i8xx_enable_vblank;
> dev->driver->disable_vblank = i8xx_disable_vblank;
> - } else if (IS_I945GM(dev_priv)) {
> + } else if (IS_I945GM(i915)) {
> dev->driver->irq_preinstall = i915_irq_reset;
> dev->driver->irq_postinstall = i915_irq_postinstall;
> dev->driver->irq_uninstall = i915_irq_reset;
> dev->driver->irq_handler = i915_irq_handler;
> dev->driver->enable_vblank = i945gm_enable_vblank;
> dev->driver->disable_vblank = i945gm_disable_vblank;
> - } else if (IS_GEN(dev_priv, 3)) {
> + } else if (IS_GEN(i915, 3)) {
> dev->driver->irq_preinstall = i915_irq_reset;
> dev->driver->irq_postinstall = i915_irq_postinstall;
> dev->driver->irq_uninstall = i915_irq_reset;
> @@ -4899,8 +4899,8 @@ void intel_irq_init(struct drm_i915_private *dev_priv)
> dev->driver->enable_vblank = i965_enable_vblank;
> dev->driver->disable_vblank = i965_disable_vblank;
> }
> - if (I915_HAS_HOTPLUG(dev_priv))
> - dev_priv->display.hpd_irq_setup = i915_hpd_irq_setup;
> + if (I915_HAS_HOTPLUG(i915))
> + i915->display.hpd_irq_setup = i915_hpd_irq_setup;
> }
> }
>
> @@ -4923,7 +4923,7 @@ void intel_irq_fini(struct drm_i915_private *i915)
>
> /**
> * intel_irq_install - enables the hardware interrupt
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> *
> * This function enables the hardware interrupt handling, but leaves the hotplug
> * handling still disabled. It is called after intel_irq_init().
> @@ -4932,56 +4932,56 @@ void intel_irq_fini(struct drm_i915_private *i915)
> * but don't want to deal with the hassle of concurrent probe and hotplug
> * workers. Hence the split into this two-stage approach.
> */
> -int intel_irq_install(struct drm_i915_private *dev_priv)
> +int intel_irq_install(struct drm_i915_private *i915)
> {
> /*
> * We enable some interrupt sources in our postinstall hooks, so mark
> * interrupts as enabled _before_ actually enabling them to avoid
> * special cases in our ordering checks.
> */
> - dev_priv->runtime_pm.irqs_enabled = true;
> + i915->runtime_pm.irqs_enabled = true;
>
> - return drm_irq_install(&dev_priv->drm, dev_priv->drm.pdev->irq);
> + return drm_irq_install(&i915->drm, i915->drm.pdev->irq);
> }
>
> /**
> * intel_irq_uninstall - finilizes all irq handling
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> *
> * This stops interrupt and hotplug handling and unregisters and frees all
> * resources acquired in the init functions.
> */
> -void intel_irq_uninstall(struct drm_i915_private *dev_priv)
> +void intel_irq_uninstall(struct drm_i915_private *i915)
> {
> - drm_irq_uninstall(&dev_priv->drm);
> - intel_hpd_cancel_work(dev_priv);
> - dev_priv->runtime_pm.irqs_enabled = false;
> + drm_irq_uninstall(&i915->drm);
> + intel_hpd_cancel_work(i915);
> + i915->runtime_pm.irqs_enabled = false;
> }
>
> /**
> * intel_runtime_pm_disable_interrupts - runtime interrupt disabling
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> *
> * This function is used to disable interrupts at runtime, both in the runtime
> * pm and the system suspend/resume code.
> */
> -void intel_runtime_pm_disable_interrupts(struct drm_i915_private *dev_priv)
> +void intel_runtime_pm_disable_interrupts(struct drm_i915_private *i915)
> {
> - dev_priv->drm.driver->irq_uninstall(&dev_priv->drm);
> - dev_priv->runtime_pm.irqs_enabled = false;
> - synchronize_irq(dev_priv->drm.irq);
> + i915->drm.driver->irq_uninstall(&i915->drm);
> + i915->runtime_pm.irqs_enabled = false;
> + synchronize_irq(i915->drm.irq);
> }
>
> /**
> * intel_runtime_pm_enable_interrupts - runtime interrupt enabling
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> *
> * This function is used to enable interrupts at runtime, both in the runtime
> * pm and the system suspend/resume code.
> */
> -void intel_runtime_pm_enable_interrupts(struct drm_i915_private *dev_priv)
> +void intel_runtime_pm_enable_interrupts(struct drm_i915_private *i915)
> {
> - dev_priv->runtime_pm.irqs_enabled = true;
> - dev_priv->drm.driver->irq_preinstall(&dev_priv->drm);
> - dev_priv->drm.driver->irq_postinstall(&dev_priv->drm);
> + i915->runtime_pm.irqs_enabled = true;
> + i915->drm.driver->irq_preinstall(&i915->drm);
> + i915->drm.driver->irq_postinstall(&i915->drm);
> }
> diff --git a/drivers/gpu/drm/i915/i915_irq.h b/drivers/gpu/drm/i915/i915_irq.h
> index cb25dd213308..604504993513 100644
> --- a/drivers/gpu/drm/i915/i915_irq.h
> +++ b/drivers/gpu/drm/i915/i915_irq.h
> @@ -13,77 +13,77 @@
> struct drm_i915_private;
> struct intel_crtc;
>
> -extern void intel_irq_init(struct drm_i915_private *dev_priv);
> -extern void intel_irq_fini(struct drm_i915_private *dev_priv);
> -int intel_irq_install(struct drm_i915_private *dev_priv);
> -void intel_irq_uninstall(struct drm_i915_private *dev_priv);
> +extern void intel_irq_init(struct drm_i915_private *i915);
> +extern void intel_irq_fini(struct drm_i915_private *i915);
> +int intel_irq_install(struct drm_i915_private *i915);
> +void intel_irq_uninstall(struct drm_i915_private *i915);
>
> -u32 i915_pipestat_enable_mask(struct drm_i915_private *dev_priv,
> +u32 i915_pipestat_enable_mask(struct drm_i915_private *i915,
> enum pipe pipe);
> void
> -i915_enable_pipestat(struct drm_i915_private *dev_priv, enum pipe pipe,
> +i915_enable_pipestat(struct drm_i915_private *i915, enum pipe pipe,
> u32 status_mask);
>
> void
> -i915_disable_pipestat(struct drm_i915_private *dev_priv, enum pipe pipe,
> +i915_disable_pipestat(struct drm_i915_private *i915, enum pipe pipe,
> u32 status_mask);
>
> -void valleyview_enable_display_irqs(struct drm_i915_private *dev_priv);
> -void valleyview_disable_display_irqs(struct drm_i915_private *dev_priv);
> +void valleyview_enable_display_irqs(struct drm_i915_private *i915);
> +void valleyview_disable_display_irqs(struct drm_i915_private *i915);
>
> -void i915_hotplug_interrupt_update(struct drm_i915_private *dev_priv,
> +void i915_hotplug_interrupt_update(struct drm_i915_private *i915,
> u32 mask,
> u32 bits);
> -void ilk_update_display_irq(struct drm_i915_private *dev_priv,
> +void ilk_update_display_irq(struct drm_i915_private *i915,
> u32 interrupt_mask,
> u32 enabled_irq_mask);
> static inline void
> -ilk_enable_display_irq(struct drm_i915_private *dev_priv, u32 bits)
> +ilk_enable_display_irq(struct drm_i915_private *i915, u32 bits)
> {
> - ilk_update_display_irq(dev_priv, bits, bits);
> + ilk_update_display_irq(i915, bits, bits);
> }
> static inline void
> -ilk_disable_display_irq(struct drm_i915_private *dev_priv, u32 bits)
> +ilk_disable_display_irq(struct drm_i915_private *i915, u32 bits)
> {
> - ilk_update_display_irq(dev_priv, bits, 0);
> + ilk_update_display_irq(i915, bits, 0);
> }
> -void bdw_update_pipe_irq(struct drm_i915_private *dev_priv,
> +void bdw_update_pipe_irq(struct drm_i915_private *i915,
> enum pipe pipe,
> u32 interrupt_mask,
> u32 enabled_irq_mask);
> -static inline void bdw_enable_pipe_irq(struct drm_i915_private *dev_priv,
> +static inline void bdw_enable_pipe_irq(struct drm_i915_private *i915,
> enum pipe pipe, u32 bits)
> {
> - bdw_update_pipe_irq(dev_priv, pipe, bits, bits);
> + bdw_update_pipe_irq(i915, pipe, bits, bits);
> }
> -static inline void bdw_disable_pipe_irq(struct drm_i915_private *dev_priv,
> +static inline void bdw_disable_pipe_irq(struct drm_i915_private *i915,
> enum pipe pipe, u32 bits)
> {
> - bdw_update_pipe_irq(dev_priv, pipe, bits, 0);
> + bdw_update_pipe_irq(i915, pipe, bits, 0);
> }
> -void ibx_display_interrupt_update(struct drm_i915_private *dev_priv,
> +void ibx_display_interrupt_update(struct drm_i915_private *i915,
> u32 interrupt_mask,
> u32 enabled_irq_mask);
> static inline void
> -ibx_enable_display_interrupt(struct drm_i915_private *dev_priv, u32 bits)
> +ibx_enable_display_interrupt(struct drm_i915_private *i915, u32 bits)
> {
> - ibx_display_interrupt_update(dev_priv, bits, bits);
> + ibx_display_interrupt_update(i915, bits, bits);
> }
> static inline void
> -ibx_disable_display_interrupt(struct drm_i915_private *dev_priv, u32 bits)
> +ibx_disable_display_interrupt(struct drm_i915_private *i915, u32 bits)
> {
> - ibx_display_interrupt_update(dev_priv, bits, 0);
> + ibx_display_interrupt_update(i915, bits, 0);
> }
>
> -void gen5_enable_gt_irq(struct drm_i915_private *dev_priv, u32 mask);
> -void gen5_disable_gt_irq(struct drm_i915_private *dev_priv, u32 mask);
> -void gen6_mask_pm_irq(struct drm_i915_private *dev_priv, u32 mask);
> -void gen6_unmask_pm_irq(struct drm_i915_private *dev_priv, u32 mask);
> -void gen11_reset_rps_interrupts(struct drm_i915_private *dev_priv);
> -void gen6_reset_rps_interrupts(struct drm_i915_private *dev_priv);
> -void gen6_enable_rps_interrupts(struct drm_i915_private *dev_priv);
> -void gen6_disable_rps_interrupts(struct drm_i915_private *dev_priv);
> -void gen6_rps_reset_ei(struct drm_i915_private *dev_priv);
> +void gen5_enable_gt_irq(struct drm_i915_private *i915, u32 mask);
> +void gen5_disable_gt_irq(struct drm_i915_private *i915, u32 mask);
> +void gen6_mask_pm_irq(struct drm_i915_private *i915, u32 mask);
> +void gen6_unmask_pm_irq(struct drm_i915_private *i915, u32 mask);
> +void gen11_reset_rps_interrupts(struct drm_i915_private *i915);
> +void gen6_reset_rps_interrupts(struct drm_i915_private *i915);
> +void gen6_enable_rps_interrupts(struct drm_i915_private *i915);
> +void gen6_disable_rps_interrupts(struct drm_i915_private *i915);
> +void gen6_rps_reset_ei(struct drm_i915_private *i915);
>
> static inline u32 gen6_sanitize_rps_pm_mask(const struct drm_i915_private *i915,
> u32 mask)
> @@ -91,25 +91,25 @@ static inline u32 gen6_sanitize_rps_pm_mask(const struct drm_i915_private *i915,
> return mask & ~i915->gt_pm.rps.pm_intrmsk_mbz;
> }
>
> -void intel_runtime_pm_disable_interrupts(struct drm_i915_private *dev_priv);
> -void intel_runtime_pm_enable_interrupts(struct drm_i915_private *dev_priv);
> -static inline bool intel_irqs_enabled(struct drm_i915_private *dev_priv)
> +void intel_runtime_pm_disable_interrupts(struct drm_i915_private *i915);
> +void intel_runtime_pm_enable_interrupts(struct drm_i915_private *i915);
> +static inline bool intel_irqs_enabled(struct drm_i915_private *i915)
> {
> /*
> * We only use drm_irq_uninstall() at unload and VT switch, so
> * this is the only thing we need to check.
> */
> - return dev_priv->runtime_pm.irqs_enabled;
> + return i915->runtime_pm.irqs_enabled;
> }
>
> int intel_get_crtc_scanline(struct intel_crtc *crtc);
> -void gen8_irq_power_well_post_enable(struct drm_i915_private *dev_priv,
> +void gen8_irq_power_well_post_enable(struct drm_i915_private *i915,
> u8 pipe_mask);
> -void gen8_irq_power_well_pre_disable(struct drm_i915_private *dev_priv,
> +void gen8_irq_power_well_pre_disable(struct drm_i915_private *i915,
> u8 pipe_mask);
> -void gen9_reset_guc_interrupts(struct drm_i915_private *dev_priv);
> -void gen9_enable_guc_interrupts(struct drm_i915_private *dev_priv);
> -void gen9_disable_guc_interrupts(struct drm_i915_private *dev_priv);
> +void gen9_reset_guc_interrupts(struct drm_i915_private *i915);
> +void gen9_enable_guc_interrupts(struct drm_i915_private *i915);
> +void gen9_disable_guc_interrupts(struct drm_i915_private *i915);
> void gen11_reset_guc_interrupts(struct drm_i915_private *i915);
> void gen11_enable_guc_interrupts(struct drm_i915_private *i915);
> void gen11_disable_guc_interrupts(struct drm_i915_private *i915);
> diff --git a/drivers/gpu/drm/i915/i915_memcpy.c b/drivers/gpu/drm/i915/i915_memcpy.c
> index 79f8ec756362..2bd664c16ba0 100644
> --- a/drivers/gpu/drm/i915/i915_memcpy.c
> +++ b/drivers/gpu/drm/i915/i915_memcpy.c
> @@ -94,7 +94,7 @@ bool i915_memcpy_from_wc(void *dst, const void *src, unsigned long len)
> return false;
> }
>
> -void i915_memcpy_init_early(struct drm_i915_private *dev_priv)
> +void i915_memcpy_init_early(struct drm_i915_private *i915)
> {
> /*
> * Some hypervisors (e.g. KVM) don't support VEX-prefix instructions
> diff --git a/drivers/gpu/drm/i915/i915_oa_bdw.c b/drivers/gpu/drm/i915/i915_oa_bdw.c
> index 4acdb94555b7..7c97cd90e973 100644
> --- a/drivers/gpu/drm/i915/i915_oa_bdw.c
> +++ b/drivers/gpu/drm/i915/i915_oa_bdw.c
> @@ -64,28 +64,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf)
> }
>
> void
> -i915_perf_load_test_config_bdw(struct drm_i915_private *dev_priv)
> +i915_perf_load_test_config_bdw(struct drm_i915_private *i915)
> {
> - strlcpy(dev_priv->perf.oa.test_config.uuid,
> + strlcpy(i915->perf.oa.test_config.uuid,
> "d6de6f55-e526-4f79-a6a6-d7315c09044e",
> - sizeof(dev_priv->perf.oa.test_config.uuid));
> - dev_priv->perf.oa.test_config.id = 1;
> + sizeof(i915->perf.oa.test_config.uuid));
> + i915->perf.oa.test_config.id = 1;
>
> - dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa;
> - dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> + i915->perf.oa.test_config.mux_regs = mux_config_test_oa;
> + i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
>
> - dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> - dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> + i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> + i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
>
> - dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> - dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> + i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> + i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
>
> - dev_priv->perf.oa.test_config.sysfs_metric.name = "d6de6f55-e526-4f79-a6a6-d7315c09044e";
> - dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs;
> + i915->perf.oa.test_config.sysfs_metric.name = "d6de6f55-e526-4f79-a6a6-d7315c09044e";
> + i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs;
>
> - dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr;
> + i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr;
>
> - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> - dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> + i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> + i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> + i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> }
> diff --git a/drivers/gpu/drm/i915/i915_oa_bdw.h b/drivers/gpu/drm/i915/i915_oa_bdw.h
> index 0e667f1a8aa1..69479e5b436f 100644
> --- a/drivers/gpu/drm/i915/i915_oa_bdw.h
> +++ b/drivers/gpu/drm/i915/i915_oa_bdw.h
> @@ -10,6 +10,6 @@
> #ifndef __I915_OA_BDW_H__
> #define __I915_OA_BDW_H__
>
> -extern void i915_perf_load_test_config_bdw(struct drm_i915_private *dev_priv);
> +extern void i915_perf_load_test_config_bdw(struct drm_i915_private *i915);
>
> #endif
> diff --git a/drivers/gpu/drm/i915/i915_oa_bxt.c b/drivers/gpu/drm/i915/i915_oa_bxt.c
> index a44195c39923..5ec236b33ecc 100644
> --- a/drivers/gpu/drm/i915/i915_oa_bxt.c
> +++ b/drivers/gpu/drm/i915/i915_oa_bxt.c
> @@ -62,28 +62,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf)
> }
>
> void
> -i915_perf_load_test_config_bxt(struct drm_i915_private *dev_priv)
> +i915_perf_load_test_config_bxt(struct drm_i915_private *i915)
> {
> - strlcpy(dev_priv->perf.oa.test_config.uuid,
> + strlcpy(i915->perf.oa.test_config.uuid,
> "5ee72f5c-092f-421e-8b70-225f7c3e9612",
> - sizeof(dev_priv->perf.oa.test_config.uuid));
> - dev_priv->perf.oa.test_config.id = 1;
> + sizeof(i915->perf.oa.test_config.uuid));
> + i915->perf.oa.test_config.id = 1;
>
> - dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa;
> - dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> + i915->perf.oa.test_config.mux_regs = mux_config_test_oa;
> + i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
>
> - dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> - dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> + i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> + i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
>
> - dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> - dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> + i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> + i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
>
> - dev_priv->perf.oa.test_config.sysfs_metric.name = "5ee72f5c-092f-421e-8b70-225f7c3e9612";
> - dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs;
> + i915->perf.oa.test_config.sysfs_metric.name = "5ee72f5c-092f-421e-8b70-225f7c3e9612";
> + i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs;
>
> - dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr;
> + i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr;
>
> - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> - dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> + i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> + i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> + i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> }
> diff --git a/drivers/gpu/drm/i915/i915_oa_bxt.h b/drivers/gpu/drm/i915/i915_oa_bxt.h
> index 679e92cf4f1d..47d8327f38a1 100644
> --- a/drivers/gpu/drm/i915/i915_oa_bxt.h
> +++ b/drivers/gpu/drm/i915/i915_oa_bxt.h
> @@ -10,6 +10,6 @@
> #ifndef __I915_OA_BXT_H__
> #define __I915_OA_BXT_H__
>
> -extern void i915_perf_load_test_config_bxt(struct drm_i915_private *dev_priv);
> +extern void i915_perf_load_test_config_bxt(struct drm_i915_private *i915);
>
> #endif
> diff --git a/drivers/gpu/drm/i915/i915_oa_cflgt2.c b/drivers/gpu/drm/i915/i915_oa_cflgt2.c
> index 7f60d51b8761..3972cd7dcbfc 100644
> --- a/drivers/gpu/drm/i915/i915_oa_cflgt2.c
> +++ b/drivers/gpu/drm/i915/i915_oa_cflgt2.c
> @@ -63,28 +63,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf)
> }
>
> void
> -i915_perf_load_test_config_cflgt2(struct drm_i915_private *dev_priv)
> +i915_perf_load_test_config_cflgt2(struct drm_i915_private *i915)
> {
> - strlcpy(dev_priv->perf.oa.test_config.uuid,
> + strlcpy(i915->perf.oa.test_config.uuid,
> "74fb4902-d3d3-4237-9e90-cbdc68d0a446",
> - sizeof(dev_priv->perf.oa.test_config.uuid));
> - dev_priv->perf.oa.test_config.id = 1;
> + sizeof(i915->perf.oa.test_config.uuid));
> + i915->perf.oa.test_config.id = 1;
>
> - dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa;
> - dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> + i915->perf.oa.test_config.mux_regs = mux_config_test_oa;
> + i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
>
> - dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> - dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> + i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> + i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
>
> - dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> - dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> + i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> + i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
>
> - dev_priv->perf.oa.test_config.sysfs_metric.name = "74fb4902-d3d3-4237-9e90-cbdc68d0a446";
> - dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs;
> + i915->perf.oa.test_config.sysfs_metric.name = "74fb4902-d3d3-4237-9e90-cbdc68d0a446";
> + i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs;
>
> - dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr;
> + i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr;
>
> - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> - dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> + i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> + i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> + i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> }
> diff --git a/drivers/gpu/drm/i915/i915_oa_cflgt2.h b/drivers/gpu/drm/i915/i915_oa_cflgt2.h
> index 4d6025559bbe..434545c9c5d9 100644
> --- a/drivers/gpu/drm/i915/i915_oa_cflgt2.h
> +++ b/drivers/gpu/drm/i915/i915_oa_cflgt2.h
> @@ -10,6 +10,6 @@
> #ifndef __I915_OA_CFLGT2_H__
> #define __I915_OA_CFLGT2_H__
>
> -extern void i915_perf_load_test_config_cflgt2(struct drm_i915_private *dev_priv);
> +extern void i915_perf_load_test_config_cflgt2(struct drm_i915_private *i915);
>
> #endif
> diff --git a/drivers/gpu/drm/i915/i915_oa_cflgt3.c b/drivers/gpu/drm/i915/i915_oa_cflgt3.c
> index a92c38e3a0ce..c00b9d9d6ddf 100644
> --- a/drivers/gpu/drm/i915/i915_oa_cflgt3.c
> +++ b/drivers/gpu/drm/i915/i915_oa_cflgt3.c
> @@ -63,28 +63,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf)
> }
>
> void
> -i915_perf_load_test_config_cflgt3(struct drm_i915_private *dev_priv)
> +i915_perf_load_test_config_cflgt3(struct drm_i915_private *i915)
> {
> - strlcpy(dev_priv->perf.oa.test_config.uuid,
> + strlcpy(i915->perf.oa.test_config.uuid,
> "577e8e2c-3fa0-4875-8743-3538d585e3b0",
> - sizeof(dev_priv->perf.oa.test_config.uuid));
> - dev_priv->perf.oa.test_config.id = 1;
> + sizeof(i915->perf.oa.test_config.uuid));
> + i915->perf.oa.test_config.id = 1;
>
> - dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa;
> - dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> + i915->perf.oa.test_config.mux_regs = mux_config_test_oa;
> + i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
>
> - dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> - dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> + i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> + i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
>
> - dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> - dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> + i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> + i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
>
> - dev_priv->perf.oa.test_config.sysfs_metric.name = "577e8e2c-3fa0-4875-8743-3538d585e3b0";
> - dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs;
> + i915->perf.oa.test_config.sysfs_metric.name = "577e8e2c-3fa0-4875-8743-3538d585e3b0";
> + i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs;
>
> - dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr;
> + i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr;
>
> - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> - dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> + i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> + i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> + i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> }
> diff --git a/drivers/gpu/drm/i915/i915_oa_cflgt3.h b/drivers/gpu/drm/i915/i915_oa_cflgt3.h
> index 0697f4077402..2c23eda056aa 100644
> --- a/drivers/gpu/drm/i915/i915_oa_cflgt3.h
> +++ b/drivers/gpu/drm/i915/i915_oa_cflgt3.h
> @@ -10,6 +10,6 @@
> #ifndef __I915_OA_CFLGT3_H__
> #define __I915_OA_CFLGT3_H__
>
> -extern void i915_perf_load_test_config_cflgt3(struct drm_i915_private *dev_priv);
> +extern void i915_perf_load_test_config_cflgt3(struct drm_i915_private *i915);
>
> #endif
> diff --git a/drivers/gpu/drm/i915/i915_oa_chv.c b/drivers/gpu/drm/i915/i915_oa_chv.c
> index 71ec889a0114..5829d24d0f7c 100644
> --- a/drivers/gpu/drm/i915/i915_oa_chv.c
> +++ b/drivers/gpu/drm/i915/i915_oa_chv.c
> @@ -63,28 +63,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf)
> }
>
> void
> -i915_perf_load_test_config_chv(struct drm_i915_private *dev_priv)
> +i915_perf_load_test_config_chv(struct drm_i915_private *i915)
> {
> - strlcpy(dev_priv->perf.oa.test_config.uuid,
> + strlcpy(i915->perf.oa.test_config.uuid,
> "4a534b07-cba3-414d-8d60-874830e883aa",
> - sizeof(dev_priv->perf.oa.test_config.uuid));
> - dev_priv->perf.oa.test_config.id = 1;
> + sizeof(i915->perf.oa.test_config.uuid));
> + i915->perf.oa.test_config.id = 1;
>
> - dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa;
> - dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> + i915->perf.oa.test_config.mux_regs = mux_config_test_oa;
> + i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
>
> - dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> - dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> + i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> + i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
>
> - dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> - dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> + i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> + i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
>
> - dev_priv->perf.oa.test_config.sysfs_metric.name = "4a534b07-cba3-414d-8d60-874830e883aa";
> - dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs;
> + i915->perf.oa.test_config.sysfs_metric.name = "4a534b07-cba3-414d-8d60-874830e883aa";
> + i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs;
>
> - dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr;
> + i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr;
>
> - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> - dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> + i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> + i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> + i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> }
> diff --git a/drivers/gpu/drm/i915/i915_oa_chv.h b/drivers/gpu/drm/i915/i915_oa_chv.h
> index 0986eae3135f..4bb3b6be7bd2 100644
> --- a/drivers/gpu/drm/i915/i915_oa_chv.h
> +++ b/drivers/gpu/drm/i915/i915_oa_chv.h
> @@ -10,6 +10,6 @@
> #ifndef __I915_OA_CHV_H__
> #define __I915_OA_CHV_H__
>
> -extern void i915_perf_load_test_config_chv(struct drm_i915_private *dev_priv);
> +extern void i915_perf_load_test_config_chv(struct drm_i915_private *i915);
>
> #endif
> diff --git a/drivers/gpu/drm/i915/i915_oa_cnl.c b/drivers/gpu/drm/i915/i915_oa_cnl.c
> index 5c23d883d6c9..fcf08b9e7c73 100644
> --- a/drivers/gpu/drm/i915/i915_oa_cnl.c
> +++ b/drivers/gpu/drm/i915/i915_oa_cnl.c
> @@ -75,28 +75,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf)
> }
>
> void
> -i915_perf_load_test_config_cnl(struct drm_i915_private *dev_priv)
> +i915_perf_load_test_config_cnl(struct drm_i915_private *i915)
> {
> - strlcpy(dev_priv->perf.oa.test_config.uuid,
> + strlcpy(i915->perf.oa.test_config.uuid,
> "db41edd4-d8e7-4730-ad11-b9a2d6833503",
> - sizeof(dev_priv->perf.oa.test_config.uuid));
> - dev_priv->perf.oa.test_config.id = 1;
> + sizeof(i915->perf.oa.test_config.uuid));
> + i915->perf.oa.test_config.id = 1;
>
> - dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa;
> - dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> + i915->perf.oa.test_config.mux_regs = mux_config_test_oa;
> + i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
>
> - dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> - dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> + i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> + i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
>
> - dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> - dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> + i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> + i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
>
> - dev_priv->perf.oa.test_config.sysfs_metric.name = "db41edd4-d8e7-4730-ad11-b9a2d6833503";
> - dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs;
> + i915->perf.oa.test_config.sysfs_metric.name = "db41edd4-d8e7-4730-ad11-b9a2d6833503";
> + i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs;
>
> - dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr;
> + i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr;
>
> - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> - dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> + i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> + i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> + i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> }
> diff --git a/drivers/gpu/drm/i915/i915_oa_cnl.h b/drivers/gpu/drm/i915/i915_oa_cnl.h
> index e830a406aff2..c70101655049 100644
> --- a/drivers/gpu/drm/i915/i915_oa_cnl.h
> +++ b/drivers/gpu/drm/i915/i915_oa_cnl.h
> @@ -10,6 +10,6 @@
> #ifndef __I915_OA_CNL_H__
> #define __I915_OA_CNL_H__
>
> -extern void i915_perf_load_test_config_cnl(struct drm_i915_private *dev_priv);
> +extern void i915_perf_load_test_config_cnl(struct drm_i915_private *i915);
>
> #endif
> diff --git a/drivers/gpu/drm/i915/i915_oa_glk.c b/drivers/gpu/drm/i915/i915_oa_glk.c
> index 4bdda66df7d2..fdb38b9ea9bb 100644
> --- a/drivers/gpu/drm/i915/i915_oa_glk.c
> +++ b/drivers/gpu/drm/i915/i915_oa_glk.c
> @@ -62,28 +62,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf)
> }
>
> void
> -i915_perf_load_test_config_glk(struct drm_i915_private *dev_priv)
> +i915_perf_load_test_config_glk(struct drm_i915_private *i915)
> {
> - strlcpy(dev_priv->perf.oa.test_config.uuid,
> + strlcpy(i915->perf.oa.test_config.uuid,
> "dd3fd789-e783-4204-8cd0-b671bbccb0cf",
> - sizeof(dev_priv->perf.oa.test_config.uuid));
> - dev_priv->perf.oa.test_config.id = 1;
> + sizeof(i915->perf.oa.test_config.uuid));
> + i915->perf.oa.test_config.id = 1;
>
> - dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa;
> - dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> + i915->perf.oa.test_config.mux_regs = mux_config_test_oa;
> + i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
>
> - dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> - dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> + i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> + i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
>
> - dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> - dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> + i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> + i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
>
> - dev_priv->perf.oa.test_config.sysfs_metric.name = "dd3fd789-e783-4204-8cd0-b671bbccb0cf";
> - dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs;
> + i915->perf.oa.test_config.sysfs_metric.name = "dd3fd789-e783-4204-8cd0-b671bbccb0cf";
> + i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs;
>
> - dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr;
> + i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr;
>
> - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> - dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> + i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> + i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> + i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> }
> diff --git a/drivers/gpu/drm/i915/i915_oa_glk.h b/drivers/gpu/drm/i915/i915_oa_glk.h
> index 06dedf991edb..5d07986827c8 100644
> --- a/drivers/gpu/drm/i915/i915_oa_glk.h
> +++ b/drivers/gpu/drm/i915/i915_oa_glk.h
> @@ -10,6 +10,6 @@
> #ifndef __I915_OA_GLK_H__
> #define __I915_OA_GLK_H__
>
> -extern void i915_perf_load_test_config_glk(struct drm_i915_private *dev_priv);
> +extern void i915_perf_load_test_config_glk(struct drm_i915_private *i915);
>
> #endif
> diff --git a/drivers/gpu/drm/i915/i915_oa_hsw.c b/drivers/gpu/drm/i915/i915_oa_hsw.c
> index cc6526fdd2bd..640301c9fb9d 100644
> --- a/drivers/gpu/drm/i915/i915_oa_hsw.c
> +++ b/drivers/gpu/drm/i915/i915_oa_hsw.c
> @@ -92,28 +92,28 @@ show_render_basic_id(struct device *kdev, struct device_attribute *attr, char *b
> }
>
> void
> -i915_perf_load_test_config_hsw(struct drm_i915_private *dev_priv)
> +i915_perf_load_test_config_hsw(struct drm_i915_private *i915)
> {
> - strlcpy(dev_priv->perf.oa.test_config.uuid,
> + strlcpy(i915->perf.oa.test_config.uuid,
> "403d8832-1a27-4aa6-a64e-f5389ce7b212",
> - sizeof(dev_priv->perf.oa.test_config.uuid));
> - dev_priv->perf.oa.test_config.id = 1;
> + sizeof(i915->perf.oa.test_config.uuid));
> + i915->perf.oa.test_config.id = 1;
>
> - dev_priv->perf.oa.test_config.mux_regs = mux_config_render_basic;
> - dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_render_basic);
> + i915->perf.oa.test_config.mux_regs = mux_config_render_basic;
> + i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_render_basic);
>
> - dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_render_basic;
> - dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_render_basic);
> + i915->perf.oa.test_config.b_counter_regs = b_counter_config_render_basic;
> + i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_render_basic);
>
> - dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_render_basic;
> - dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_render_basic);
> + i915->perf.oa.test_config.flex_regs = flex_eu_config_render_basic;
> + i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_render_basic);
>
> - dev_priv->perf.oa.test_config.sysfs_metric.name = "403d8832-1a27-4aa6-a64e-f5389ce7b212";
> - dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs;
> + i915->perf.oa.test_config.sysfs_metric.name = "403d8832-1a27-4aa6-a64e-f5389ce7b212";
> + i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs;
>
> - dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr;
> + i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr;
>
> - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> - dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_render_basic_id;
> + i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> + i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> + i915->perf.oa.test_config.sysfs_metric_id.show = show_render_basic_id;
> }
> diff --git a/drivers/gpu/drm/i915/i915_oa_hsw.h b/drivers/gpu/drm/i915/i915_oa_hsw.h
> index 3d0c870cd0bd..7812a9df8647 100644
> --- a/drivers/gpu/drm/i915/i915_oa_hsw.h
> +++ b/drivers/gpu/drm/i915/i915_oa_hsw.h
> @@ -10,6 +10,6 @@
> #ifndef __I915_OA_HSW_H__
> #define __I915_OA_HSW_H__
>
> -extern void i915_perf_load_test_config_hsw(struct drm_i915_private *dev_priv);
> +extern void i915_perf_load_test_config_hsw(struct drm_i915_private *i915);
>
> #endif
> diff --git a/drivers/gpu/drm/i915/i915_oa_icl.c b/drivers/gpu/drm/i915/i915_oa_icl.c
> index baa51427a543..2a8c6de619ba 100644
> --- a/drivers/gpu/drm/i915/i915_oa_icl.c
> +++ b/drivers/gpu/drm/i915/i915_oa_icl.c
> @@ -72,28 +72,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf)
> }
>
> void
> -i915_perf_load_test_config_icl(struct drm_i915_private *dev_priv)
> +i915_perf_load_test_config_icl(struct drm_i915_private *i915)
> {
> - strlcpy(dev_priv->perf.oa.test_config.uuid,
> + strlcpy(i915->perf.oa.test_config.uuid,
> "a291665e-244b-4b76-9b9a-01de9d3c8068",
> - sizeof(dev_priv->perf.oa.test_config.uuid));
> - dev_priv->perf.oa.test_config.id = 1;
> + sizeof(i915->perf.oa.test_config.uuid));
> + i915->perf.oa.test_config.id = 1;
>
> - dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa;
> - dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> + i915->perf.oa.test_config.mux_regs = mux_config_test_oa;
> + i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
>
> - dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> - dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> + i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> + i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
>
> - dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> - dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> + i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> + i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
>
> - dev_priv->perf.oa.test_config.sysfs_metric.name = "a291665e-244b-4b76-9b9a-01de9d3c8068";
> - dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs;
> + i915->perf.oa.test_config.sysfs_metric.name = "a291665e-244b-4b76-9b9a-01de9d3c8068";
> + i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs;
>
> - dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr;
> + i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr;
>
> - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> - dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> + i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> + i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> + i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> }
> diff --git a/drivers/gpu/drm/i915/i915_oa_icl.h b/drivers/gpu/drm/i915/i915_oa_icl.h
> index 24eaa97d61ba..a7c50d31e678 100644
> --- a/drivers/gpu/drm/i915/i915_oa_icl.h
> +++ b/drivers/gpu/drm/i915/i915_oa_icl.h
> @@ -10,6 +10,6 @@
> #ifndef __I915_OA_ICL_H__
> #define __I915_OA_ICL_H__
>
> -extern void i915_perf_load_test_config_icl(struct drm_i915_private *dev_priv);
> +extern void i915_perf_load_test_config_icl(struct drm_i915_private *i915);
>
> #endif
> diff --git a/drivers/gpu/drm/i915/i915_oa_kblgt2.c b/drivers/gpu/drm/i915/i915_oa_kblgt2.c
> index 168e49ab0d4d..975311df3d62 100644
> --- a/drivers/gpu/drm/i915/i915_oa_kblgt2.c
> +++ b/drivers/gpu/drm/i915/i915_oa_kblgt2.c
> @@ -63,28 +63,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf)
> }
>
> void
> -i915_perf_load_test_config_kblgt2(struct drm_i915_private *dev_priv)
> +i915_perf_load_test_config_kblgt2(struct drm_i915_private *i915)
> {
> - strlcpy(dev_priv->perf.oa.test_config.uuid,
> + strlcpy(i915->perf.oa.test_config.uuid,
> "baa3c7e4-52b6-4b85-801e-465a94b746dd",
> - sizeof(dev_priv->perf.oa.test_config.uuid));
> - dev_priv->perf.oa.test_config.id = 1;
> + sizeof(i915->perf.oa.test_config.uuid));
> + i915->perf.oa.test_config.id = 1;
>
> - dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa;
> - dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> + i915->perf.oa.test_config.mux_regs = mux_config_test_oa;
> + i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
>
> - dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> - dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> + i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> + i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
>
> - dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> - dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> + i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> + i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
>
> - dev_priv->perf.oa.test_config.sysfs_metric.name = "baa3c7e4-52b6-4b85-801e-465a94b746dd";
> - dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs;
> + i915->perf.oa.test_config.sysfs_metric.name = "baa3c7e4-52b6-4b85-801e-465a94b746dd";
> + i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs;
>
> - dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr;
> + i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr;
>
> - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> - dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> + i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> + i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> + i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> }
> diff --git a/drivers/gpu/drm/i915/i915_oa_kblgt2.h b/drivers/gpu/drm/i915/i915_oa_kblgt2.h
> index a55398a904de..53655de7b976 100644
> --- a/drivers/gpu/drm/i915/i915_oa_kblgt2.h
> +++ b/drivers/gpu/drm/i915/i915_oa_kblgt2.h
> @@ -10,6 +10,6 @@
> #ifndef __I915_OA_KBLGT2_H__
> #define __I915_OA_KBLGT2_H__
>
> -extern void i915_perf_load_test_config_kblgt2(struct drm_i915_private *dev_priv);
> +extern void i915_perf_load_test_config_kblgt2(struct drm_i915_private *i915);
>
> #endif
> diff --git a/drivers/gpu/drm/i915/i915_oa_kblgt3.c b/drivers/gpu/drm/i915/i915_oa_kblgt3.c
> index 6ffa553c388e..ed39f8dd9af2 100644
> --- a/drivers/gpu/drm/i915/i915_oa_kblgt3.c
> +++ b/drivers/gpu/drm/i915/i915_oa_kblgt3.c
> @@ -63,28 +63,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf)
> }
>
> void
> -i915_perf_load_test_config_kblgt3(struct drm_i915_private *dev_priv)
> +i915_perf_load_test_config_kblgt3(struct drm_i915_private *i915)
> {
> - strlcpy(dev_priv->perf.oa.test_config.uuid,
> + strlcpy(i915->perf.oa.test_config.uuid,
> "f1792f32-6db2-4b50-b4b2-557128f1688d",
> - sizeof(dev_priv->perf.oa.test_config.uuid));
> - dev_priv->perf.oa.test_config.id = 1;
> + sizeof(i915->perf.oa.test_config.uuid));
> + i915->perf.oa.test_config.id = 1;
>
> - dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa;
> - dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> + i915->perf.oa.test_config.mux_regs = mux_config_test_oa;
> + i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
>
> - dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> - dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> + i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> + i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
>
> - dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> - dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> + i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> + i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
>
> - dev_priv->perf.oa.test_config.sysfs_metric.name = "f1792f32-6db2-4b50-b4b2-557128f1688d";
> - dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs;
> + i915->perf.oa.test_config.sysfs_metric.name = "f1792f32-6db2-4b50-b4b2-557128f1688d";
> + i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs;
>
> - dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr;
> + i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr;
>
> - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> - dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> + i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> + i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> + i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> }
> diff --git a/drivers/gpu/drm/i915/i915_oa_kblgt3.h b/drivers/gpu/drm/i915/i915_oa_kblgt3.h
> index 3ddd3483b7cc..60ebe2b20ada 100644
> --- a/drivers/gpu/drm/i915/i915_oa_kblgt3.h
> +++ b/drivers/gpu/drm/i915/i915_oa_kblgt3.h
> @@ -10,6 +10,6 @@
> #ifndef __I915_OA_KBLGT3_H__
> #define __I915_OA_KBLGT3_H__
>
> -extern void i915_perf_load_test_config_kblgt3(struct drm_i915_private *dev_priv);
> +extern void i915_perf_load_test_config_kblgt3(struct drm_i915_private *i915);
>
> #endif
> diff --git a/drivers/gpu/drm/i915/i915_oa_sklgt2.c b/drivers/gpu/drm/i915/i915_oa_sklgt2.c
> index 7ce6ee851d43..60b4dbcd4c59 100644
> --- a/drivers/gpu/drm/i915/i915_oa_sklgt2.c
> +++ b/drivers/gpu/drm/i915/i915_oa_sklgt2.c
> @@ -62,28 +62,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf)
> }
>
> void
> -i915_perf_load_test_config_sklgt2(struct drm_i915_private *dev_priv)
> +i915_perf_load_test_config_sklgt2(struct drm_i915_private *i915)
> {
> - strlcpy(dev_priv->perf.oa.test_config.uuid,
> + strlcpy(i915->perf.oa.test_config.uuid,
> "1651949f-0ac0-4cb1-a06f-dafd74a407d1",
> - sizeof(dev_priv->perf.oa.test_config.uuid));
> - dev_priv->perf.oa.test_config.id = 1;
> + sizeof(i915->perf.oa.test_config.uuid));
> + i915->perf.oa.test_config.id = 1;
>
> - dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa;
> - dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> + i915->perf.oa.test_config.mux_regs = mux_config_test_oa;
> + i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
>
> - dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> - dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> + i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> + i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
>
> - dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> - dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> + i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> + i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
>
> - dev_priv->perf.oa.test_config.sysfs_metric.name = "1651949f-0ac0-4cb1-a06f-dafd74a407d1";
> - dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs;
> + i915->perf.oa.test_config.sysfs_metric.name = "1651949f-0ac0-4cb1-a06f-dafd74a407d1";
> + i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs;
>
> - dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr;
> + i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr;
>
> - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> - dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> + i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> + i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> + i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> }
> diff --git a/drivers/gpu/drm/i915/i915_oa_sklgt2.h b/drivers/gpu/drm/i915/i915_oa_sklgt2.h
> index be6256037239..ee6f747502ea 100644
> --- a/drivers/gpu/drm/i915/i915_oa_sklgt2.h
> +++ b/drivers/gpu/drm/i915/i915_oa_sklgt2.h
> @@ -10,6 +10,6 @@
> #ifndef __I915_OA_SKLGT2_H__
> #define __I915_OA_SKLGT2_H__
>
> -extern void i915_perf_load_test_config_sklgt2(struct drm_i915_private *dev_priv);
> +extern void i915_perf_load_test_config_sklgt2(struct drm_i915_private *i915);
>
> #endif
> diff --git a/drivers/gpu/drm/i915/i915_oa_sklgt3.c b/drivers/gpu/drm/i915/i915_oa_sklgt3.c
> index 086ca2631e1c..8c23cf712a7a 100644
> --- a/drivers/gpu/drm/i915/i915_oa_sklgt3.c
> +++ b/drivers/gpu/drm/i915/i915_oa_sklgt3.c
> @@ -63,28 +63,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf)
> }
>
> void
> -i915_perf_load_test_config_sklgt3(struct drm_i915_private *dev_priv)
> +i915_perf_load_test_config_sklgt3(struct drm_i915_private *i915)
> {
> - strlcpy(dev_priv->perf.oa.test_config.uuid,
> + strlcpy(i915->perf.oa.test_config.uuid,
> "2b985803-d3c9-4629-8a4f-634bfecba0e8",
> - sizeof(dev_priv->perf.oa.test_config.uuid));
> - dev_priv->perf.oa.test_config.id = 1;
> + sizeof(i915->perf.oa.test_config.uuid));
> + i915->perf.oa.test_config.id = 1;
>
> - dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa;
> - dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> + i915->perf.oa.test_config.mux_regs = mux_config_test_oa;
> + i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
>
> - dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> - dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> + i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> + i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
>
> - dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> - dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> + i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> + i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
>
> - dev_priv->perf.oa.test_config.sysfs_metric.name = "2b985803-d3c9-4629-8a4f-634bfecba0e8";
> - dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs;
> + i915->perf.oa.test_config.sysfs_metric.name = "2b985803-d3c9-4629-8a4f-634bfecba0e8";
> + i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs;
>
> - dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr;
> + i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr;
>
> - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> - dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> + i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> + i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> + i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> }
> diff --git a/drivers/gpu/drm/i915/i915_oa_sklgt3.h b/drivers/gpu/drm/i915/i915_oa_sklgt3.h
> index 650beb068e56..35ccb4125d71 100644
> --- a/drivers/gpu/drm/i915/i915_oa_sklgt3.h
> +++ b/drivers/gpu/drm/i915/i915_oa_sklgt3.h
> @@ -10,6 +10,6 @@
> #ifndef __I915_OA_SKLGT3_H__
> #define __I915_OA_SKLGT3_H__
>
> -extern void i915_perf_load_test_config_sklgt3(struct drm_i915_private *dev_priv);
> +extern void i915_perf_load_test_config_sklgt3(struct drm_i915_private *i915);
>
> #endif
> diff --git a/drivers/gpu/drm/i915/i915_oa_sklgt4.c b/drivers/gpu/drm/i915/i915_oa_sklgt4.c
> index b291a6eb8a87..b846da6198f8 100644
> --- a/drivers/gpu/drm/i915/i915_oa_sklgt4.c
> +++ b/drivers/gpu/drm/i915/i915_oa_sklgt4.c
> @@ -63,28 +63,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf)
> }
>
> void
> -i915_perf_load_test_config_sklgt4(struct drm_i915_private *dev_priv)
> +i915_perf_load_test_config_sklgt4(struct drm_i915_private *i915)
> {
> - strlcpy(dev_priv->perf.oa.test_config.uuid,
> + strlcpy(i915->perf.oa.test_config.uuid,
> "882fa433-1f4a-4a67-a962-c741888fe5f5",
> - sizeof(dev_priv->perf.oa.test_config.uuid));
> - dev_priv->perf.oa.test_config.id = 1;
> + sizeof(i915->perf.oa.test_config.uuid));
> + i915->perf.oa.test_config.id = 1;
>
> - dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa;
> - dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> + i915->perf.oa.test_config.mux_regs = mux_config_test_oa;
> + i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
>
> - dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> - dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> + i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> + i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
>
> - dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> - dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> + i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> + i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
>
> - dev_priv->perf.oa.test_config.sysfs_metric.name = "882fa433-1f4a-4a67-a962-c741888fe5f5";
> - dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs;
> + i915->perf.oa.test_config.sysfs_metric.name = "882fa433-1f4a-4a67-a962-c741888fe5f5";
> + i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs;
>
> - dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr;
> + i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr;
>
> - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> - dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> + i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> + i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> + i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> }
> diff --git a/drivers/gpu/drm/i915/i915_oa_sklgt4.h b/drivers/gpu/drm/i915/i915_oa_sklgt4.h
> index 8dcf849d131e..f3f6d7bf28e0 100644
> --- a/drivers/gpu/drm/i915/i915_oa_sklgt4.h
> +++ b/drivers/gpu/drm/i915/i915_oa_sklgt4.h
> @@ -10,6 +10,6 @@
> #ifndef __I915_OA_SKLGT4_H__
> #define __I915_OA_SKLGT4_H__
>
> -extern void i915_perf_load_test_config_sklgt4(struct drm_i915_private *dev_priv);
> +extern void i915_perf_load_test_config_sklgt4(struct drm_i915_private *i915);
>
> #endif
> diff --git a/drivers/gpu/drm/i915/i915_perf.c b/drivers/gpu/drm/i915/i915_perf.c
> index d92ddfada262..e0ef77ec4da3 100644
> --- a/drivers/gpu/drm/i915/i915_perf.c
> +++ b/drivers/gpu/drm/i915/i915_perf.c
> @@ -366,7 +366,7 @@ struct perf_open_properties {
> int oa_period_exponent;
> };
>
> -static void free_oa_config(struct drm_i915_private *dev_priv,
> +static void free_oa_config(struct drm_i915_private *i915,
> struct i915_oa_config *oa_config)
> {
> if (!PTR_ERR(oa_config->flex_regs))
> @@ -378,48 +378,48 @@ static void free_oa_config(struct drm_i915_private *dev_priv,
> kfree(oa_config);
> }
>
> -static void put_oa_config(struct drm_i915_private *dev_priv,
> +static void put_oa_config(struct drm_i915_private *i915,
> struct i915_oa_config *oa_config)
> {
> if (!atomic_dec_and_test(&oa_config->ref_count))
> return;
>
> - free_oa_config(dev_priv, oa_config);
> + free_oa_config(i915, oa_config);
> }
>
> -static int get_oa_config(struct drm_i915_private *dev_priv,
> +static int get_oa_config(struct drm_i915_private *i915,
> int metrics_set,
> struct i915_oa_config **out_config)
> {
> int ret;
>
> if (metrics_set == 1) {
> - *out_config = &dev_priv->perf.oa.test_config;
> - atomic_inc(&dev_priv->perf.oa.test_config.ref_count);
> + *out_config = &i915->perf.oa.test_config;
> + atomic_inc(&i915->perf.oa.test_config.ref_count);
> return 0;
> }
>
> - ret = mutex_lock_interruptible(&dev_priv->perf.metrics_lock);
> + ret = mutex_lock_interruptible(&i915->perf.metrics_lock);
> if (ret)
> return ret;
>
> - *out_config = idr_find(&dev_priv->perf.metrics_idr, metrics_set);
> + *out_config = idr_find(&i915->perf.metrics_idr, metrics_set);
> if (!*out_config)
> ret = -EINVAL;
> else
> atomic_inc(&(*out_config)->ref_count);
>
> - mutex_unlock(&dev_priv->perf.metrics_lock);
> + mutex_unlock(&i915->perf.metrics_lock);
>
> return ret;
> }
>
> -static u32 gen8_oa_hw_tail_read(struct drm_i915_private *dev_priv)
> +static u32 gen8_oa_hw_tail_read(struct drm_i915_private *i915)
> {
> return I915_READ(GEN8_OATAILPTR) & GEN8_OATAILPTR_MASK;
> }
>
> -static u32 gen7_oa_hw_tail_read(struct drm_i915_private *dev_priv)
> +static u32 gen7_oa_hw_tail_read(struct drm_i915_private *i915)
> {
> u32 oastatus1 = I915_READ(GEN7_OASTATUS1);
>
> @@ -428,7 +428,7 @@ static u32 gen7_oa_hw_tail_read(struct drm_i915_private *dev_priv)
>
> /**
> * oa_buffer_check_unlocked - check for data and update tail ptr state
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> *
> * This is either called via fops (for blocking reads in user ctx) or the poll
> * check hrtimer (atomic ctx) to check the OA buffer tail pointer and check
> @@ -450,9 +450,9 @@ static u32 gen7_oa_hw_tail_read(struct drm_i915_private *dev_priv)
> *
> * Returns: %true if the OA buffer contains data, else %false
> */
> -static bool oa_buffer_check_unlocked(struct drm_i915_private *dev_priv)
> +static bool oa_buffer_check_unlocked(struct drm_i915_private *i915)
> {
> - int report_size = dev_priv->perf.oa.oa_buffer.format_size;
> + int report_size = i915->perf.oa.oa_buffer.format_size;
> unsigned long flags;
> unsigned int aged_idx;
> u32 head, hw_tail, aged_tail, aging_tail;
> @@ -462,19 +462,19 @@ static bool oa_buffer_check_unlocked(struct drm_i915_private *dev_priv)
> * could result in an OA buffer reset which might reset the head,
> * tails[] and aged_tail state.
> */
> - spin_lock_irqsave(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags);
> + spin_lock_irqsave(&i915->perf.oa.oa_buffer.ptr_lock, flags);
>
> /* NB: The head we observe here might effectively be a little out of
> * date (between head and tails[aged_idx].offset if there is currently
> * a read() in progress.
> */
> - head = dev_priv->perf.oa.oa_buffer.head;
> + head = i915->perf.oa.oa_buffer.head;
>
> - aged_idx = dev_priv->perf.oa.oa_buffer.aged_tail_idx;
> - aged_tail = dev_priv->perf.oa.oa_buffer.tails[aged_idx].offset;
> - aging_tail = dev_priv->perf.oa.oa_buffer.tails[!aged_idx].offset;
> + aged_idx = i915->perf.oa.oa_buffer.aged_tail_idx;
> + aged_tail = i915->perf.oa.oa_buffer.tails[aged_idx].offset;
> + aging_tail = i915->perf.oa.oa_buffer.tails[!aged_idx].offset;
>
> - hw_tail = dev_priv->perf.oa.ops.oa_hw_tail_read(dev_priv);
> + hw_tail = i915->perf.oa.ops.oa_hw_tail_read(i915);
>
> /* The tail pointer increases in 64 byte increments,
> * not in report_size steps...
> @@ -494,16 +494,16 @@ static bool oa_buffer_check_unlocked(struct drm_i915_private *dev_priv)
> * available) without needing to wait for a later hrtimer callback.
> */
> if (aging_tail != INVALID_TAIL_PTR &&
> - ((now - dev_priv->perf.oa.oa_buffer.aging_timestamp) >
> + ((now - i915->perf.oa.oa_buffer.aging_timestamp) >
> OA_TAIL_MARGIN_NSEC)) {
>
> aged_idx ^= 1;
> - dev_priv->perf.oa.oa_buffer.aged_tail_idx = aged_idx;
> + i915->perf.oa.oa_buffer.aged_tail_idx = aged_idx;
>
> aged_tail = aging_tail;
>
> /* Mark that we need a new pointer to start aging... */
> - dev_priv->perf.oa.oa_buffer.tails[!aged_idx].offset = INVALID_TAIL_PTR;
> + i915->perf.oa.oa_buffer.tails[!aged_idx].offset = INVALID_TAIL_PTR;
> aging_tail = INVALID_TAIL_PTR;
> }
>
> @@ -518,7 +518,7 @@ static bool oa_buffer_check_unlocked(struct drm_i915_private *dev_priv)
> if (aging_tail == INVALID_TAIL_PTR &&
> (aged_tail == INVALID_TAIL_PTR ||
> OA_TAKEN(hw_tail, aged_tail) >= report_size)) {
> - struct i915_vma *vma = dev_priv->perf.oa.oa_buffer.vma;
> + struct i915_vma *vma = i915->perf.oa.oa_buffer.vma;
> u32 gtt_offset = i915_ggtt_offset(vma);
>
> /* Be paranoid and do a bounds check on the pointer read back
> @@ -527,16 +527,16 @@ static bool oa_buffer_check_unlocked(struct drm_i915_private *dev_priv)
> */
> if (hw_tail >= gtt_offset &&
> hw_tail < (gtt_offset + OA_BUFFER_SIZE)) {
> - dev_priv->perf.oa.oa_buffer.tails[!aged_idx].offset =
> + i915->perf.oa.oa_buffer.tails[!aged_idx].offset =
> aging_tail = hw_tail;
> - dev_priv->perf.oa.oa_buffer.aging_timestamp = now;
> + i915->perf.oa.oa_buffer.aging_timestamp = now;
> } else {
> DRM_ERROR("Ignoring spurious out of range OA buffer tail pointer = %u\n",
> hw_tail);
> }
> }
>
> - spin_unlock_irqrestore(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags);
> + spin_unlock_irqrestore(&i915->perf.oa.oa_buffer.ptr_lock, flags);
>
> return aged_tail == INVALID_TAIL_PTR ?
> false : OA_TAKEN(aged_tail, head) >= report_size;
> @@ -599,8 +599,8 @@ static int append_oa_sample(struct i915_perf_stream *stream,
> size_t *offset,
> const u8 *report)
> {
> - struct drm_i915_private *dev_priv = stream->dev_priv;
> - int report_size = dev_priv->perf.oa.oa_buffer.format_size;
> + struct drm_i915_private *i915 = stream->i915;
> + int report_size = i915->perf.oa.oa_buffer.format_size;
> struct drm_i915_perf_record_header header;
> u32 sample_flags = stream->sample_flags;
>
> @@ -651,10 +651,10 @@ static int gen8_append_oa_reports(struct i915_perf_stream *stream,
> size_t count,
> size_t *offset)
> {
> - struct drm_i915_private *dev_priv = stream->dev_priv;
> - int report_size = dev_priv->perf.oa.oa_buffer.format_size;
> - u8 *oa_buf_base = dev_priv->perf.oa.oa_buffer.vaddr;
> - u32 gtt_offset = i915_ggtt_offset(dev_priv->perf.oa.oa_buffer.vma);
> + struct drm_i915_private *i915 = stream->i915;
> + int report_size = i915->perf.oa.oa_buffer.format_size;
> + u8 *oa_buf_base = i915->perf.oa.oa_buffer.vaddr;
> + u32 gtt_offset = i915_ggtt_offset(i915->perf.oa.oa_buffer.vma);
> u32 mask = (OA_BUFFER_SIZE - 1);
> size_t start_offset = *offset;
> unsigned long flags;
> @@ -666,13 +666,13 @@ static int gen8_append_oa_reports(struct i915_perf_stream *stream,
> if (WARN_ON(!stream->enabled))
> return -EIO;
>
> - spin_lock_irqsave(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags);
> + spin_lock_irqsave(&i915->perf.oa.oa_buffer.ptr_lock, flags);
>
> - head = dev_priv->perf.oa.oa_buffer.head;
> - aged_tail_idx = dev_priv->perf.oa.oa_buffer.aged_tail_idx;
> - tail = dev_priv->perf.oa.oa_buffer.tails[aged_tail_idx].offset;
> + head = i915->perf.oa.oa_buffer.head;
> + aged_tail_idx = i915->perf.oa.oa_buffer.aged_tail_idx;
> + tail = i915->perf.oa.oa_buffer.tails[aged_tail_idx].offset;
>
> - spin_unlock_irqrestore(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags);
> + spin_unlock_irqrestore(&i915->perf.oa.oa_buffer.ptr_lock, flags);
>
> /*
> * An invalid tail pointer here means we're still waiting for the poll
> @@ -736,12 +736,12 @@ static int gen8_append_oa_reports(struct i915_perf_stream *stream,
> reason = ((report32[0] >> OAREPORT_REASON_SHIFT) &
> OAREPORT_REASON_MASK);
> if (reason == 0) {
> - if (__ratelimit(&dev_priv->perf.oa.spurious_report_rs))
> + if (__ratelimit(&i915->perf.oa.spurious_report_rs))
> DRM_NOTE("Skipping spurious, invalid OA report\n");
> continue;
> }
>
> - ctx_id = report32[2] & dev_priv->perf.oa.specific_ctx_id_mask;
> + ctx_id = report32[2] & i915->perf.oa.specific_ctx_id_mask;
>
> /*
> * Squash whatever is in the CTX_ID field if it's marked as
> @@ -751,7 +751,7 @@ static int gen8_append_oa_reports(struct i915_perf_stream *stream,
> * Note: that we don't clear the valid_ctx_bit so userspace can
> * understand that the ID has been squashed by the kernel.
> */
> - if (!(report32[0] & dev_priv->perf.oa.gen8_valid_ctx_bit))
> + if (!(report32[0] & i915->perf.oa.gen8_valid_ctx_bit))
> ctx_id = report32[2] = INVALID_CTX_ID;
>
> /*
> @@ -785,18 +785,18 @@ static int gen8_append_oa_reports(struct i915_perf_stream *stream,
> * switches since it's not-uncommon for periodic samples to
> * identify a switch before any 'context switch' report.
> */
> - if (!dev_priv->perf.oa.exclusive_stream->ctx ||
> - dev_priv->perf.oa.specific_ctx_id == ctx_id ||
> - (dev_priv->perf.oa.oa_buffer.last_ctx_id ==
> - dev_priv->perf.oa.specific_ctx_id) ||
> + if (!i915->perf.oa.exclusive_stream->ctx ||
> + i915->perf.oa.specific_ctx_id == ctx_id ||
> + (i915->perf.oa.oa_buffer.last_ctx_id ==
> + i915->perf.oa.specific_ctx_id) ||
> reason & OAREPORT_REASON_CTX_SWITCH) {
>
> /*
> * While filtering for a single context we avoid
> * leaking the IDs of other contexts.
> */
> - if (dev_priv->perf.oa.exclusive_stream->ctx &&
> - dev_priv->perf.oa.specific_ctx_id != ctx_id) {
> + if (i915->perf.oa.exclusive_stream->ctx &&
> + i915->perf.oa.specific_ctx_id != ctx_id) {
> report32[2] = INVALID_CTX_ID;
> }
>
> @@ -805,7 +805,7 @@ static int gen8_append_oa_reports(struct i915_perf_stream *stream,
> if (ret)
> break;
>
> - dev_priv->perf.oa.oa_buffer.last_ctx_id = ctx_id;
> + i915->perf.oa.oa_buffer.last_ctx_id = ctx_id;
> }
>
> /*
> @@ -819,7 +819,7 @@ static int gen8_append_oa_reports(struct i915_perf_stream *stream,
> }
>
> if (start_offset != *offset) {
> - spin_lock_irqsave(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags);
> + spin_lock_irqsave(&i915->perf.oa.oa_buffer.ptr_lock, flags);
>
> /*
> * We removed the gtt_offset for the copy loop above, indexing
> @@ -828,9 +828,9 @@ static int gen8_append_oa_reports(struct i915_perf_stream *stream,
> head += gtt_offset;
>
> I915_WRITE(GEN8_OAHEADPTR, head & GEN8_OAHEADPTR_MASK);
> - dev_priv->perf.oa.oa_buffer.head = head;
> + i915->perf.oa.oa_buffer.head = head;
>
> - spin_unlock_irqrestore(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags);
> + spin_unlock_irqrestore(&i915->perf.oa.oa_buffer.ptr_lock, flags);
> }
>
> return ret;
> @@ -861,11 +861,11 @@ static int gen8_oa_read(struct i915_perf_stream *stream,
> size_t count,
> size_t *offset)
> {
> - struct drm_i915_private *dev_priv = stream->dev_priv;
> + struct drm_i915_private *i915 = stream->i915;
> u32 oastatus;
> int ret;
>
> - if (WARN_ON(!dev_priv->perf.oa.oa_buffer.vaddr))
> + if (WARN_ON(!i915->perf.oa.oa_buffer.vaddr))
> return -EIO;
>
> oastatus = I915_READ(GEN8_OASTATUS);
> @@ -891,10 +891,10 @@ static int gen8_oa_read(struct i915_perf_stream *stream,
> return ret;
>
> DRM_DEBUG("OA buffer overflow (exponent = %d): force restart\n",
> - dev_priv->perf.oa.period_exponent);
> + i915->perf.oa.period_exponent);
>
> - dev_priv->perf.oa.ops.oa_disable(stream);
> - dev_priv->perf.oa.ops.oa_enable(stream);
> + i915->perf.oa.ops.oa_disable(stream);
> + i915->perf.oa.ops.oa_enable(stream);
>
> /*
> * Note: .oa_enable() is expected to re-init the oabuffer and
> @@ -940,10 +940,10 @@ static int gen7_append_oa_reports(struct i915_perf_stream *stream,
> size_t count,
> size_t *offset)
> {
> - struct drm_i915_private *dev_priv = stream->dev_priv;
> - int report_size = dev_priv->perf.oa.oa_buffer.format_size;
> - u8 *oa_buf_base = dev_priv->perf.oa.oa_buffer.vaddr;
> - u32 gtt_offset = i915_ggtt_offset(dev_priv->perf.oa.oa_buffer.vma);
> + struct drm_i915_private *i915 = stream->i915;
> + int report_size = i915->perf.oa.oa_buffer.format_size;
> + u8 *oa_buf_base = i915->perf.oa.oa_buffer.vaddr;
> + u32 gtt_offset = i915_ggtt_offset(i915->perf.oa.oa_buffer.vma);
> u32 mask = (OA_BUFFER_SIZE - 1);
> size_t start_offset = *offset;
> unsigned long flags;
> @@ -955,13 +955,13 @@ static int gen7_append_oa_reports(struct i915_perf_stream *stream,
> if (WARN_ON(!stream->enabled))
> return -EIO;
>
> - spin_lock_irqsave(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags);
> + spin_lock_irqsave(&i915->perf.oa.oa_buffer.ptr_lock, flags);
>
> - head = dev_priv->perf.oa.oa_buffer.head;
> - aged_tail_idx = dev_priv->perf.oa.oa_buffer.aged_tail_idx;
> - tail = dev_priv->perf.oa.oa_buffer.tails[aged_tail_idx].offset;
> + head = i915->perf.oa.oa_buffer.head;
> + aged_tail_idx = i915->perf.oa.oa_buffer.aged_tail_idx;
> + tail = i915->perf.oa.oa_buffer.tails[aged_tail_idx].offset;
>
> - spin_unlock_irqrestore(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags);
> + spin_unlock_irqrestore(&i915->perf.oa.oa_buffer.ptr_lock, flags);
>
> /* An invalid tail pointer here means we're still waiting for the poll
> * hrtimer callback to give us a pointer
> @@ -1014,7 +1014,7 @@ static int gen7_append_oa_reports(struct i915_perf_stream *stream,
> * copying it to userspace...
> */
> if (report32[0] == 0) {
> - if (__ratelimit(&dev_priv->perf.oa.spurious_report_rs))
> + if (__ratelimit(&i915->perf.oa.spurious_report_rs))
> DRM_NOTE("Skipping spurious, invalid OA report\n");
> continue;
> }
> @@ -1033,7 +1033,7 @@ static int gen7_append_oa_reports(struct i915_perf_stream *stream,
> }
>
> if (start_offset != *offset) {
> - spin_lock_irqsave(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags);
> + spin_lock_irqsave(&i915->perf.oa.oa_buffer.ptr_lock, flags);
>
> /* We removed the gtt_offset for the copy loop above, indexing
> * relative to oa_buf_base so put back here...
> @@ -1043,9 +1043,9 @@ static int gen7_append_oa_reports(struct i915_perf_stream *stream,
> I915_WRITE(GEN7_OASTATUS2,
> ((head & GEN7_OASTATUS2_HEAD_MASK) |
> GEN7_OASTATUS2_MEM_SELECT_GGTT));
> - dev_priv->perf.oa.oa_buffer.head = head;
> + i915->perf.oa.oa_buffer.head = head;
>
> - spin_unlock_irqrestore(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags);
> + spin_unlock_irqrestore(&i915->perf.oa.oa_buffer.ptr_lock, flags);
> }
>
> return ret;
> @@ -1072,11 +1072,11 @@ static int gen7_oa_read(struct i915_perf_stream *stream,
> size_t count,
> size_t *offset)
> {
> - struct drm_i915_private *dev_priv = stream->dev_priv;
> + struct drm_i915_private *i915 = stream->i915;
> u32 oastatus1;
> int ret;
>
> - if (WARN_ON(!dev_priv->perf.oa.oa_buffer.vaddr))
> + if (WARN_ON(!i915->perf.oa.oa_buffer.vaddr))
> return -EIO;
>
> oastatus1 = I915_READ(GEN7_OASTATUS1);
> @@ -1086,7 +1086,7 @@ static int gen7_oa_read(struct i915_perf_stream *stream,
> * may be updated asynchronously) so we ignore status bits
> * that have already been reported to userspace.
> */
> - oastatus1 &= ~dev_priv->perf.oa.gen7_latched_oastatus1;
> + oastatus1 &= ~i915->perf.oa.gen7_latched_oastatus1;
>
> /* We treat OABUFFER_OVERFLOW as a significant error:
> *
> @@ -1115,10 +1115,10 @@ static int gen7_oa_read(struct i915_perf_stream *stream,
> return ret;
>
> DRM_DEBUG("OA buffer overflow (exponent = %d): force restart\n",
> - dev_priv->perf.oa.period_exponent);
> + i915->perf.oa.period_exponent);
>
> - dev_priv->perf.oa.ops.oa_disable(stream);
> - dev_priv->perf.oa.ops.oa_enable(stream);
> + i915->perf.oa.ops.oa_disable(stream);
> + i915->perf.oa.ops.oa_enable(stream);
>
> oastatus1 = I915_READ(GEN7_OASTATUS1);
> }
> @@ -1128,7 +1128,7 @@ static int gen7_oa_read(struct i915_perf_stream *stream,
> DRM_I915_PERF_RECORD_OA_REPORT_LOST);
> if (ret)
> return ret;
> - dev_priv->perf.oa.gen7_latched_oastatus1 |=
> + i915->perf.oa.gen7_latched_oastatus1 |=
> GEN7_OASTATUS1_REPORT_LOST;
> }
>
> @@ -1151,14 +1151,14 @@ static int gen7_oa_read(struct i915_perf_stream *stream,
> */
> static int i915_oa_wait_unlocked(struct i915_perf_stream *stream)
> {
> - struct drm_i915_private *dev_priv = stream->dev_priv;
> + struct drm_i915_private *i915 = stream->i915;
>
> /* We would wait indefinitely if periodic sampling is not enabled */
> - if (!dev_priv->perf.oa.periodic)
> + if (!i915->perf.oa.periodic)
> return -EIO;
>
> - return wait_event_interruptible(dev_priv->perf.oa.poll_wq,
> - oa_buffer_check_unlocked(dev_priv));
> + return wait_event_interruptible(i915->perf.oa.poll_wq,
> + oa_buffer_check_unlocked(i915));
> }
>
> /**
> @@ -1175,9 +1175,9 @@ static void i915_oa_poll_wait(struct i915_perf_stream *stream,
> struct file *file,
> poll_table *wait)
> {
> - struct drm_i915_private *dev_priv = stream->dev_priv;
> + struct drm_i915_private *i915 = stream->i915;
>
> - poll_wait(file, &dev_priv->perf.oa.poll_wq, wait);
> + poll_wait(file, &i915->perf.oa.poll_wq, wait);
> }
>
> /**
> @@ -1197,9 +1197,9 @@ static int i915_oa_read(struct i915_perf_stream *stream,
> size_t count,
> size_t *offset)
> {
> - struct drm_i915_private *dev_priv = stream->dev_priv;
> + struct drm_i915_private *i915 = stream->i915;
>
> - return dev_priv->perf.oa.ops.read(stream, buf, count, offset);
> + return i915->perf.oa.ops.read(stream, buf, count, offset);
> }
>
> static struct intel_context *oa_pin_context(struct drm_i915_private *i915,
> @@ -1248,7 +1248,7 @@ static struct intel_context *oa_pin_context(struct drm_i915_private *i915,
> */
> static int oa_get_render_ctx_id(struct i915_perf_stream *stream)
> {
> - struct drm_i915_private *i915 = stream->dev_priv;
> + struct drm_i915_private *i915 = stream->i915;
> struct intel_context *ce;
>
> ce = oa_pin_context(i915, stream->ctx);
> @@ -1330,17 +1330,17 @@ static int oa_get_render_ctx_id(struct i915_perf_stream *stream)
> */
> static void oa_put_render_ctx_id(struct i915_perf_stream *stream)
> {
> - struct drm_i915_private *dev_priv = stream->dev_priv;
> + struct drm_i915_private *i915 = stream->i915;
> struct intel_context *ce;
>
> - dev_priv->perf.oa.specific_ctx_id = INVALID_CTX_ID;
> - dev_priv->perf.oa.specific_ctx_id_mask = 0;
> + i915->perf.oa.specific_ctx_id = INVALID_CTX_ID;
> + i915->perf.oa.specific_ctx_id_mask = 0;
>
> - ce = fetch_and_zero(&dev_priv->perf.oa.pinned_ctx);
> + ce = fetch_and_zero(&i915->perf.oa.pinned_ctx);
> if (ce) {
> - mutex_lock(&dev_priv->drm.struct_mutex);
> + mutex_lock(&i915->drm.struct_mutex);
> intel_context_unpin(ce);
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + mutex_unlock(&i915->drm.struct_mutex);
> }
> }
>
> @@ -1359,64 +1359,64 @@ free_oa_buffer(struct drm_i915_private *i915)
>
> static void i915_oa_stream_destroy(struct i915_perf_stream *stream)
> {
> - struct drm_i915_private *dev_priv = stream->dev_priv;
> + struct drm_i915_private *i915 = stream->i915;
>
> - BUG_ON(stream != dev_priv->perf.oa.exclusive_stream);
> + BUG_ON(stream != i915->perf.oa.exclusive_stream);
>
> /*
> * Unset exclusive_stream first, it will be checked while disabling
> * the metric set on gen8+.
> */
> - mutex_lock(&dev_priv->drm.struct_mutex);
> - dev_priv->perf.oa.exclusive_stream = NULL;
> - dev_priv->perf.oa.ops.disable_metric_set(dev_priv);
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + mutex_lock(&i915->drm.struct_mutex);
> + i915->perf.oa.exclusive_stream = NULL;
> + i915->perf.oa.ops.disable_metric_set(i915);
> + mutex_unlock(&i915->drm.struct_mutex);
>
> - free_oa_buffer(dev_priv);
> + free_oa_buffer(i915);
>
> - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> - intel_runtime_pm_put(dev_priv, stream->wakeref);
> + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> + intel_runtime_pm_put(i915, stream->wakeref);
>
> if (stream->ctx)
> oa_put_render_ctx_id(stream);
>
> - put_oa_config(dev_priv, stream->oa_config);
> + put_oa_config(i915, stream->oa_config);
>
> - if (dev_priv->perf.oa.spurious_report_rs.missed) {
> + if (i915->perf.oa.spurious_report_rs.missed) {
> DRM_NOTE("%d spurious OA report notices suppressed due to ratelimiting\n",
> - dev_priv->perf.oa.spurious_report_rs.missed);
> + i915->perf.oa.spurious_report_rs.missed);
> }
> }
>
> -static void gen7_init_oa_buffer(struct drm_i915_private *dev_priv)
> +static void gen7_init_oa_buffer(struct drm_i915_private *i915)
> {
> - u32 gtt_offset = i915_ggtt_offset(dev_priv->perf.oa.oa_buffer.vma);
> + u32 gtt_offset = i915_ggtt_offset(i915->perf.oa.oa_buffer.vma);
> unsigned long flags;
>
> - spin_lock_irqsave(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags);
> + spin_lock_irqsave(&i915->perf.oa.oa_buffer.ptr_lock, flags);
>
> /* Pre-DevBDW: OABUFFER must be set with counters off,
> * before OASTATUS1, but after OASTATUS2
> */
> I915_WRITE(GEN7_OASTATUS2,
> gtt_offset | GEN7_OASTATUS2_MEM_SELECT_GGTT); /* head */
> - dev_priv->perf.oa.oa_buffer.head = gtt_offset;
> + i915->perf.oa.oa_buffer.head = gtt_offset;
>
> I915_WRITE(GEN7_OABUFFER, gtt_offset);
>
> I915_WRITE(GEN7_OASTATUS1, gtt_offset | OABUFFER_SIZE_16M); /* tail */
>
> /* Mark that we need updated tail pointers to read from... */
> - dev_priv->perf.oa.oa_buffer.tails[0].offset = INVALID_TAIL_PTR;
> - dev_priv->perf.oa.oa_buffer.tails[1].offset = INVALID_TAIL_PTR;
> + i915->perf.oa.oa_buffer.tails[0].offset = INVALID_TAIL_PTR;
> + i915->perf.oa.oa_buffer.tails[1].offset = INVALID_TAIL_PTR;
>
> - spin_unlock_irqrestore(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags);
> + spin_unlock_irqrestore(&i915->perf.oa.oa_buffer.ptr_lock, flags);
>
> /* On Haswell we have to track which OASTATUS1 flags we've
> * already seen since they can't be cleared while periodic
> * sampling is enabled.
> */
> - dev_priv->perf.oa.gen7_latched_oastatus1 = 0;
> + i915->perf.oa.gen7_latched_oastatus1 = 0;
>
> /* NB: although the OA buffer will initially be allocated
> * zeroed via shmfs (and so this memset is redundant when
> @@ -1429,24 +1429,24 @@ static void gen7_init_oa_buffer(struct drm_i915_private *dev_priv)
> * the assumption that new reports are being written to zeroed
> * memory...
> */
> - memset(dev_priv->perf.oa.oa_buffer.vaddr, 0, OA_BUFFER_SIZE);
> + memset(i915->perf.oa.oa_buffer.vaddr, 0, OA_BUFFER_SIZE);
>
> /* Maybe make ->pollin per-stream state if we support multiple
> * concurrent streams in the future.
> */
> - dev_priv->perf.oa.pollin = false;
> + i915->perf.oa.pollin = false;
> }
>
> -static void gen8_init_oa_buffer(struct drm_i915_private *dev_priv)
> +static void gen8_init_oa_buffer(struct drm_i915_private *i915)
> {
> - u32 gtt_offset = i915_ggtt_offset(dev_priv->perf.oa.oa_buffer.vma);
> + u32 gtt_offset = i915_ggtt_offset(i915->perf.oa.oa_buffer.vma);
> unsigned long flags;
>
> - spin_lock_irqsave(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags);
> + spin_lock_irqsave(&i915->perf.oa.oa_buffer.ptr_lock, flags);
>
> I915_WRITE(GEN8_OASTATUS, 0);
> I915_WRITE(GEN8_OAHEADPTR, gtt_offset);
> - dev_priv->perf.oa.oa_buffer.head = gtt_offset;
> + i915->perf.oa.oa_buffer.head = gtt_offset;
>
> I915_WRITE(GEN8_OABUFFER_UDW, 0);
>
> @@ -1463,17 +1463,17 @@ static void gen8_init_oa_buffer(struct drm_i915_private *dev_priv)
> I915_WRITE(GEN8_OATAILPTR, gtt_offset & GEN8_OATAILPTR_MASK);
>
> /* Mark that we need updated tail pointers to read from... */
> - dev_priv->perf.oa.oa_buffer.tails[0].offset = INVALID_TAIL_PTR;
> - dev_priv->perf.oa.oa_buffer.tails[1].offset = INVALID_TAIL_PTR;
> + i915->perf.oa.oa_buffer.tails[0].offset = INVALID_TAIL_PTR;
> + i915->perf.oa.oa_buffer.tails[1].offset = INVALID_TAIL_PTR;
>
> /*
> * Reset state used to recognise context switches, affecting which
> * reports we will forward to userspace while filtering for a single
> * context.
> */
> - dev_priv->perf.oa.oa_buffer.last_ctx_id = INVALID_CTX_ID;
> + i915->perf.oa.oa_buffer.last_ctx_id = INVALID_CTX_ID;
>
> - spin_unlock_irqrestore(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags);
> + spin_unlock_irqrestore(&i915->perf.oa.oa_buffer.ptr_lock, flags);
>
> /*
> * NB: although the OA buffer will initially be allocated
> @@ -1487,32 +1487,32 @@ static void gen8_init_oa_buffer(struct drm_i915_private *dev_priv)
> * the assumption that new reports are being written to zeroed
> * memory...
> */
> - memset(dev_priv->perf.oa.oa_buffer.vaddr, 0, OA_BUFFER_SIZE);
> + memset(i915->perf.oa.oa_buffer.vaddr, 0, OA_BUFFER_SIZE);
>
> /*
> * Maybe make ->pollin per-stream state if we support multiple
> * concurrent streams in the future.
> */
> - dev_priv->perf.oa.pollin = false;
> + i915->perf.oa.pollin = false;
> }
>
> -static int alloc_oa_buffer(struct drm_i915_private *dev_priv)
> +static int alloc_oa_buffer(struct drm_i915_private *i915)
> {
> struct drm_i915_gem_object *bo;
> struct i915_vma *vma;
> int ret;
>
> - if (WARN_ON(dev_priv->perf.oa.oa_buffer.vma))
> + if (WARN_ON(i915->perf.oa.oa_buffer.vma))
> return -ENODEV;
>
> - ret = i915_mutex_lock_interruptible(&dev_priv->drm);
> + ret = i915_mutex_lock_interruptible(&i915->drm);
> if (ret)
> return ret;
>
> BUILD_BUG_ON_NOT_POWER_OF_2(OA_BUFFER_SIZE);
> BUILD_BUG_ON(OA_BUFFER_SIZE < SZ_128K || OA_BUFFER_SIZE > SZ_16M);
>
> - bo = i915_gem_object_create_shmem(dev_priv, OA_BUFFER_SIZE);
> + bo = i915_gem_object_create_shmem(i915, OA_BUFFER_SIZE);
> if (IS_ERR(bo)) {
> DRM_ERROR("Failed to allocate OA buffer\n");
> ret = PTR_ERR(bo);
> @@ -1527,18 +1527,18 @@ static int alloc_oa_buffer(struct drm_i915_private *dev_priv)
> ret = PTR_ERR(vma);
> goto err_unref;
> }
> - dev_priv->perf.oa.oa_buffer.vma = vma;
> + i915->perf.oa.oa_buffer.vma = vma;
>
> - dev_priv->perf.oa.oa_buffer.vaddr =
> + i915->perf.oa.oa_buffer.vaddr =
> i915_gem_object_pin_map(bo, I915_MAP_WB);
> - if (IS_ERR(dev_priv->perf.oa.oa_buffer.vaddr)) {
> - ret = PTR_ERR(dev_priv->perf.oa.oa_buffer.vaddr);
> + if (IS_ERR(i915->perf.oa.oa_buffer.vaddr)) {
> + ret = PTR_ERR(i915->perf.oa.oa_buffer.vaddr);
> goto err_unpin;
> }
>
> DRM_DEBUG_DRIVER("OA Buffer initialized, gtt offset = 0x%x, vaddr = %p\n",
> - i915_ggtt_offset(dev_priv->perf.oa.oa_buffer.vma),
> - dev_priv->perf.oa.oa_buffer.vaddr);
> + i915_ggtt_offset(i915->perf.oa.oa_buffer.vma),
> + i915->perf.oa.oa_buffer.vaddr);
>
> goto unlock;
>
> @@ -1548,15 +1548,15 @@ static int alloc_oa_buffer(struct drm_i915_private *dev_priv)
> err_unref:
> i915_gem_object_put(bo);
>
> - dev_priv->perf.oa.oa_buffer.vaddr = NULL;
> - dev_priv->perf.oa.oa_buffer.vma = NULL;
> + i915->perf.oa.oa_buffer.vaddr = NULL;
> + i915->perf.oa.oa_buffer.vma = NULL;
>
> unlock:
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + mutex_unlock(&i915->drm.struct_mutex);
> return ret;
> }
>
> -static void config_oa_regs(struct drm_i915_private *dev_priv,
> +static void config_oa_regs(struct drm_i915_private *i915,
> const struct i915_oa_reg *regs,
> u32 n_regs)
> {
> @@ -1571,7 +1571,7 @@ static void config_oa_regs(struct drm_i915_private *dev_priv,
>
> static int hsw_enable_metric_set(struct i915_perf_stream *stream)
> {
> - struct drm_i915_private *dev_priv = stream->dev_priv;
> + struct drm_i915_private *i915 = stream->i915;
> const struct i915_oa_config *oa_config = stream->oa_config;
>
> /* PRM:
> @@ -1588,7 +1588,7 @@ static int hsw_enable_metric_set(struct i915_perf_stream *stream)
> I915_WRITE(GEN6_UCGCTL1, (I915_READ(GEN6_UCGCTL1) |
> GEN6_CSUNIT_CLOCK_GATE_DISABLE));
>
> - config_oa_regs(dev_priv, oa_config->mux_regs, oa_config->mux_regs_len);
> + config_oa_regs(i915, oa_config->mux_regs, oa_config->mux_regs_len);
>
> /* It apparently takes a fairly long time for a new MUX
> * configuration to be be applied after these register writes.
> @@ -1613,13 +1613,13 @@ static int hsw_enable_metric_set(struct i915_perf_stream *stream)
> */
> usleep_range(15000, 20000);
>
> - config_oa_regs(dev_priv, oa_config->b_counter_regs,
> + config_oa_regs(i915, oa_config->b_counter_regs,
> oa_config->b_counter_regs_len);
>
> return 0;
> }
>
> -static void hsw_disable_metric_set(struct drm_i915_private *dev_priv)
> +static void hsw_disable_metric_set(struct drm_i915_private *i915)
> {
> I915_WRITE(GEN6_UCGCTL1, (I915_READ(GEN6_UCGCTL1) &
> ~GEN6_CSUNIT_CLOCK_GATE_DISABLE));
> @@ -1716,15 +1716,15 @@ gen8_update_reg_state_unlocked(struct intel_context *ce,
> *
> * Note: it's only the RCS/Render context that has any OA state.
> */
> -static int gen8_configure_all_contexts(struct drm_i915_private *dev_priv,
> +static int gen8_configure_all_contexts(struct drm_i915_private *i915,
> const struct i915_oa_config *oa_config)
> {
> - unsigned int map_type = i915_coherent_map_type(dev_priv);
> + unsigned int map_type = i915_coherent_map_type(i915);
> struct i915_gem_context *ctx;
> struct i915_request *rq;
> int ret;
>
> - lockdep_assert_held(&dev_priv->drm.struct_mutex);
> + lockdep_assert_held(&i915->drm.struct_mutex);
>
> /*
> * The OA register config is setup through the context image. This image
> @@ -1739,14 +1739,14 @@ static int gen8_configure_all_contexts(struct drm_i915_private *dev_priv,
> * So far the best way to work around this issue seems to be draining
> * the GPU from any submitted work.
> */
> - ret = i915_gem_wait_for_idle(dev_priv,
> + ret = i915_gem_wait_for_idle(i915,
> I915_WAIT_LOCKED,
> MAX_SCHEDULE_TIMEOUT);
> if (ret)
> return ret;
>
> /* Update all contexts now that we've stalled the submission. */
> - list_for_each_entry(ctx, &dev_priv->contexts.list, link) {
> + list_for_each_entry(ctx, &i915->contexts.list, link) {
> struct i915_gem_engines_iter it;
> struct intel_context *ce;
>
> @@ -1783,7 +1783,7 @@ static int gen8_configure_all_contexts(struct drm_i915_private *dev_priv,
> * Apply the configuration by doing one context restore of the edited
> * context image.
> */
> - rq = i915_request_create(dev_priv->engine[RCS0]->kernel_context);
> + rq = i915_request_create(i915->engine[RCS0]->kernel_context);
> if (IS_ERR(rq))
> return PTR_ERR(rq);
>
> @@ -1794,7 +1794,7 @@ static int gen8_configure_all_contexts(struct drm_i915_private *dev_priv,
>
> static int gen8_enable_metric_set(struct i915_perf_stream *stream)
> {
> - struct drm_i915_private *dev_priv = stream->dev_priv;
> + struct drm_i915_private *i915 = stream->i915;
> const struct i915_oa_config *oa_config = stream->oa_config;
> int ret;
>
> @@ -1821,7 +1821,7 @@ static int gen8_enable_metric_set(struct i915_perf_stream *stream)
> * be read back from automatically triggered reports, as part of the
> * RPT_ID field.
> */
> - if (IS_GEN_RANGE(dev_priv, 9, 11)) {
> + if (IS_GEN_RANGE(i915, 9, 11)) {
> I915_WRITE(GEN8_OA_DEBUG,
> _MASKED_BIT_ENABLE(GEN9_OA_DEBUG_DISABLE_CLK_RATIO_REPORTS |
> GEN9_OA_DEBUG_INCLUDE_CLK_RATIO));
> @@ -1832,31 +1832,31 @@ static int gen8_enable_metric_set(struct i915_perf_stream *stream)
> * to make sure all slices/subslices are ON before writing to NOA
> * registers.
> */
> - ret = gen8_configure_all_contexts(dev_priv, oa_config);
> + ret = gen8_configure_all_contexts(i915, oa_config);
> if (ret)
> return ret;
>
> - config_oa_regs(dev_priv, oa_config->mux_regs, oa_config->mux_regs_len);
> + config_oa_regs(i915, oa_config->mux_regs, oa_config->mux_regs_len);
>
> - config_oa_regs(dev_priv, oa_config->b_counter_regs,
> + config_oa_regs(i915, oa_config->b_counter_regs,
> oa_config->b_counter_regs_len);
>
> return 0;
> }
>
> -static void gen8_disable_metric_set(struct drm_i915_private *dev_priv)
> +static void gen8_disable_metric_set(struct drm_i915_private *i915)
> {
> /* Reset all contexts' slices/subslices configurations. */
> - gen8_configure_all_contexts(dev_priv, NULL);
> + gen8_configure_all_contexts(i915, NULL);
>
> I915_WRITE(GDT_CHICKEN_BITS, (I915_READ(GDT_CHICKEN_BITS) &
> ~GT_NOA_ENABLE));
> }
>
> -static void gen10_disable_metric_set(struct drm_i915_private *dev_priv)
> +static void gen10_disable_metric_set(struct drm_i915_private *i915)
> {
> /* Reset all contexts' slices/subslices configurations. */
> - gen8_configure_all_contexts(dev_priv, NULL);
> + gen8_configure_all_contexts(i915, NULL);
>
> /* Make sure we disable noa to save power. */
> I915_WRITE(RPM_CONFIG1,
> @@ -1865,12 +1865,12 @@ static void gen10_disable_metric_set(struct drm_i915_private *dev_priv)
>
> static void gen7_oa_enable(struct i915_perf_stream *stream)
> {
> - struct drm_i915_private *dev_priv = stream->dev_priv;
> + struct drm_i915_private *i915 = stream->i915;
> struct i915_gem_context *ctx = stream->ctx;
> - u32 ctx_id = dev_priv->perf.oa.specific_ctx_id;
> - bool periodic = dev_priv->perf.oa.periodic;
> - u32 period_exponent = dev_priv->perf.oa.period_exponent;
> - u32 report_format = dev_priv->perf.oa.oa_buffer.format;
> + u32 ctx_id = i915->perf.oa.specific_ctx_id;
> + bool periodic = i915->perf.oa.periodic;
> + u32 period_exponent = i915->perf.oa.period_exponent;
> + u32 report_format = i915->perf.oa.oa_buffer.format;
>
> /*
> * Reset buf pointers so we don't forward reports from before now.
> @@ -1881,7 +1881,7 @@ static void gen7_oa_enable(struct i915_perf_stream *stream)
> * on the assumption that certain fields are written to zeroed
> * memory which this helps maintains.
> */
> - gen7_init_oa_buffer(dev_priv);
> + gen7_init_oa_buffer(i915);
>
> I915_WRITE(GEN7_OACONTROL,
> (ctx_id & GEN7_OACONTROL_CTX_MASK) |
> @@ -1895,8 +1895,8 @@ static void gen7_oa_enable(struct i915_perf_stream *stream)
>
> static void gen8_oa_enable(struct i915_perf_stream *stream)
> {
> - struct drm_i915_private *dev_priv = stream->dev_priv;
> - u32 report_format = dev_priv->perf.oa.oa_buffer.format;
> + struct drm_i915_private *i915 = stream->i915;
> + u32 report_format = i915->perf.oa.oa_buffer.format;
>
> /*
> * Reset buf pointers so we don't forward reports from before now.
> @@ -1907,7 +1907,7 @@ static void gen8_oa_enable(struct i915_perf_stream *stream)
> * on the assumption that certain fields are written to zeroed
> * memory which this helps maintains.
> */
> - gen8_init_oa_buffer(dev_priv);
> + gen8_init_oa_buffer(i915);
>
> /*
> * Note: we don't rely on the hardware to perform single context
> @@ -1930,19 +1930,19 @@ static void gen8_oa_enable(struct i915_perf_stream *stream)
> */
> static void i915_oa_stream_enable(struct i915_perf_stream *stream)
> {
> - struct drm_i915_private *dev_priv = stream->dev_priv;
> + struct drm_i915_private *i915 = stream->i915;
>
> - dev_priv->perf.oa.ops.oa_enable(stream);
> + i915->perf.oa.ops.oa_enable(stream);
>
> - if (dev_priv->perf.oa.periodic)
> - hrtimer_start(&dev_priv->perf.oa.poll_check_timer,
> + if (i915->perf.oa.periodic)
> + hrtimer_start(&i915->perf.oa.poll_check_timer,
> ns_to_ktime(POLL_PERIOD),
> HRTIMER_MODE_REL_PINNED);
> }
>
> static void gen7_oa_disable(struct i915_perf_stream *stream)
> {
> - struct intel_uncore *uncore = &stream->dev_priv->uncore;
> + struct intel_uncore *uncore = &stream->i915->uncore;
>
> intel_uncore_write(uncore, GEN7_OACONTROL, 0);
> if (intel_wait_for_register(uncore,
> @@ -1953,7 +1953,7 @@ static void gen7_oa_disable(struct i915_perf_stream *stream)
>
> static void gen8_oa_disable(struct i915_perf_stream *stream)
> {
> - struct intel_uncore *uncore = &stream->dev_priv->uncore;
> + struct intel_uncore *uncore = &stream->i915->uncore;
>
> intel_uncore_write(uncore, GEN8_OACONTROL, 0);
> if (intel_wait_for_register(uncore,
> @@ -1972,12 +1972,12 @@ static void gen8_oa_disable(struct i915_perf_stream *stream)
> */
> static void i915_oa_stream_disable(struct i915_perf_stream *stream)
> {
> - struct drm_i915_private *dev_priv = stream->dev_priv;
> + struct drm_i915_private *i915 = stream->i915;
>
> - dev_priv->perf.oa.ops.oa_disable(stream);
> + i915->perf.oa.ops.oa_disable(stream);
>
> - if (dev_priv->perf.oa.periodic)
> - hrtimer_cancel(&dev_priv->perf.oa.poll_check_timer);
> + if (i915->perf.oa.periodic)
> + hrtimer_cancel(&i915->perf.oa.poll_check_timer);
> }
>
> static const struct i915_perf_stream_ops i915_oa_stream_ops = {
> @@ -2011,7 +2011,7 @@ static int i915_oa_stream_init(struct i915_perf_stream *stream,
> struct drm_i915_perf_open_param *param,
> struct perf_open_properties *props)
> {
> - struct drm_i915_private *dev_priv = stream->dev_priv;
> + struct drm_i915_private *i915 = stream->i915;
> int format_size;
> int ret;
>
> @@ -2019,7 +2019,7 @@ static int i915_oa_stream_init(struct i915_perf_stream *stream,
> * reason then don't let userspace try their luck with config
> * IDs
> */
> - if (!dev_priv->perf.metrics_kobj) {
> + if (!i915->perf.metrics_kobj) {
> DRM_DEBUG("OA metrics weren't advertised via sysfs\n");
> return -EINVAL;
> }
> @@ -2029,7 +2029,7 @@ static int i915_oa_stream_init(struct i915_perf_stream *stream,
> return -EINVAL;
> }
>
> - if (!dev_priv->perf.oa.ops.enable_metric_set) {
> + if (!i915->perf.oa.ops.enable_metric_set) {
> DRM_DEBUG("OA unit not supported\n");
> return -ENODEV;
> }
> @@ -2038,7 +2038,7 @@ static int i915_oa_stream_init(struct i915_perf_stream *stream,
> * counter reports and marshal to the appropriate client
> * we currently only allow exclusive access
> */
> - if (dev_priv->perf.oa.exclusive_stream) {
> + if (i915->perf.oa.exclusive_stream) {
> DRM_DEBUG("OA unit already in use\n");
> return -EBUSY;
> }
> @@ -2059,32 +2059,32 @@ static int i915_oa_stream_init(struct i915_perf_stream *stream,
> *
> * Using the same limiting factors as printk_ratelimit()
> */
> - ratelimit_state_init(&dev_priv->perf.oa.spurious_report_rs,
> + ratelimit_state_init(&i915->perf.oa.spurious_report_rs,
> 5 * HZ, 10);
> /* Since we use a DRM_NOTE for spurious reports it would be
> * inconsistent to let __ratelimit() automatically print a warning for
> * throttling.
> */
> - ratelimit_set_flags(&dev_priv->perf.oa.spurious_report_rs,
> + ratelimit_set_flags(&i915->perf.oa.spurious_report_rs,
> RATELIMIT_MSG_ON_RELEASE);
>
> stream->sample_size = sizeof(struct drm_i915_perf_record_header);
>
> - format_size = dev_priv->perf.oa.oa_formats[props->oa_format].size;
> + format_size = i915->perf.oa.oa_formats[props->oa_format].size;
>
> stream->sample_flags |= SAMPLE_OA_REPORT;
> stream->sample_size += format_size;
>
> - dev_priv->perf.oa.oa_buffer.format_size = format_size;
> - if (WARN_ON(dev_priv->perf.oa.oa_buffer.format_size == 0))
> + i915->perf.oa.oa_buffer.format_size = format_size;
> + if (WARN_ON(i915->perf.oa.oa_buffer.format_size == 0))
> return -EINVAL;
>
> - dev_priv->perf.oa.oa_buffer.format =
> - dev_priv->perf.oa.oa_formats[props->oa_format].format;
> + i915->perf.oa.oa_buffer.format =
> + i915->perf.oa.oa_formats[props->oa_format].format;
>
> - dev_priv->perf.oa.periodic = props->oa_periodic;
> - if (dev_priv->perf.oa.periodic)
> - dev_priv->perf.oa.period_exponent = props->oa_period_exponent;
> + i915->perf.oa.periodic = props->oa_periodic;
> + if (i915->perf.oa.periodic)
> + i915->perf.oa.period_exponent = props->oa_period_exponent;
>
> if (stream->ctx) {
> ret = oa_get_render_ctx_id(stream);
> @@ -2094,7 +2094,7 @@ static int i915_oa_stream_init(struct i915_perf_stream *stream,
> }
> }
>
> - ret = get_oa_config(dev_priv, props->metrics_set, &stream->oa_config);
> + ret = get_oa_config(i915, props->metrics_set, &stream->oa_config);
> if (ret) {
> DRM_DEBUG("Invalid OA config id=%i\n", props->metrics_set);
> goto err_config;
> @@ -2112,43 +2112,43 @@ static int i915_oa_stream_init(struct i915_perf_stream *stream,
> * In our case we are expecting that taking pm + FORCEWAKE
> * references will effectively disable RC6.
> */
> - stream->wakeref = intel_runtime_pm_get(dev_priv);
> - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> + stream->wakeref = intel_runtime_pm_get(i915);
> + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
>
> - ret = alloc_oa_buffer(dev_priv);
> + ret = alloc_oa_buffer(i915);
> if (ret)
> goto err_oa_buf_alloc;
>
> - ret = i915_mutex_lock_interruptible(&dev_priv->drm);
> + ret = i915_mutex_lock_interruptible(&i915->drm);
> if (ret)
> goto err_lock;
>
> stream->ops = &i915_oa_stream_ops;
> - dev_priv->perf.oa.exclusive_stream = stream;
> + i915->perf.oa.exclusive_stream = stream;
>
> - ret = dev_priv->perf.oa.ops.enable_metric_set(stream);
> + ret = i915->perf.oa.ops.enable_metric_set(stream);
> if (ret) {
> DRM_DEBUG("Unable to enable metric set\n");
> goto err_enable;
> }
>
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + mutex_unlock(&i915->drm.struct_mutex);
>
> return 0;
>
> err_enable:
> - dev_priv->perf.oa.exclusive_stream = NULL;
> - dev_priv->perf.oa.ops.disable_metric_set(dev_priv);
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + i915->perf.oa.exclusive_stream = NULL;
> + i915->perf.oa.ops.disable_metric_set(i915);
> + mutex_unlock(&i915->drm.struct_mutex);
>
> err_lock:
> - free_oa_buffer(dev_priv);
> + free_oa_buffer(i915);
>
> err_oa_buf_alloc:
> - put_oa_config(dev_priv, stream->oa_config);
> + put_oa_config(i915, stream->oa_config);
>
> - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> - intel_runtime_pm_put(dev_priv, stream->wakeref);
> + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> + intel_runtime_pm_put(i915, stream->wakeref);
>
> err_config:
> if (stream->ctx)
> @@ -2238,7 +2238,7 @@ static ssize_t i915_perf_read(struct file *file,
> loff_t *ppos)
> {
> struct i915_perf_stream *stream = file->private_data;
> - struct drm_i915_private *dev_priv = stream->dev_priv;
> + struct drm_i915_private *i915 = stream->i915;
> ssize_t ret;
>
> /* To ensure it's handled consistently we simply treat all reads of a
> @@ -2261,15 +2261,15 @@ static ssize_t i915_perf_read(struct file *file,
> if (ret)
> return ret;
>
> - mutex_lock(&dev_priv->perf.lock);
> + mutex_lock(&i915->perf.lock);
> ret = i915_perf_read_locked(stream, file,
> buf, count, ppos);
> - mutex_unlock(&dev_priv->perf.lock);
> + mutex_unlock(&i915->perf.lock);
> } while (ret == -EAGAIN);
> } else {
> - mutex_lock(&dev_priv->perf.lock);
> + mutex_lock(&i915->perf.lock);
> ret = i915_perf_read_locked(stream, file, buf, count, ppos);
> - mutex_unlock(&dev_priv->perf.lock);
> + mutex_unlock(&i915->perf.lock);
> }
>
> /* We allow the poll checking to sometimes report false positive EPOLLIN
> @@ -2284,7 +2284,7 @@ static ssize_t i915_perf_read(struct file *file,
> /* Maybe make ->pollin per-stream state if we support multiple
> * concurrent streams in the future.
> */
> - dev_priv->perf.oa.pollin = false;
> + i915->perf.oa.pollin = false;
> }
>
> return ret;
> @@ -2292,13 +2292,13 @@ static ssize_t i915_perf_read(struct file *file,
>
> static enum hrtimer_restart oa_poll_check_timer_cb(struct hrtimer *hrtimer)
> {
> - struct drm_i915_private *dev_priv =
> - container_of(hrtimer, typeof(*dev_priv),
> + struct drm_i915_private *i915 =
> + container_of(hrtimer, typeof(*i915),
> perf.oa.poll_check_timer);
>
> - if (oa_buffer_check_unlocked(dev_priv)) {
> - dev_priv->perf.oa.pollin = true;
> - wake_up(&dev_priv->perf.oa.poll_wq);
> + if (oa_buffer_check_unlocked(i915)) {
> + i915->perf.oa.pollin = true;
> + wake_up(&i915->perf.oa.poll_wq);
> }
>
> hrtimer_forward_now(hrtimer, ns_to_ktime(POLL_PERIOD));
> @@ -2308,7 +2308,7 @@ static enum hrtimer_restart oa_poll_check_timer_cb(struct hrtimer *hrtimer)
>
> /**
> * i915_perf_poll_locked - poll_wait() with a suitable wait queue for stream
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> * @stream: An i915 perf stream
> * @file: An i915 perf stream file
> * @wait: poll() state table
> @@ -2322,7 +2322,7 @@ static enum hrtimer_restart oa_poll_check_timer_cb(struct hrtimer *hrtimer)
> *
> * Returns: any poll events that are ready without sleeping
> */
> -static __poll_t i915_perf_poll_locked(struct drm_i915_private *dev_priv,
> +static __poll_t i915_perf_poll_locked(struct drm_i915_private *i915,
> struct i915_perf_stream *stream,
> struct file *file,
> poll_table *wait)
> @@ -2337,7 +2337,7 @@ static __poll_t i915_perf_poll_locked(struct drm_i915_private *dev_priv,
> * the hrtimer/oa_poll_check_timer_cb to notify us when there are
> * samples to read.
> */
> - if (dev_priv->perf.oa.pollin)
> + if (i915->perf.oa.pollin)
> events |= EPOLLIN;
>
> return events;
> @@ -2359,12 +2359,12 @@ static __poll_t i915_perf_poll_locked(struct drm_i915_private *dev_priv,
> static __poll_t i915_perf_poll(struct file *file, poll_table *wait)
> {
> struct i915_perf_stream *stream = file->private_data;
> - struct drm_i915_private *dev_priv = stream->dev_priv;
> + struct drm_i915_private *i915 = stream->i915;
> __poll_t ret;
>
> - mutex_lock(&dev_priv->perf.lock);
> - ret = i915_perf_poll_locked(dev_priv, stream, file, wait);
> - mutex_unlock(&dev_priv->perf.lock);
> + mutex_lock(&i915->perf.lock);
> + ret = i915_perf_poll_locked(i915, stream, file, wait);
> + mutex_unlock(&i915->perf.lock);
>
> return ret;
> }
> @@ -2461,12 +2461,12 @@ static long i915_perf_ioctl(struct file *file,
> unsigned long arg)
> {
> struct i915_perf_stream *stream = file->private_data;
> - struct drm_i915_private *dev_priv = stream->dev_priv;
> + struct drm_i915_private *i915 = stream->i915;
> long ret;
>
> - mutex_lock(&dev_priv->perf.lock);
> + mutex_lock(&i915->perf.lock);
> ret = i915_perf_ioctl_locked(stream, cmd, arg);
> - mutex_unlock(&dev_priv->perf.lock);
> + mutex_unlock(&i915->perf.lock);
>
> return ret;
> }
> @@ -2511,11 +2511,11 @@ static void i915_perf_destroy_locked(struct i915_perf_stream *stream)
> static int i915_perf_release(struct inode *inode, struct file *file)
> {
> struct i915_perf_stream *stream = file->private_data;
> - struct drm_i915_private *dev_priv = stream->dev_priv;
> + struct drm_i915_private *i915 = stream->i915;
>
> - mutex_lock(&dev_priv->perf.lock);
> + mutex_lock(&i915->perf.lock);
> i915_perf_destroy_locked(stream);
> - mutex_unlock(&dev_priv->perf.lock);
> + mutex_unlock(&i915->perf.lock);
>
> return 0;
> }
> @@ -2537,7 +2537,7 @@ static const struct file_operations fops = {
>
> /**
> * i915_perf_open_ioctl_locked - DRM ioctl() for userspace to open a stream FD
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> * @param: The open parameters passed to 'DRM_I915_PERF_OPEN`
> * @props: individually validated u64 property value pairs
> * @file: drm file
> @@ -2560,7 +2560,7 @@ static const struct file_operations fops = {
> * Returns: zero on success or a negative error code.
> */
> static int
> -i915_perf_open_ioctl_locked(struct drm_i915_private *dev_priv,
> +i915_perf_open_ioctl_locked(struct drm_i915_private *i915,
> struct drm_i915_perf_open_param *param,
> struct perf_open_properties *props,
> struct drm_file *file)
> @@ -2599,7 +2599,7 @@ i915_perf_open_ioctl_locked(struct drm_i915_private *dev_priv,
> * MI_REPORT_PERF_COUNT commands and so consider it a privileged op to
> * enable the OA unit by default.
> */
> - if (IS_HASWELL(dev_priv) && specific_ctx)
> + if (IS_HASWELL(i915) && specific_ctx)
> privileged_op = false;
>
> /* Similar to perf's kernel.perf_paranoid_cpu sysctl option
> @@ -2620,7 +2620,7 @@ i915_perf_open_ioctl_locked(struct drm_i915_private *dev_priv,
> goto err_ctx;
> }
>
> - stream->dev_priv = dev_priv;
> + stream->i915 = i915;
> stream->ctx = specific_ctx;
>
> ret = i915_oa_stream_init(stream, param, props);
> @@ -2636,7 +2636,7 @@ i915_perf_open_ioctl_locked(struct drm_i915_private *dev_priv,
> goto err_flags;
> }
>
> - list_add(&stream->link, &dev_priv->perf.streams);
> + list_add(&stream->link, &i915->perf.streams);
>
> if (param->flags & I915_PERF_FLAG_FD_CLOEXEC)
> f_flags |= O_CLOEXEC;
> @@ -2668,15 +2668,15 @@ i915_perf_open_ioctl_locked(struct drm_i915_private *dev_priv,
> return ret;
> }
>
> -static u64 oa_exponent_to_ns(struct drm_i915_private *dev_priv, int exponent)
> +static u64 oa_exponent_to_ns(struct drm_i915_private *i915, int exponent)
> {
> return div64_u64(1000000000ULL * (2ULL << exponent),
> - 1000ULL * RUNTIME_INFO(dev_priv)->cs_timestamp_frequency_khz);
> + 1000ULL * RUNTIME_INFO(i915)->cs_timestamp_frequency_khz);
> }
>
> /**
> * read_properties_unlocked - validate + copy userspace stream open properties
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> * @uprops: The array of u64 key value pairs given by userspace
> * @n_props: The number of key value pairs expected in @uprops
> * @props: The stream configuration built up while validating properties
> @@ -2689,7 +2689,7 @@ static u64 oa_exponent_to_ns(struct drm_i915_private *dev_priv, int exponent)
> * we shouldn't validate or assume anything about ordering here. This doesn't
> * rule out defining new properties with ordering requirements in the future.
> */
> -static int read_properties_unlocked(struct drm_i915_private *dev_priv,
> +static int read_properties_unlocked(struct drm_i915_private *i915,
> u64 __user *uprops,
> u32 n_props,
> struct perf_open_properties *props)
> @@ -2755,7 +2755,7 @@ static int read_properties_unlocked(struct drm_i915_private *dev_priv,
> value);
> return -EINVAL;
> }
> - if (!dev_priv->perf.oa.oa_formats[value].size) {
> + if (!i915->perf.oa.oa_formats[value].size) {
> DRM_DEBUG("Unsupported OA report format %llu\n",
> value);
> return -EINVAL;
> @@ -2776,7 +2776,7 @@ static int read_properties_unlocked(struct drm_i915_private *dev_priv,
> */
>
> BUILD_BUG_ON(sizeof(oa_period) != 8);
> - oa_period = oa_exponent_to_ns(dev_priv, value);
> + oa_period = oa_exponent_to_ns(i915, value);
>
> /* This check is primarily to ensure that oa_period <=
> * UINT32_MAX (before passing to do_div which only
> @@ -2839,13 +2839,13 @@ static int read_properties_unlocked(struct drm_i915_private *dev_priv,
> int i915_perf_open_ioctl(struct drm_device *dev, void *data,
> struct drm_file *file)
> {
> - struct drm_i915_private *dev_priv = dev->dev_private;
> + struct drm_i915_private *i915 = dev->dev_private;
> struct drm_i915_perf_open_param *param = data;
> struct perf_open_properties props;
> u32 known_open_flags;
> int ret;
>
> - if (!dev_priv->perf.initialized) {
> + if (!i915->perf.initialized) {
> DRM_DEBUG("i915 perf interface not available for this system\n");
> return -ENOTSUPP;
> }
> @@ -2858,124 +2858,124 @@ int i915_perf_open_ioctl(struct drm_device *dev, void *data,
> return -EINVAL;
> }
>
> - ret = read_properties_unlocked(dev_priv,
> + ret = read_properties_unlocked(i915,
> u64_to_user_ptr(param->properties_ptr),
> param->num_properties,
> &props);
> if (ret)
> return ret;
>
> - mutex_lock(&dev_priv->perf.lock);
> - ret = i915_perf_open_ioctl_locked(dev_priv, param, &props, file);
> - mutex_unlock(&dev_priv->perf.lock);
> + mutex_lock(&i915->perf.lock);
> + ret = i915_perf_open_ioctl_locked(i915, param, &props, file);
> + mutex_unlock(&i915->perf.lock);
>
> return ret;
> }
>
> /**
> * i915_perf_register - exposes i915-perf to userspace
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> *
> * In particular OA metric sets are advertised under a sysfs metrics/
> * directory allowing userspace to enumerate valid IDs that can be
> * used to open an i915-perf stream.
> */
> -void i915_perf_register(struct drm_i915_private *dev_priv)
> +void i915_perf_register(struct drm_i915_private *i915)
> {
> int ret;
>
> - if (!dev_priv->perf.initialized)
> + if (!i915->perf.initialized)
> return;
>
> /* To be sure we're synchronized with an attempted
> * i915_perf_open_ioctl(); considering that we register after
> * being exposed to userspace.
> */
> - mutex_lock(&dev_priv->perf.lock);
> + mutex_lock(&i915->perf.lock);
>
> - dev_priv->perf.metrics_kobj =
> + i915->perf.metrics_kobj =
> kobject_create_and_add("metrics",
> - &dev_priv->drm.primary->kdev->kobj);
> - if (!dev_priv->perf.metrics_kobj)
> + &i915->drm.primary->kdev->kobj);
> + if (!i915->perf.metrics_kobj)
> goto exit;
>
> - sysfs_attr_init(&dev_priv->perf.oa.test_config.sysfs_metric_id.attr);
> -
> - if (INTEL_GEN(dev_priv) >= 11) {
> - i915_perf_load_test_config_icl(dev_priv);
> - } else if (IS_CANNONLAKE(dev_priv)) {
> - i915_perf_load_test_config_cnl(dev_priv);
> - } else if (IS_COFFEELAKE(dev_priv)) {
> - if (IS_CFL_GT2(dev_priv))
> - i915_perf_load_test_config_cflgt2(dev_priv);
> - if (IS_CFL_GT3(dev_priv))
> - i915_perf_load_test_config_cflgt3(dev_priv);
> - } else if (IS_GEMINILAKE(dev_priv)) {
> - i915_perf_load_test_config_glk(dev_priv);
> - } else if (IS_KABYLAKE(dev_priv)) {
> - if (IS_KBL_GT2(dev_priv))
> - i915_perf_load_test_config_kblgt2(dev_priv);
> - else if (IS_KBL_GT3(dev_priv))
> - i915_perf_load_test_config_kblgt3(dev_priv);
> - } else if (IS_BROXTON(dev_priv)) {
> - i915_perf_load_test_config_bxt(dev_priv);
> - } else if (IS_SKYLAKE(dev_priv)) {
> - if (IS_SKL_GT2(dev_priv))
> - i915_perf_load_test_config_sklgt2(dev_priv);
> - else if (IS_SKL_GT3(dev_priv))
> - i915_perf_load_test_config_sklgt3(dev_priv);
> - else if (IS_SKL_GT4(dev_priv))
> - i915_perf_load_test_config_sklgt4(dev_priv);
> - } else if (IS_CHERRYVIEW(dev_priv)) {
> - i915_perf_load_test_config_chv(dev_priv);
> - } else if (IS_BROADWELL(dev_priv)) {
> - i915_perf_load_test_config_bdw(dev_priv);
> - } else if (IS_HASWELL(dev_priv)) {
> - i915_perf_load_test_config_hsw(dev_priv);
> -}
> -
> - if (dev_priv->perf.oa.test_config.id == 0)
> + sysfs_attr_init(&i915->perf.oa.test_config.sysfs_metric_id.attr);
> +
> + if (INTEL_GEN(i915) >= 11) {
> + i915_perf_load_test_config_icl(i915);
> + } else if (IS_CANNONLAKE(i915)) {
> + i915_perf_load_test_config_cnl(i915);
> + } else if (IS_COFFEELAKE(i915)) {
> + if (IS_CFL_GT2(i915))
> + i915_perf_load_test_config_cflgt2(i915);
> + if (IS_CFL_GT3(i915))
> + i915_perf_load_test_config_cflgt3(i915);
> + } else if (IS_GEMINILAKE(i915)) {
> + i915_perf_load_test_config_glk(i915);
> + } else if (IS_KABYLAKE(i915)) {
> + if (IS_KBL_GT2(i915))
> + i915_perf_load_test_config_kblgt2(i915);
> + else if (IS_KBL_GT3(i915))
> + i915_perf_load_test_config_kblgt3(i915);
> + } else if (IS_BROXTON(i915)) {
> + i915_perf_load_test_config_bxt(i915);
> + } else if (IS_SKYLAKE(i915)) {
> + if (IS_SKL_GT2(i915))
> + i915_perf_load_test_config_sklgt2(i915);
> + else if (IS_SKL_GT3(i915))
> + i915_perf_load_test_config_sklgt3(i915);
> + else if (IS_SKL_GT4(i915))
> + i915_perf_load_test_config_sklgt4(i915);
> + } else if (IS_CHERRYVIEW(i915)) {
> + i915_perf_load_test_config_chv(i915);
> + } else if (IS_BROADWELL(i915)) {
> + i915_perf_load_test_config_bdw(i915);
> + } else if (IS_HASWELL(i915)) {
> + i915_perf_load_test_config_hsw(i915);
> +}
> +
> + if (i915->perf.oa.test_config.id == 0)
> goto sysfs_error;
>
> - ret = sysfs_create_group(dev_priv->perf.metrics_kobj,
> - &dev_priv->perf.oa.test_config.sysfs_metric);
> + ret = sysfs_create_group(i915->perf.metrics_kobj,
> + &i915->perf.oa.test_config.sysfs_metric);
> if (ret)
> goto sysfs_error;
>
> - atomic_set(&dev_priv->perf.oa.test_config.ref_count, 1);
> + atomic_set(&i915->perf.oa.test_config.ref_count, 1);
>
> goto exit;
>
> sysfs_error:
> - kobject_put(dev_priv->perf.metrics_kobj);
> - dev_priv->perf.metrics_kobj = NULL;
> + kobject_put(i915->perf.metrics_kobj);
> + i915->perf.metrics_kobj = NULL;
>
> exit:
> - mutex_unlock(&dev_priv->perf.lock);
> + mutex_unlock(&i915->perf.lock);
> }
>
> /**
> * i915_perf_unregister - hide i915-perf from userspace
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> *
> * i915-perf state cleanup is split up into an 'unregister' and
> * 'deinit' phase where the interface is first hidden from
> * userspace by i915_perf_unregister() before cleaning up
> * remaining state in i915_perf_fini().
> */
> -void i915_perf_unregister(struct drm_i915_private *dev_priv)
> +void i915_perf_unregister(struct drm_i915_private *i915)
> {
> - if (!dev_priv->perf.metrics_kobj)
> + if (!i915->perf.metrics_kobj)
> return;
>
> - sysfs_remove_group(dev_priv->perf.metrics_kobj,
> - &dev_priv->perf.oa.test_config.sysfs_metric);
> + sysfs_remove_group(i915->perf.metrics_kobj,
> + &i915->perf.oa.test_config.sysfs_metric);
>
> - kobject_put(dev_priv->perf.metrics_kobj);
> - dev_priv->perf.metrics_kobj = NULL;
> + kobject_put(i915->perf.metrics_kobj);
> + i915->perf.metrics_kobj = NULL;
> }
>
> -static bool gen8_is_valid_flex_addr(struct drm_i915_private *dev_priv, u32 addr)
> +static bool gen8_is_valid_flex_addr(struct drm_i915_private *i915, u32 addr)
> {
> static const i915_reg_t flex_eu_regs[] = {
> EU_PERF_CNTL0,
> @@ -2995,7 +2995,7 @@ static bool gen8_is_valid_flex_addr(struct drm_i915_private *dev_priv, u32 addr)
> return false;
> }
>
> -static bool gen7_is_valid_b_counter_addr(struct drm_i915_private *dev_priv, u32 addr)
> +static bool gen7_is_valid_b_counter_addr(struct drm_i915_private *i915, u32 addr)
> {
> return (addr >= i915_mmio_reg_offset(OASTARTTRIG1) &&
> addr <= i915_mmio_reg_offset(OASTARTTRIG8)) ||
> @@ -3005,7 +3005,7 @@ static bool gen7_is_valid_b_counter_addr(struct drm_i915_private *dev_priv, u32
> addr <= i915_mmio_reg_offset(OACEC7_1));
> }
>
> -static bool gen7_is_valid_mux_addr(struct drm_i915_private *dev_priv, u32 addr)
> +static bool gen7_is_valid_mux_addr(struct drm_i915_private *i915, u32 addr)
> {
> return addr == i915_mmio_reg_offset(HALF_SLICE_CHICKEN2) ||
> (addr >= i915_mmio_reg_offset(MICRO_BP0_0) &&
> @@ -3016,34 +3016,34 @@ static bool gen7_is_valid_mux_addr(struct drm_i915_private *dev_priv, u32 addr)
> addr <= i915_mmio_reg_offset(OA_PERFMATRIX_HI));
> }
>
> -static bool gen8_is_valid_mux_addr(struct drm_i915_private *dev_priv, u32 addr)
> +static bool gen8_is_valid_mux_addr(struct drm_i915_private *i915, u32 addr)
> {
> - return gen7_is_valid_mux_addr(dev_priv, addr) ||
> + return gen7_is_valid_mux_addr(i915, addr) ||
> addr == i915_mmio_reg_offset(WAIT_FOR_RC6_EXIT) ||
> (addr >= i915_mmio_reg_offset(RPM_CONFIG0) &&
> addr <= i915_mmio_reg_offset(NOA_CONFIG(8)));
> }
>
> -static bool gen10_is_valid_mux_addr(struct drm_i915_private *dev_priv, u32 addr)
> +static bool gen10_is_valid_mux_addr(struct drm_i915_private *i915, u32 addr)
> {
> - return gen8_is_valid_mux_addr(dev_priv, addr) ||
> + return gen8_is_valid_mux_addr(i915, addr) ||
> addr == i915_mmio_reg_offset(GEN10_NOA_WRITE_HIGH) ||
> (addr >= i915_mmio_reg_offset(OA_PERFCNT3_LO) &&
> addr <= i915_mmio_reg_offset(OA_PERFCNT4_HI));
> }
>
> -static bool hsw_is_valid_mux_addr(struct drm_i915_private *dev_priv, u32 addr)
> +static bool hsw_is_valid_mux_addr(struct drm_i915_private *i915, u32 addr)
> {
> - return gen7_is_valid_mux_addr(dev_priv, addr) ||
> + return gen7_is_valid_mux_addr(i915, addr) ||
> (addr >= 0x25100 && addr <= 0x2FF90) ||
> (addr >= i915_mmio_reg_offset(HSW_MBVID2_NOA0) &&
> addr <= i915_mmio_reg_offset(HSW_MBVID2_NOA9)) ||
> addr == i915_mmio_reg_offset(HSW_MBVID2_MISR0);
> }
>
> -static bool chv_is_valid_mux_addr(struct drm_i915_private *dev_priv, u32 addr)
> +static bool chv_is_valid_mux_addr(struct drm_i915_private *i915, u32 addr)
> {
> - return gen7_is_valid_mux_addr(dev_priv, addr) ||
> + return gen7_is_valid_mux_addr(i915, addr) ||
> (addr >= 0x182300 && addr <= 0x1823A4);
> }
>
> @@ -3066,8 +3066,8 @@ static u32 mask_reg_value(u32 reg, u32 val)
> return val;
> }
>
> -static struct i915_oa_reg *alloc_oa_regs(struct drm_i915_private *dev_priv,
> - bool (*is_valid)(struct drm_i915_private *dev_priv, u32 addr),
> +static struct i915_oa_reg *alloc_oa_regs(struct drm_i915_private *i915,
> + bool (*is_valid)(struct drm_i915_private *i915, u32 addr),
> u32 __user *regs,
> u32 n_regs)
> {
> @@ -3097,7 +3097,7 @@ static struct i915_oa_reg *alloc_oa_regs(struct drm_i915_private *dev_priv,
> if (err)
> goto addr_err;
>
> - if (!is_valid(dev_priv, addr)) {
> + if (!is_valid(i915, addr)) {
> DRM_DEBUG("Invalid oa_reg address: %X\n", addr);
> err = -EINVAL;
> goto addr_err;
> @@ -3130,7 +3130,7 @@ static ssize_t show_dynamic_id(struct device *dev,
> return sprintf(buf, "%d\n", oa_config->id);
> }
>
> -static int create_dynamic_oa_sysfs_entry(struct drm_i915_private *dev_priv,
> +static int create_dynamic_oa_sysfs_entry(struct drm_i915_private *i915,
> struct i915_oa_config *oa_config)
> {
> sysfs_attr_init(&oa_config->sysfs_metric_id.attr);
> @@ -3145,7 +3145,7 @@ static int create_dynamic_oa_sysfs_entry(struct drm_i915_private *dev_priv,
> oa_config->sysfs_metric.name = oa_config->uuid;
> oa_config->sysfs_metric.attrs = oa_config->attrs;
>
> - return sysfs_create_group(dev_priv->perf.metrics_kobj,
> + return sysfs_create_group(i915->perf.metrics_kobj,
> &oa_config->sysfs_metric);
> }
>
> @@ -3165,17 +3165,17 @@ static int create_dynamic_oa_sysfs_entry(struct drm_i915_private *dev_priv,
> int i915_perf_add_config_ioctl(struct drm_device *dev, void *data,
> struct drm_file *file)
> {
> - struct drm_i915_private *dev_priv = dev->dev_private;
> + struct drm_i915_private *i915 = dev->dev_private;
> struct drm_i915_perf_oa_config *args = data;
> struct i915_oa_config *oa_config, *tmp;
> int err, id;
>
> - if (!dev_priv->perf.initialized) {
> + if (!i915->perf.initialized) {
> DRM_DEBUG("i915 perf interface not available for this system\n");
> return -ENOTSUPP;
> }
>
> - if (!dev_priv->perf.metrics_kobj) {
> + if (!i915->perf.metrics_kobj) {
> DRM_DEBUG("OA metrics weren't advertised via sysfs\n");
> return -EINVAL;
> }
> @@ -3213,8 +3213,8 @@ int i915_perf_add_config_ioctl(struct drm_device *dev, void *data,
>
> oa_config->mux_regs_len = args->n_mux_regs;
> oa_config->mux_regs =
> - alloc_oa_regs(dev_priv,
> - dev_priv->perf.oa.ops.is_valid_mux_reg,
> + alloc_oa_regs(i915,
> + i915->perf.oa.ops.is_valid_mux_reg,
> u64_to_user_ptr(args->mux_regs_ptr),
> args->n_mux_regs);
>
> @@ -3226,8 +3226,8 @@ int i915_perf_add_config_ioctl(struct drm_device *dev, void *data,
>
> oa_config->b_counter_regs_len = args->n_boolean_regs;
> oa_config->b_counter_regs =
> - alloc_oa_regs(dev_priv,
> - dev_priv->perf.oa.ops.is_valid_b_counter_reg,
> + alloc_oa_regs(i915,
> + i915->perf.oa.ops.is_valid_b_counter_reg,
> u64_to_user_ptr(args->boolean_regs_ptr),
> args->n_boolean_regs);
>
> @@ -3237,7 +3237,7 @@ int i915_perf_add_config_ioctl(struct drm_device *dev, void *data,
> goto reg_err;
> }
>
> - if (INTEL_GEN(dev_priv) < 8) {
> + if (INTEL_GEN(i915) < 8) {
> if (args->n_flex_regs != 0) {
> err = -EINVAL;
> goto reg_err;
> @@ -3245,8 +3245,8 @@ int i915_perf_add_config_ioctl(struct drm_device *dev, void *data,
> } else {
> oa_config->flex_regs_len = args->n_flex_regs;
> oa_config->flex_regs =
> - alloc_oa_regs(dev_priv,
> - dev_priv->perf.oa.ops.is_valid_flex_reg,
> + alloc_oa_regs(i915,
> + i915->perf.oa.ops.is_valid_flex_reg,
> u64_to_user_ptr(args->flex_regs_ptr),
> args->n_flex_regs);
>
> @@ -3257,14 +3257,14 @@ int i915_perf_add_config_ioctl(struct drm_device *dev, void *data,
> }
> }
>
> - err = mutex_lock_interruptible(&dev_priv->perf.metrics_lock);
> + err = mutex_lock_interruptible(&i915->perf.metrics_lock);
> if (err)
> goto reg_err;
>
> /* We shouldn't have too many configs, so this iteration shouldn't be
> * too costly.
> */
> - idr_for_each_entry(&dev_priv->perf.metrics_idr, tmp, id) {
> + idr_for_each_entry(&i915->perf.metrics_idr, tmp, id) {
> if (!strcmp(tmp->uuid, oa_config->uuid)) {
> DRM_DEBUG("OA config already exists with this uuid\n");
> err = -EADDRINUSE;
> @@ -3272,14 +3272,14 @@ int i915_perf_add_config_ioctl(struct drm_device *dev, void *data,
> }
> }
>
> - err = create_dynamic_oa_sysfs_entry(dev_priv, oa_config);
> + err = create_dynamic_oa_sysfs_entry(i915, oa_config);
> if (err) {
> DRM_DEBUG("Failed to create sysfs entry for OA config\n");
> goto sysfs_err;
> }
>
> /* Config id 0 is invalid, id 1 for kernel stored test config. */
> - oa_config->id = idr_alloc(&dev_priv->perf.metrics_idr,
> + oa_config->id = idr_alloc(&i915->perf.metrics_idr,
> oa_config, 2,
> 0, GFP_KERNEL);
> if (oa_config->id < 0) {
> @@ -3288,16 +3288,16 @@ int i915_perf_add_config_ioctl(struct drm_device *dev, void *data,
> goto sysfs_err;
> }
>
> - mutex_unlock(&dev_priv->perf.metrics_lock);
> + mutex_unlock(&i915->perf.metrics_lock);
>
> DRM_DEBUG("Added config %s id=%i\n", oa_config->uuid, oa_config->id);
>
> return oa_config->id;
>
> sysfs_err:
> - mutex_unlock(&dev_priv->perf.metrics_lock);
> + mutex_unlock(&i915->perf.metrics_lock);
> reg_err:
> - put_oa_config(dev_priv, oa_config);
> + put_oa_config(i915, oa_config);
> DRM_DEBUG("Failed to add new OA config\n");
> return err;
> }
> @@ -3316,12 +3316,12 @@ int i915_perf_add_config_ioctl(struct drm_device *dev, void *data,
> int i915_perf_remove_config_ioctl(struct drm_device *dev, void *data,
> struct drm_file *file)
> {
> - struct drm_i915_private *dev_priv = dev->dev_private;
> + struct drm_i915_private *i915 = dev->dev_private;
> u64 *arg = data;
> struct i915_oa_config *oa_config;
> int ret;
>
> - if (!dev_priv->perf.initialized) {
> + if (!i915->perf.initialized) {
> DRM_DEBUG("i915 perf interface not available for this system\n");
> return -ENOTSUPP;
> }
> @@ -3331,11 +3331,11 @@ int i915_perf_remove_config_ioctl(struct drm_device *dev, void *data,
> return -EACCES;
> }
>
> - ret = mutex_lock_interruptible(&dev_priv->perf.metrics_lock);
> + ret = mutex_lock_interruptible(&i915->perf.metrics_lock);
> if (ret)
> goto lock_err;
>
> - oa_config = idr_find(&dev_priv->perf.metrics_idr, *arg);
> + oa_config = idr_find(&i915->perf.metrics_idr, *arg);
> if (!oa_config) {
> DRM_DEBUG("Failed to remove unknown OA config\n");
> ret = -ENOENT;
> @@ -3344,17 +3344,17 @@ int i915_perf_remove_config_ioctl(struct drm_device *dev, void *data,
>
> GEM_BUG_ON(*arg != oa_config->id);
>
> - sysfs_remove_group(dev_priv->perf.metrics_kobj,
> + sysfs_remove_group(i915->perf.metrics_kobj,
> &oa_config->sysfs_metric);
>
> - idr_remove(&dev_priv->perf.metrics_idr, *arg);
> + idr_remove(&i915->perf.metrics_idr, *arg);
>
> DRM_DEBUG("Removed config %s id=%i\n", oa_config->uuid, oa_config->id);
>
> - put_oa_config(dev_priv, oa_config);
> + put_oa_config(i915, oa_config);
>
> config_err:
> - mutex_unlock(&dev_priv->perf.metrics_lock);
> + mutex_unlock(&i915->perf.metrics_lock);
> lock_err:
> return ret;
> }
> @@ -3403,135 +3403,135 @@ static struct ctl_table dev_root[] = {
>
> /**
> * i915_perf_init - initialize i915-perf state on module load
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> *
> * Initializes i915-perf state without exposing anything to userspace.
> *
> * Note: i915-perf initialization is split into an 'init' and 'register'
> * phase with the i915_perf_register() exposing state to userspace.
> */
> -void i915_perf_init(struct drm_i915_private *dev_priv)
> +void i915_perf_init(struct drm_i915_private *i915)
> {
> - if (IS_HASWELL(dev_priv)) {
> - dev_priv->perf.oa.ops.is_valid_b_counter_reg =
> + if (IS_HASWELL(i915)) {
> + i915->perf.oa.ops.is_valid_b_counter_reg =
> gen7_is_valid_b_counter_addr;
> - dev_priv->perf.oa.ops.is_valid_mux_reg =
> + i915->perf.oa.ops.is_valid_mux_reg =
> hsw_is_valid_mux_addr;
> - dev_priv->perf.oa.ops.is_valid_flex_reg = NULL;
> - dev_priv->perf.oa.ops.enable_metric_set = hsw_enable_metric_set;
> - dev_priv->perf.oa.ops.disable_metric_set = hsw_disable_metric_set;
> - dev_priv->perf.oa.ops.oa_enable = gen7_oa_enable;
> - dev_priv->perf.oa.ops.oa_disable = gen7_oa_disable;
> - dev_priv->perf.oa.ops.read = gen7_oa_read;
> - dev_priv->perf.oa.ops.oa_hw_tail_read =
> + i915->perf.oa.ops.is_valid_flex_reg = NULL;
> + i915->perf.oa.ops.enable_metric_set = hsw_enable_metric_set;
> + i915->perf.oa.ops.disable_metric_set = hsw_disable_metric_set;
> + i915->perf.oa.ops.oa_enable = gen7_oa_enable;
> + i915->perf.oa.ops.oa_disable = gen7_oa_disable;
> + i915->perf.oa.ops.read = gen7_oa_read;
> + i915->perf.oa.ops.oa_hw_tail_read =
> gen7_oa_hw_tail_read;
>
> - dev_priv->perf.oa.oa_formats = hsw_oa_formats;
> - } else if (HAS_LOGICAL_RING_CONTEXTS(dev_priv)) {
> + i915->perf.oa.oa_formats = hsw_oa_formats;
> + } else if (HAS_LOGICAL_RING_CONTEXTS(i915)) {
> /* Note: that although we could theoretically also support the
> * legacy ringbuffer mode on BDW (and earlier iterations of
> * this driver, before upstreaming did this) it didn't seem
> * worth the complexity to maintain now that BDW+ enable
> * execlist mode by default.
> */
> - dev_priv->perf.oa.oa_formats = gen8_plus_oa_formats;
> + i915->perf.oa.oa_formats = gen8_plus_oa_formats;
>
> - dev_priv->perf.oa.ops.oa_enable = gen8_oa_enable;
> - dev_priv->perf.oa.ops.oa_disable = gen8_oa_disable;
> - dev_priv->perf.oa.ops.read = gen8_oa_read;
> - dev_priv->perf.oa.ops.oa_hw_tail_read = gen8_oa_hw_tail_read;
> + i915->perf.oa.ops.oa_enable = gen8_oa_enable;
> + i915->perf.oa.ops.oa_disable = gen8_oa_disable;
> + i915->perf.oa.ops.read = gen8_oa_read;
> + i915->perf.oa.ops.oa_hw_tail_read = gen8_oa_hw_tail_read;
>
> - if (IS_GEN_RANGE(dev_priv, 8, 9)) {
> - dev_priv->perf.oa.ops.is_valid_b_counter_reg =
> + if (IS_GEN_RANGE(i915, 8, 9)) {
> + i915->perf.oa.ops.is_valid_b_counter_reg =
> gen7_is_valid_b_counter_addr;
> - dev_priv->perf.oa.ops.is_valid_mux_reg =
> + i915->perf.oa.ops.is_valid_mux_reg =
> gen8_is_valid_mux_addr;
> - dev_priv->perf.oa.ops.is_valid_flex_reg =
> + i915->perf.oa.ops.is_valid_flex_reg =
> gen8_is_valid_flex_addr;
>
> - if (IS_CHERRYVIEW(dev_priv)) {
> - dev_priv->perf.oa.ops.is_valid_mux_reg =
> + if (IS_CHERRYVIEW(i915)) {
> + i915->perf.oa.ops.is_valid_mux_reg =
> chv_is_valid_mux_addr;
> }
>
> - dev_priv->perf.oa.ops.enable_metric_set = gen8_enable_metric_set;
> - dev_priv->perf.oa.ops.disable_metric_set = gen8_disable_metric_set;
> + i915->perf.oa.ops.enable_metric_set = gen8_enable_metric_set;
> + i915->perf.oa.ops.disable_metric_set = gen8_disable_metric_set;
>
> - if (IS_GEN(dev_priv, 8)) {
> - dev_priv->perf.oa.ctx_oactxctrl_offset = 0x120;
> - dev_priv->perf.oa.ctx_flexeu0_offset = 0x2ce;
> + if (IS_GEN(i915, 8)) {
> + i915->perf.oa.ctx_oactxctrl_offset = 0x120;
> + i915->perf.oa.ctx_flexeu0_offset = 0x2ce;
>
> - dev_priv->perf.oa.gen8_valid_ctx_bit = (1<<25);
> + i915->perf.oa.gen8_valid_ctx_bit = (1<<25);
> } else {
> - dev_priv->perf.oa.ctx_oactxctrl_offset = 0x128;
> - dev_priv->perf.oa.ctx_flexeu0_offset = 0x3de;
> + i915->perf.oa.ctx_oactxctrl_offset = 0x128;
> + i915->perf.oa.ctx_flexeu0_offset = 0x3de;
>
> - dev_priv->perf.oa.gen8_valid_ctx_bit = (1<<16);
> + i915->perf.oa.gen8_valid_ctx_bit = (1<<16);
> }
> - } else if (IS_GEN_RANGE(dev_priv, 10, 11)) {
> - dev_priv->perf.oa.ops.is_valid_b_counter_reg =
> + } else if (IS_GEN_RANGE(i915, 10, 11)) {
> + i915->perf.oa.ops.is_valid_b_counter_reg =
> gen7_is_valid_b_counter_addr;
> - dev_priv->perf.oa.ops.is_valid_mux_reg =
> + i915->perf.oa.ops.is_valid_mux_reg =
> gen10_is_valid_mux_addr;
> - dev_priv->perf.oa.ops.is_valid_flex_reg =
> + i915->perf.oa.ops.is_valid_flex_reg =
> gen8_is_valid_flex_addr;
>
> - dev_priv->perf.oa.ops.enable_metric_set = gen8_enable_metric_set;
> - dev_priv->perf.oa.ops.disable_metric_set = gen10_disable_metric_set;
> + i915->perf.oa.ops.enable_metric_set = gen8_enable_metric_set;
> + i915->perf.oa.ops.disable_metric_set = gen10_disable_metric_set;
>
> - dev_priv->perf.oa.ctx_oactxctrl_offset = 0x128;
> - dev_priv->perf.oa.ctx_flexeu0_offset = 0x3de;
> + i915->perf.oa.ctx_oactxctrl_offset = 0x128;
> + i915->perf.oa.ctx_flexeu0_offset = 0x3de;
>
> - dev_priv->perf.oa.gen8_valid_ctx_bit = (1<<16);
> + i915->perf.oa.gen8_valid_ctx_bit = (1<<16);
> }
> }
>
> - if (dev_priv->perf.oa.ops.enable_metric_set) {
> - hrtimer_init(&dev_priv->perf.oa.poll_check_timer,
> + if (i915->perf.oa.ops.enable_metric_set) {
> + hrtimer_init(&i915->perf.oa.poll_check_timer,
> CLOCK_MONOTONIC, HRTIMER_MODE_REL);
> - dev_priv->perf.oa.poll_check_timer.function = oa_poll_check_timer_cb;
> - init_waitqueue_head(&dev_priv->perf.oa.poll_wq);
> + i915->perf.oa.poll_check_timer.function = oa_poll_check_timer_cb;
> + init_waitqueue_head(&i915->perf.oa.poll_wq);
>
> - INIT_LIST_HEAD(&dev_priv->perf.streams);
> - mutex_init(&dev_priv->perf.lock);
> - spin_lock_init(&dev_priv->perf.oa.oa_buffer.ptr_lock);
> + INIT_LIST_HEAD(&i915->perf.streams);
> + mutex_init(&i915->perf.lock);
> + spin_lock_init(&i915->perf.oa.oa_buffer.ptr_lock);
>
> oa_sample_rate_hard_limit = 1000 *
> - (RUNTIME_INFO(dev_priv)->cs_timestamp_frequency_khz / 2);
> - dev_priv->perf.sysctl_header = register_sysctl_table(dev_root);
> + (RUNTIME_INFO(i915)->cs_timestamp_frequency_khz / 2);
> + i915->perf.sysctl_header = register_sysctl_table(dev_root);
>
> - mutex_init(&dev_priv->perf.metrics_lock);
> - idr_init(&dev_priv->perf.metrics_idr);
> + mutex_init(&i915->perf.metrics_lock);
> + idr_init(&i915->perf.metrics_idr);
>
> - dev_priv->perf.initialized = true;
> + i915->perf.initialized = true;
> }
> }
>
> static int destroy_config(int id, void *p, void *data)
> {
> - struct drm_i915_private *dev_priv = data;
> + struct drm_i915_private *i915 = data;
> struct i915_oa_config *oa_config = p;
>
> - put_oa_config(dev_priv, oa_config);
> + put_oa_config(i915, oa_config);
>
> return 0;
> }
>
> /**
> * i915_perf_fini - Counter part to i915_perf_init()
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> */
> -void i915_perf_fini(struct drm_i915_private *dev_priv)
> +void i915_perf_fini(struct drm_i915_private *i915)
> {
> - if (!dev_priv->perf.initialized)
> + if (!i915->perf.initialized)
> return;
>
> - idr_for_each(&dev_priv->perf.metrics_idr, destroy_config, dev_priv);
> - idr_destroy(&dev_priv->perf.metrics_idr);
> + idr_for_each(&i915->perf.metrics_idr, destroy_config, i915);
> + idr_destroy(&i915->perf.metrics_idr);
>
> - unregister_sysctl_table(dev_priv->perf.sysctl_header);
> + unregister_sysctl_table(i915->perf.sysctl_header);
>
> - memset(&dev_priv->perf.oa.ops, 0, sizeof(dev_priv->perf.oa.ops));
> + memset(&i915->perf.oa.ops, 0, sizeof(i915->perf.oa.ops));
>
> - dev_priv->perf.initialized = false;
> + i915->perf.initialized = false;
> }
> diff --git a/drivers/gpu/drm/i915/i915_pmu.c b/drivers/gpu/drm/i915/i915_pmu.c
> index eb9c0e0e545c..c4861c12100c 100644
> --- a/drivers/gpu/drm/i915/i915_pmu.c
> +++ b/drivers/gpu/drm/i915/i915_pmu.c
> @@ -159,24 +159,24 @@ add_sample(struct i915_pmu_sample *sample, u32 val)
> }
>
> static void
> -engines_sample(struct drm_i915_private *dev_priv, unsigned int period_ns)
> +engines_sample(struct drm_i915_private *i915, unsigned int period_ns)
> {
> struct intel_engine_cs *engine;
> enum intel_engine_id id;
> intel_wakeref_t wakeref;
> unsigned long flags;
>
> - if ((dev_priv->pmu.enable & ENGINE_SAMPLE_MASK) == 0)
> + if ((i915->pmu.enable & ENGINE_SAMPLE_MASK) == 0)
> return;
>
> wakeref = 0;
> - if (READ_ONCE(dev_priv->gt.awake))
> - wakeref = intel_runtime_pm_get_if_in_use(dev_priv);
> + if (READ_ONCE(i915->gt.awake))
> + wakeref = intel_runtime_pm_get_if_in_use(i915);
> if (!wakeref)
> return;
>
> - spin_lock_irqsave(&dev_priv->uncore.lock, flags);
> - for_each_engine(engine, dev_priv, id) {
> + spin_lock_irqsave(&i915->uncore.lock, flags);
> + for_each_engine(engine, i915, id) {
> struct intel_engine_pmu *pmu = &engine->pmu;
> bool busy;
> u32 val;
> @@ -205,9 +205,9 @@ engines_sample(struct drm_i915_private *dev_priv, unsigned int period_ns)
> if (busy)
> add_sample(&pmu->sample[I915_SAMPLE_BUSY], period_ns);
> }
> - spin_unlock_irqrestore(&dev_priv->uncore.lock, flags);
> + spin_unlock_irqrestore(&i915->uncore.lock, flags);
>
> - intel_runtime_pm_put(dev_priv, wakeref);
> + intel_runtime_pm_put(i915, wakeref);
> }
>
> static void
> @@ -217,33 +217,33 @@ add_sample_mult(struct i915_pmu_sample *sample, u32 val, u32 mul)
> }
>
> static void
> -frequency_sample(struct drm_i915_private *dev_priv, unsigned int period_ns)
> +frequency_sample(struct drm_i915_private *i915, unsigned int period_ns)
> {
> - if (dev_priv->pmu.enable &
> + if (i915->pmu.enable &
> config_enabled_mask(I915_PMU_ACTUAL_FREQUENCY)) {
> u32 val;
>
> - val = dev_priv->gt_pm.rps.cur_freq;
> - if (dev_priv->gt.awake) {
> + val = i915->gt_pm.rps.cur_freq;
> + if (i915->gt.awake) {
> intel_wakeref_t wakeref;
>
> - with_intel_runtime_pm_if_in_use(dev_priv, wakeref) {
> - val = intel_uncore_read_notrace(&dev_priv->uncore,
> + with_intel_runtime_pm_if_in_use(i915, wakeref) {
> + val = intel_uncore_read_notrace(&i915->uncore,
> GEN6_RPSTAT1);
> - val = intel_get_cagf(dev_priv, val);
> + val = intel_get_cagf(i915, val);
> }
> }
>
> - add_sample_mult(&dev_priv->pmu.sample[__I915_SAMPLE_FREQ_ACT],
> - intel_gpu_freq(dev_priv, val),
> + add_sample_mult(&i915->pmu.sample[__I915_SAMPLE_FREQ_ACT],
> + intel_gpu_freq(i915, val),
> period_ns / 1000);
> }
>
> - if (dev_priv->pmu.enable &
> + if (i915->pmu.enable &
> config_enabled_mask(I915_PMU_REQUESTED_FREQUENCY)) {
> - add_sample_mult(&dev_priv->pmu.sample[__I915_SAMPLE_FREQ_REQ],
> - intel_gpu_freq(dev_priv,
> - dev_priv->gt_pm.rps.cur_freq),
> + add_sample_mult(&i915->pmu.sample[__I915_SAMPLE_FREQ_REQ],
> + intel_gpu_freq(i915,
> + i915->gt_pm.rps.cur_freq),
> period_ns / 1000);
> }
> }
> diff --git a/drivers/gpu/drm/i915/i915_query.c b/drivers/gpu/drm/i915/i915_query.c
> index 7b7016171057..87606362b56f 100644
> --- a/drivers/gpu/drm/i915/i915_query.c
> +++ b/drivers/gpu/drm/i915/i915_query.c
> @@ -31,10 +31,10 @@ static int copy_query_item(void *query_hdr, size_t query_sz,
> return 0;
> }
>
> -static int query_topology_info(struct drm_i915_private *dev_priv,
> +static int query_topology_info(struct drm_i915_private *i915,
> struct drm_i915_query_item *query_item)
> {
> - const struct sseu_dev_info *sseu = &RUNTIME_INFO(dev_priv)->sseu;
> + const struct sseu_dev_info *sseu = &RUNTIME_INFO(i915)->sseu;
> struct drm_i915_query_topology_info topo;
> u32 slice_length, subslice_length, eu_length, total_length;
> u8 subslice_stride = GEN_SSEU_STRIDE(sseu->max_subslices);
> @@ -143,7 +143,7 @@ query_engine_info(struct drm_i915_private *i915,
> return len;
> }
>
> -static int (* const i915_query_funcs[])(struct drm_i915_private *dev_priv,
> +static int (* const i915_query_funcs[])(struct drm_i915_private *i915,
> struct drm_i915_query_item *query_item) = {
> query_topology_info,
> query_engine_info,
> @@ -151,7 +151,7 @@ static int (* const i915_query_funcs[])(struct drm_i915_private *dev_priv,
>
> int i915_query_ioctl(struct drm_device *dev, void *data, struct drm_file *file)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct drm_i915_query *args = data;
> struct drm_i915_query_item __user *user_item_ptr =
> u64_to_user_ptr(args->items_ptr);
> @@ -180,7 +180,7 @@ int i915_query_ioctl(struct drm_device *dev, void *data, struct drm_file *file)
> if (func_idx < ARRAY_SIZE(i915_query_funcs)) {
> func_idx = array_index_nospec(func_idx,
> ARRAY_SIZE(i915_query_funcs));
> - ret = i915_query_funcs[func_idx](dev_priv, &item);
> + ret = i915_query_funcs[func_idx](i915, &item);
> }
>
> /* Only write the length back to userspace if they differ. */
> diff --git a/drivers/gpu/drm/i915/i915_reg.h b/drivers/gpu/drm/i915/i915_reg.h
> index edf9f93934a1..b522970bdf74 100644
> --- a/drivers/gpu/drm/i915/i915_reg.h
> +++ b/drivers/gpu/drm/i915/i915_reg.h
> @@ -205,7 +205,7 @@ static inline bool i915_mmio_reg_valid(i915_reg_t reg)
> #define VLV_MIPI_BASE VLV_DISPLAY_BASE
> #define BXT_MIPI_BASE 0x60000
>
> -#define DISPLAY_MMIO_BASE(dev_priv) (INTEL_INFO(dev_priv)->display_mmio_offset)
> +#define DISPLAY_MMIO_BASE(i915) (INTEL_INFO(i915)->display_mmio_offset)
>
> /*
> * Given the first two numbers __a and __b of arbitrarily many evenly spaced
> @@ -247,15 +247,15 @@ static inline bool i915_mmio_reg_valid(i915_reg_t reg)
> * Device info offset array based helpers for groups of registers with unevenly
> * spaced base offsets.
> */
> -#define _MMIO_PIPE2(pipe, reg) _MMIO(INTEL_INFO(dev_priv)->pipe_offsets[pipe] - \
> - INTEL_INFO(dev_priv)->pipe_offsets[PIPE_A] + (reg) + \
> - DISPLAY_MMIO_BASE(dev_priv))
> -#define _MMIO_TRANS2(pipe, reg) _MMIO(INTEL_INFO(dev_priv)->trans_offsets[(pipe)] - \
> - INTEL_INFO(dev_priv)->trans_offsets[TRANSCODER_A] + (reg) + \
> - DISPLAY_MMIO_BASE(dev_priv))
> -#define _CURSOR2(pipe, reg) _MMIO(INTEL_INFO(dev_priv)->cursor_offsets[(pipe)] - \
> - INTEL_INFO(dev_priv)->cursor_offsets[PIPE_A] + (reg) + \
> - DISPLAY_MMIO_BASE(dev_priv))
> +#define _MMIO_PIPE2(pipe, reg) _MMIO(INTEL_INFO(i915)->pipe_offsets[pipe] - \
> + INTEL_INFO(i915)->pipe_offsets[PIPE_A] + (reg) + \
> + DISPLAY_MMIO_BASE(i915))
> +#define _MMIO_TRANS2(pipe, reg) _MMIO(INTEL_INFO(i915)->trans_offsets[(pipe)] - \
> + INTEL_INFO(i915)->trans_offsets[TRANSCODER_A] + (reg) + \
> + DISPLAY_MMIO_BASE(i915))
> +#define _CURSOR2(pipe, reg) _MMIO(INTEL_INFO(i915)->cursor_offsets[(pipe)] - \
> + INTEL_INFO(i915)->cursor_offsets[PIPE_A] + (reg) + \
> + DISPLAY_MMIO_BASE(i915))
>
> #define __MASKED_FIELD(mask, value) ((mask) << 16 | (value))
> #define _MASKED_FIELD(mask, value) ({ \
> @@ -1307,7 +1307,7 @@ enum i915_power_well_id {
> #define DPIO_CMNRST (1 << 0)
>
> #define DPIO_PHY(pipe) ((pipe) >> 1)
> -#define DPIO_PHY_IOSF_PORT(phy) (dev_priv->dpio_phy_iosf_port[phy])
> +#define DPIO_PHY_IOSF_PORT(phy) (i915->dpio_phy_iosf_port[phy])
>
> /*
> * Per pipe/PLL DPIO regs
> @@ -3004,9 +3004,9 @@ enum i915_power_well_id {
> #define PM_VEBOX_CS_ERROR_INTERRUPT (1 << 12) /* hsw+ */
> #define PM_VEBOX_USER_INTERRUPT (1 << 10) /* hsw+ */
>
> -#define GT_PARITY_ERROR(dev_priv) \
> +#define GT_PARITY_ERROR(i915) \
> (GT_RENDER_L3_PARITY_ERROR_INTERRUPT | \
> - (IS_HASWELL(dev_priv) ? GT_RENDER_L3_PARITY_ERROR_INTERRUPT_S1 : 0))
> + (IS_HASWELL(i915) ? GT_RENDER_L3_PARITY_ERROR_INTERRUPT_S1 : 0))
>
> /* These are all the "old" interrupts */
> #define ILK_BSD_USER_INTERRUPT (1 << 5)
> @@ -3195,7 +3195,7 @@ enum i915_power_well_id {
> /*
> * GPIO regs
> */
> -#define GPIO(gpio) _MMIO(dev_priv->gpio_mmio_base + 0x5010 + \
> +#define GPIO(gpio) _MMIO(i915->gpio_mmio_base + 0x5010 + \
> 4 * (gpio))
>
> # define GPIO_CLOCK_DIR_MASK (1 << 0)
> @@ -3213,7 +3213,7 @@ enum i915_power_well_id {
> # define GPIO_DATA_VAL_IN (1 << 12)
> # define GPIO_DATA_PULLUP_DISABLE (1 << 13)
>
> -#define GMBUS0 _MMIO(dev_priv->gpio_mmio_base + 0x5100) /* clock/port select */
> +#define GMBUS0 _MMIO(i915->gpio_mmio_base + 0x5100) /* clock/port select */
> #define GMBUS_AKSV_SELECT (1 << 11)
> #define GMBUS_RATE_100KHZ (0 << 8)
> #define GMBUS_RATE_50KHZ (1 << 8)
> @@ -3240,7 +3240,7 @@ enum i915_power_well_id {
> #define GMBUS_PIN_12_TC4_ICP 12
>
> #define GMBUS_NUM_PINS 13 /* including 0 */
> -#define GMBUS1 _MMIO(dev_priv->gpio_mmio_base + 0x5104) /* command/status */
> +#define GMBUS1 _MMIO(i915->gpio_mmio_base + 0x5104) /* command/status */
> #define GMBUS_SW_CLR_INT (1 << 31)
> #define GMBUS_SW_RDY (1 << 30)
> #define GMBUS_ENT (1 << 29) /* enable timeout */
> @@ -3255,7 +3255,7 @@ enum i915_power_well_id {
> #define GMBUS_SLAVE_ADDR_SHIFT 1
> #define GMBUS_SLAVE_READ (1 << 0)
> #define GMBUS_SLAVE_WRITE (0 << 0)
> -#define GMBUS2 _MMIO(dev_priv->gpio_mmio_base + 0x5108) /* status */
> +#define GMBUS2 _MMIO(i915->gpio_mmio_base + 0x5108) /* status */
> #define GMBUS_INUSE (1 << 15)
> #define GMBUS_HW_WAIT_PHASE (1 << 14)
> #define GMBUS_STALL_TIMEOUT (1 << 13)
> @@ -3263,22 +3263,22 @@ enum i915_power_well_id {
> #define GMBUS_HW_RDY (1 << 11)
> #define GMBUS_SATOER (1 << 10)
> #define GMBUS_ACTIVE (1 << 9)
> -#define GMBUS3 _MMIO(dev_priv->gpio_mmio_base + 0x510c) /* data buffer bytes 3-0 */
> -#define GMBUS4 _MMIO(dev_priv->gpio_mmio_base + 0x5110) /* interrupt mask (Pineview+) */
> +#define GMBUS3 _MMIO(i915->gpio_mmio_base + 0x510c) /* data buffer bytes 3-0 */
> +#define GMBUS4 _MMIO(i915->gpio_mmio_base + 0x5110) /* interrupt mask (Pineview+) */
> #define GMBUS_SLAVE_TIMEOUT_EN (1 << 4)
> #define GMBUS_NAK_EN (1 << 3)
> #define GMBUS_IDLE_EN (1 << 2)
> #define GMBUS_HW_WAIT_EN (1 << 1)
> #define GMBUS_HW_RDY_EN (1 << 0)
> -#define GMBUS5 _MMIO(dev_priv->gpio_mmio_base + 0x5120) /* byte index */
> +#define GMBUS5 _MMIO(i915->gpio_mmio_base + 0x5120) /* byte index */
> #define GMBUS_2BYTE_INDEX_EN (1 << 31)
>
> /*
> * Clock control & power management
> */
> -#define _DPLL_A (DISPLAY_MMIO_BASE(dev_priv) + 0x6014)
> -#define _DPLL_B (DISPLAY_MMIO_BASE(dev_priv) + 0x6018)
> -#define _CHV_DPLL_C (DISPLAY_MMIO_BASE(dev_priv) + 0x6030)
> +#define _DPLL_A (DISPLAY_MMIO_BASE(i915) + 0x6014)
> +#define _DPLL_B (DISPLAY_MMIO_BASE(i915) + 0x6018)
> +#define _CHV_DPLL_C (DISPLAY_MMIO_BASE(i915) + 0x6030)
> #define DPLL(pipe) _MMIO_PIPE3((pipe), _DPLL_A, _DPLL_B, _CHV_DPLL_C)
>
> #define VGA0 _MMIO(0x6000)
> @@ -3375,9 +3375,9 @@ enum i915_power_well_id {
> #define SDVO_MULTIPLIER_SHIFT_HIRES 4
> #define SDVO_MULTIPLIER_SHIFT_VGA 0
>
> -#define _DPLL_A_MD (DISPLAY_MMIO_BASE(dev_priv) + 0x601c)
> -#define _DPLL_B_MD (DISPLAY_MMIO_BASE(dev_priv) + 0x6020)
> -#define _CHV_DPLL_C_MD (DISPLAY_MMIO_BASE(dev_priv) + 0x603c)
> +#define _DPLL_A_MD (DISPLAY_MMIO_BASE(i915) + 0x601c)
> +#define _DPLL_B_MD (DISPLAY_MMIO_BASE(i915) + 0x6020)
> +#define _CHV_DPLL_C_MD (DISPLAY_MMIO_BASE(i915) + 0x603c)
> #define DPLL_MD(pipe) _MMIO_PIPE3((pipe), _DPLL_A_MD, _DPLL_B_MD, _CHV_DPLL_C_MD)
>
> /*
> @@ -3449,7 +3449,7 @@ enum i915_power_well_id {
> #define DSTATE_PLL_D3_OFF (1 << 3)
> #define DSTATE_GFX_CLOCK_GATING (1 << 1)
> #define DSTATE_DOT_CLOCK_GATING (1 << 0)
> -#define DSPCLK_GATE_D _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x6200)
> +#define DSPCLK_GATE_D _MMIO(DISPLAY_MMIO_BASE(i915) + 0x6200)
> # define DPUNIT_B_CLOCK_GATE_DISABLE (1 << 30) /* 965 */
> # define VSUNIT_CLOCK_GATE_DISABLE (1 << 29) /* 965 */
> # define VRHUNIT_CLOCK_GATE_DISABLE (1 << 28) /* 965 */
> @@ -3589,7 +3589,7 @@ enum i915_power_well_id {
> #define _PALETTE_A 0xa000
> #define _PALETTE_B 0xa800
> #define _CHV_PALETTE_C 0xc000
> -#define PALETTE(pipe, i) _MMIO(DISPLAY_MMIO_BASE(dev_priv) + \
> +#define PALETTE(pipe, i) _MMIO(DISPLAY_MMIO_BASE(i915) + \
> _PICK((pipe), _PALETTE_A, \
> _PALETTE_B, _CHV_PALETTE_C) + \
> (i) * 4)
> @@ -3936,8 +3936,8 @@ enum i915_power_well_id {
> #define INTERVAL_1_28_US(us) roundup(((us) * 100) >> 7, 25)
> #define INTERVAL_1_33_US(us) (((us) * 3) >> 2)
> #define INTERVAL_0_833_US(us) (((us) * 6) / 5)
> -#define GT_INTERVAL_FROM_US(dev_priv, us) (INTEL_GEN(dev_priv) >= 9 ? \
> - (IS_GEN9_LP(dev_priv) ? \
> +#define GT_INTERVAL_FROM_US(i915, us) (INTEL_GEN(i915) >= 9 ? \
> + (IS_GEN9_LP(i915) ? \
> INTERVAL_0_833_US(us) : \
> INTERVAL_1_33_US(us)) : \
> INTERVAL_1_28_US(us))
> @@ -3945,8 +3945,8 @@ enum i915_power_well_id {
> #define INTERVAL_1_28_TO_US(interval) (((interval) << 7) / 100)
> #define INTERVAL_1_33_TO_US(interval) (((interval) << 2) / 3)
> #define INTERVAL_0_833_TO_US(interval) (((interval) * 5) / 6)
> -#define GT_PM_INTERVAL_TO_US(dev_priv, interval) (INTEL_GEN(dev_priv) >= 9 ? \
> - (IS_GEN9_LP(dev_priv) ? \
> +#define GT_PM_INTERVAL_TO_US(i915, interval) (INTEL_GEN(i915) >= 9 ? \
> + (IS_GEN9_LP(i915) ? \
> INTERVAL_0_833_TO_US(interval) : \
> INTERVAL_1_33_TO_US(interval)) : \
> INTERVAL_1_28_TO_US(interval))
> @@ -4219,7 +4219,7 @@ enum {
> /* HSW+ eDP PSR registers */
> #define HSW_EDP_PSR_BASE 0x64800
> #define BDW_EDP_PSR_BASE 0x6f800
> -#define EDP_PSR_CTL _MMIO(dev_priv->psr_mmio_base + 0)
> +#define EDP_PSR_CTL _MMIO(i915->psr_mmio_base + 0)
> #define EDP_PSR_ENABLE (1 << 31)
> #define BDW_PSR_SINGLE_FRAME (1 << 30)
> #define EDP_PSR_RESTORE_PSR_ACTIVE_CTX_MASK (1 << 29) /* SW can't modify */
> @@ -4256,16 +4256,16 @@ enum {
> #define EDP_PSR_TRANSCODER_A_SHIFT 8
> #define EDP_PSR_TRANSCODER_EDP_SHIFT 0
>
> -#define EDP_PSR_AUX_CTL _MMIO(dev_priv->psr_mmio_base + 0x10)
> +#define EDP_PSR_AUX_CTL _MMIO(i915->psr_mmio_base + 0x10)
> #define EDP_PSR_AUX_CTL_TIME_OUT_MASK (3 << 26)
> #define EDP_PSR_AUX_CTL_MESSAGE_SIZE_MASK (0x1f << 20)
> #define EDP_PSR_AUX_CTL_PRECHARGE_2US_MASK (0xf << 16)
> #define EDP_PSR_AUX_CTL_ERROR_INTERRUPT (1 << 11)
> #define EDP_PSR_AUX_CTL_BIT_CLOCK_2X_MASK (0x7ff)
>
> -#define EDP_PSR_AUX_DATA(i) _MMIO(dev_priv->psr_mmio_base + 0x14 + (i) * 4) /* 5 registers */
> +#define EDP_PSR_AUX_DATA(i) _MMIO(i915->psr_mmio_base + 0x14 + (i) * 4) /* 5 registers */
>
> -#define EDP_PSR_STATUS _MMIO(dev_priv->psr_mmio_base + 0x40)
> +#define EDP_PSR_STATUS _MMIO(i915->psr_mmio_base + 0x40)
> #define EDP_PSR_STATUS_STATE_MASK (7 << 29)
> #define EDP_PSR_STATUS_STATE_SHIFT 29
> #define EDP_PSR_STATUS_STATE_IDLE (0 << 29)
> @@ -4290,10 +4290,10 @@ enum {
> #define EDP_PSR_STATUS_SENDING_TP1 (1 << 4)
> #define EDP_PSR_STATUS_IDLE_MASK 0xf
>
> -#define EDP_PSR_PERF_CNT _MMIO(dev_priv->psr_mmio_base + 0x44)
> +#define EDP_PSR_PERF_CNT _MMIO(i915->psr_mmio_base + 0x44)
> #define EDP_PSR_PERF_CNT_MASK 0xffffff
>
> -#define EDP_PSR_DEBUG _MMIO(dev_priv->psr_mmio_base + 0x60) /* PSR_MASK on SKL+ */
> +#define EDP_PSR_DEBUG _MMIO(i915->psr_mmio_base + 0x60) /* PSR_MASK on SKL+ */
> #define EDP_PSR_DEBUG_MASK_MAX_SLEEP (1 << 28)
> #define EDP_PSR_DEBUG_MASK_LPSP (1 << 27)
> #define EDP_PSR_DEBUG_MASK_MEMUP (1 << 26)
> @@ -4405,7 +4405,7 @@ enum {
>
>
> /* Hotplug control (945+ only) */
> -#define PORT_HOTPLUG_EN _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x61110)
> +#define PORT_HOTPLUG_EN _MMIO(DISPLAY_MMIO_BASE(i915) + 0x61110)
> #define PORTB_HOTPLUG_INT_EN (1 << 29)
> #define PORTC_HOTPLUG_INT_EN (1 << 28)
> #define PORTD_HOTPLUG_INT_EN (1 << 27)
> @@ -4435,7 +4435,7 @@ enum {
> #define CRT_HOTPLUG_DETECT_VOLTAGE_325MV (0 << 2)
> #define CRT_HOTPLUG_DETECT_VOLTAGE_475MV (1 << 2)
>
> -#define PORT_HOTPLUG_STAT _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x61114)
> +#define PORT_HOTPLUG_STAT _MMIO(DISPLAY_MMIO_BASE(i915) + 0x61114)
> /*
> * HDMI/DP bits are g4x+
> *
> @@ -4517,7 +4517,7 @@ enum {
>
> #define PORT_DFT_I9XX _MMIO(0x61150)
> #define DC_BALANCE_RESET (1 << 25)
> -#define PORT_DFT2_G4X _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x61154)
> +#define PORT_DFT2_G4X _MMIO(DISPLAY_MMIO_BASE(i915) + 0x61154)
> #define DC_BALANCE_RESET_VLV (1 << 31)
> #define PIPE_SCRAMBLE_RESET_MASK ((1 << 14) | (0x3 << 0))
> #define PIPE_C_SCRAMBLE_RESET (1 << 14) /* chv */
> @@ -4719,7 +4719,7 @@ enum {
> #define VLV_PPS_BASE (VLV_DISPLAY_BASE + PPS_BASE)
> #define PCH_PPS_BASE 0xC7200
>
> -#define _MMIO_PPS(pps_idx, reg) _MMIO(dev_priv->pps_mmio_base - \
> +#define _MMIO_PPS(pps_idx, reg) _MMIO(i915->pps_mmio_base - \
> PPS_BASE + (reg) + \
> (pps_idx) * 0x100)
>
> @@ -4792,7 +4792,7 @@ enum {
> #define PANEL_POWER_CYCLE_DELAY_MASK REG_GENMASK(4, 0)
>
> /* Panel fitting */
> -#define PFIT_CONTROL _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x61230)
> +#define PFIT_CONTROL _MMIO(DISPLAY_MMIO_BASE(i915) + 0x61230)
> #define PFIT_ENABLE (1 << 31)
> #define PFIT_PIPE_MASK (3 << 29)
> #define PFIT_PIPE_SHIFT 29
> @@ -4810,7 +4810,7 @@ enum {
> #define PFIT_SCALING_PROGRAMMED (1 << 26)
> #define PFIT_SCALING_PILLAR (2 << 26)
> #define PFIT_SCALING_LETTER (3 << 26)
> -#define PFIT_PGM_RATIOS _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x61234)
> +#define PFIT_PGM_RATIOS _MMIO(DISPLAY_MMIO_BASE(i915) + 0x61234)
> /* Pre-965 */
> #define PFIT_VERT_SCALE_SHIFT 20
> #define PFIT_VERT_SCALE_MASK 0xfff00000
> @@ -4822,25 +4822,25 @@ enum {
> #define PFIT_HORIZ_SCALE_SHIFT_965 0
> #define PFIT_HORIZ_SCALE_MASK_965 0x00001fff
>
> -#define PFIT_AUTO_RATIOS _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x61238)
> +#define PFIT_AUTO_RATIOS _MMIO(DISPLAY_MMIO_BASE(i915) + 0x61238)
>
> -#define _VLV_BLC_PWM_CTL2_A (DISPLAY_MMIO_BASE(dev_priv) + 0x61250)
> -#define _VLV_BLC_PWM_CTL2_B (DISPLAY_MMIO_BASE(dev_priv) + 0x61350)
> +#define _VLV_BLC_PWM_CTL2_A (DISPLAY_MMIO_BASE(i915) + 0x61250)
> +#define _VLV_BLC_PWM_CTL2_B (DISPLAY_MMIO_BASE(i915) + 0x61350)
> #define VLV_BLC_PWM_CTL2(pipe) _MMIO_PIPE(pipe, _VLV_BLC_PWM_CTL2_A, \
> _VLV_BLC_PWM_CTL2_B)
>
> -#define _VLV_BLC_PWM_CTL_A (DISPLAY_MMIO_BASE(dev_priv) + 0x61254)
> -#define _VLV_BLC_PWM_CTL_B (DISPLAY_MMIO_BASE(dev_priv) + 0x61354)
> +#define _VLV_BLC_PWM_CTL_A (DISPLAY_MMIO_BASE(i915) + 0x61254)
> +#define _VLV_BLC_PWM_CTL_B (DISPLAY_MMIO_BASE(i915) + 0x61354)
> #define VLV_BLC_PWM_CTL(pipe) _MMIO_PIPE(pipe, _VLV_BLC_PWM_CTL_A, \
> _VLV_BLC_PWM_CTL_B)
>
> -#define _VLV_BLC_HIST_CTL_A (DISPLAY_MMIO_BASE(dev_priv) + 0x61260)
> -#define _VLV_BLC_HIST_CTL_B (DISPLAY_MMIO_BASE(dev_priv) + 0x61360)
> +#define _VLV_BLC_HIST_CTL_A (DISPLAY_MMIO_BASE(i915) + 0x61260)
> +#define _VLV_BLC_HIST_CTL_B (DISPLAY_MMIO_BASE(i915) + 0x61360)
> #define VLV_BLC_HIST_CTL(pipe) _MMIO_PIPE(pipe, _VLV_BLC_HIST_CTL_A, \
> _VLV_BLC_HIST_CTL_B)
>
> /* Backlight control */
> -#define BLC_PWM_CTL2 _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x61250) /* 965+ only */
> +#define BLC_PWM_CTL2 _MMIO(DISPLAY_MMIO_BASE(i915) + 0x61250) /* 965+ only */
> #define BLM_PWM_ENABLE (1 << 31)
> #define BLM_COMBINATION_MODE (1 << 30) /* gen4 only */
> #define BLM_PIPE_SELECT (1 << 29)
> @@ -4863,7 +4863,7 @@ enum {
> #define BLM_PHASE_IN_COUNT_MASK (0xff << 8)
> #define BLM_PHASE_IN_INCR_SHIFT (0)
> #define BLM_PHASE_IN_INCR_MASK (0xff << 0)
> -#define BLC_PWM_CTL _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x61254)
> +#define BLC_PWM_CTL _MMIO(DISPLAY_MMIO_BASE(i915) + 0x61254)
> /*
> * This is the most significant 15 bits of the number of backlight cycles in a
> * complete cycle of the modulated backlight control.
> @@ -4885,7 +4885,7 @@ enum {
> #define BACKLIGHT_DUTY_CYCLE_MASK_PNV (0xfffe)
> #define BLM_POLARITY_PNV (1 << 0) /* pnv only */
>
> -#define BLC_HIST_CTL _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x61260)
> +#define BLC_HIST_CTL _MMIO(DISPLAY_MMIO_BASE(i915) + 0x61260)
> #define BLM_HISTOGRAM_ENABLE (1 << 31)
>
> /* New registers for PCH-split platforms. Safe where new bits show up, the
> @@ -5510,47 +5510,47 @@ enum {
> * is 20 bytes in each direction, hence the 5 fixed
> * data registers
> */
> -#define _DPA_AUX_CH_CTL (DISPLAY_MMIO_BASE(dev_priv) + 0x64010)
> -#define _DPA_AUX_CH_DATA1 (DISPLAY_MMIO_BASE(dev_priv) + 0x64014)
> -#define _DPA_AUX_CH_DATA2 (DISPLAY_MMIO_BASE(dev_priv) + 0x64018)
> -#define _DPA_AUX_CH_DATA3 (DISPLAY_MMIO_BASE(dev_priv) + 0x6401c)
> -#define _DPA_AUX_CH_DATA4 (DISPLAY_MMIO_BASE(dev_priv) + 0x64020)
> -#define _DPA_AUX_CH_DATA5 (DISPLAY_MMIO_BASE(dev_priv) + 0x64024)
> -
> -#define _DPB_AUX_CH_CTL (DISPLAY_MMIO_BASE(dev_priv) + 0x64110)
> -#define _DPB_AUX_CH_DATA1 (DISPLAY_MMIO_BASE(dev_priv) + 0x64114)
> -#define _DPB_AUX_CH_DATA2 (DISPLAY_MMIO_BASE(dev_priv) + 0x64118)
> -#define _DPB_AUX_CH_DATA3 (DISPLAY_MMIO_BASE(dev_priv) + 0x6411c)
> -#define _DPB_AUX_CH_DATA4 (DISPLAY_MMIO_BASE(dev_priv) + 0x64120)
> -#define _DPB_AUX_CH_DATA5 (DISPLAY_MMIO_BASE(dev_priv) + 0x64124)
> -
> -#define _DPC_AUX_CH_CTL (DISPLAY_MMIO_BASE(dev_priv) + 0x64210)
> -#define _DPC_AUX_CH_DATA1 (DISPLAY_MMIO_BASE(dev_priv) + 0x64214)
> -#define _DPC_AUX_CH_DATA2 (DISPLAY_MMIO_BASE(dev_priv) + 0x64218)
> -#define _DPC_AUX_CH_DATA3 (DISPLAY_MMIO_BASE(dev_priv) + 0x6421c)
> -#define _DPC_AUX_CH_DATA4 (DISPLAY_MMIO_BASE(dev_priv) + 0x64220)
> -#define _DPC_AUX_CH_DATA5 (DISPLAY_MMIO_BASE(dev_priv) + 0x64224)
> -
> -#define _DPD_AUX_CH_CTL (DISPLAY_MMIO_BASE(dev_priv) + 0x64310)
> -#define _DPD_AUX_CH_DATA1 (DISPLAY_MMIO_BASE(dev_priv) + 0x64314)
> -#define _DPD_AUX_CH_DATA2 (DISPLAY_MMIO_BASE(dev_priv) + 0x64318)
> -#define _DPD_AUX_CH_DATA3 (DISPLAY_MMIO_BASE(dev_priv) + 0x6431c)
> -#define _DPD_AUX_CH_DATA4 (DISPLAY_MMIO_BASE(dev_priv) + 0x64320)
> -#define _DPD_AUX_CH_DATA5 (DISPLAY_MMIO_BASE(dev_priv) + 0x64324)
> -
> -#define _DPE_AUX_CH_CTL (DISPLAY_MMIO_BASE(dev_priv) + 0x64410)
> -#define _DPE_AUX_CH_DATA1 (DISPLAY_MMIO_BASE(dev_priv) + 0x64414)
> -#define _DPE_AUX_CH_DATA2 (DISPLAY_MMIO_BASE(dev_priv) + 0x64418)
> -#define _DPE_AUX_CH_DATA3 (DISPLAY_MMIO_BASE(dev_priv) + 0x6441c)
> -#define _DPE_AUX_CH_DATA4 (DISPLAY_MMIO_BASE(dev_priv) + 0x64420)
> -#define _DPE_AUX_CH_DATA5 (DISPLAY_MMIO_BASE(dev_priv) + 0x64424)
> -
> -#define _DPF_AUX_CH_CTL (DISPLAY_MMIO_BASE(dev_priv) + 0x64510)
> -#define _DPF_AUX_CH_DATA1 (DISPLAY_MMIO_BASE(dev_priv) + 0x64514)
> -#define _DPF_AUX_CH_DATA2 (DISPLAY_MMIO_BASE(dev_priv) + 0x64518)
> -#define _DPF_AUX_CH_DATA3 (DISPLAY_MMIO_BASE(dev_priv) + 0x6451c)
> -#define _DPF_AUX_CH_DATA4 (DISPLAY_MMIO_BASE(dev_priv) + 0x64520)
> -#define _DPF_AUX_CH_DATA5 (DISPLAY_MMIO_BASE(dev_priv) + 0x64524)
> +#define _DPA_AUX_CH_CTL (DISPLAY_MMIO_BASE(i915) + 0x64010)
> +#define _DPA_AUX_CH_DATA1 (DISPLAY_MMIO_BASE(i915) + 0x64014)
> +#define _DPA_AUX_CH_DATA2 (DISPLAY_MMIO_BASE(i915) + 0x64018)
> +#define _DPA_AUX_CH_DATA3 (DISPLAY_MMIO_BASE(i915) + 0x6401c)
> +#define _DPA_AUX_CH_DATA4 (DISPLAY_MMIO_BASE(i915) + 0x64020)
> +#define _DPA_AUX_CH_DATA5 (DISPLAY_MMIO_BASE(i915) + 0x64024)
> +
> +#define _DPB_AUX_CH_CTL (DISPLAY_MMIO_BASE(i915) + 0x64110)
> +#define _DPB_AUX_CH_DATA1 (DISPLAY_MMIO_BASE(i915) + 0x64114)
> +#define _DPB_AUX_CH_DATA2 (DISPLAY_MMIO_BASE(i915) + 0x64118)
> +#define _DPB_AUX_CH_DATA3 (DISPLAY_MMIO_BASE(i915) + 0x6411c)
> +#define _DPB_AUX_CH_DATA4 (DISPLAY_MMIO_BASE(i915) + 0x64120)
> +#define _DPB_AUX_CH_DATA5 (DISPLAY_MMIO_BASE(i915) + 0x64124)
> +
> +#define _DPC_AUX_CH_CTL (DISPLAY_MMIO_BASE(i915) + 0x64210)
> +#define _DPC_AUX_CH_DATA1 (DISPLAY_MMIO_BASE(i915) + 0x64214)
> +#define _DPC_AUX_CH_DATA2 (DISPLAY_MMIO_BASE(i915) + 0x64218)
> +#define _DPC_AUX_CH_DATA3 (DISPLAY_MMIO_BASE(i915) + 0x6421c)
> +#define _DPC_AUX_CH_DATA4 (DISPLAY_MMIO_BASE(i915) + 0x64220)
> +#define _DPC_AUX_CH_DATA5 (DISPLAY_MMIO_BASE(i915) + 0x64224)
> +
> +#define _DPD_AUX_CH_CTL (DISPLAY_MMIO_BASE(i915) + 0x64310)
> +#define _DPD_AUX_CH_DATA1 (DISPLAY_MMIO_BASE(i915) + 0x64314)
> +#define _DPD_AUX_CH_DATA2 (DISPLAY_MMIO_BASE(i915) + 0x64318)
> +#define _DPD_AUX_CH_DATA3 (DISPLAY_MMIO_BASE(i915) + 0x6431c)
> +#define _DPD_AUX_CH_DATA4 (DISPLAY_MMIO_BASE(i915) + 0x64320)
> +#define _DPD_AUX_CH_DATA5 (DISPLAY_MMIO_BASE(i915) + 0x64324)
> +
> +#define _DPE_AUX_CH_CTL (DISPLAY_MMIO_BASE(i915) + 0x64410)
> +#define _DPE_AUX_CH_DATA1 (DISPLAY_MMIO_BASE(i915) + 0x64414)
> +#define _DPE_AUX_CH_DATA2 (DISPLAY_MMIO_BASE(i915) + 0x64418)
> +#define _DPE_AUX_CH_DATA3 (DISPLAY_MMIO_BASE(i915) + 0x6441c)
> +#define _DPE_AUX_CH_DATA4 (DISPLAY_MMIO_BASE(i915) + 0x64420)
> +#define _DPE_AUX_CH_DATA5 (DISPLAY_MMIO_BASE(i915) + 0x64424)
> +
> +#define _DPF_AUX_CH_CTL (DISPLAY_MMIO_BASE(i915) + 0x64510)
> +#define _DPF_AUX_CH_DATA1 (DISPLAY_MMIO_BASE(i915) + 0x64514)
> +#define _DPF_AUX_CH_DATA2 (DISPLAY_MMIO_BASE(i915) + 0x64518)
> +#define _DPF_AUX_CH_DATA3 (DISPLAY_MMIO_BASE(i915) + 0x6451c)
> +#define _DPF_AUX_CH_DATA4 (DISPLAY_MMIO_BASE(i915) + 0x64520)
> +#define _DPF_AUX_CH_DATA5 (DISPLAY_MMIO_BASE(i915) + 0x64524)
>
> #define DP_AUX_CH_CTL(aux_ch) _MMIO_PORT(aux_ch, _DPA_AUX_CH_CTL, _DPB_AUX_CH_CTL)
> #define DP_AUX_CH_DATA(aux_ch, i) _MMIO(_PORT(aux_ch, _DPA_AUX_CH_DATA1, _DPB_AUX_CH_DATA1) + (i) * 4) /* 5 registers */
> @@ -5843,7 +5843,7 @@ enum {
> #define DPINVGTT_STATUS_MASK 0xff
> #define DPINVGTT_STATUS_MASK_CHV 0xfff
>
> -#define DSPARB _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x70030)
> +#define DSPARB _MMIO(DISPLAY_MMIO_BASE(i915) + 0x70030)
> #define DSPARB_CSTART_MASK (0x7f << 7)
> #define DSPARB_CSTART_SHIFT 7
> #define DSPARB_BSTART_MASK (0x7f)
> @@ -5878,7 +5878,7 @@ enum {
> #define DSPARB_SPRITEF_MASK_VLV (0xff << 8)
>
> /* pnv/gen4/g4x/vlv/chv */
> -#define DSPFW1 _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x70034)
> +#define DSPFW1 _MMIO(DISPLAY_MMIO_BASE(i915) + 0x70034)
> #define DSPFW_SR_SHIFT 23
> #define DSPFW_SR_MASK (0x1ff << 23)
> #define DSPFW_CURSORB_SHIFT 16
> @@ -5889,7 +5889,7 @@ enum {
> #define DSPFW_PLANEA_SHIFT 0
> #define DSPFW_PLANEA_MASK (0x7f << 0)
> #define DSPFW_PLANEA_MASK_VLV (0xff << 0) /* vlv/chv */
> -#define DSPFW2 _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x70038)
> +#define DSPFW2 _MMIO(DISPLAY_MMIO_BASE(i915) + 0x70038)
> #define DSPFW_FBC_SR_EN (1 << 31) /* g4x */
> #define DSPFW_FBC_SR_SHIFT 28
> #define DSPFW_FBC_SR_MASK (0x7 << 28) /* g4x */
> @@ -5905,7 +5905,7 @@ enum {
> #define DSPFW_SPRITEA_SHIFT 0
> #define DSPFW_SPRITEA_MASK (0x7f << 0) /* g4x */
> #define DSPFW_SPRITEA_MASK_VLV (0xff << 0) /* vlv/chv */
> -#define DSPFW3 _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x7003c)
> +#define DSPFW3 _MMIO(DISPLAY_MMIO_BASE(i915) + 0x7003c)
> #define DSPFW_HPLL_SR_EN (1 << 31)
> #define PINEVIEW_SELF_REFRESH_EN (1 << 30)
> #define DSPFW_CURSOR_SR_SHIFT 24
> @@ -6322,35 +6322,35 @@ enum {
> * [10:1f] all
> * [30:32] all
> */
> -#define SWF0(i) _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x70410 + (i) * 4)
> -#define SWF1(i) _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x71410 + (i) * 4)
> -#define SWF3(i) _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x72414 + (i) * 4)
> +#define SWF0(i) _MMIO(DISPLAY_MMIO_BASE(i915) + 0x70410 + (i) * 4)
> +#define SWF1(i) _MMIO(DISPLAY_MMIO_BASE(i915) + 0x71410 + (i) * 4)
> +#define SWF3(i) _MMIO(DISPLAY_MMIO_BASE(i915) + 0x72414 + (i) * 4)
> #define SWF_ILK(i) _MMIO(0x4F000 + (i) * 4)
>
> /* Pipe B */
> -#define _PIPEBDSL (DISPLAY_MMIO_BASE(dev_priv) + 0x71000)
> -#define _PIPEBCONF (DISPLAY_MMIO_BASE(dev_priv) + 0x71008)
> -#define _PIPEBSTAT (DISPLAY_MMIO_BASE(dev_priv) + 0x71024)
> +#define _PIPEBDSL (DISPLAY_MMIO_BASE(i915) + 0x71000)
> +#define _PIPEBCONF (DISPLAY_MMIO_BASE(i915) + 0x71008)
> +#define _PIPEBSTAT (DISPLAY_MMIO_BASE(i915) + 0x71024)
> #define _PIPEBFRAMEHIGH 0x71040
> #define _PIPEBFRAMEPIXEL 0x71044
> -#define _PIPEB_FRMCOUNT_G4X (DISPLAY_MMIO_BASE(dev_priv) + 0x71040)
> -#define _PIPEB_FLIPCOUNT_G4X (DISPLAY_MMIO_BASE(dev_priv) + 0x71044)
> +#define _PIPEB_FRMCOUNT_G4X (DISPLAY_MMIO_BASE(i915) + 0x71040)
> +#define _PIPEB_FLIPCOUNT_G4X (DISPLAY_MMIO_BASE(i915) + 0x71044)
>
>
> /* Display B control */
> -#define _DSPBCNTR (DISPLAY_MMIO_BASE(dev_priv) + 0x71180)
> +#define _DSPBCNTR (DISPLAY_MMIO_BASE(i915) + 0x71180)
> #define DISPPLANE_ALPHA_TRANS_ENABLE (1 << 15)
> #define DISPPLANE_ALPHA_TRANS_DISABLE 0
> #define DISPPLANE_SPRITE_ABOVE_DISPLAY 0
> #define DISPPLANE_SPRITE_ABOVE_OVERLAY (1)
> -#define _DSPBADDR (DISPLAY_MMIO_BASE(dev_priv) + 0x71184)
> -#define _DSPBSTRIDE (DISPLAY_MMIO_BASE(dev_priv) + 0x71188)
> -#define _DSPBPOS (DISPLAY_MMIO_BASE(dev_priv) + 0x7118C)
> -#define _DSPBSIZE (DISPLAY_MMIO_BASE(dev_priv) + 0x71190)
> -#define _DSPBSURF (DISPLAY_MMIO_BASE(dev_priv) + 0x7119C)
> -#define _DSPBTILEOFF (DISPLAY_MMIO_BASE(dev_priv) + 0x711A4)
> -#define _DSPBOFFSET (DISPLAY_MMIO_BASE(dev_priv) + 0x711A4)
> -#define _DSPBSURFLIVE (DISPLAY_MMIO_BASE(dev_priv) + 0x711AC)
> +#define _DSPBADDR (DISPLAY_MMIO_BASE(i915) + 0x71184)
> +#define _DSPBSTRIDE (DISPLAY_MMIO_BASE(i915) + 0x71188)
> +#define _DSPBPOS (DISPLAY_MMIO_BASE(i915) + 0x7118C)
> +#define _DSPBSIZE (DISPLAY_MMIO_BASE(i915) + 0x71190)
> +#define _DSPBSURF (DISPLAY_MMIO_BASE(i915) + 0x7119C)
> +#define _DSPBTILEOFF (DISPLAY_MMIO_BASE(i915) + 0x711A4)
> +#define _DSPBOFFSET (DISPLAY_MMIO_BASE(i915) + 0x711A4)
> +#define _DSPBSURFLIVE (DISPLAY_MMIO_BASE(i915) + 0x711AC)
>
> /* ICL DSI 0 and 1 */
> #define _PIPEDSI0CONF 0x7b008
> @@ -8946,7 +8946,7 @@ enum {
> #define GEN9_ENABLE_GPGPU_PREEMPTION (1 << 2)
>
> /* Audio */
> -#define G4X_AUD_VID_DID _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x62020)
> +#define G4X_AUD_VID_DID _MMIO(DISPLAY_MMIO_BASE(i915) + 0x62020)
> #define INTEL_AUDIO_DEVCL 0x808629FB
> #define INTEL_AUDIO_DEVBLC 0x80862801
> #define INTEL_AUDIO_DEVCTG 0x80862802
> @@ -10469,8 +10469,8 @@ enum skl_power_gate {
>
> /* MIPI DSI Controller and D-PHY registers */
>
> -#define _MIPIA_DEVICE_READY (dev_priv->mipi_mmio_base + 0xb000)
> -#define _MIPIC_DEVICE_READY (dev_priv->mipi_mmio_base + 0xb800)
> +#define _MIPIA_DEVICE_READY (i915->mipi_mmio_base + 0xb000)
> +#define _MIPIC_DEVICE_READY (i915->mipi_mmio_base + 0xb800)
> #define MIPI_DEVICE_READY(port) _MMIO_MIPI(port, _MIPIA_DEVICE_READY, _MIPIC_DEVICE_READY)
> #define BUS_POSSESSION (1 << 3) /* set to give bus to receiver */
> #define ULPS_STATE_MASK (3 << 1)
> @@ -10479,11 +10479,11 @@ enum skl_power_gate {
> #define ULPS_STATE_NORMAL_OPERATION (0 << 1)
> #define DEVICE_READY (1 << 0)
>
> -#define _MIPIA_INTR_STAT (dev_priv->mipi_mmio_base + 0xb004)
> -#define _MIPIC_INTR_STAT (dev_priv->mipi_mmio_base + 0xb804)
> +#define _MIPIA_INTR_STAT (i915->mipi_mmio_base + 0xb004)
> +#define _MIPIC_INTR_STAT (i915->mipi_mmio_base + 0xb804)
> #define MIPI_INTR_STAT(port) _MMIO_MIPI(port, _MIPIA_INTR_STAT, _MIPIC_INTR_STAT)
> -#define _MIPIA_INTR_EN (dev_priv->mipi_mmio_base + 0xb008)
> -#define _MIPIC_INTR_EN (dev_priv->mipi_mmio_base + 0xb808)
> +#define _MIPIA_INTR_EN (i915->mipi_mmio_base + 0xb008)
> +#define _MIPIC_INTR_EN (i915->mipi_mmio_base + 0xb808)
> #define MIPI_INTR_EN(port) _MMIO_MIPI(port, _MIPIA_INTR_EN, _MIPIC_INTR_EN)
> #define TEARING_EFFECT (1 << 31)
> #define SPL_PKT_SENT_INTERRUPT (1 << 30)
> @@ -10518,8 +10518,8 @@ enum skl_power_gate {
> #define RXSOT_SYNC_ERROR (1 << 1)
> #define RXSOT_ERROR (1 << 0)
>
> -#define _MIPIA_DSI_FUNC_PRG (dev_priv->mipi_mmio_base + 0xb00c)
> -#define _MIPIC_DSI_FUNC_PRG (dev_priv->mipi_mmio_base + 0xb80c)
> +#define _MIPIA_DSI_FUNC_PRG (i915->mipi_mmio_base + 0xb00c)
> +#define _MIPIC_DSI_FUNC_PRG (i915->mipi_mmio_base + 0xb80c)
> #define MIPI_DSI_FUNC_PRG(port) _MMIO_MIPI(port, _MIPIA_DSI_FUNC_PRG, _MIPIC_DSI_FUNC_PRG)
> #define CMD_MODE_DATA_WIDTH_MASK (7 << 13)
> #define CMD_MODE_NOT_SUPPORTED (0 << 13)
> @@ -10541,78 +10541,78 @@ enum skl_power_gate {
> #define DATA_LANES_PRG_REG_SHIFT 0
> #define DATA_LANES_PRG_REG_MASK (7 << 0)
>
> -#define _MIPIA_HS_TX_TIMEOUT (dev_priv->mipi_mmio_base + 0xb010)
> -#define _MIPIC_HS_TX_TIMEOUT (dev_priv->mipi_mmio_base + 0xb810)
> +#define _MIPIA_HS_TX_TIMEOUT (i915->mipi_mmio_base + 0xb010)
> +#define _MIPIC_HS_TX_TIMEOUT (i915->mipi_mmio_base + 0xb810)
> #define MIPI_HS_TX_TIMEOUT(port) _MMIO_MIPI(port, _MIPIA_HS_TX_TIMEOUT, _MIPIC_HS_TX_TIMEOUT)
> #define HIGH_SPEED_TX_TIMEOUT_COUNTER_MASK 0xffffff
>
> -#define _MIPIA_LP_RX_TIMEOUT (dev_priv->mipi_mmio_base + 0xb014)
> -#define _MIPIC_LP_RX_TIMEOUT (dev_priv->mipi_mmio_base + 0xb814)
> +#define _MIPIA_LP_RX_TIMEOUT (i915->mipi_mmio_base + 0xb014)
> +#define _MIPIC_LP_RX_TIMEOUT (i915->mipi_mmio_base + 0xb814)
> #define MIPI_LP_RX_TIMEOUT(port) _MMIO_MIPI(port, _MIPIA_LP_RX_TIMEOUT, _MIPIC_LP_RX_TIMEOUT)
> #define LOW_POWER_RX_TIMEOUT_COUNTER_MASK 0xffffff
>
> -#define _MIPIA_TURN_AROUND_TIMEOUT (dev_priv->mipi_mmio_base + 0xb018)
> -#define _MIPIC_TURN_AROUND_TIMEOUT (dev_priv->mipi_mmio_base + 0xb818)
> +#define _MIPIA_TURN_AROUND_TIMEOUT (i915->mipi_mmio_base + 0xb018)
> +#define _MIPIC_TURN_AROUND_TIMEOUT (i915->mipi_mmio_base + 0xb818)
> #define MIPI_TURN_AROUND_TIMEOUT(port) _MMIO_MIPI(port, _MIPIA_TURN_AROUND_TIMEOUT, _MIPIC_TURN_AROUND_TIMEOUT)
> #define TURN_AROUND_TIMEOUT_MASK 0x3f
>
> -#define _MIPIA_DEVICE_RESET_TIMER (dev_priv->mipi_mmio_base + 0xb01c)
> -#define _MIPIC_DEVICE_RESET_TIMER (dev_priv->mipi_mmio_base + 0xb81c)
> +#define _MIPIA_DEVICE_RESET_TIMER (i915->mipi_mmio_base + 0xb01c)
> +#define _MIPIC_DEVICE_RESET_TIMER (i915->mipi_mmio_base + 0xb81c)
> #define MIPI_DEVICE_RESET_TIMER(port) _MMIO_MIPI(port, _MIPIA_DEVICE_RESET_TIMER, _MIPIC_DEVICE_RESET_TIMER)
> #define DEVICE_RESET_TIMER_MASK 0xffff
>
> -#define _MIPIA_DPI_RESOLUTION (dev_priv->mipi_mmio_base + 0xb020)
> -#define _MIPIC_DPI_RESOLUTION (dev_priv->mipi_mmio_base + 0xb820)
> +#define _MIPIA_DPI_RESOLUTION (i915->mipi_mmio_base + 0xb020)
> +#define _MIPIC_DPI_RESOLUTION (i915->mipi_mmio_base + 0xb820)
> #define MIPI_DPI_RESOLUTION(port) _MMIO_MIPI(port, _MIPIA_DPI_RESOLUTION, _MIPIC_DPI_RESOLUTION)
> #define VERTICAL_ADDRESS_SHIFT 16
> #define VERTICAL_ADDRESS_MASK (0xffff << 16)
> #define HORIZONTAL_ADDRESS_SHIFT 0
> #define HORIZONTAL_ADDRESS_MASK 0xffff
>
> -#define _MIPIA_DBI_FIFO_THROTTLE (dev_priv->mipi_mmio_base + 0xb024)
> -#define _MIPIC_DBI_FIFO_THROTTLE (dev_priv->mipi_mmio_base + 0xb824)
> +#define _MIPIA_DBI_FIFO_THROTTLE (i915->mipi_mmio_base + 0xb024)
> +#define _MIPIC_DBI_FIFO_THROTTLE (i915->mipi_mmio_base + 0xb824)
> #define MIPI_DBI_FIFO_THROTTLE(port) _MMIO_MIPI(port, _MIPIA_DBI_FIFO_THROTTLE, _MIPIC_DBI_FIFO_THROTTLE)
> #define DBI_FIFO_EMPTY_HALF (0 << 0)
> #define DBI_FIFO_EMPTY_QUARTER (1 << 0)
> #define DBI_FIFO_EMPTY_7_LOCATIONS (2 << 0)
>
> /* regs below are bits 15:0 */
> -#define _MIPIA_HSYNC_PADDING_COUNT (dev_priv->mipi_mmio_base + 0xb028)
> -#define _MIPIC_HSYNC_PADDING_COUNT (dev_priv->mipi_mmio_base + 0xb828)
> +#define _MIPIA_HSYNC_PADDING_COUNT (i915->mipi_mmio_base + 0xb028)
> +#define _MIPIC_HSYNC_PADDING_COUNT (i915->mipi_mmio_base + 0xb828)
> #define MIPI_HSYNC_PADDING_COUNT(port) _MMIO_MIPI(port, _MIPIA_HSYNC_PADDING_COUNT, _MIPIC_HSYNC_PADDING_COUNT)
>
> -#define _MIPIA_HBP_COUNT (dev_priv->mipi_mmio_base + 0xb02c)
> -#define _MIPIC_HBP_COUNT (dev_priv->mipi_mmio_base + 0xb82c)
> +#define _MIPIA_HBP_COUNT (i915->mipi_mmio_base + 0xb02c)
> +#define _MIPIC_HBP_COUNT (i915->mipi_mmio_base + 0xb82c)
> #define MIPI_HBP_COUNT(port) _MMIO_MIPI(port, _MIPIA_HBP_COUNT, _MIPIC_HBP_COUNT)
>
> -#define _MIPIA_HFP_COUNT (dev_priv->mipi_mmio_base + 0xb030)
> -#define _MIPIC_HFP_COUNT (dev_priv->mipi_mmio_base + 0xb830)
> +#define _MIPIA_HFP_COUNT (i915->mipi_mmio_base + 0xb030)
> +#define _MIPIC_HFP_COUNT (i915->mipi_mmio_base + 0xb830)
> #define MIPI_HFP_COUNT(port) _MMIO_MIPI(port, _MIPIA_HFP_COUNT, _MIPIC_HFP_COUNT)
>
> -#define _MIPIA_HACTIVE_AREA_COUNT (dev_priv->mipi_mmio_base + 0xb034)
> -#define _MIPIC_HACTIVE_AREA_COUNT (dev_priv->mipi_mmio_base + 0xb834)
> +#define _MIPIA_HACTIVE_AREA_COUNT (i915->mipi_mmio_base + 0xb034)
> +#define _MIPIC_HACTIVE_AREA_COUNT (i915->mipi_mmio_base + 0xb834)
> #define MIPI_HACTIVE_AREA_COUNT(port) _MMIO_MIPI(port, _MIPIA_HACTIVE_AREA_COUNT, _MIPIC_HACTIVE_AREA_COUNT)
>
> -#define _MIPIA_VSYNC_PADDING_COUNT (dev_priv->mipi_mmio_base + 0xb038)
> -#define _MIPIC_VSYNC_PADDING_COUNT (dev_priv->mipi_mmio_base + 0xb838)
> +#define _MIPIA_VSYNC_PADDING_COUNT (i915->mipi_mmio_base + 0xb038)
> +#define _MIPIC_VSYNC_PADDING_COUNT (i915->mipi_mmio_base + 0xb838)
> #define MIPI_VSYNC_PADDING_COUNT(port) _MMIO_MIPI(port, _MIPIA_VSYNC_PADDING_COUNT, _MIPIC_VSYNC_PADDING_COUNT)
>
> -#define _MIPIA_VBP_COUNT (dev_priv->mipi_mmio_base + 0xb03c)
> -#define _MIPIC_VBP_COUNT (dev_priv->mipi_mmio_base + 0xb83c)
> +#define _MIPIA_VBP_COUNT (i915->mipi_mmio_base + 0xb03c)
> +#define _MIPIC_VBP_COUNT (i915->mipi_mmio_base + 0xb83c)
> #define MIPI_VBP_COUNT(port) _MMIO_MIPI(port, _MIPIA_VBP_COUNT, _MIPIC_VBP_COUNT)
>
> -#define _MIPIA_VFP_COUNT (dev_priv->mipi_mmio_base + 0xb040)
> -#define _MIPIC_VFP_COUNT (dev_priv->mipi_mmio_base + 0xb840)
> +#define _MIPIA_VFP_COUNT (i915->mipi_mmio_base + 0xb040)
> +#define _MIPIC_VFP_COUNT (i915->mipi_mmio_base + 0xb840)
> #define MIPI_VFP_COUNT(port) _MMIO_MIPI(port, _MIPIA_VFP_COUNT, _MIPIC_VFP_COUNT)
>
> -#define _MIPIA_HIGH_LOW_SWITCH_COUNT (dev_priv->mipi_mmio_base + 0xb044)
> -#define _MIPIC_HIGH_LOW_SWITCH_COUNT (dev_priv->mipi_mmio_base + 0xb844)
> +#define _MIPIA_HIGH_LOW_SWITCH_COUNT (i915->mipi_mmio_base + 0xb044)
> +#define _MIPIC_HIGH_LOW_SWITCH_COUNT (i915->mipi_mmio_base + 0xb844)
> #define MIPI_HIGH_LOW_SWITCH_COUNT(port) _MMIO_MIPI(port, _MIPIA_HIGH_LOW_SWITCH_COUNT, _MIPIC_HIGH_LOW_SWITCH_COUNT)
>
> /* regs above are bits 15:0 */
>
> -#define _MIPIA_DPI_CONTROL (dev_priv->mipi_mmio_base + 0xb048)
> -#define _MIPIC_DPI_CONTROL (dev_priv->mipi_mmio_base + 0xb848)
> +#define _MIPIA_DPI_CONTROL (i915->mipi_mmio_base + 0xb048)
> +#define _MIPIC_DPI_CONTROL (i915->mipi_mmio_base + 0xb848)
> #define MIPI_DPI_CONTROL(port) _MMIO_MIPI(port, _MIPIA_DPI_CONTROL, _MIPIC_DPI_CONTROL)
> #define DPI_LP_MODE (1 << 6)
> #define BACKLIGHT_OFF (1 << 5)
> @@ -10622,27 +10622,27 @@ enum skl_power_gate {
> #define TURN_ON (1 << 1)
> #define SHUTDOWN (1 << 0)
>
> -#define _MIPIA_DPI_DATA (dev_priv->mipi_mmio_base + 0xb04c)
> -#define _MIPIC_DPI_DATA (dev_priv->mipi_mmio_base + 0xb84c)
> +#define _MIPIA_DPI_DATA (i915->mipi_mmio_base + 0xb04c)
> +#define _MIPIC_DPI_DATA (i915->mipi_mmio_base + 0xb84c)
> #define MIPI_DPI_DATA(port) _MMIO_MIPI(port, _MIPIA_DPI_DATA, _MIPIC_DPI_DATA)
> #define COMMAND_BYTE_SHIFT 0
> #define COMMAND_BYTE_MASK (0x3f << 0)
>
> -#define _MIPIA_INIT_COUNT (dev_priv->mipi_mmio_base + 0xb050)
> -#define _MIPIC_INIT_COUNT (dev_priv->mipi_mmio_base + 0xb850)
> +#define _MIPIA_INIT_COUNT (i915->mipi_mmio_base + 0xb050)
> +#define _MIPIC_INIT_COUNT (i915->mipi_mmio_base + 0xb850)
> #define MIPI_INIT_COUNT(port) _MMIO_MIPI(port, _MIPIA_INIT_COUNT, _MIPIC_INIT_COUNT)
> #define MASTER_INIT_TIMER_SHIFT 0
> #define MASTER_INIT_TIMER_MASK (0xffff << 0)
>
> -#define _MIPIA_MAX_RETURN_PKT_SIZE (dev_priv->mipi_mmio_base + 0xb054)
> -#define _MIPIC_MAX_RETURN_PKT_SIZE (dev_priv->mipi_mmio_base + 0xb854)
> +#define _MIPIA_MAX_RETURN_PKT_SIZE (i915->mipi_mmio_base + 0xb054)
> +#define _MIPIC_MAX_RETURN_PKT_SIZE (i915->mipi_mmio_base + 0xb854)
> #define MIPI_MAX_RETURN_PKT_SIZE(port) _MMIO_MIPI(port, \
> _MIPIA_MAX_RETURN_PKT_SIZE, _MIPIC_MAX_RETURN_PKT_SIZE)
> #define MAX_RETURN_PKT_SIZE_SHIFT 0
> #define MAX_RETURN_PKT_SIZE_MASK (0x3ff << 0)
>
> -#define _MIPIA_VIDEO_MODE_FORMAT (dev_priv->mipi_mmio_base + 0xb058)
> -#define _MIPIC_VIDEO_MODE_FORMAT (dev_priv->mipi_mmio_base + 0xb858)
> +#define _MIPIA_VIDEO_MODE_FORMAT (i915->mipi_mmio_base + 0xb058)
> +#define _MIPIC_VIDEO_MODE_FORMAT (i915->mipi_mmio_base + 0xb858)
> #define MIPI_VIDEO_MODE_FORMAT(port) _MMIO_MIPI(port, _MIPIA_VIDEO_MODE_FORMAT, _MIPIC_VIDEO_MODE_FORMAT)
> #define RANDOM_DPI_DISPLAY_RESOLUTION (1 << 4)
> #define DISABLE_VIDEO_BTA (1 << 3)
> @@ -10651,8 +10651,8 @@ enum skl_power_gate {
> #define VIDEO_MODE_NON_BURST_WITH_SYNC_EVENTS (2 << 0)
> #define VIDEO_MODE_BURST (3 << 0)
>
> -#define _MIPIA_EOT_DISABLE (dev_priv->mipi_mmio_base + 0xb05c)
> -#define _MIPIC_EOT_DISABLE (dev_priv->mipi_mmio_base + 0xb85c)
> +#define _MIPIA_EOT_DISABLE (i915->mipi_mmio_base + 0xb05c)
> +#define _MIPIC_EOT_DISABLE (i915->mipi_mmio_base + 0xb85c)
> #define MIPI_EOT_DISABLE(port) _MMIO_MIPI(port, _MIPIA_EOT_DISABLE, _MIPIC_EOT_DISABLE)
> #define BXT_DEFEATURE_DPI_FIFO_CTR (1 << 9)
> #define BXT_DPHY_DEFEATURE_EN (1 << 8)
> @@ -10665,35 +10665,35 @@ enum skl_power_gate {
> #define CLOCKSTOP (1 << 1)
> #define EOT_DISABLE (1 << 0)
>
> -#define _MIPIA_LP_BYTECLK (dev_priv->mipi_mmio_base + 0xb060)
> -#define _MIPIC_LP_BYTECLK (dev_priv->mipi_mmio_base + 0xb860)
> +#define _MIPIA_LP_BYTECLK (i915->mipi_mmio_base + 0xb060)
> +#define _MIPIC_LP_BYTECLK (i915->mipi_mmio_base + 0xb860)
> #define MIPI_LP_BYTECLK(port) _MMIO_MIPI(port, _MIPIA_LP_BYTECLK, _MIPIC_LP_BYTECLK)
> #define LP_BYTECLK_SHIFT 0
> #define LP_BYTECLK_MASK (0xffff << 0)
>
> -#define _MIPIA_TLPX_TIME_COUNT (dev_priv->mipi_mmio_base + 0xb0a4)
> -#define _MIPIC_TLPX_TIME_COUNT (dev_priv->mipi_mmio_base + 0xb8a4)
> +#define _MIPIA_TLPX_TIME_COUNT (i915->mipi_mmio_base + 0xb0a4)
> +#define _MIPIC_TLPX_TIME_COUNT (i915->mipi_mmio_base + 0xb8a4)
> #define MIPI_TLPX_TIME_COUNT(port) _MMIO_MIPI(port, _MIPIA_TLPX_TIME_COUNT, _MIPIC_TLPX_TIME_COUNT)
>
> -#define _MIPIA_CLK_LANE_TIMING (dev_priv->mipi_mmio_base + 0xb098)
> -#define _MIPIC_CLK_LANE_TIMING (dev_priv->mipi_mmio_base + 0xb898)
> +#define _MIPIA_CLK_LANE_TIMING (i915->mipi_mmio_base + 0xb098)
> +#define _MIPIC_CLK_LANE_TIMING (i915->mipi_mmio_base + 0xb898)
> #define MIPI_CLK_LANE_TIMING(port) _MMIO_MIPI(port, _MIPIA_CLK_LANE_TIMING, _MIPIC_CLK_LANE_TIMING)
>
> /* bits 31:0 */
> -#define _MIPIA_LP_GEN_DATA (dev_priv->mipi_mmio_base + 0xb064)
> -#define _MIPIC_LP_GEN_DATA (dev_priv->mipi_mmio_base + 0xb864)
> +#define _MIPIA_LP_GEN_DATA (i915->mipi_mmio_base + 0xb064)
> +#define _MIPIC_LP_GEN_DATA (i915->mipi_mmio_base + 0xb864)
> #define MIPI_LP_GEN_DATA(port) _MMIO_MIPI(port, _MIPIA_LP_GEN_DATA, _MIPIC_LP_GEN_DATA)
>
> /* bits 31:0 */
> -#define _MIPIA_HS_GEN_DATA (dev_priv->mipi_mmio_base + 0xb068)
> -#define _MIPIC_HS_GEN_DATA (dev_priv->mipi_mmio_base + 0xb868)
> +#define _MIPIA_HS_GEN_DATA (i915->mipi_mmio_base + 0xb068)
> +#define _MIPIC_HS_GEN_DATA (i915->mipi_mmio_base + 0xb868)
> #define MIPI_HS_GEN_DATA(port) _MMIO_MIPI(port, _MIPIA_HS_GEN_DATA, _MIPIC_HS_GEN_DATA)
>
> -#define _MIPIA_LP_GEN_CTRL (dev_priv->mipi_mmio_base + 0xb06c)
> -#define _MIPIC_LP_GEN_CTRL (dev_priv->mipi_mmio_base + 0xb86c)
> +#define _MIPIA_LP_GEN_CTRL (i915->mipi_mmio_base + 0xb06c)
> +#define _MIPIC_LP_GEN_CTRL (i915->mipi_mmio_base + 0xb86c)
> #define MIPI_LP_GEN_CTRL(port) _MMIO_MIPI(port, _MIPIA_LP_GEN_CTRL, _MIPIC_LP_GEN_CTRL)
> -#define _MIPIA_HS_GEN_CTRL (dev_priv->mipi_mmio_base + 0xb070)
> -#define _MIPIC_HS_GEN_CTRL (dev_priv->mipi_mmio_base + 0xb870)
> +#define _MIPIA_HS_GEN_CTRL (i915->mipi_mmio_base + 0xb070)
> +#define _MIPIC_HS_GEN_CTRL (i915->mipi_mmio_base + 0xb870)
> #define MIPI_HS_GEN_CTRL(port) _MMIO_MIPI(port, _MIPIA_HS_GEN_CTRL, _MIPIC_HS_GEN_CTRL)
> #define LONG_PACKET_WORD_COUNT_SHIFT 8
> #define LONG_PACKET_WORD_COUNT_MASK (0xffff << 8)
> @@ -10705,8 +10705,8 @@ enum skl_power_gate {
> #define DATA_TYPE_MASK (0x3f << 0)
> /* data type values, see include/video/mipi_display.h */
>
> -#define _MIPIA_GEN_FIFO_STAT (dev_priv->mipi_mmio_base + 0xb074)
> -#define _MIPIC_GEN_FIFO_STAT (dev_priv->mipi_mmio_base + 0xb874)
> +#define _MIPIA_GEN_FIFO_STAT (i915->mipi_mmio_base + 0xb074)
> +#define _MIPIC_GEN_FIFO_STAT (i915->mipi_mmio_base + 0xb874)
> #define MIPI_GEN_FIFO_STAT(port) _MMIO_MIPI(port, _MIPIA_GEN_FIFO_STAT, _MIPIC_GEN_FIFO_STAT)
> #define DPI_FIFO_EMPTY (1 << 28)
> #define DBI_FIFO_EMPTY (1 << 27)
> @@ -10723,15 +10723,15 @@ enum skl_power_gate {
> #define HS_DATA_FIFO_HALF_EMPTY (1 << 1)
> #define HS_DATA_FIFO_FULL (1 << 0)
>
> -#define _MIPIA_HS_LS_DBI_ENABLE (dev_priv->mipi_mmio_base + 0xb078)
> -#define _MIPIC_HS_LS_DBI_ENABLE (dev_priv->mipi_mmio_base + 0xb878)
> +#define _MIPIA_HS_LS_DBI_ENABLE (i915->mipi_mmio_base + 0xb078)
> +#define _MIPIC_HS_LS_DBI_ENABLE (i915->mipi_mmio_base + 0xb878)
> #define MIPI_HS_LP_DBI_ENABLE(port) _MMIO_MIPI(port, _MIPIA_HS_LS_DBI_ENABLE, _MIPIC_HS_LS_DBI_ENABLE)
> #define DBI_HS_LP_MODE_MASK (1 << 0)
> #define DBI_LP_MODE (1 << 0)
> #define DBI_HS_MODE (0 << 0)
>
> -#define _MIPIA_DPHY_PARAM (dev_priv->mipi_mmio_base + 0xb080)
> -#define _MIPIC_DPHY_PARAM (dev_priv->mipi_mmio_base + 0xb880)
> +#define _MIPIA_DPHY_PARAM (i915->mipi_mmio_base + 0xb080)
> +#define _MIPIC_DPHY_PARAM (i915->mipi_mmio_base + 0xb880)
> #define MIPI_DPHY_PARAM(port) _MMIO_MIPI(port, _MIPIA_DPHY_PARAM, _MIPIC_DPHY_PARAM)
> #define EXIT_ZERO_COUNT_SHIFT 24
> #define EXIT_ZERO_COUNT_MASK (0x3f << 24)
> @@ -10978,34 +10978,34 @@ enum skl_power_gate {
> #define TA_TIMEOUT_VALUE(x) ((x) << 0)
>
> /* bits 31:0 */
> -#define _MIPIA_DBI_BW_CTRL (dev_priv->mipi_mmio_base + 0xb084)
> -#define _MIPIC_DBI_BW_CTRL (dev_priv->mipi_mmio_base + 0xb884)
> +#define _MIPIA_DBI_BW_CTRL (i915->mipi_mmio_base + 0xb084)
> +#define _MIPIC_DBI_BW_CTRL (i915->mipi_mmio_base + 0xb884)
> #define MIPI_DBI_BW_CTRL(port) _MMIO_MIPI(port, _MIPIA_DBI_BW_CTRL, _MIPIC_DBI_BW_CTRL)
>
> -#define _MIPIA_CLK_LANE_SWITCH_TIME_CNT (dev_priv->mipi_mmio_base + 0xb088)
> -#define _MIPIC_CLK_LANE_SWITCH_TIME_CNT (dev_priv->mipi_mmio_base + 0xb888)
> +#define _MIPIA_CLK_LANE_SWITCH_TIME_CNT (i915->mipi_mmio_base + 0xb088)
> +#define _MIPIC_CLK_LANE_SWITCH_TIME_CNT (i915->mipi_mmio_base + 0xb888)
> #define MIPI_CLK_LANE_SWITCH_TIME_CNT(port) _MMIO_MIPI(port, _MIPIA_CLK_LANE_SWITCH_TIME_CNT, _MIPIC_CLK_LANE_SWITCH_TIME_CNT)
> #define LP_HS_SSW_CNT_SHIFT 16
> #define LP_HS_SSW_CNT_MASK (0xffff << 16)
> #define HS_LP_PWR_SW_CNT_SHIFT 0
> #define HS_LP_PWR_SW_CNT_MASK (0xffff << 0)
>
> -#define _MIPIA_STOP_STATE_STALL (dev_priv->mipi_mmio_base + 0xb08c)
> -#define _MIPIC_STOP_STATE_STALL (dev_priv->mipi_mmio_base + 0xb88c)
> +#define _MIPIA_STOP_STATE_STALL (i915->mipi_mmio_base + 0xb08c)
> +#define _MIPIC_STOP_STATE_STALL (i915->mipi_mmio_base + 0xb88c)
> #define MIPI_STOP_STATE_STALL(port) _MMIO_MIPI(port, _MIPIA_STOP_STATE_STALL, _MIPIC_STOP_STATE_STALL)
> #define STOP_STATE_STALL_COUNTER_SHIFT 0
> #define STOP_STATE_STALL_COUNTER_MASK (0xff << 0)
>
> -#define _MIPIA_INTR_STAT_REG_1 (dev_priv->mipi_mmio_base + 0xb090)
> -#define _MIPIC_INTR_STAT_REG_1 (dev_priv->mipi_mmio_base + 0xb890)
> +#define _MIPIA_INTR_STAT_REG_1 (i915->mipi_mmio_base + 0xb090)
> +#define _MIPIC_INTR_STAT_REG_1 (i915->mipi_mmio_base + 0xb890)
> #define MIPI_INTR_STAT_REG_1(port) _MMIO_MIPI(port, _MIPIA_INTR_STAT_REG_1, _MIPIC_INTR_STAT_REG_1)
> -#define _MIPIA_INTR_EN_REG_1 (dev_priv->mipi_mmio_base + 0xb094)
> -#define _MIPIC_INTR_EN_REG_1 (dev_priv->mipi_mmio_base + 0xb894)
> +#define _MIPIA_INTR_EN_REG_1 (i915->mipi_mmio_base + 0xb094)
> +#define _MIPIC_INTR_EN_REG_1 (i915->mipi_mmio_base + 0xb894)
> #define MIPI_INTR_EN_REG_1(port) _MMIO_MIPI(port, _MIPIA_INTR_EN_REG_1, _MIPIC_INTR_EN_REG_1)
> #define RX_CONTENTION_DETECTED (1 << 0)
>
> /* XXX: only pipe A ?!? */
> -#define MIPIA_DBI_TYPEC_CTRL (dev_priv->mipi_mmio_base + 0xb100)
> +#define MIPIA_DBI_TYPEC_CTRL (i915->mipi_mmio_base + 0xb100)
> #define DBI_TYPEC_ENABLE (1 << 31)
> #define DBI_TYPEC_WIP (1 << 30)
> #define DBI_TYPEC_OPTION_SHIFT 28
> @@ -11019,8 +11019,8 @@ enum skl_power_gate {
>
> /* MIPI adapter registers */
>
> -#define _MIPIA_CTRL (dev_priv->mipi_mmio_base + 0xb104)
> -#define _MIPIC_CTRL (dev_priv->mipi_mmio_base + 0xb904)
> +#define _MIPIA_CTRL (i915->mipi_mmio_base + 0xb104)
> +#define _MIPIC_CTRL (i915->mipi_mmio_base + 0xb904)
> #define MIPI_CTRL(port) _MMIO_MIPI(port, _MIPIA_CTRL, _MIPIC_CTRL)
> #define ESCAPE_CLOCK_DIVIDER_SHIFT 5 /* A only */
> #define ESCAPE_CLOCK_DIVIDER_MASK (3 << 5)
> @@ -11052,21 +11052,21 @@ enum skl_power_gate {
> #define GLK_MIPIIO_PORT_POWERED (1 << 1) /* RO */
> #define GLK_MIPIIO_ENABLE (1 << 0)
>
> -#define _MIPIA_DATA_ADDRESS (dev_priv->mipi_mmio_base + 0xb108)
> -#define _MIPIC_DATA_ADDRESS (dev_priv->mipi_mmio_base + 0xb908)
> +#define _MIPIA_DATA_ADDRESS (i915->mipi_mmio_base + 0xb108)
> +#define _MIPIC_DATA_ADDRESS (i915->mipi_mmio_base + 0xb908)
> #define MIPI_DATA_ADDRESS(port) _MMIO_MIPI(port, _MIPIA_DATA_ADDRESS, _MIPIC_DATA_ADDRESS)
> #define DATA_MEM_ADDRESS_SHIFT 5
> #define DATA_MEM_ADDRESS_MASK (0x7ffffff << 5)
> #define DATA_VALID (1 << 0)
>
> -#define _MIPIA_DATA_LENGTH (dev_priv->mipi_mmio_base + 0xb10c)
> -#define _MIPIC_DATA_LENGTH (dev_priv->mipi_mmio_base + 0xb90c)
> +#define _MIPIA_DATA_LENGTH (i915->mipi_mmio_base + 0xb10c)
> +#define _MIPIC_DATA_LENGTH (i915->mipi_mmio_base + 0xb90c)
> #define MIPI_DATA_LENGTH(port) _MMIO_MIPI(port, _MIPIA_DATA_LENGTH, _MIPIC_DATA_LENGTH)
> #define DATA_LENGTH_SHIFT 0
> #define DATA_LENGTH_MASK (0xfffff << 0)
>
> -#define _MIPIA_COMMAND_ADDRESS (dev_priv->mipi_mmio_base + 0xb110)
> -#define _MIPIC_COMMAND_ADDRESS (dev_priv->mipi_mmio_base + 0xb910)
> +#define _MIPIA_COMMAND_ADDRESS (i915->mipi_mmio_base + 0xb110)
> +#define _MIPIC_COMMAND_ADDRESS (i915->mipi_mmio_base + 0xb910)
> #define MIPI_COMMAND_ADDRESS(port) _MMIO_MIPI(port, _MIPIA_COMMAND_ADDRESS, _MIPIC_COMMAND_ADDRESS)
> #define COMMAND_MEM_ADDRESS_SHIFT 5
> #define COMMAND_MEM_ADDRESS_MASK (0x7ffffff << 5)
> @@ -11074,18 +11074,18 @@ enum skl_power_gate {
> #define MEMORY_WRITE_DATA_FROM_PIPE_RENDERING (1 << 1)
> #define COMMAND_VALID (1 << 0)
>
> -#define _MIPIA_COMMAND_LENGTH (dev_priv->mipi_mmio_base + 0xb114)
> -#define _MIPIC_COMMAND_LENGTH (dev_priv->mipi_mmio_base + 0xb914)
> +#define _MIPIA_COMMAND_LENGTH (i915->mipi_mmio_base + 0xb114)
> +#define _MIPIC_COMMAND_LENGTH (i915->mipi_mmio_base + 0xb914)
> #define MIPI_COMMAND_LENGTH(port) _MMIO_MIPI(port, _MIPIA_COMMAND_LENGTH, _MIPIC_COMMAND_LENGTH)
> #define COMMAND_LENGTH_SHIFT(n) (8 * (n)) /* n: 0...3 */
> #define COMMAND_LENGTH_MASK(n) (0xff << (8 * (n)))
>
> -#define _MIPIA_READ_DATA_RETURN0 (dev_priv->mipi_mmio_base + 0xb118)
> -#define _MIPIC_READ_DATA_RETURN0 (dev_priv->mipi_mmio_base + 0xb918)
> +#define _MIPIA_READ_DATA_RETURN0 (i915->mipi_mmio_base + 0xb118)
> +#define _MIPIC_READ_DATA_RETURN0 (i915->mipi_mmio_base + 0xb918)
> #define MIPI_READ_DATA_RETURN(port, n) _MMIO(_MIPI(port, _MIPIA_READ_DATA_RETURN0, _MIPIC_READ_DATA_RETURN0) + 4 * (n)) /* n: 0...7 */
>
> -#define _MIPIA_READ_DATA_VALID (dev_priv->mipi_mmio_base + 0xb138)
> -#define _MIPIC_READ_DATA_VALID (dev_priv->mipi_mmio_base + 0xb938)
> +#define _MIPIA_READ_DATA_VALID (i915->mipi_mmio_base + 0xb138)
> +#define _MIPIC_READ_DATA_VALID (i915->mipi_mmio_base + 0xb938)
> #define MIPI_READ_DATA_VALID(port) _MMIO_MIPI(port, _MIPIA_READ_DATA_VALID, _MIPIC_READ_DATA_VALID)
> #define READ_DATA_VALID(n) (1 << (n))
>
> diff --git a/drivers/gpu/drm/i915/i915_suspend.c b/drivers/gpu/drm/i915/i915_suspend.c
> index 581201bcb81a..5b3ad7c9240e 100644
> --- a/drivers/gpu/drm/i915/i915_suspend.c
> +++ b/drivers/gpu/drm/i915/i915_suspend.c
> @@ -31,121 +31,121 @@
> #include "intel_fbc.h"
> #include "intel_gmbus.h"
>
> -static void i915_save_display(struct drm_i915_private *dev_priv)
> +static void i915_save_display(struct drm_i915_private *i915)
> {
> /* Display arbitration control */
> - if (INTEL_GEN(dev_priv) <= 4)
> - dev_priv->regfile.saveDSPARB = I915_READ(DSPARB);
> + if (INTEL_GEN(i915) <= 4)
> + i915->regfile.saveDSPARB = I915_READ(DSPARB);
>
> /* save FBC interval */
> - if (HAS_FBC(dev_priv) && INTEL_GEN(dev_priv) <= 4 && !IS_G4X(dev_priv))
> - dev_priv->regfile.saveFBC_CONTROL = I915_READ(FBC_CONTROL);
> + if (HAS_FBC(i915) && INTEL_GEN(i915) <= 4 && !IS_G4X(i915))
> + i915->regfile.saveFBC_CONTROL = I915_READ(FBC_CONTROL);
> }
>
> -static void i915_restore_display(struct drm_i915_private *dev_priv)
> +static void i915_restore_display(struct drm_i915_private *i915)
> {
> /* Display arbitration */
> - if (INTEL_GEN(dev_priv) <= 4)
> - I915_WRITE(DSPARB, dev_priv->regfile.saveDSPARB);
> + if (INTEL_GEN(i915) <= 4)
> + I915_WRITE(DSPARB, i915->regfile.saveDSPARB);
>
> /* only restore FBC info on the platform that supports FBC*/
> - intel_fbc_global_disable(dev_priv);
> + intel_fbc_global_disable(i915);
>
> /* restore FBC interval */
> - if (HAS_FBC(dev_priv) && INTEL_GEN(dev_priv) <= 4 && !IS_G4X(dev_priv))
> - I915_WRITE(FBC_CONTROL, dev_priv->regfile.saveFBC_CONTROL);
> + if (HAS_FBC(i915) && INTEL_GEN(i915) <= 4 && !IS_G4X(i915))
> + I915_WRITE(FBC_CONTROL, i915->regfile.saveFBC_CONTROL);
>
> - i915_redisable_vga(dev_priv);
> + i915_redisable_vga(i915);
> }
>
> -int i915_save_state(struct drm_i915_private *dev_priv)
> +int i915_save_state(struct drm_i915_private *i915)
> {
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct pci_dev *pdev = i915->drm.pdev;
> int i;
>
> - mutex_lock(&dev_priv->drm.struct_mutex);
> + mutex_lock(&i915->drm.struct_mutex);
>
> - i915_save_display(dev_priv);
> + i915_save_display(i915);
>
> - if (IS_GEN(dev_priv, 4))
> + if (IS_GEN(i915, 4))
> pci_read_config_word(pdev, GCDGMBUS,
> - &dev_priv->regfile.saveGCDGMBUS);
> + &i915->regfile.saveGCDGMBUS);
>
> /* Cache mode state */
> - if (INTEL_GEN(dev_priv) < 7)
> - dev_priv->regfile.saveCACHE_MODE_0 = I915_READ(CACHE_MODE_0);
> + if (INTEL_GEN(i915) < 7)
> + i915->regfile.saveCACHE_MODE_0 = I915_READ(CACHE_MODE_0);
>
> /* Memory Arbitration state */
> - dev_priv->regfile.saveMI_ARB_STATE = I915_READ(MI_ARB_STATE);
> + i915->regfile.saveMI_ARB_STATE = I915_READ(MI_ARB_STATE);
>
> /* Scratch space */
> - if (IS_GEN(dev_priv, 2) && IS_MOBILE(dev_priv)) {
> + if (IS_GEN(i915, 2) && IS_MOBILE(i915)) {
> for (i = 0; i < 7; i++) {
> - dev_priv->regfile.saveSWF0[i] = I915_READ(SWF0(i));
> - dev_priv->regfile.saveSWF1[i] = I915_READ(SWF1(i));
> + i915->regfile.saveSWF0[i] = I915_READ(SWF0(i));
> + i915->regfile.saveSWF1[i] = I915_READ(SWF1(i));
> }
> for (i = 0; i < 3; i++)
> - dev_priv->regfile.saveSWF3[i] = I915_READ(SWF3(i));
> - } else if (IS_GEN(dev_priv, 2)) {
> + i915->regfile.saveSWF3[i] = I915_READ(SWF3(i));
> + } else if (IS_GEN(i915, 2)) {
> for (i = 0; i < 7; i++)
> - dev_priv->regfile.saveSWF1[i] = I915_READ(SWF1(i));
> - } else if (HAS_GMCH(dev_priv)) {
> + i915->regfile.saveSWF1[i] = I915_READ(SWF1(i));
> + } else if (HAS_GMCH(i915)) {
> for (i = 0; i < 16; i++) {
> - dev_priv->regfile.saveSWF0[i] = I915_READ(SWF0(i));
> - dev_priv->regfile.saveSWF1[i] = I915_READ(SWF1(i));
> + i915->regfile.saveSWF0[i] = I915_READ(SWF0(i));
> + i915->regfile.saveSWF1[i] = I915_READ(SWF1(i));
> }
> for (i = 0; i < 3; i++)
> - dev_priv->regfile.saveSWF3[i] = I915_READ(SWF3(i));
> + i915->regfile.saveSWF3[i] = I915_READ(SWF3(i));
> }
>
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + mutex_unlock(&i915->drm.struct_mutex);
>
> return 0;
> }
>
> -int i915_restore_state(struct drm_i915_private *dev_priv)
> +int i915_restore_state(struct drm_i915_private *i915)
> {
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct pci_dev *pdev = i915->drm.pdev;
> int i;
>
> - mutex_lock(&dev_priv->drm.struct_mutex);
> + mutex_lock(&i915->drm.struct_mutex);
>
> - if (IS_GEN(dev_priv, 4))
> + if (IS_GEN(i915, 4))
> pci_write_config_word(pdev, GCDGMBUS,
> - dev_priv->regfile.saveGCDGMBUS);
> - i915_restore_display(dev_priv);
> + i915->regfile.saveGCDGMBUS);
> + i915_restore_display(i915);
>
> /* Cache mode state */
> - if (INTEL_GEN(dev_priv) < 7)
> - I915_WRITE(CACHE_MODE_0, dev_priv->regfile.saveCACHE_MODE_0 |
> + if (INTEL_GEN(i915) < 7)
> + I915_WRITE(CACHE_MODE_0, i915->regfile.saveCACHE_MODE_0 |
> 0xffff0000);
>
> /* Memory arbitration state */
> - I915_WRITE(MI_ARB_STATE, dev_priv->regfile.saveMI_ARB_STATE | 0xffff0000);
> + I915_WRITE(MI_ARB_STATE, i915->regfile.saveMI_ARB_STATE | 0xffff0000);
>
> /* Scratch space */
> - if (IS_GEN(dev_priv, 2) && IS_MOBILE(dev_priv)) {
> + if (IS_GEN(i915, 2) && IS_MOBILE(i915)) {
> for (i = 0; i < 7; i++) {
> - I915_WRITE(SWF0(i), dev_priv->regfile.saveSWF0[i]);
> - I915_WRITE(SWF1(i), dev_priv->regfile.saveSWF1[i]);
> + I915_WRITE(SWF0(i), i915->regfile.saveSWF0[i]);
> + I915_WRITE(SWF1(i), i915->regfile.saveSWF1[i]);
> }
> for (i = 0; i < 3; i++)
> - I915_WRITE(SWF3(i), dev_priv->regfile.saveSWF3[i]);
> - } else if (IS_GEN(dev_priv, 2)) {
> + I915_WRITE(SWF3(i), i915->regfile.saveSWF3[i]);
> + } else if (IS_GEN(i915, 2)) {
> for (i = 0; i < 7; i++)
> - I915_WRITE(SWF1(i), dev_priv->regfile.saveSWF1[i]);
> - } else if (HAS_GMCH(dev_priv)) {
> + I915_WRITE(SWF1(i), i915->regfile.saveSWF1[i]);
> + } else if (HAS_GMCH(i915)) {
> for (i = 0; i < 16; i++) {
> - I915_WRITE(SWF0(i), dev_priv->regfile.saveSWF0[i]);
> - I915_WRITE(SWF1(i), dev_priv->regfile.saveSWF1[i]);
> + I915_WRITE(SWF0(i), i915->regfile.saveSWF0[i]);
> + I915_WRITE(SWF1(i), i915->regfile.saveSWF1[i]);
> }
> for (i = 0; i < 3; i++)
> - I915_WRITE(SWF3(i), dev_priv->regfile.saveSWF3[i]);
> + I915_WRITE(SWF3(i), i915->regfile.saveSWF3[i]);
> }
>
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + mutex_unlock(&i915->drm.struct_mutex);
>
> - intel_gmbus_reset(dev_priv);
> + intel_gmbus_reset(i915);
>
> return 0;
> }
> diff --git a/drivers/gpu/drm/i915/i915_sysfs.c b/drivers/gpu/drm/i915/i915_sysfs.c
> index 3ef07b987d40..9c89be3ec4c1 100644
> --- a/drivers/gpu/drm/i915/i915_sysfs.c
> +++ b/drivers/gpu/drm/i915/i915_sysfs.c
> @@ -42,14 +42,14 @@ static inline struct drm_i915_private *kdev_minor_to_i915(struct device *kdev)
> }
>
> #ifdef CONFIG_PM
> -static u32 calc_residency(struct drm_i915_private *dev_priv,
> +static u32 calc_residency(struct drm_i915_private *i915,
> i915_reg_t reg)
> {
> intel_wakeref_t wakeref;
> u64 res = 0;
>
> - with_intel_runtime_pm(dev_priv, wakeref)
> - res = intel_rc6_residency_us(dev_priv, reg);
> + with_intel_runtime_pm(i915, wakeref)
> + res = intel_rc6_residency_us(i915, reg);
>
> return DIV_ROUND_CLOSEST_ULL(res, 1000);
> }
> @@ -57,15 +57,15 @@ static u32 calc_residency(struct drm_i915_private *dev_priv,
> static ssize_t
> show_rc6_mask(struct device *kdev, struct device_attribute *attr, char *buf)
> {
> - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> unsigned int mask;
>
> mask = 0;
> - if (HAS_RC6(dev_priv))
> + if (HAS_RC6(i915))
> mask |= BIT(0);
> - if (HAS_RC6p(dev_priv))
> + if (HAS_RC6p(i915))
> mask |= BIT(1);
> - if (HAS_RC6pp(dev_priv))
> + if (HAS_RC6pp(i915))
> mask |= BIT(2);
>
> return snprintf(buf, PAGE_SIZE, "%x\n", mask);
> @@ -74,32 +74,32 @@ show_rc6_mask(struct device *kdev, struct device_attribute *attr, char *buf)
> static ssize_t
> show_rc6_ms(struct device *kdev, struct device_attribute *attr, char *buf)
> {
> - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> - u32 rc6_residency = calc_residency(dev_priv, GEN6_GT_GFX_RC6);
> + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> + u32 rc6_residency = calc_residency(i915, GEN6_GT_GFX_RC6);
> return snprintf(buf, PAGE_SIZE, "%u\n", rc6_residency);
> }
>
> static ssize_t
> show_rc6p_ms(struct device *kdev, struct device_attribute *attr, char *buf)
> {
> - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> - u32 rc6p_residency = calc_residency(dev_priv, GEN6_GT_GFX_RC6p);
> + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> + u32 rc6p_residency = calc_residency(i915, GEN6_GT_GFX_RC6p);
> return snprintf(buf, PAGE_SIZE, "%u\n", rc6p_residency);
> }
>
> static ssize_t
> show_rc6pp_ms(struct device *kdev, struct device_attribute *attr, char *buf)
> {
> - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> - u32 rc6pp_residency = calc_residency(dev_priv, GEN6_GT_GFX_RC6pp);
> + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> + u32 rc6pp_residency = calc_residency(i915, GEN6_GT_GFX_RC6pp);
> return snprintf(buf, PAGE_SIZE, "%u\n", rc6pp_residency);
> }
>
> static ssize_t
> show_media_rc6_ms(struct device *kdev, struct device_attribute *attr, char *buf)
> {
> - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> - u32 rc6_residency = calc_residency(dev_priv, VLV_GT_MEDIA_RC6);
> + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> + u32 rc6_residency = calc_residency(i915, VLV_GT_MEDIA_RC6);
> return snprintf(buf, PAGE_SIZE, "%u\n", rc6_residency);
> }
>
> @@ -142,9 +142,9 @@ static const struct attribute_group media_rc6_attr_group = {
> };
> #endif
>
> -static int l3_access_valid(struct drm_i915_private *dev_priv, loff_t offset)
> +static int l3_access_valid(struct drm_i915_private *i915, loff_t offset)
> {
> - if (!HAS_L3_DPF(dev_priv))
> + if (!HAS_L3_DPF(i915))
> return -EPERM;
>
> if (offset % 4 != 0)
> @@ -162,14 +162,14 @@ i915_l3_read(struct file *filp, struct kobject *kobj,
> loff_t offset, size_t count)
> {
> struct device *kdev = kobj_to_dev(kobj);
> - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> + struct drm_device *dev = &i915->drm;
> int slice = (int)(uintptr_t)attr->private;
> int ret;
>
> count = round_down(count, 4);
>
> - ret = l3_access_valid(dev_priv, offset);
> + ret = l3_access_valid(i915, offset);
> if (ret)
> return ret;
>
> @@ -179,9 +179,9 @@ i915_l3_read(struct file *filp, struct kobject *kobj,
> if (ret)
> return ret;
>
> - if (dev_priv->l3_parity.remap_info[slice])
> + if (i915->l3_parity.remap_info[slice])
> memcpy(buf,
> - dev_priv->l3_parity.remap_info[slice] + (offset/4),
> + i915->l3_parity.remap_info[slice] + (offset/4),
> count);
> else
> memset(buf, 0, count);
> @@ -197,14 +197,14 @@ i915_l3_write(struct file *filp, struct kobject *kobj,
> loff_t offset, size_t count)
> {
> struct device *kdev = kobj_to_dev(kobj);
> - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> + struct drm_device *dev = &i915->drm;
> struct i915_gem_context *ctx;
> int slice = (int)(uintptr_t)attr->private;
> u32 **remap_info;
> int ret;
>
> - ret = l3_access_valid(dev_priv, offset);
> + ret = l3_access_valid(i915, offset);
> if (ret)
> return ret;
>
> @@ -212,7 +212,7 @@ i915_l3_write(struct file *filp, struct kobject *kobj,
> if (ret)
> return ret;
>
> - remap_info = &dev_priv->l3_parity.remap_info[slice];
> + remap_info = &i915->l3_parity.remap_info[slice];
> if (!*remap_info) {
> *remap_info = kzalloc(GEN7_L3LOG_SIZE, GFP_KERNEL);
> if (!*remap_info) {
> @@ -228,7 +228,7 @@ i915_l3_write(struct file *filp, struct kobject *kobj,
> memcpy(*remap_info + (offset/4), buf, count);
>
> /* NB: We defer the remapping until we switch to the context */
> - list_for_each_entry(ctx, &dev_priv->contexts.list, link)
> + list_for_each_entry(ctx, &i915->contexts.list, link)
> ctx->remap_slice |= (1<<slice);
>
> ret = count;
> @@ -260,52 +260,52 @@ static const struct bin_attribute dpf_attrs_1 = {
> static ssize_t gt_act_freq_mhz_show(struct device *kdev,
> struct device_attribute *attr, char *buf)
> {
> - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> intel_wakeref_t wakeref;
> u32 freq;
>
> - wakeref = intel_runtime_pm_get(dev_priv);
> + wakeref = intel_runtime_pm_get(i915);
>
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> - vlv_punit_get(dev_priv);
> - freq = vlv_punit_read(dev_priv, PUNIT_REG_GPU_FREQ_STS);
> - vlv_punit_put(dev_priv);
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> + vlv_punit_get(i915);
> + freq = vlv_punit_read(i915, PUNIT_REG_GPU_FREQ_STS);
> + vlv_punit_put(i915);
>
> freq = (freq >> 8) & 0xff;
> } else {
> - freq = intel_get_cagf(dev_priv, I915_READ(GEN6_RPSTAT1));
> + freq = intel_get_cagf(i915, I915_READ(GEN6_RPSTAT1));
> }
>
> - intel_runtime_pm_put(dev_priv, wakeref);
> + intel_runtime_pm_put(i915, wakeref);
>
> - return snprintf(buf, PAGE_SIZE, "%d\n", intel_gpu_freq(dev_priv, freq));
> + return snprintf(buf, PAGE_SIZE, "%d\n", intel_gpu_freq(i915, freq));
> }
>
> static ssize_t gt_cur_freq_mhz_show(struct device *kdev,
> struct device_attribute *attr, char *buf)
> {
> - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
>
> return snprintf(buf, PAGE_SIZE, "%d\n",
> - intel_gpu_freq(dev_priv,
> - dev_priv->gt_pm.rps.cur_freq));
> + intel_gpu_freq(i915,
> + i915->gt_pm.rps.cur_freq));
> }
>
> static ssize_t gt_boost_freq_mhz_show(struct device *kdev, struct device_attribute *attr, char *buf)
> {
> - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
>
> return snprintf(buf, PAGE_SIZE, "%d\n",
> - intel_gpu_freq(dev_priv,
> - dev_priv->gt_pm.rps.boost_freq));
> + intel_gpu_freq(i915,
> + i915->gt_pm.rps.boost_freq));
> }
>
> static ssize_t gt_boost_freq_mhz_store(struct device *kdev,
> struct device_attribute *attr,
> const char *buf, size_t count)
> {
> - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> + struct intel_rps *rps = &i915->gt_pm.rps;
> bool boost = false;
> ssize_t ret;
> u32 val;
> @@ -315,7 +315,7 @@ static ssize_t gt_boost_freq_mhz_store(struct device *kdev,
> return ret;
>
> /* Validate against (static) hardware limits */
> - val = intel_freq_opcode(dev_priv, val);
> + val = intel_freq_opcode(i915, val);
> if (val < rps->min_freq || val > rps->max_freq)
> return -EINVAL;
>
> @@ -334,28 +334,28 @@ static ssize_t gt_boost_freq_mhz_store(struct device *kdev,
> static ssize_t vlv_rpe_freq_mhz_show(struct device *kdev,
> struct device_attribute *attr, char *buf)
> {
> - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
>
> return snprintf(buf, PAGE_SIZE, "%d\n",
> - intel_gpu_freq(dev_priv,
> - dev_priv->gt_pm.rps.efficient_freq));
> + intel_gpu_freq(i915,
> + i915->gt_pm.rps.efficient_freq));
> }
>
> static ssize_t gt_max_freq_mhz_show(struct device *kdev, struct device_attribute *attr, char *buf)
> {
> - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
>
> return snprintf(buf, PAGE_SIZE, "%d\n",
> - intel_gpu_freq(dev_priv,
> - dev_priv->gt_pm.rps.max_freq_softlimit));
> + intel_gpu_freq(i915,
> + i915->gt_pm.rps.max_freq_softlimit));
> }
>
> static ssize_t gt_max_freq_mhz_store(struct device *kdev,
> struct device_attribute *attr,
> const char *buf, size_t count)
> {
> - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> + struct intel_rps *rps = &i915->gt_pm.rps;
> intel_wakeref_t wakeref;
> u32 val;
> ssize_t ret;
> @@ -364,10 +364,10 @@ static ssize_t gt_max_freq_mhz_store(struct device *kdev,
> if (ret)
> return ret;
>
> - wakeref = intel_runtime_pm_get(dev_priv);
> + wakeref = intel_runtime_pm_get(i915);
> mutex_lock(&rps->lock);
>
> - val = intel_freq_opcode(dev_priv, val);
> + val = intel_freq_opcode(i915, val);
> if (val < rps->min_freq ||
> val > rps->max_freq ||
> val < rps->min_freq_softlimit) {
> @@ -377,7 +377,7 @@ static ssize_t gt_max_freq_mhz_store(struct device *kdev,
>
> if (val > rps->rp0_freq)
> DRM_DEBUG("User requested overclocking to %d\n",
> - intel_gpu_freq(dev_priv, val));
> + intel_gpu_freq(i915, val));
>
> rps->max_freq_softlimit = val;
>
> @@ -388,30 +388,30 @@ static ssize_t gt_max_freq_mhz_store(struct device *kdev,
> /* We still need *_set_rps to process the new max_delay and
> * update the interrupt limits and PMINTRMSK even though
> * frequency request may be unchanged. */
> - ret = intel_set_rps(dev_priv, val);
> + ret = intel_set_rps(i915, val);
>
> unlock:
> mutex_unlock(&rps->lock);
> - intel_runtime_pm_put(dev_priv, wakeref);
> + intel_runtime_pm_put(i915, wakeref);
>
> return ret ?: count;
> }
>
> static ssize_t gt_min_freq_mhz_show(struct device *kdev, struct device_attribute *attr, char *buf)
> {
> - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
>
> return snprintf(buf, PAGE_SIZE, "%d\n",
> - intel_gpu_freq(dev_priv,
> - dev_priv->gt_pm.rps.min_freq_softlimit));
> + intel_gpu_freq(i915,
> + i915->gt_pm.rps.min_freq_softlimit));
> }
>
> static ssize_t gt_min_freq_mhz_store(struct device *kdev,
> struct device_attribute *attr,
> const char *buf, size_t count)
> {
> - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> + struct intel_rps *rps = &i915->gt_pm.rps;
> intel_wakeref_t wakeref;
> u32 val;
> ssize_t ret;
> @@ -420,10 +420,10 @@ static ssize_t gt_min_freq_mhz_store(struct device *kdev,
> if (ret)
> return ret;
>
> - wakeref = intel_runtime_pm_get(dev_priv);
> + wakeref = intel_runtime_pm_get(i915);
> mutex_lock(&rps->lock);
>
> - val = intel_freq_opcode(dev_priv, val);
> + val = intel_freq_opcode(i915, val);
> if (val < rps->min_freq ||
> val > rps->max_freq ||
> val > rps->max_freq_softlimit) {
> @@ -440,11 +440,11 @@ static ssize_t gt_min_freq_mhz_store(struct device *kdev,
> /* We still need *_set_rps to process the new min_delay and
> * update the interrupt limits and PMINTRMSK even though
> * frequency request may be unchanged. */
> - ret = intel_set_rps(dev_priv, val);
> + ret = intel_set_rps(i915, val);
>
> unlock:
> mutex_unlock(&rps->lock);
> - intel_runtime_pm_put(dev_priv, wakeref);
> + intel_runtime_pm_put(i915, wakeref);
>
> return ret ?: count;
> }
> @@ -465,16 +465,16 @@ static DEVICE_ATTR(gt_RPn_freq_mhz, S_IRUGO, gt_rp_mhz_show, NULL);
> /* For now we have a static number of RP states */
> static ssize_t gt_rp_mhz_show(struct device *kdev, struct device_attribute *attr, char *buf)
> {
> - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> + struct intel_rps *rps = &i915->gt_pm.rps;
> u32 val;
>
> if (attr == &dev_attr_gt_RP0_freq_mhz)
> - val = intel_gpu_freq(dev_priv, rps->rp0_freq);
> + val = intel_gpu_freq(i915, rps->rp0_freq);
> else if (attr == &dev_attr_gt_RP1_freq_mhz)
> - val = intel_gpu_freq(dev_priv, rps->rp1_freq);
> + val = intel_gpu_freq(i915, rps->rp1_freq);
> else if (attr == &dev_attr_gt_RPn_freq_mhz)
> - val = intel_gpu_freq(dev_priv, rps->min_freq);
> + val = intel_gpu_freq(i915, rps->min_freq);
> else
> BUG();
>
> @@ -540,10 +540,10 @@ static ssize_t error_state_write(struct file *file, struct kobject *kobj,
> loff_t off, size_t count)
> {
> struct device *kdev = kobj_to_dev(kobj);
> - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
>
> DRM_DEBUG_DRIVER("Resetting error state\n");
> - i915_reset_error_state(dev_priv);
> + i915_reset_error_state(i915);
>
> return count;
> }
> @@ -571,37 +571,37 @@ static void i915_setup_error_capture(struct device *kdev) {}
> static void i915_teardown_error_capture(struct device *kdev) {}
> #endif
>
> -void i915_setup_sysfs(struct drm_i915_private *dev_priv)
> +void i915_setup_sysfs(struct drm_i915_private *i915)
> {
> - struct device *kdev = dev_priv->drm.primary->kdev;
> + struct device *kdev = i915->drm.primary->kdev;
> int ret;
>
> #ifdef CONFIG_PM
> - if (HAS_RC6(dev_priv)) {
> + if (HAS_RC6(i915)) {
> ret = sysfs_merge_group(&kdev->kobj,
> &rc6_attr_group);
> if (ret)
> DRM_ERROR("RC6 residency sysfs setup failed\n");
> }
> - if (HAS_RC6p(dev_priv)) {
> + if (HAS_RC6p(i915)) {
> ret = sysfs_merge_group(&kdev->kobj,
> &rc6p_attr_group);
> if (ret)
> DRM_ERROR("RC6p residency sysfs setup failed\n");
> }
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> ret = sysfs_merge_group(&kdev->kobj,
> &media_rc6_attr_group);
> if (ret)
> DRM_ERROR("Media RC6 residency sysfs setup failed\n");
> }
> #endif
> - if (HAS_L3_DPF(dev_priv)) {
> + if (HAS_L3_DPF(i915)) {
> ret = device_create_bin_file(kdev, &dpf_attrs);
> if (ret)
> DRM_ERROR("l3 parity sysfs setup failed\n");
>
> - if (NUM_L3_SLICES(dev_priv) > 1) {
> + if (NUM_L3_SLICES(i915) > 1) {
> ret = device_create_bin_file(kdev,
> &dpf_attrs_1);
> if (ret)
> @@ -610,9 +610,9 @@ void i915_setup_sysfs(struct drm_i915_private *dev_priv)
> }
>
> ret = 0;
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> ret = sysfs_create_files(&kdev->kobj, vlv_attrs);
> - else if (INTEL_GEN(dev_priv) >= 6)
> + else if (INTEL_GEN(i915) >= 6)
> ret = sysfs_create_files(&kdev->kobj, gen6_attrs);
> if (ret)
> DRM_ERROR("RPS sysfs setup failed\n");
> @@ -620,13 +620,13 @@ void i915_setup_sysfs(struct drm_i915_private *dev_priv)
> i915_setup_error_capture(kdev);
> }
>
> -void i915_teardown_sysfs(struct drm_i915_private *dev_priv)
> +void i915_teardown_sysfs(struct drm_i915_private *i915)
> {
> - struct device *kdev = dev_priv->drm.primary->kdev;
> + struct device *kdev = i915->drm.primary->kdev;
>
> i915_teardown_error_capture(kdev);
>
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> sysfs_remove_files(&kdev->kobj, vlv_attrs);
> else
> sysfs_remove_files(&kdev->kobj, gen6_attrs);
> diff --git a/drivers/gpu/drm/i915/i915_trace.h b/drivers/gpu/drm/i915/i915_trace.h
> index 5c8cfaa70d72..9548b9fe4199 100644
> --- a/drivers/gpu/drm/i915/i915_trace.h
> +++ b/drivers/gpu/drm/i915/i915_trace.h
> @@ -21,8 +21,8 @@
> /* watermark/fifo updates */
>
> TRACE_EVENT(intel_pipe_enable,
> - TP_PROTO(struct drm_i915_private *dev_priv, enum pipe pipe),
> - TP_ARGS(dev_priv, pipe),
> + TP_PROTO(struct drm_i915_private *i915, enum pipe pipe),
> + TP_ARGS(i915, pipe),
>
> TP_STRUCT__entry(
> __array(u32, frame, 3)
> @@ -32,11 +32,11 @@ TRACE_EVENT(intel_pipe_enable,
>
> TP_fast_assign(
> enum pipe _pipe;
> - for_each_pipe(dev_priv, _pipe) {
> + for_each_pipe(i915, _pipe) {
> __entry->frame[_pipe] =
> - dev_priv->drm.driver->get_vblank_counter(&dev_priv->drm, _pipe);
> + i915->drm.driver->get_vblank_counter(&i915->drm, _pipe);
> __entry->scanline[_pipe] =
> - intel_get_crtc_scanline(intel_get_crtc_for_pipe(dev_priv, _pipe));
> + intel_get_crtc_scanline(intel_get_crtc_for_pipe(i915, _pipe));
> }
> __entry->pipe = pipe;
> ),
> @@ -49,8 +49,8 @@ TRACE_EVENT(intel_pipe_enable,
> );
>
> TRACE_EVENT(intel_pipe_disable,
> - TP_PROTO(struct drm_i915_private *dev_priv, enum pipe pipe),
> - TP_ARGS(dev_priv, pipe),
> + TP_PROTO(struct drm_i915_private *i915, enum pipe pipe),
> + TP_ARGS(i915, pipe),
>
> TP_STRUCT__entry(
> __array(u32, frame, 3)
> @@ -60,11 +60,11 @@ TRACE_EVENT(intel_pipe_disable,
>
> TP_fast_assign(
> enum pipe _pipe;
> - for_each_pipe(dev_priv, _pipe) {
> + for_each_pipe(i915, _pipe) {
> __entry->frame[_pipe] =
> - dev_priv->drm.driver->get_vblank_counter(&dev_priv->drm, _pipe);
> + i915->drm.driver->get_vblank_counter(&i915->drm, _pipe);
> __entry->scanline[_pipe] =
> - intel_get_crtc_scanline(intel_get_crtc_for_pipe(dev_priv, _pipe));
> + intel_get_crtc_scanline(intel_get_crtc_for_pipe(i915, _pipe));
> }
> __entry->pipe = pipe;
> ),
> @@ -102,8 +102,8 @@ TRACE_EVENT(intel_pipe_crc,
> );
>
> TRACE_EVENT(intel_cpu_fifo_underrun,
> - TP_PROTO(struct drm_i915_private *dev_priv, enum pipe pipe),
> - TP_ARGS(dev_priv, pipe),
> + TP_PROTO(struct drm_i915_private *i915, enum pipe pipe),
> + TP_ARGS(i915, pipe),
>
> TP_STRUCT__entry(
> __field(enum pipe, pipe)
> @@ -113,8 +113,8 @@ TRACE_EVENT(intel_cpu_fifo_underrun,
>
> TP_fast_assign(
> __entry->pipe = pipe;
> - __entry->frame = dev_priv->drm.driver->get_vblank_counter(&dev_priv->drm, pipe);
> - __entry->scanline = intel_get_crtc_scanline(intel_get_crtc_for_pipe(dev_priv, pipe));
> + __entry->frame = i915->drm.driver->get_vblank_counter(&i915->drm, pipe);
> + __entry->scanline = intel_get_crtc_scanline(intel_get_crtc_for_pipe(i915, pipe));
> ),
>
> TP_printk("pipe %c, frame=%u, scanline=%u",
> @@ -123,8 +123,8 @@ TRACE_EVENT(intel_cpu_fifo_underrun,
> );
>
> TRACE_EVENT(intel_pch_fifo_underrun,
> - TP_PROTO(struct drm_i915_private *dev_priv, enum pipe pch_transcoder),
> - TP_ARGS(dev_priv, pch_transcoder),
> + TP_PROTO(struct drm_i915_private *i915, enum pipe pch_transcoder),
> + TP_ARGS(i915, pch_transcoder),
>
> TP_STRUCT__entry(
> __field(enum pipe, pipe)
> @@ -135,8 +135,8 @@ TRACE_EVENT(intel_pch_fifo_underrun,
> TP_fast_assign(
> enum pipe pipe = pch_transcoder;
> __entry->pipe = pipe;
> - __entry->frame = dev_priv->drm.driver->get_vblank_counter(&dev_priv->drm, pipe);
> - __entry->scanline = intel_get_crtc_scanline(intel_get_crtc_for_pipe(dev_priv, pipe));
> + __entry->frame = i915->drm.driver->get_vblank_counter(&i915->drm, pipe);
> + __entry->scanline = intel_get_crtc_scanline(intel_get_crtc_for_pipe(i915, pipe));
> ),
>
> TP_printk("pch transcoder %c, frame=%u, scanline=%u",
> @@ -145,8 +145,8 @@ TRACE_EVENT(intel_pch_fifo_underrun,
> );
>
> TRACE_EVENT(intel_memory_cxsr,
> - TP_PROTO(struct drm_i915_private *dev_priv, bool old, bool new),
> - TP_ARGS(dev_priv, old, new),
> + TP_PROTO(struct drm_i915_private *i915, bool old, bool new),
> + TP_ARGS(i915, old, new),
>
> TP_STRUCT__entry(
> __array(u32, frame, 3)
> @@ -157,11 +157,11 @@ TRACE_EVENT(intel_memory_cxsr,
>
> TP_fast_assign(
> enum pipe pipe;
> - for_each_pipe(dev_priv, pipe) {
> + for_each_pipe(i915, pipe) {
> __entry->frame[pipe] =
> - dev_priv->drm.driver->get_vblank_counter(&dev_priv->drm, pipe);
> + i915->drm.driver->get_vblank_counter(&i915->drm, pipe);
> __entry->scanline[pipe] =
> - intel_get_crtc_scanline(intel_get_crtc_for_pipe(dev_priv, pipe));
> + intel_get_crtc_scanline(intel_get_crtc_for_pipe(i915, pipe));
> }
> __entry->old = old;
> __entry->new = new;
> diff --git a/drivers/gpu/drm/i915/i915_vgpu.c b/drivers/gpu/drm/i915/i915_vgpu.c
> index 94d3992b599d..96874a4819a9 100644
> --- a/drivers/gpu/drm/i915/i915_vgpu.c
> +++ b/drivers/gpu/drm/i915/i915_vgpu.c
> @@ -53,14 +53,14 @@
>
> /**
> * i915_check_vgpu - detect virtual GPU
> - * @dev_priv: i915 device private
> + * @i915: i915 device private
> *
> * This function is called at the initialization stage, to detect whether
> * running on a vGPU.
> */
> -void i915_check_vgpu(struct drm_i915_private *dev_priv)
> +void i915_check_vgpu(struct drm_i915_private *i915)
> {
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct intel_uncore *uncore = &i915->uncore;
> u64 magic;
> u16 version_major;
>
> @@ -76,15 +76,15 @@ void i915_check_vgpu(struct drm_i915_private *dev_priv)
> return;
> }
>
> - dev_priv->vgpu.caps = __raw_uncore_read32(uncore, vgtif_reg(vgt_caps));
> + i915->vgpu.caps = __raw_uncore_read32(uncore, vgtif_reg(vgt_caps));
>
> - dev_priv->vgpu.active = true;
> + i915->vgpu.active = true;
> DRM_INFO("Virtual GPU for Intel GVT-g detected.\n");
> }
>
> -bool intel_vgpu_has_full_ppgtt(struct drm_i915_private *dev_priv)
> +bool intel_vgpu_has_full_ppgtt(struct drm_i915_private *i915)
> {
> - return dev_priv->vgpu.caps & VGT_CAPS_FULL_PPGTT;
> + return i915->vgpu.caps & VGT_CAPS_FULL_PPGTT;
> }
>
> struct _balloon_info_ {
> @@ -112,22 +112,22 @@ static void vgt_deballoon_space(struct i915_ggtt *ggtt,
>
> /**
> * intel_vgt_deballoon - deballoon reserved graphics address trunks
> - * @dev_priv: i915 device private data
> + * @i915: i915 device private data
> *
> * This function is called to deallocate the ballooned-out graphic memory, when
> * driver is unloaded or when ballooning fails.
> */
> -void intel_vgt_deballoon(struct drm_i915_private *dev_priv)
> +void intel_vgt_deballoon(struct drm_i915_private *i915)
> {
> int i;
>
> - if (!intel_vgpu_active(dev_priv))
> + if (!intel_vgpu_active(i915))
> return;
>
> DRM_DEBUG("VGT deballoon.\n");
>
> for (i = 0; i < 4; i++)
> - vgt_deballoon_space(&dev_priv->ggtt, &bl_info.space[i]);
> + vgt_deballoon_space(&i915->ggtt, &bl_info.space[i]);
> }
>
> static int vgt_balloon_space(struct i915_ggtt *ggtt,
> @@ -153,7 +153,7 @@ static int vgt_balloon_space(struct i915_ggtt *ggtt,
>
> /**
> * intel_vgt_balloon - balloon out reserved graphics address trunks
> - * @dev_priv: i915 device private data
> + * @i915: i915 device private data
> *
> * This function is called at the initialization stage, to balloon out the
> * graphic address space allocated to other vGPUs, by marking these spaces as
> @@ -195,16 +195,16 @@ static int vgt_balloon_space(struct i915_ggtt *ggtt,
> * Returns:
> * zero on success, non-zero if configuration invalid or ballooning failed
> */
> -int intel_vgt_balloon(struct drm_i915_private *dev_priv)
> +int intel_vgt_balloon(struct drm_i915_private *i915)
> {
> - struct i915_ggtt *ggtt = &dev_priv->ggtt;
> + struct i915_ggtt *ggtt = &i915->ggtt;
> unsigned long ggtt_end = ggtt->vm.total;
>
> unsigned long mappable_base, mappable_size, mappable_end;
> unsigned long unmappable_base, unmappable_size, unmappable_end;
> int ret;
>
> - if (!intel_vgpu_active(dev_priv))
> + if (!intel_vgpu_active(i915))
> return 0;
>
> mappable_base = I915_READ(vgtif_reg(avail_rs.mappable_gmadr.base));
> diff --git a/drivers/gpu/drm/i915/i915_vgpu.h b/drivers/gpu/drm/i915/i915_vgpu.h
> index ebe1b7bced98..0fa0bfb88bed 100644
> --- a/drivers/gpu/drm/i915/i915_vgpu.h
> +++ b/drivers/gpu/drm/i915/i915_vgpu.h
> @@ -26,23 +26,23 @@
>
> #include "i915_pvinfo.h"
>
> -void i915_check_vgpu(struct drm_i915_private *dev_priv);
> +void i915_check_vgpu(struct drm_i915_private *i915);
>
> -bool intel_vgpu_has_full_ppgtt(struct drm_i915_private *dev_priv);
> +bool intel_vgpu_has_full_ppgtt(struct drm_i915_private *i915);
>
> static inline bool
> -intel_vgpu_has_hwsp_emulation(struct drm_i915_private *dev_priv)
> +intel_vgpu_has_hwsp_emulation(struct drm_i915_private *i915)
> {
> - return dev_priv->vgpu.caps & VGT_CAPS_HWSP_EMULATION;
> + return i915->vgpu.caps & VGT_CAPS_HWSP_EMULATION;
> }
>
> static inline bool
> -intel_vgpu_has_huge_gtt(struct drm_i915_private *dev_priv)
> +intel_vgpu_has_huge_gtt(struct drm_i915_private *i915)
> {
> - return dev_priv->vgpu.caps & VGT_CAPS_HUGE_GTT;
> + return i915->vgpu.caps & VGT_CAPS_HUGE_GTT;
> }
>
> -int intel_vgt_balloon(struct drm_i915_private *dev_priv);
> -void intel_vgt_deballoon(struct drm_i915_private *dev_priv);
> +int intel_vgt_balloon(struct drm_i915_private *i915);
> +void intel_vgt_deballoon(struct drm_i915_private *i915);
>
> #endif /* _I915_VGPU_H_ */
> diff --git a/drivers/gpu/drm/i915/i915_vma.c b/drivers/gpu/drm/i915/i915_vma.c
> index cb341e4acf99..e3f4d7006a1b 100644
> --- a/drivers/gpu/drm/i915/i915_vma.c
> +++ b/drivers/gpu/drm/i915/i915_vma.c
> @@ -555,7 +555,7 @@ static void assert_bind_count(const struct drm_i915_gem_object *obj)
> static int
> i915_vma_insert(struct i915_vma *vma, u64 size, u64 alignment, u64 flags)
> {
> - struct drm_i915_private *dev_priv = vma->vm->i915;
> + struct drm_i915_private *i915 = vma->vm->i915;
> unsigned int cache_level;
> u64 start, end;
> int ret;
> @@ -581,7 +581,7 @@ i915_vma_insert(struct i915_vma *vma, u64 size, u64 alignment, u64 flags)
>
> end = vma->vm->total;
> if (flags & PIN_MAPPABLE)
> - end = min_t(u64, end, dev_priv->ggtt.mappable_end);
> + end = min_t(u64, end, i915->ggtt.mappable_end);
> if (flags & PIN_ZONE_4G)
> end = min_t(u64, end, (1ULL << 32) - I915_GTT_PAGE_SIZE);
> GEM_BUG_ON(!IS_ALIGNED(end, I915_GTT_PAGE_SIZE));
> diff --git a/drivers/gpu/drm/i915/icl_dsi.c b/drivers/gpu/drm/i915/icl_dsi.c
> index 74448e6bf749..d0192e093f6f 100644
> --- a/drivers/gpu/drm/i915/icl_dsi.c
> +++ b/drivers/gpu/drm/i915/icl_dsi.c
> @@ -35,32 +35,32 @@
> #include "intel_dsi.h"
> #include "intel_panel.h"
>
> -static inline int header_credits_available(struct drm_i915_private *dev_priv,
> +static inline int header_credits_available(struct drm_i915_private *i915,
> enum transcoder dsi_trans)
> {
> return (I915_READ(DSI_CMD_TXCTL(dsi_trans)) & FREE_HEADER_CREDIT_MASK)
> >> FREE_HEADER_CREDIT_SHIFT;
> }
>
> -static inline int payload_credits_available(struct drm_i915_private *dev_priv,
> +static inline int payload_credits_available(struct drm_i915_private *i915,
> enum transcoder dsi_trans)
> {
> return (I915_READ(DSI_CMD_TXCTL(dsi_trans)) & FREE_PLOAD_CREDIT_MASK)
> >> FREE_PLOAD_CREDIT_SHIFT;
> }
>
> -static void wait_for_header_credits(struct drm_i915_private *dev_priv,
> +static void wait_for_header_credits(struct drm_i915_private *i915,
> enum transcoder dsi_trans)
> {
> - if (wait_for_us(header_credits_available(dev_priv, dsi_trans) >=
> + if (wait_for_us(header_credits_available(i915, dsi_trans) >=
> MAX_HEADER_CREDIT, 100))
> DRM_ERROR("DSI header credits not released\n");
> }
>
> -static void wait_for_payload_credits(struct drm_i915_private *dev_priv,
> +static void wait_for_payload_credits(struct drm_i915_private *i915,
> enum transcoder dsi_trans)
> {
> - if (wait_for_us(payload_credits_available(dev_priv, dsi_trans) >=
> + if (wait_for_us(payload_credits_available(i915, dsi_trans) >=
> MAX_PLOAD_CREDIT, 100))
> DRM_ERROR("DSI payload credits not released\n");
> }
> @@ -75,7 +75,7 @@ static enum transcoder dsi_port_to_transcoder(enum port port)
>
> static void wait_for_cmds_dispatched_to_panel(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> struct mipi_dsi_device *dsi;
> enum port port;
> @@ -85,8 +85,8 @@ static void wait_for_cmds_dispatched_to_panel(struct intel_encoder *encoder)
> /* wait for header/payload credits to be released */
> for_each_dsi_port(port, intel_dsi->ports) {
> dsi_trans = dsi_port_to_transcoder(port);
> - wait_for_header_credits(dev_priv, dsi_trans);
> - wait_for_payload_credits(dev_priv, dsi_trans);
> + wait_for_header_credits(i915, dsi_trans);
> + wait_for_payload_credits(i915, dsi_trans);
> }
>
> /* send nop DCS command */
> @@ -102,7 +102,7 @@ static void wait_for_cmds_dispatched_to_panel(struct intel_encoder *encoder)
> /* wait for header credits to be released */
> for_each_dsi_port(port, intel_dsi->ports) {
> dsi_trans = dsi_port_to_transcoder(port);
> - wait_for_header_credits(dev_priv, dsi_trans);
> + wait_for_header_credits(i915, dsi_trans);
> }
>
> /* wait for LP TX in progress bit to be cleared */
> @@ -118,7 +118,7 @@ static bool add_payld_to_queue(struct intel_dsi_host *host, const u8 *data,
> u32 len)
> {
> struct intel_dsi *intel_dsi = host->intel_dsi;
> - struct drm_i915_private *dev_priv = to_i915(intel_dsi->base.base.dev);
> + struct drm_i915_private *i915 = to_i915(intel_dsi->base.base.dev);
> enum transcoder dsi_trans = dsi_port_to_transcoder(host->port);
> int free_credits;
> int i, j;
> @@ -126,7 +126,7 @@ static bool add_payld_to_queue(struct intel_dsi_host *host, const u8 *data,
> for (i = 0; i < len; i += 4) {
> u32 tmp = 0;
>
> - free_credits = payload_credits_available(dev_priv, dsi_trans);
> + free_credits = payload_credits_available(i915, dsi_trans);
> if (free_credits < 1) {
> DRM_ERROR("Payload credit not available\n");
> return false;
> @@ -145,13 +145,13 @@ static int dsi_send_pkt_hdr(struct intel_dsi_host *host,
> struct mipi_dsi_packet pkt, bool enable_lpdt)
> {
> struct intel_dsi *intel_dsi = host->intel_dsi;
> - struct drm_i915_private *dev_priv = to_i915(intel_dsi->base.base.dev);
> + struct drm_i915_private *i915 = to_i915(intel_dsi->base.base.dev);
> enum transcoder dsi_trans = dsi_port_to_transcoder(host->port);
> u32 tmp;
> int free_credits;
>
> /* check if header credit available */
> - free_credits = header_credits_available(dev_priv, dsi_trans);
> + free_credits = header_credits_available(i915, dsi_trans);
> if (free_credits < 1) {
> DRM_ERROR("send pkt header failed, not enough hdr credits\n");
> return -1;
> @@ -200,7 +200,7 @@ static int dsi_send_pkt_payld(struct intel_dsi_host *host,
>
> static void dsi_program_swing_and_deemphasis(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> enum port port;
> u32 tmp;
> @@ -266,7 +266,7 @@ static void dsi_program_swing_and_deemphasis(struct intel_encoder *encoder)
> static void configure_dual_link_mode(struct intel_encoder *encoder,
> const struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> u32 dss_ctl1;
>
> @@ -304,7 +304,7 @@ static void configure_dual_link_mode(struct intel_encoder *encoder,
>
> static void gen11_dsi_program_esc_clk_div(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> enum port port;
> u32 bpp = mipi_dsi_pixel_format_to_bpp(intel_dsi->pixel_format);
> @@ -329,7 +329,7 @@ static void gen11_dsi_program_esc_clk_div(struct intel_encoder *encoder)
> }
> }
>
> -static void get_dsi_io_power_domains(struct drm_i915_private *dev_priv,
> +static void get_dsi_io_power_domains(struct drm_i915_private *i915,
> struct intel_dsi *intel_dsi)
> {
> enum port port;
> @@ -337,7 +337,7 @@ static void get_dsi_io_power_domains(struct drm_i915_private *dev_priv,
> for_each_dsi_port(port, intel_dsi->ports) {
> WARN_ON(intel_dsi->io_wakeref[port]);
> intel_dsi->io_wakeref[port] =
> - intel_display_power_get(dev_priv,
> + intel_display_power_get(i915,
> port == PORT_A ?
> POWER_DOMAIN_PORT_DDI_A_IO :
> POWER_DOMAIN_PORT_DDI_B_IO);
> @@ -346,7 +346,7 @@ static void get_dsi_io_power_domains(struct drm_i915_private *dev_priv,
>
> static void gen11_dsi_enable_io_power(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> enum port port;
> u32 tmp;
> @@ -357,23 +357,23 @@ static void gen11_dsi_enable_io_power(struct intel_encoder *encoder)
> I915_WRITE(ICL_DSI_IO_MODECTL(port), tmp);
> }
>
> - get_dsi_io_power_domains(dev_priv, intel_dsi);
> + get_dsi_io_power_domains(i915, intel_dsi);
> }
>
> static void gen11_dsi_power_up_lanes(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> enum port port;
>
> for_each_dsi_port(port, intel_dsi->ports)
> - intel_combo_phy_power_up_lanes(dev_priv, port, true,
> + intel_combo_phy_power_up_lanes(i915, port, true,
> intel_dsi->lane_count, false);
> }
>
> static void gen11_dsi_config_phy_lanes_sequence(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> enum port port;
> u32 tmp;
> @@ -409,7 +409,7 @@ static void gen11_dsi_config_phy_lanes_sequence(struct intel_encoder *encoder)
>
> static void gen11_dsi_voltage_swing_program_seq(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> u32 tmp;
> enum port port;
> @@ -461,7 +461,7 @@ static void gen11_dsi_voltage_swing_program_seq(struct intel_encoder *encoder)
>
> static void gen11_dsi_enable_ddi_buffer(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> u32 tmp;
> enum port port;
> @@ -480,7 +480,7 @@ static void gen11_dsi_enable_ddi_buffer(struct intel_encoder *encoder)
>
> static void gen11_dsi_setup_dphy_timings(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> u32 tmp;
> enum port port;
> @@ -535,48 +535,48 @@ static void gen11_dsi_setup_dphy_timings(struct intel_encoder *encoder)
>
> static void gen11_dsi_gate_clocks(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> u32 tmp;
> enum port port;
>
> - mutex_lock(&dev_priv->dpll_lock);
> + mutex_lock(&i915->dpll_lock);
> tmp = I915_READ(DPCLKA_CFGCR0_ICL);
> for_each_dsi_port(port, intel_dsi->ports) {
> tmp |= DPCLKA_CFGCR0_DDI_CLK_OFF(port);
> }
>
> I915_WRITE(DPCLKA_CFGCR0_ICL, tmp);
> - mutex_unlock(&dev_priv->dpll_lock);
> + mutex_unlock(&i915->dpll_lock);
> }
>
> static void gen11_dsi_ungate_clocks(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> u32 tmp;
> enum port port;
>
> - mutex_lock(&dev_priv->dpll_lock);
> + mutex_lock(&i915->dpll_lock);
> tmp = I915_READ(DPCLKA_CFGCR0_ICL);
> for_each_dsi_port(port, intel_dsi->ports) {
> tmp &= ~DPCLKA_CFGCR0_DDI_CLK_OFF(port);
> }
>
> I915_WRITE(DPCLKA_CFGCR0_ICL, tmp);
> - mutex_unlock(&dev_priv->dpll_lock);
> + mutex_unlock(&i915->dpll_lock);
> }
>
> static void gen11_dsi_map_pll(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> struct intel_shared_dpll *pll = crtc_state->shared_dpll;
> enum port port;
> u32 val;
>
> - mutex_lock(&dev_priv->dpll_lock);
> + mutex_lock(&i915->dpll_lock);
>
> val = I915_READ(DPCLKA_CFGCR0_ICL);
> for_each_dsi_port(port, intel_dsi->ports) {
> @@ -592,14 +592,14 @@ static void gen11_dsi_map_pll(struct intel_encoder *encoder,
>
> POSTING_READ(DPCLKA_CFGCR0_ICL);
>
> - mutex_unlock(&dev_priv->dpll_lock);
> + mutex_unlock(&i915->dpll_lock);
> }
>
> static void
> gen11_dsi_configure_transcoder(struct intel_encoder *encoder,
> const struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> struct intel_crtc *intel_crtc = to_intel_crtc(pipe_config->base.crtc);
> enum pipe pipe = intel_crtc->pipe;
> @@ -736,7 +736,7 @@ static void
> gen11_dsi_set_transcoder_timings(struct intel_encoder *encoder,
> const struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> const struct drm_display_mode *adjusted_mode =
> &pipe_config->base.adjusted_mode;
> @@ -849,7 +849,7 @@ gen11_dsi_set_transcoder_timings(struct intel_encoder *encoder,
>
> static void gen11_dsi_enable_transcoder(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> enum port port;
> enum transcoder dsi_trans;
> @@ -862,7 +862,7 @@ static void gen11_dsi_enable_transcoder(struct intel_encoder *encoder)
> I915_WRITE(PIPECONF(dsi_trans), tmp);
>
> /* wait for transcoder to be enabled */
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> PIPECONF(dsi_trans),
> I965_PIPECONF_ACTIVE,
> I965_PIPECONF_ACTIVE, 10))
> @@ -872,7 +872,7 @@ static void gen11_dsi_enable_transcoder(struct intel_encoder *encoder)
>
> static void gen11_dsi_setup_timeouts(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> enum port port;
> enum transcoder dsi_trans;
> @@ -950,7 +950,7 @@ gen11_dsi_enable_port_and_phy(struct intel_encoder *encoder,
>
> static void gen11_dsi_powerup_panel(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> struct mipi_dsi_device *dsi;
> enum port port;
> @@ -1026,7 +1026,7 @@ static void gen11_dsi_pre_enable(struct intel_encoder *encoder,
>
> static void gen11_dsi_disable_transcoder(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> enum port port;
> enum transcoder dsi_trans;
> @@ -1041,7 +1041,7 @@ static void gen11_dsi_disable_transcoder(struct intel_encoder *encoder)
> I915_WRITE(PIPECONF(dsi_trans), tmp);
>
> /* wait for transcoder to be disabled */
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> PIPECONF(dsi_trans),
> I965_PIPECONF_ACTIVE, 0, 50))
> DRM_ERROR("DSI trancoder not disabled\n");
> @@ -1062,7 +1062,7 @@ static void gen11_dsi_powerdown_panel(struct intel_encoder *encoder)
>
> static void gen11_dsi_deconfigure_trancoder(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> enum port port;
> enum transcoder dsi_trans;
> @@ -1103,7 +1103,7 @@ static void gen11_dsi_deconfigure_trancoder(struct intel_encoder *encoder)
>
> static void gen11_dsi_disable_port(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> u32 tmp;
> enum port port;
> @@ -1125,7 +1125,7 @@ static void gen11_dsi_disable_port(struct intel_encoder *encoder)
>
> static void gen11_dsi_disable_io_power(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> enum port port;
> u32 tmp;
> @@ -1134,7 +1134,7 @@ static void gen11_dsi_disable_io_power(struct intel_encoder *encoder)
> intel_wakeref_t wakeref;
>
> wakeref = fetch_and_zero(&intel_dsi->io_wakeref[port]);
> - intel_display_power_put(dev_priv,
> + intel_display_power_put(i915,
> port == PORT_A ?
> POWER_DOMAIN_PORT_DDI_A_IO :
> POWER_DOMAIN_PORT_DDI_B_IO,
> @@ -1205,13 +1205,13 @@ static void gen11_dsi_get_timings(struct intel_encoder *encoder,
> static void gen11_dsi_get_config(struct intel_encoder *encoder,
> struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crtc *crtc = to_intel_crtc(pipe_config->base.crtc);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
>
> /* FIXME: adapt icl_ddi_clock_get() for DSI and use that? */
> pipe_config->port_clock =
> - cnl_calc_wrpll_link(dev_priv, &pipe_config->dpll_hw_state);
> + cnl_calc_wrpll_link(i915, &pipe_config->dpll_hw_state);
>
> pipe_config->base.adjusted_mode.crtc_clock = intel_dsi->pclk;
> if (intel_dsi->dual_link)
> @@ -1263,7 +1263,7 @@ static void gen11_dsi_get_power_domains(struct intel_encoder *encoder,
> static bool gen11_dsi_get_hw_state(struct intel_encoder *encoder,
> enum pipe *pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> enum transcoder dsi_trans;
> intel_wakeref_t wakeref;
> @@ -1271,7 +1271,7 @@ static bool gen11_dsi_get_hw_state(struct intel_encoder *encoder,
> bool ret = false;
> u32 tmp;
>
> - wakeref = intel_display_power_get_if_enabled(dev_priv,
> + wakeref = intel_display_power_get_if_enabled(i915,
> encoder->power_domain);
> if (!wakeref)
> return false;
> @@ -1298,7 +1298,7 @@ static bool gen11_dsi_get_hw_state(struct intel_encoder *encoder,
> ret = tmp & PIPECONF_ENABLE;
> }
> out:
> - intel_display_power_put(dev_priv, encoder->power_domain, wakeref);
> + intel_display_power_put(i915, encoder->power_domain, wakeref);
> return ret;
> }
>
> @@ -1391,8 +1391,8 @@ static const struct mipi_dsi_host_ops gen11_dsi_host_ops = {
> static void icl_dphy_param_init(struct intel_dsi *intel_dsi)
> {
> struct drm_device *dev = intel_dsi->base.base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct mipi_config *mipi_config = dev_priv->vbt.dsi.config;
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct mipi_config *mipi_config = i915->vbt.dsi.config;
> u32 tlpx_ns;
> u32 prepare_cnt, exit_zero_cnt, clk_zero_cnt, trail_cnt;
> u32 ths_prepare_ns, tclk_trail_ns;
> @@ -1487,9 +1487,9 @@ static void icl_dphy_param_init(struct intel_dsi *intel_dsi)
> intel_dsi_log_params(intel_dsi);
> }
>
> -void icl_dsi_init(struct drm_i915_private *dev_priv)
> +void icl_dsi_init(struct drm_i915_private *i915)
> {
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_device *dev = &i915->drm;
> struct intel_dsi *intel_dsi;
> struct intel_encoder *encoder;
> struct intel_connector *intel_connector;
> @@ -1497,7 +1497,7 @@ void icl_dsi_init(struct drm_i915_private *dev_priv)
> struct drm_display_mode *fixed_mode;
> enum port port;
>
> - if (!intel_bios_is_dsi_present(dev_priv, &port))
> + if (!intel_bios_is_dsi_present(i915, &port))
> return;
>
> intel_dsi = kzalloc(sizeof(*intel_dsi), GFP_KERNEL);
> @@ -1556,13 +1556,13 @@ void icl_dsi_init(struct drm_i915_private *dev_priv)
> intel_panel_init(&intel_connector->panel, fixed_mode, NULL);
> intel_panel_setup_backlight(connector, INVALID_PIPE);
>
> - if (dev_priv->vbt.dsi.config->dual_link)
> + if (i915->vbt.dsi.config->dual_link)
> intel_dsi->ports = BIT(PORT_A) | BIT(PORT_B);
> else
> intel_dsi->ports = BIT(port);
>
> - intel_dsi->dcs_backlight_ports = dev_priv->vbt.dsi.bl_ports;
> - intel_dsi->dcs_cabc_ports = dev_priv->vbt.dsi.cabc_ports;
> + intel_dsi->dcs_backlight_ports = i915->vbt.dsi.bl_ports;
> + intel_dsi->dcs_cabc_ports = i915->vbt.dsi.cabc_ports;
>
> for_each_dsi_port(port, intel_dsi->ports) {
> struct intel_dsi_host *host;
> diff --git a/drivers/gpu/drm/i915/intel_atomic.c b/drivers/gpu/drm/i915/intel_atomic.c
> index 6b985e895a97..45782e6ff0f4 100644
> --- a/drivers/gpu/drm/i915/intel_atomic.c
> +++ b/drivers/gpu/drm/i915/intel_atomic.c
> @@ -54,13 +54,13 @@ int intel_digital_connector_atomic_get_property(struct drm_connector *connector,
> u64 *val)
> {
> struct drm_device *dev = connector->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_digital_connector_state *intel_conn_state =
> to_intel_digital_connector_state(state);
>
> - if (property == dev_priv->force_audio_property)
> + if (property == i915->force_audio_property)
> *val = intel_conn_state->force_audio;
> - else if (property == dev_priv->broadcast_rgb_property)
> + else if (property == i915->broadcast_rgb_property)
> *val = intel_conn_state->broadcast_rgb;
> else {
> DRM_DEBUG_ATOMIC("Unknown property [PROP:%d:%s]\n",
> @@ -86,16 +86,16 @@ int intel_digital_connector_atomic_set_property(struct drm_connector *connector,
> u64 val)
> {
> struct drm_device *dev = connector->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_digital_connector_state *intel_conn_state =
> to_intel_digital_connector_state(state);
>
> - if (property == dev_priv->force_audio_property) {
> + if (property == i915->force_audio_property) {
> intel_conn_state->force_audio = val;
> return 0;
> }
>
> - if (property == dev_priv->broadcast_rgb_property) {
> + if (property == i915->broadcast_rgb_property) {
> intel_conn_state->broadcast_rgb = val;
> return 0;
> }
> @@ -227,7 +227,7 @@ static void intel_atomic_setup_scaler(struct intel_crtc_scaler_state *scaler_sta
> struct intel_plane_state *plane_state,
> int *scaler_id)
> {
> - struct drm_i915_private *dev_priv = to_i915(intel_crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(intel_crtc->base.dev);
> int j;
> u32 mode;
>
> @@ -251,10 +251,10 @@ static void intel_atomic_setup_scaler(struct intel_crtc_scaler_state *scaler_sta
> plane_state->base.fb->format->is_yuv &&
> plane_state->base.fb->format->num_planes > 1) {
> struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> - if (IS_GEN(dev_priv, 9) &&
> - !IS_GEMINILAKE(dev_priv)) {
> + if (IS_GEN(i915, 9) &&
> + !IS_GEMINILAKE(i915)) {
> mode = SKL_PS_SCALER_MODE_NV12;
> - } else if (icl_is_hdr_plane(dev_priv, plane->id)) {
> + } else if (icl_is_hdr_plane(i915, plane->id)) {
> /*
> * On gen11+'s HDR planes we only use the scaler for
> * scaling. They have a dedicated chroma upsampler, so
> @@ -267,7 +267,7 @@ static void intel_atomic_setup_scaler(struct intel_crtc_scaler_state *scaler_sta
> if (plane_state->linked_plane)
> mode |= PS_PLANE_Y_SEL(plane_state->linked_plane->id);
> }
> - } else if (INTEL_GEN(dev_priv) > 9 || IS_GEMINILAKE(dev_priv)) {
> + } else if (INTEL_GEN(i915) > 9 || IS_GEMINILAKE(i915)) {
> mode = PS_SCALER_MODE_NORMAL;
> } else if (num_scalers_need == 1 && intel_crtc->num_scalers > 1) {
> /*
> @@ -290,7 +290,7 @@ static void intel_atomic_setup_scaler(struct intel_crtc_scaler_state *scaler_sta
>
> /**
> * intel_atomic_setup_scalers() - setup scalers for crtc per staged requests
> - * @dev_priv: i915 device
> + * @i915: i915 device
> * @intel_crtc: intel crtc
> * @crtc_state: incoming crtc_state to validate and setup scalers
> *
> @@ -305,7 +305,7 @@ static void intel_atomic_setup_scaler(struct intel_crtc_scaler_state *scaler_sta
> * 0 - scalers were setup succesfully
> * error code - otherwise
> */
> -int intel_atomic_setup_scalers(struct drm_i915_private *dev_priv,
> +int intel_atomic_setup_scalers(struct drm_i915_private *i915,
> struct intel_crtc *intel_crtc,
> struct intel_crtc_state *crtc_state)
> {
> @@ -369,7 +369,7 @@ int intel_atomic_setup_scalers(struct drm_i915_private *dev_priv,
> */
> if (!plane) {
> struct drm_plane_state *state;
> - plane = drm_plane_from_index(&dev_priv->drm, i);
> + plane = drm_plane_from_index(&i915->drm, i);
> state = drm_atomic_get_plane_state(drm_state, plane);
> if (IS_ERR(state)) {
> DRM_DEBUG_KMS("Failed to add [PLANE:%d] to drm_state\n",
> diff --git a/drivers/gpu/drm/i915/intel_atomic.h b/drivers/gpu/drm/i915/intel_atomic.h
> index 1c8507da1a69..5482fd04d639 100644
> --- a/drivers/gpu/drm/i915/intel_atomic.h
> +++ b/drivers/gpu/drm/i915/intel_atomic.h
> @@ -42,7 +42,7 @@ struct intel_crtc_state *
> intel_atomic_get_crtc_state(struct drm_atomic_state *state,
> struct intel_crtc *crtc);
>
> -int intel_atomic_setup_scalers(struct drm_i915_private *dev_priv,
> +int intel_atomic_setup_scalers(struct drm_i915_private *i915,
> struct intel_crtc *intel_crtc,
> struct intel_crtc_state *crtc_state);
>
> diff --git a/drivers/gpu/drm/i915/intel_audio.c b/drivers/gpu/drm/i915/intel_audio.c
> index 840daff12246..f463d8452fe9 100644
> --- a/drivers/gpu/drm/i915/intel_audio.c
> +++ b/drivers/gpu/drm/i915/intel_audio.c
> @@ -230,7 +230,7 @@ static bool intel_eld_uptodate(struct drm_connector *connector,
> i915_reg_t reg_elda, u32 bits_elda,
> i915_reg_t reg_edid)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->dev);
> + struct drm_i915_private *i915 = to_i915(connector->dev);
> const u8 *eld = connector->eld;
> u32 tmp;
> int i;
> @@ -256,7 +256,7 @@ static void g4x_audio_codec_disable(struct intel_encoder *encoder,
> const struct intel_crtc_state *old_crtc_state,
> const struct drm_connector_state *old_conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> u32 eldv, tmp;
>
> DRM_DEBUG_KMS("Disable audio codec\n");
> @@ -277,7 +277,7 @@ static void g4x_audio_codec_enable(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state,
> const struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct drm_connector *connector = conn_state->connector;
> const u8 *eld = connector->eld;
> u32 eldv;
> @@ -317,8 +317,8 @@ static void
> hsw_dp_audio_config_update(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> - struct i915_audio_component *acomp = dev_priv->audio_component;
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> + struct i915_audio_component *acomp = i915->audio_component;
> enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> enum port port = encoder->port;
> const struct dp_aud_n_m *nm;
> @@ -364,8 +364,8 @@ static void
> hsw_hdmi_audio_config_update(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> - struct i915_audio_component *acomp = dev_priv->audio_component;
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> + struct i915_audio_component *acomp = i915->audio_component;
> enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> enum port port = encoder->port;
> int n, rate;
> @@ -416,14 +416,14 @@ static void hsw_audio_codec_disable(struct intel_encoder *encoder,
> const struct intel_crtc_state *old_crtc_state,
> const struct drm_connector_state *old_conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> enum transcoder cpu_transcoder = old_crtc_state->cpu_transcoder;
> u32 tmp;
>
> DRM_DEBUG_KMS("Disable audio codec on transcoder %s\n",
> transcoder_name(cpu_transcoder));
>
> - mutex_lock(&dev_priv->av_mutex);
> + mutex_lock(&i915->av_mutex);
>
> /* Disable timestamps */
> tmp = I915_READ(HSW_AUD_CFG(cpu_transcoder));
> @@ -441,14 +441,14 @@ static void hsw_audio_codec_disable(struct intel_encoder *encoder,
> tmp &= ~AUDIO_OUTPUT_ENABLE(cpu_transcoder);
> I915_WRITE(HSW_AUD_PIN_ELD_CP_VLD, tmp);
>
> - mutex_unlock(&dev_priv->av_mutex);
> + mutex_unlock(&i915->av_mutex);
> }
>
> static void hsw_audio_codec_enable(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state,
> const struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct drm_connector *connector = conn_state->connector;
> enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> const u8 *eld = connector->eld;
> @@ -458,7 +458,7 @@ static void hsw_audio_codec_enable(struct intel_encoder *encoder,
> DRM_DEBUG_KMS("Enable audio codec on transcoder %s, %u bytes ELD\n",
> transcoder_name(cpu_transcoder), drm_eld_size(eld));
>
> - mutex_lock(&dev_priv->av_mutex);
> + mutex_lock(&i915->av_mutex);
>
> /* Enable audio presence detect, invalidate ELD */
> tmp = I915_READ(HSW_AUD_PIN_ELD_CP_VLD);
> @@ -491,14 +491,14 @@ static void hsw_audio_codec_enable(struct intel_encoder *encoder,
> /* Enable timestamps */
> hsw_audio_config_update(encoder, crtc_state);
>
> - mutex_unlock(&dev_priv->av_mutex);
> + mutex_unlock(&i915->av_mutex);
> }
>
> static void ilk_audio_codec_disable(struct intel_encoder *encoder,
> const struct intel_crtc_state *old_crtc_state,
> const struct drm_connector_state *old_conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc);
> enum pipe pipe = crtc->pipe;
> enum port port = encoder->port;
> @@ -511,10 +511,10 @@ static void ilk_audio_codec_disable(struct intel_encoder *encoder,
> if (WARN_ON(port == PORT_A))
> return;
>
> - if (HAS_PCH_IBX(dev_priv)) {
> + if (HAS_PCH_IBX(i915)) {
> aud_config = IBX_AUD_CFG(pipe);
> aud_cntrl_st2 = IBX_AUD_CNTL_ST2;
> - } else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> + } else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> aud_config = VLV_AUD_CFG(pipe);
> aud_cntrl_st2 = VLV_AUD_CNTL_ST2;
> } else {
> @@ -544,7 +544,7 @@ static void ilk_audio_codec_enable(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state,
> const struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> struct drm_connector *connector = conn_state->connector;
> enum pipe pipe = crtc->pipe;
> @@ -567,13 +567,13 @@ static void ilk_audio_codec_enable(struct intel_encoder *encoder,
> * infrastructure is not there yet.
> */
>
> - if (HAS_PCH_IBX(dev_priv)) {
> + if (HAS_PCH_IBX(i915)) {
> hdmiw_hdmiedid = IBX_HDMIW_HDMIEDID(pipe);
> aud_config = IBX_AUD_CFG(pipe);
> aud_cntl_st = IBX_AUD_CNTL_ST(pipe);
> aud_cntrl_st2 = IBX_AUD_CNTL_ST2;
> - } else if (IS_VALLEYVIEW(dev_priv) ||
> - IS_CHERRYVIEW(dev_priv)) {
> + } else if (IS_VALLEYVIEW(i915) ||
> + IS_CHERRYVIEW(i915)) {
> hdmiw_hdmiedid = VLV_HDMIW_HDMIEDID(pipe);
> aud_config = VLV_AUD_CFG(pipe);
> aud_cntl_st = VLV_AUD_CNTL_ST(pipe);
> @@ -632,8 +632,8 @@ void intel_audio_codec_enable(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state,
> const struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> - struct i915_audio_component *acomp = dev_priv->audio_component;
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> + struct i915_audio_component *acomp = i915->audio_component;
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> struct drm_connector *connector = conn_state->connector;
> const struct drm_display_mode *adjusted_mode =
> @@ -654,17 +654,17 @@ void intel_audio_codec_enable(struct intel_encoder *encoder,
>
> connector->eld[6] = drm_av_sync_delay(connector, adjusted_mode) / 2;
>
> - if (dev_priv->display.audio_codec_enable)
> - dev_priv->display.audio_codec_enable(encoder,
> + if (i915->display.audio_codec_enable)
> + i915->display.audio_codec_enable(encoder,
> crtc_state,
> conn_state);
>
> - mutex_lock(&dev_priv->av_mutex);
> + mutex_lock(&i915->av_mutex);
> encoder->audio_connector = connector;
>
> /* referred in audio callbacks */
> - dev_priv->av_enc_map[pipe] = encoder;
> - mutex_unlock(&dev_priv->av_mutex);
> + i915->av_enc_map[pipe] = encoder;
> + mutex_unlock(&i915->av_mutex);
>
> if (acomp && acomp->base.audio_ops &&
> acomp->base.audio_ops->pin_eld_notify) {
> @@ -675,7 +675,7 @@ void intel_audio_codec_enable(struct intel_encoder *encoder,
> (int) port, (int) pipe);
> }
>
> - intel_lpe_audio_notify(dev_priv, pipe, port, connector->eld,
> + intel_lpe_audio_notify(i915, pipe, port, connector->eld,
> crtc_state->port_clock,
> intel_crtc_has_dp_encoder(crtc_state));
> }
> @@ -693,21 +693,21 @@ void intel_audio_codec_disable(struct intel_encoder *encoder,
> const struct intel_crtc_state *old_crtc_state,
> const struct drm_connector_state *old_conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> - struct i915_audio_component *acomp = dev_priv->audio_component;
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> + struct i915_audio_component *acomp = i915->audio_component;
> struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc);
> enum port port = encoder->port;
> enum pipe pipe = crtc->pipe;
>
> - if (dev_priv->display.audio_codec_disable)
> - dev_priv->display.audio_codec_disable(encoder,
> + if (i915->display.audio_codec_disable)
> + i915->display.audio_codec_disable(encoder,
> old_crtc_state,
> old_conn_state);
>
> - mutex_lock(&dev_priv->av_mutex);
> + mutex_lock(&i915->av_mutex);
> encoder->audio_connector = NULL;
> - dev_priv->av_enc_map[pipe] = NULL;
> - mutex_unlock(&dev_priv->av_mutex);
> + i915->av_enc_map[pipe] = NULL;
> + mutex_unlock(&i915->av_mutex);
>
> if (acomp && acomp->base.audio_ops &&
> acomp->base.audio_ops->pin_eld_notify) {
> @@ -718,31 +718,31 @@ void intel_audio_codec_disable(struct intel_encoder *encoder,
> (int) port, (int) pipe);
> }
>
> - intel_lpe_audio_notify(dev_priv, pipe, port, NULL, 0, false);
> + intel_lpe_audio_notify(i915, pipe, port, NULL, 0, false);
> }
>
> /**
> * intel_init_audio_hooks - Set up chip specific audio hooks
> - * @dev_priv: device private
> + * @i915: device private
> */
> -void intel_init_audio_hooks(struct drm_i915_private *dev_priv)
> +void intel_init_audio_hooks(struct drm_i915_private *i915)
> {
> - if (IS_G4X(dev_priv)) {
> - dev_priv->display.audio_codec_enable = g4x_audio_codec_enable;
> - dev_priv->display.audio_codec_disable = g4x_audio_codec_disable;
> - } else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> - dev_priv->display.audio_codec_enable = ilk_audio_codec_enable;
> - dev_priv->display.audio_codec_disable = ilk_audio_codec_disable;
> - } else if (IS_HASWELL(dev_priv) || INTEL_GEN(dev_priv) >= 8) {
> - dev_priv->display.audio_codec_enable = hsw_audio_codec_enable;
> - dev_priv->display.audio_codec_disable = hsw_audio_codec_disable;
> - } else if (HAS_PCH_SPLIT(dev_priv)) {
> - dev_priv->display.audio_codec_enable = ilk_audio_codec_enable;
> - dev_priv->display.audio_codec_disable = ilk_audio_codec_disable;
> + if (IS_G4X(i915)) {
> + i915->display.audio_codec_enable = g4x_audio_codec_enable;
> + i915->display.audio_codec_disable = g4x_audio_codec_disable;
> + } else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> + i915->display.audio_codec_enable = ilk_audio_codec_enable;
> + i915->display.audio_codec_disable = ilk_audio_codec_disable;
> + } else if (IS_HASWELL(i915) || INTEL_GEN(i915) >= 8) {
> + i915->display.audio_codec_enable = hsw_audio_codec_enable;
> + i915->display.audio_codec_disable = hsw_audio_codec_disable;
> + } else if (HAS_PCH_SPLIT(i915)) {
> + i915->display.audio_codec_enable = ilk_audio_codec_enable;
> + i915->display.audio_codec_disable = ilk_audio_codec_disable;
> }
> }
>
> -static void glk_force_audio_cdclk(struct drm_i915_private *dev_priv,
> +static void glk_force_audio_cdclk(struct drm_i915_private *i915,
> bool enable)
> {
> struct drm_modeset_acquire_ctx ctx;
> @@ -750,7 +750,7 @@ static void glk_force_audio_cdclk(struct drm_i915_private *dev_priv,
> int ret;
>
> drm_modeset_acquire_init(&ctx, 0);
> - state = drm_atomic_state_alloc(&dev_priv->drm);
> + state = drm_atomic_state_alloc(&i915->drm);
> if (WARN_ON(!state))
> return;
>
> @@ -762,11 +762,11 @@ static void glk_force_audio_cdclk(struct drm_i915_private *dev_priv,
> enable ? 2 * 96000 : 0;
>
> /*
> - * Protects dev_priv->cdclk.force_min_cdclk
> + * Protects i915->cdclk.force_min_cdclk
> * Need to lock this here in case we have no active pipes
> * and thus wouldn't lock it during the commit otherwise.
> */
> - ret = drm_modeset_lock(&dev_priv->drm.mode_config.connection_mutex,
> + ret = drm_modeset_lock(&i915->drm.mode_config.connection_mutex,
> &ctx);
> if (!ret)
> ret = drm_atomic_commit(state);
> @@ -787,18 +787,18 @@ static void glk_force_audio_cdclk(struct drm_i915_private *dev_priv,
>
> static unsigned long i915_audio_component_get_power(struct device *kdev)
> {
> - struct drm_i915_private *dev_priv = kdev_to_i915(kdev);
> + struct drm_i915_private *i915 = kdev_to_i915(kdev);
> intel_wakeref_t ret;
>
> /* Catch potential impedance mismatches before they occur! */
> BUILD_BUG_ON(sizeof(intel_wakeref_t) > sizeof(unsigned long));
>
> - ret = intel_display_power_get(dev_priv, POWER_DOMAIN_AUDIO);
> + ret = intel_display_power_get(i915, POWER_DOMAIN_AUDIO);
>
> /* Force CDCLK to 2*BCLK as long as we need audio to be powered. */
> - if (dev_priv->audio_power_refcount++ == 0)
> - if (IS_CANNONLAKE(dev_priv) || IS_GEMINILAKE(dev_priv))
> - glk_force_audio_cdclk(dev_priv, true);
> + if (i915->audio_power_refcount++ == 0)
> + if (IS_CANNONLAKE(i915) || IS_GEMINILAKE(i915))
> + glk_force_audio_cdclk(i915, true);
>
> return ret;
> }
> @@ -806,24 +806,24 @@ static unsigned long i915_audio_component_get_power(struct device *kdev)
> static void i915_audio_component_put_power(struct device *kdev,
> unsigned long cookie)
> {
> - struct drm_i915_private *dev_priv = kdev_to_i915(kdev);
> + struct drm_i915_private *i915 = kdev_to_i915(kdev);
>
> /* Stop forcing CDCLK to 2*BCLK if no need for audio to be powered. */
> - if (--dev_priv->audio_power_refcount == 0)
> - if (IS_CANNONLAKE(dev_priv) || IS_GEMINILAKE(dev_priv))
> - glk_force_audio_cdclk(dev_priv, false);
> + if (--i915->audio_power_refcount == 0)
> + if (IS_CANNONLAKE(i915) || IS_GEMINILAKE(i915))
> + glk_force_audio_cdclk(i915, false);
>
> - intel_display_power_put(dev_priv, POWER_DOMAIN_AUDIO, cookie);
> + intel_display_power_put(i915, POWER_DOMAIN_AUDIO, cookie);
> }
>
> static void i915_audio_component_codec_wake_override(struct device *kdev,
> bool enable)
> {
> - struct drm_i915_private *dev_priv = kdev_to_i915(kdev);
> + struct drm_i915_private *i915 = kdev_to_i915(kdev);
> unsigned long cookie;
> u32 tmp;
>
> - if (!IS_GEN(dev_priv, 9))
> + if (!IS_GEN(i915, 9))
> return;
>
> cookie = i915_audio_component_get_power(kdev);
> @@ -850,12 +850,12 @@ static void i915_audio_component_codec_wake_override(struct device *kdev,
> /* Get CDCLK in kHz */
> static int i915_audio_component_get_cdclk_freq(struct device *kdev)
> {
> - struct drm_i915_private *dev_priv = kdev_to_i915(kdev);
> + struct drm_i915_private *i915 = kdev_to_i915(kdev);
>
> - if (WARN_ON_ONCE(!HAS_DDI(dev_priv)))
> + if (WARN_ON_ONCE(!HAS_DDI(i915)))
> return -ENODEV;
>
> - return dev_priv->cdclk.hw.cdclk;
> + return i915->cdclk.hw.cdclk;
> }
>
> /*
> @@ -868,17 +868,17 @@ static int i915_audio_component_get_cdclk_freq(struct device *kdev)
> * will get the right intel_encoder with port matched
> * Non-MST & (pipe < 0): get the right intel_encoder with port matched
> */
> -static struct intel_encoder *get_saved_enc(struct drm_i915_private *dev_priv,
> +static struct intel_encoder *get_saved_enc(struct drm_i915_private *i915,
> int port, int pipe)
> {
> struct intel_encoder *encoder;
>
> /* MST */
> if (pipe >= 0) {
> - if (WARN_ON(pipe >= ARRAY_SIZE(dev_priv->av_enc_map)))
> + if (WARN_ON(pipe >= ARRAY_SIZE(i915->av_enc_map)))
> return NULL;
>
> - encoder = dev_priv->av_enc_map[pipe];
> + encoder = i915->av_enc_map[pipe];
> /*
> * when bootup, audio driver may not know it is
> * MST or not. So it will poll all the port & pipe
> @@ -893,8 +893,8 @@ static struct intel_encoder *get_saved_enc(struct drm_i915_private *dev_priv,
> if (pipe > 0)
> return NULL;
>
> - for_each_pipe(dev_priv, pipe) {
> - encoder = dev_priv->av_enc_map[pipe];
> + for_each_pipe(i915, pipe) {
> + encoder = i915->av_enc_map[pipe];
> if (encoder == NULL)
> continue;
>
> @@ -911,21 +911,21 @@ static struct intel_encoder *get_saved_enc(struct drm_i915_private *dev_priv,
> static int i915_audio_component_sync_audio_rate(struct device *kdev, int port,
> int pipe, int rate)
> {
> - struct drm_i915_private *dev_priv = kdev_to_i915(kdev);
> - struct i915_audio_component *acomp = dev_priv->audio_component;
> + struct drm_i915_private *i915 = kdev_to_i915(kdev);
> + struct i915_audio_component *acomp = i915->audio_component;
> struct intel_encoder *encoder;
> struct intel_crtc *crtc;
> unsigned long cookie;
> int err = 0;
>
> - if (!HAS_DDI(dev_priv))
> + if (!HAS_DDI(i915))
> return 0;
>
> cookie = i915_audio_component_get_power(kdev);
> - mutex_lock(&dev_priv->av_mutex);
> + mutex_lock(&i915->av_mutex);
>
> /* 1. get the pipe */
> - encoder = get_saved_enc(dev_priv, port, pipe);
> + encoder = get_saved_enc(i915, port, pipe);
> if (!encoder || !encoder->base.crtc) {
> DRM_DEBUG_KMS("Not valid for port %c\n", port_name(port));
> err = -ENODEV;
> @@ -940,7 +940,7 @@ static int i915_audio_component_sync_audio_rate(struct device *kdev, int port,
> hsw_audio_config_update(encoder, crtc->config);
>
> unlock:
> - mutex_unlock(&dev_priv->av_mutex);
> + mutex_unlock(&i915->av_mutex);
> i915_audio_component_put_power(kdev, cookie);
> return err;
> }
> @@ -949,17 +949,17 @@ static int i915_audio_component_get_eld(struct device *kdev, int port,
> int pipe, bool *enabled,
> unsigned char *buf, int max_bytes)
> {
> - struct drm_i915_private *dev_priv = kdev_to_i915(kdev);
> + struct drm_i915_private *i915 = kdev_to_i915(kdev);
> struct intel_encoder *intel_encoder;
> const u8 *eld;
> int ret = -EINVAL;
>
> - mutex_lock(&dev_priv->av_mutex);
> + mutex_lock(&i915->av_mutex);
>
> - intel_encoder = get_saved_enc(dev_priv, port, pipe);
> + intel_encoder = get_saved_enc(i915, port, pipe);
> if (!intel_encoder) {
> DRM_DEBUG_KMS("Not valid for port %c\n", port_name(port));
> - mutex_unlock(&dev_priv->av_mutex);
> + mutex_unlock(&i915->av_mutex);
> return ret;
> }
>
> @@ -971,7 +971,7 @@ static int i915_audio_component_get_eld(struct device *kdev, int port,
> memcpy(buf, eld, min(max_bytes, ret));
> }
>
> - mutex_unlock(&dev_priv->av_mutex);
> + mutex_unlock(&i915->av_mutex);
> return ret;
> }
>
> @@ -989,7 +989,7 @@ static int i915_audio_component_bind(struct device *i915_kdev,
> struct device *hda_kdev, void *data)
> {
> struct i915_audio_component *acomp = data;
> - struct drm_i915_private *dev_priv = kdev_to_i915(i915_kdev);
> + struct drm_i915_private *i915 = kdev_to_i915(i915_kdev);
> int i;
>
> if (WARN_ON(acomp->base.ops || acomp->base.dev))
> @@ -998,14 +998,14 @@ static int i915_audio_component_bind(struct device *i915_kdev,
> if (WARN_ON(!device_link_add(hda_kdev, i915_kdev, DL_FLAG_STATELESS)))
> return -ENOMEM;
>
> - drm_modeset_lock_all(&dev_priv->drm);
> + drm_modeset_lock_all(&i915->drm);
> acomp->base.ops = &i915_audio_component_ops;
> acomp->base.dev = i915_kdev;
> BUILD_BUG_ON(MAX_PORTS != I915_MAX_PORTS);
> for (i = 0; i < ARRAY_SIZE(acomp->aud_sample_rate); i++)
> acomp->aud_sample_rate[i] = 0;
> - dev_priv->audio_component = acomp;
> - drm_modeset_unlock_all(&dev_priv->drm);
> + i915->audio_component = acomp;
> + drm_modeset_unlock_all(&i915->drm);
>
> return 0;
> }
> @@ -1014,13 +1014,13 @@ static void i915_audio_component_unbind(struct device *i915_kdev,
> struct device *hda_kdev, void *data)
> {
> struct i915_audio_component *acomp = data;
> - struct drm_i915_private *dev_priv = kdev_to_i915(i915_kdev);
> + struct drm_i915_private *i915 = kdev_to_i915(i915_kdev);
>
> - drm_modeset_lock_all(&dev_priv->drm);
> + drm_modeset_lock_all(&i915->drm);
> acomp->base.ops = NULL;
> acomp->base.dev = NULL;
> - dev_priv->audio_component = NULL;
> - drm_modeset_unlock_all(&dev_priv->drm);
> + i915->audio_component = NULL;
> + drm_modeset_unlock_all(&i915->drm);
>
> device_link_remove(hda_kdev, i915_kdev);
> }
> @@ -1032,7 +1032,7 @@ static const struct component_ops i915_audio_component_bind_ops = {
>
> /**
> * i915_audio_component_init - initialize and register the audio component
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> *
> * This will register with the component framework a child component which
> * will bind dynamically to the snd_hda_intel driver's corresponding master
> @@ -1046,11 +1046,11 @@ static const struct component_ops i915_audio_component_bind_ops = {
> * We ignore any error during registration and continue with reduced
> * functionality (i.e. without HDMI audio).
> */
> -static void i915_audio_component_init(struct drm_i915_private *dev_priv)
> +static void i915_audio_component_init(struct drm_i915_private *i915)
> {
> int ret;
>
> - ret = component_add_typed(dev_priv->drm.dev,
> + ret = component_add_typed(i915->drm.dev,
> &i915_audio_component_bind_ops,
> I915_COMPONENT_AUDIO);
> if (ret < 0) {
> @@ -1059,46 +1059,46 @@ static void i915_audio_component_init(struct drm_i915_private *dev_priv)
> return;
> }
>
> - dev_priv->audio_component_registered = true;
> + i915->audio_component_registered = true;
> }
>
> /**
> * i915_audio_component_cleanup - deregister the audio component
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> *
> * Deregisters the audio component, breaking any existing binding to the
> * corresponding snd_hda_intel driver's master component.
> */
> -static void i915_audio_component_cleanup(struct drm_i915_private *dev_priv)
> +static void i915_audio_component_cleanup(struct drm_i915_private *i915)
> {
> - if (!dev_priv->audio_component_registered)
> + if (!i915->audio_component_registered)
> return;
>
> - component_del(dev_priv->drm.dev, &i915_audio_component_bind_ops);
> - dev_priv->audio_component_registered = false;
> + component_del(i915->drm.dev, &i915_audio_component_bind_ops);
> + i915->audio_component_registered = false;
> }
>
> /**
> * intel_audio_init() - Initialize the audio driver either using
> * component framework or using lpe audio bridge
> - * @dev_priv: the i915 drm device private data
> + * @i915: the i915 drm device private data
> *
> */
> -void intel_audio_init(struct drm_i915_private *dev_priv)
> +void intel_audio_init(struct drm_i915_private *i915)
> {
> - if (intel_lpe_audio_init(dev_priv) < 0)
> - i915_audio_component_init(dev_priv);
> + if (intel_lpe_audio_init(i915) < 0)
> + i915_audio_component_init(i915);
> }
>
> /**
> * intel_audio_deinit() - deinitialize the audio driver
> - * @dev_priv: the i915 drm device private data
> + * @i915: the i915 drm device private data
> *
> */
> -void intel_audio_deinit(struct drm_i915_private *dev_priv)
> +void intel_audio_deinit(struct drm_i915_private *i915)
> {
> - if ((dev_priv)->lpe_audio.platdev != NULL)
> - intel_lpe_audio_teardown(dev_priv);
> + if ((i915)->lpe_audio.platdev != NULL)
> + intel_lpe_audio_teardown(i915);
> else
> - i915_audio_component_cleanup(dev_priv);
> + i915_audio_component_cleanup(i915);
> }
> diff --git a/drivers/gpu/drm/i915/intel_audio.h b/drivers/gpu/drm/i915/intel_audio.h
> index a3657c7a7ba2..c8fde42a9fde 100644
> --- a/drivers/gpu/drm/i915/intel_audio.h
> +++ b/drivers/gpu/drm/i915/intel_audio.h
> @@ -11,14 +11,14 @@ struct drm_i915_private;
> struct intel_crtc_state;
> struct intel_encoder;
>
> -void intel_init_audio_hooks(struct drm_i915_private *dev_priv);
> +void intel_init_audio_hooks(struct drm_i915_private *i915);
> void intel_audio_codec_enable(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state,
> const struct drm_connector_state *conn_state);
> void intel_audio_codec_disable(struct intel_encoder *encoder,
> const struct intel_crtc_state *old_crtc_state,
> const struct drm_connector_state *old_conn_state);
> -void intel_audio_init(struct drm_i915_private *dev_priv);
> -void intel_audio_deinit(struct drm_i915_private *dev_priv);
> +void intel_audio_init(struct drm_i915_private *i915);
> +void intel_audio_deinit(struct drm_i915_private *i915);
>
> #endif /* __INTEL_AUDIO_H__ */
> diff --git a/drivers/gpu/drm/i915/intel_bios.c b/drivers/gpu/drm/i915/intel_bios.c
> index 1c037dfa83f5..7bdac55d08cc 100644
> --- a/drivers/gpu/drm/i915/intel_bios.c
> +++ b/drivers/gpu/drm/i915/intel_bios.c
> @@ -202,7 +202,7 @@ get_lvds_fp_timing(const struct bdb_header *bdb,
>
> /* Try to find integrated panel data */
> static void
> -parse_lfp_panel_data(struct drm_i915_private *dev_priv,
> +parse_lfp_panel_data(struct drm_i915_private *i915,
> const struct bdb_header *bdb)
> {
> const struct bdb_lvds_options *lvds_options;
> @@ -219,9 +219,9 @@ parse_lfp_panel_data(struct drm_i915_private *dev_priv,
> if (!lvds_options)
> return;
>
> - dev_priv->vbt.lvds_dither = lvds_options->pixel_dither;
> + i915->vbt.lvds_dither = lvds_options->pixel_dither;
>
> - ret = intel_opregion_get_panel_type(dev_priv);
> + ret = intel_opregion_get_panel_type(i915);
> if (ret >= 0) {
> WARN_ON(ret > 0xf);
> panel_type = ret;
> @@ -236,7 +236,7 @@ parse_lfp_panel_data(struct drm_i915_private *dev_priv,
> DRM_DEBUG_KMS("Panel type: %d (VBT)\n", panel_type);
> }
>
> - dev_priv->vbt.panel_type = panel_type;
> + i915->vbt.panel_type = panel_type;
>
> drrs_mode = (lvds_options->dps_panel_type_bits
> >> (panel_type * 2)) & MODE_MASK;
> @@ -247,15 +247,15 @@ parse_lfp_panel_data(struct drm_i915_private *dev_priv,
> */
> switch (drrs_mode) {
> case 0:
> - dev_priv->vbt.drrs_type = STATIC_DRRS_SUPPORT;
> + i915->vbt.drrs_type = STATIC_DRRS_SUPPORT;
> DRM_DEBUG_KMS("DRRS supported mode is static\n");
> break;
> case 2:
> - dev_priv->vbt.drrs_type = SEAMLESS_DRRS_SUPPORT;
> + i915->vbt.drrs_type = SEAMLESS_DRRS_SUPPORT;
> DRM_DEBUG_KMS("DRRS supported mode is seamless\n");
> break;
> default:
> - dev_priv->vbt.drrs_type = DRRS_NOT_SUPPORTED;
> + i915->vbt.drrs_type = DRRS_NOT_SUPPORTED;
> DRM_DEBUG_KMS("DRRS not supported (VBT input)\n");
> break;
> }
> @@ -278,7 +278,7 @@ parse_lfp_panel_data(struct drm_i915_private *dev_priv,
>
> fill_detail_timing_data(panel_fixed_mode, panel_dvo_timing);
>
> - dev_priv->vbt.lfp_lvds_vbt_mode = panel_fixed_mode;
> + i915->vbt.lfp_lvds_vbt_mode = panel_fixed_mode;
>
> DRM_DEBUG_KMS("Found panel mode in BIOS VBT tables:\n");
> drm_mode_debug_printmodeline(panel_fixed_mode);
> @@ -290,20 +290,20 @@ parse_lfp_panel_data(struct drm_i915_private *dev_priv,
> /* check the resolution, just to be sure */
> if (fp_timing->x_res == panel_fixed_mode->hdisplay &&
> fp_timing->y_res == panel_fixed_mode->vdisplay) {
> - dev_priv->vbt.bios_lvds_val = fp_timing->lvds_reg_val;
> + i915->vbt.bios_lvds_val = fp_timing->lvds_reg_val;
> DRM_DEBUG_KMS("VBT initial LVDS value %x\n",
> - dev_priv->vbt.bios_lvds_val);
> + i915->vbt.bios_lvds_val);
> }
> }
> }
>
> static void
> -parse_lfp_backlight(struct drm_i915_private *dev_priv,
> +parse_lfp_backlight(struct drm_i915_private *i915,
> const struct bdb_header *bdb)
> {
> const struct bdb_lfp_backlight_data *backlight_data;
> const struct lfp_backlight_data_entry *entry;
> - int panel_type = dev_priv->vbt.panel_type;
> + int panel_type = i915->vbt.panel_type;
>
> backlight_data = find_section(bdb, BDB_LVDS_BACKLIGHT);
> if (!backlight_data)
> @@ -317,38 +317,38 @@ parse_lfp_backlight(struct drm_i915_private *dev_priv,
>
> entry = &backlight_data->data[panel_type];
>
> - dev_priv->vbt.backlight.present = entry->type == BDB_BACKLIGHT_TYPE_PWM;
> - if (!dev_priv->vbt.backlight.present) {
> + i915->vbt.backlight.present = entry->type == BDB_BACKLIGHT_TYPE_PWM;
> + if (!i915->vbt.backlight.present) {
> DRM_DEBUG_KMS("PWM backlight not present in VBT (type %u)\n",
> entry->type);
> return;
> }
>
> - dev_priv->vbt.backlight.type = INTEL_BACKLIGHT_DISPLAY_DDI;
> + i915->vbt.backlight.type = INTEL_BACKLIGHT_DISPLAY_DDI;
> if (bdb->version >= 191 &&
> get_blocksize(backlight_data) >= sizeof(*backlight_data)) {
> const struct lfp_backlight_control_method *method;
>
> method = &backlight_data->backlight_control[panel_type];
> - dev_priv->vbt.backlight.type = method->type;
> - dev_priv->vbt.backlight.controller = method->controller;
> + i915->vbt.backlight.type = method->type;
> + i915->vbt.backlight.controller = method->controller;
> }
>
> - dev_priv->vbt.backlight.pwm_freq_hz = entry->pwm_freq_hz;
> - dev_priv->vbt.backlight.active_low_pwm = entry->active_low_pwm;
> - dev_priv->vbt.backlight.min_brightness = entry->min_brightness;
> + i915->vbt.backlight.pwm_freq_hz = entry->pwm_freq_hz;
> + i915->vbt.backlight.active_low_pwm = entry->active_low_pwm;
> + i915->vbt.backlight.min_brightness = entry->min_brightness;
> DRM_DEBUG_KMS("VBT backlight PWM modulation frequency %u Hz, "
> "active %s, min brightness %u, level %u, controller %u\n",
> - dev_priv->vbt.backlight.pwm_freq_hz,
> - dev_priv->vbt.backlight.active_low_pwm ? "low" : "high",
> - dev_priv->vbt.backlight.min_brightness,
> + i915->vbt.backlight.pwm_freq_hz,
> + i915->vbt.backlight.active_low_pwm ? "low" : "high",
> + i915->vbt.backlight.min_brightness,
> backlight_data->level[panel_type],
> - dev_priv->vbt.backlight.controller);
> + i915->vbt.backlight.controller);
> }
>
> /* Try to find sdvo panel data */
> static void
> -parse_sdvo_panel_data(struct drm_i915_private *dev_priv,
> +parse_sdvo_panel_data(struct drm_i915_private *i915,
> const struct bdb_header *bdb)
> {
> const struct bdb_sdvo_panel_dtds *dtds;
> @@ -381,16 +381,16 @@ parse_sdvo_panel_data(struct drm_i915_private *dev_priv,
>
> fill_detail_timing_data(panel_fixed_mode, &dtds->dtds[index]);
>
> - dev_priv->vbt.sdvo_lvds_vbt_mode = panel_fixed_mode;
> + i915->vbt.sdvo_lvds_vbt_mode = panel_fixed_mode;
>
> DRM_DEBUG_KMS("Found SDVO panel mode in BIOS VBT tables:\n");
> drm_mode_debug_printmodeline(panel_fixed_mode);
> }
>
> -static int intel_bios_ssc_frequency(struct drm_i915_private *dev_priv,
> +static int intel_bios_ssc_frequency(struct drm_i915_private *i915,
> bool alternate)
> {
> - switch (INTEL_GEN(dev_priv)) {
> + switch (INTEL_GEN(i915)) {
> case 2:
> return alternate ? 66667 : 48000;
> case 3:
> @@ -402,7 +402,7 @@ static int intel_bios_ssc_frequency(struct drm_i915_private *dev_priv,
> }
>
> static void
> -parse_general_features(struct drm_i915_private *dev_priv,
> +parse_general_features(struct drm_i915_private *i915,
> const struct bdb_header *bdb)
> {
> const struct bdb_general_features *general;
> @@ -411,30 +411,30 @@ parse_general_features(struct drm_i915_private *dev_priv,
> if (!general)
> return;
>
> - dev_priv->vbt.int_tv_support = general->int_tv_support;
> + i915->vbt.int_tv_support = general->int_tv_support;
> /* int_crt_support can't be trusted on earlier platforms */
> if (bdb->version >= 155 &&
> - (HAS_DDI(dev_priv) || IS_VALLEYVIEW(dev_priv)))
> - dev_priv->vbt.int_crt_support = general->int_crt_support;
> - dev_priv->vbt.lvds_use_ssc = general->enable_ssc;
> - dev_priv->vbt.lvds_ssc_freq =
> - intel_bios_ssc_frequency(dev_priv, general->ssc_freq);
> - dev_priv->vbt.display_clock_mode = general->display_clock_mode;
> - dev_priv->vbt.fdi_rx_polarity_inverted = general->fdi_rx_polarity_inverted;
> + (HAS_DDI(i915) || IS_VALLEYVIEW(i915)))
> + i915->vbt.int_crt_support = general->int_crt_support;
> + i915->vbt.lvds_use_ssc = general->enable_ssc;
> + i915->vbt.lvds_ssc_freq =
> + intel_bios_ssc_frequency(i915, general->ssc_freq);
> + i915->vbt.display_clock_mode = general->display_clock_mode;
> + i915->vbt.fdi_rx_polarity_inverted = general->fdi_rx_polarity_inverted;
> if (bdb->version >= 181) {
> - dev_priv->vbt.orientation = general->rotate_180 ?
> + i915->vbt.orientation = general->rotate_180 ?
> DRM_MODE_PANEL_ORIENTATION_BOTTOM_UP :
> DRM_MODE_PANEL_ORIENTATION_NORMAL;
> } else {
> - dev_priv->vbt.orientation = DRM_MODE_PANEL_ORIENTATION_UNKNOWN;
> + i915->vbt.orientation = DRM_MODE_PANEL_ORIENTATION_UNKNOWN;
> }
> DRM_DEBUG_KMS("BDB_GENERAL_FEATURES int_tv_support %d int_crt_support %d lvds_use_ssc %d lvds_ssc_freq %d display_clock_mode %d fdi_rx_polarity_inverted %d\n",
> - dev_priv->vbt.int_tv_support,
> - dev_priv->vbt.int_crt_support,
> - dev_priv->vbt.lvds_use_ssc,
> - dev_priv->vbt.lvds_ssc_freq,
> - dev_priv->vbt.display_clock_mode,
> - dev_priv->vbt.fdi_rx_polarity_inverted);
> + i915->vbt.int_tv_support,
> + i915->vbt.int_crt_support,
> + i915->vbt.lvds_use_ssc,
> + i915->vbt.lvds_ssc_freq,
> + i915->vbt.display_clock_mode,
> + i915->vbt.fdi_rx_polarity_inverted);
> }
>
> static const struct child_device_config *
> @@ -444,7 +444,7 @@ child_device_ptr(const struct bdb_general_definitions *defs, int i)
> }
>
> static void
> -parse_sdvo_device_mapping(struct drm_i915_private *dev_priv, u8 bdb_version)
> +parse_sdvo_device_mapping(struct drm_i915_private *i915, u8 bdb_version)
> {
> struct sdvo_device_mapping *mapping;
> const struct child_device_config *child;
> @@ -454,13 +454,13 @@ parse_sdvo_device_mapping(struct drm_i915_private *dev_priv, u8 bdb_version)
> * Only parse SDVO mappings on gens that could have SDVO. This isn't
> * accurate and doesn't have to be, as long as it's not too strict.
> */
> - if (!IS_GEN_RANGE(dev_priv, 3, 7)) {
> + if (!IS_GEN_RANGE(i915, 3, 7)) {
> DRM_DEBUG_KMS("Skipping SDVO device mapping\n");
> return;
> }
>
> - for (i = 0, count = 0; i < dev_priv->vbt.child_dev_num; i++) {
> - child = dev_priv->vbt.child_dev + i;
> + for (i = 0, count = 0; i < i915->vbt.child_dev_num; i++) {
> + child = i915->vbt.child_dev + i;
>
> if (child->slave_addr != SLAVE_ADDR1 &&
> child->slave_addr != SLAVE_ADDR2) {
> @@ -481,7 +481,7 @@ parse_sdvo_device_mapping(struct drm_i915_private *dev_priv, u8 bdb_version)
> child->slave_addr,
> (child->dvo_port == DEVICE_PORT_DVOB) ?
> "SDVOB" : "SDVOC");
> - mapping = &dev_priv->vbt.sdvo_mappings[child->dvo_port - 1];
> + mapping = &i915->vbt.sdvo_mappings[child->dvo_port - 1];
> if (!mapping->initialized) {
> mapping->dvo_port = child->dvo_port;
> mapping->slave_addr = child->slave_addr;
> @@ -515,7 +515,7 @@ parse_sdvo_device_mapping(struct drm_i915_private *dev_priv, u8 bdb_version)
> }
>
> static void
> -parse_driver_features(struct drm_i915_private *dev_priv,
> +parse_driver_features(struct drm_i915_private *i915,
> const struct bdb_header *bdb)
> {
> const struct bdb_driver_features *driver;
> @@ -524,14 +524,14 @@ parse_driver_features(struct drm_i915_private *dev_priv,
> if (!driver)
> return;
>
> - if (INTEL_GEN(dev_priv) >= 5) {
> + if (INTEL_GEN(i915) >= 5) {
> /*
> * Note that we consider BDB_DRIVER_FEATURE_INT_SDVO_LVDS
> * to mean "eDP". The VBT spec doesn't agree with that
> * interpretation, but real world VBTs seem to.
> */
> if (driver->lvds_config != BDB_DRIVER_FEATURE_INT_LVDS)
> - dev_priv->vbt.int_lvds_support = 0;
> + i915->vbt.int_lvds_support = 0;
> } else {
> /*
> * FIXME it's not clear which BDB version has the LVDS config
> @@ -547,7 +547,7 @@ parse_driver_features(struct drm_i915_private *dev_priv,
> if (bdb->version >= 134 &&
> driver->lvds_config != BDB_DRIVER_FEATURE_INT_LVDS &&
> driver->lvds_config != BDB_DRIVER_FEATURE_INT_SDVO_LVDS)
> - dev_priv->vbt.int_lvds_support = 0;
> + i915->vbt.int_lvds_support = 0;
> }
>
> DRM_DEBUG_KMS("DRRS State Enabled:%d\n", driver->drrs_enabled);
> @@ -558,17 +558,17 @@ parse_driver_features(struct drm_i915_private *dev_priv,
> * driver->drrs_enabled=false
> */
> if (!driver->drrs_enabled)
> - dev_priv->vbt.drrs_type = DRRS_NOT_SUPPORTED;
> - dev_priv->vbt.psr.enable = driver->psr_enabled;
> + i915->vbt.drrs_type = DRRS_NOT_SUPPORTED;
> + i915->vbt.psr.enable = driver->psr_enabled;
> }
>
> static void
> -parse_edp(struct drm_i915_private *dev_priv, const struct bdb_header *bdb)
> +parse_edp(struct drm_i915_private *i915, const struct bdb_header *bdb)
> {
> const struct bdb_edp *edp;
> const struct edp_power_seq *edp_pps;
> const struct edp_fast_link_params *edp_link_params;
> - int panel_type = dev_priv->vbt.panel_type;
> + int panel_type = i915->vbt.panel_type;
>
> edp = find_section(bdb, BDB_EDP);
> if (!edp)
> @@ -576,13 +576,13 @@ parse_edp(struct drm_i915_private *dev_priv, const struct bdb_header *bdb)
>
> switch ((edp->color_depth >> (panel_type * 2)) & 3) {
> case EDP_18BPP:
> - dev_priv->vbt.edp.bpp = 18;
> + i915->vbt.edp.bpp = 18;
> break;
> case EDP_24BPP:
> - dev_priv->vbt.edp.bpp = 24;
> + i915->vbt.edp.bpp = 24;
> break;
> case EDP_30BPP:
> - dev_priv->vbt.edp.bpp = 30;
> + i915->vbt.edp.bpp = 30;
> break;
> }
>
> @@ -590,14 +590,14 @@ parse_edp(struct drm_i915_private *dev_priv, const struct bdb_header *bdb)
> edp_pps = &edp->power_seqs[panel_type];
> edp_link_params = &edp->fast_link_params[panel_type];
>
> - dev_priv->vbt.edp.pps = *edp_pps;
> + i915->vbt.edp.pps = *edp_pps;
>
> switch (edp_link_params->rate) {
> case EDP_RATE_1_62:
> - dev_priv->vbt.edp.rate = DP_LINK_BW_1_62;
> + i915->vbt.edp.rate = DP_LINK_BW_1_62;
> break;
> case EDP_RATE_2_7:
> - dev_priv->vbt.edp.rate = DP_LINK_BW_2_7;
> + i915->vbt.edp.rate = DP_LINK_BW_2_7;
> break;
> default:
> DRM_DEBUG_KMS("VBT has unknown eDP link rate value %u\n",
> @@ -607,13 +607,13 @@ parse_edp(struct drm_i915_private *dev_priv, const struct bdb_header *bdb)
>
> switch (edp_link_params->lanes) {
> case EDP_LANE_1:
> - dev_priv->vbt.edp.lanes = 1;
> + i915->vbt.edp.lanes = 1;
> break;
> case EDP_LANE_2:
> - dev_priv->vbt.edp.lanes = 2;
> + i915->vbt.edp.lanes = 2;
> break;
> case EDP_LANE_4:
> - dev_priv->vbt.edp.lanes = 4;
> + i915->vbt.edp.lanes = 4;
> break;
> default:
> DRM_DEBUG_KMS("VBT has unknown eDP lane count value %u\n",
> @@ -623,16 +623,16 @@ parse_edp(struct drm_i915_private *dev_priv, const struct bdb_header *bdb)
>
> switch (edp_link_params->preemphasis) {
> case EDP_PREEMPHASIS_NONE:
> - dev_priv->vbt.edp.preemphasis = DP_TRAIN_PRE_EMPH_LEVEL_0;
> + i915->vbt.edp.preemphasis = DP_TRAIN_PRE_EMPH_LEVEL_0;
> break;
> case EDP_PREEMPHASIS_3_5dB:
> - dev_priv->vbt.edp.preemphasis = DP_TRAIN_PRE_EMPH_LEVEL_1;
> + i915->vbt.edp.preemphasis = DP_TRAIN_PRE_EMPH_LEVEL_1;
> break;
> case EDP_PREEMPHASIS_6dB:
> - dev_priv->vbt.edp.preemphasis = DP_TRAIN_PRE_EMPH_LEVEL_2;
> + i915->vbt.edp.preemphasis = DP_TRAIN_PRE_EMPH_LEVEL_2;
> break;
> case EDP_PREEMPHASIS_9_5dB:
> - dev_priv->vbt.edp.preemphasis = DP_TRAIN_PRE_EMPH_LEVEL_3;
> + i915->vbt.edp.preemphasis = DP_TRAIN_PRE_EMPH_LEVEL_3;
> break;
> default:
> DRM_DEBUG_KMS("VBT has unknown eDP pre-emphasis value %u\n",
> @@ -642,16 +642,16 @@ parse_edp(struct drm_i915_private *dev_priv, const struct bdb_header *bdb)
>
> switch (edp_link_params->vswing) {
> case EDP_VSWING_0_4V:
> - dev_priv->vbt.edp.vswing = DP_TRAIN_VOLTAGE_SWING_LEVEL_0;
> + i915->vbt.edp.vswing = DP_TRAIN_VOLTAGE_SWING_LEVEL_0;
> break;
> case EDP_VSWING_0_6V:
> - dev_priv->vbt.edp.vswing = DP_TRAIN_VOLTAGE_SWING_LEVEL_1;
> + i915->vbt.edp.vswing = DP_TRAIN_VOLTAGE_SWING_LEVEL_1;
> break;
> case EDP_VSWING_0_8V:
> - dev_priv->vbt.edp.vswing = DP_TRAIN_VOLTAGE_SWING_LEVEL_2;
> + i915->vbt.edp.vswing = DP_TRAIN_VOLTAGE_SWING_LEVEL_2;
> break;
> case EDP_VSWING_1_2V:
> - dev_priv->vbt.edp.vswing = DP_TRAIN_VOLTAGE_SWING_LEVEL_3;
> + i915->vbt.edp.vswing = DP_TRAIN_VOLTAGE_SWING_LEVEL_3;
> break;
> default:
> DRM_DEBUG_KMS("VBT has unknown eDP voltage swing value %u\n",
> @@ -664,21 +664,21 @@ parse_edp(struct drm_i915_private *dev_priv, const struct bdb_header *bdb)
>
> /* Don't read from VBT if module parameter has valid value*/
> if (i915_modparams.edp_vswing) {
> - dev_priv->vbt.edp.low_vswing =
> + i915->vbt.edp.low_vswing =
> i915_modparams.edp_vswing == 1;
> } else {
> vswing = (edp->edp_vswing_preemph >> (panel_type * 4)) & 0xF;
> - dev_priv->vbt.edp.low_vswing = vswing == 0;
> + i915->vbt.edp.low_vswing = vswing == 0;
> }
> }
> }
>
> static void
> -parse_psr(struct drm_i915_private *dev_priv, const struct bdb_header *bdb)
> +parse_psr(struct drm_i915_private *i915, const struct bdb_header *bdb)
> {
> const struct bdb_psr *psr;
> const struct psr_table *psr_table;
> - int panel_type = dev_priv->vbt.panel_type;
> + int panel_type = i915->vbt.panel_type;
>
> psr = find_section(bdb, BDB_PSR);
> if (!psr) {
> @@ -688,25 +688,25 @@ parse_psr(struct drm_i915_private *dev_priv, const struct bdb_header *bdb)
>
> psr_table = &psr->psr_table[panel_type];
>
> - dev_priv->vbt.psr.full_link = psr_table->full_link;
> - dev_priv->vbt.psr.require_aux_wakeup = psr_table->require_aux_to_wakeup;
> + i915->vbt.psr.full_link = psr_table->full_link;
> + i915->vbt.psr.require_aux_wakeup = psr_table->require_aux_to_wakeup;
>
> /* Allowed VBT values goes from 0 to 15 */
> - dev_priv->vbt.psr.idle_frames = psr_table->idle_frames < 0 ? 0 :
> + i915->vbt.psr.idle_frames = psr_table->idle_frames < 0 ? 0 :
> psr_table->idle_frames > 15 ? 15 : psr_table->idle_frames;
>
> switch (psr_table->lines_to_wait) {
> case 0:
> - dev_priv->vbt.psr.lines_to_wait = PSR_0_LINES_TO_WAIT;
> + i915->vbt.psr.lines_to_wait = PSR_0_LINES_TO_WAIT;
> break;
> case 1:
> - dev_priv->vbt.psr.lines_to_wait = PSR_1_LINE_TO_WAIT;
> + i915->vbt.psr.lines_to_wait = PSR_1_LINE_TO_WAIT;
> break;
> case 2:
> - dev_priv->vbt.psr.lines_to_wait = PSR_4_LINES_TO_WAIT;
> + i915->vbt.psr.lines_to_wait = PSR_4_LINES_TO_WAIT;
> break;
> case 3:
> - dev_priv->vbt.psr.lines_to_wait = PSR_8_LINES_TO_WAIT;
> + i915->vbt.psr.lines_to_wait = PSR_8_LINES_TO_WAIT;
> break;
> default:
> DRM_DEBUG_KMS("VBT has unknown PSR lines to wait %u\n",
> @@ -719,48 +719,48 @@ parse_psr(struct drm_i915_private *dev_priv, const struct bdb_header *bdb)
> * Old decimal value is wake up time in multiples of 100 us.
> */
> if (bdb->version >= 205 &&
> - (IS_GEN9_BC(dev_priv) || IS_GEMINILAKE(dev_priv) ||
> - INTEL_GEN(dev_priv) >= 10)) {
> + (IS_GEN9_BC(i915) || IS_GEMINILAKE(i915) ||
> + INTEL_GEN(i915) >= 10)) {
> switch (psr_table->tp1_wakeup_time) {
> case 0:
> - dev_priv->vbt.psr.tp1_wakeup_time_us = 500;
> + i915->vbt.psr.tp1_wakeup_time_us = 500;
> break;
> case 1:
> - dev_priv->vbt.psr.tp1_wakeup_time_us = 100;
> + i915->vbt.psr.tp1_wakeup_time_us = 100;
> break;
> case 3:
> - dev_priv->vbt.psr.tp1_wakeup_time_us = 0;
> + i915->vbt.psr.tp1_wakeup_time_us = 0;
> break;
> default:
> DRM_DEBUG_KMS("VBT tp1 wakeup time value %d is outside range[0-3], defaulting to max value 2500us\n",
> psr_table->tp1_wakeup_time);
> /* fallthrough */
> case 2:
> - dev_priv->vbt.psr.tp1_wakeup_time_us = 2500;
> + i915->vbt.psr.tp1_wakeup_time_us = 2500;
> break;
> }
>
> switch (psr_table->tp2_tp3_wakeup_time) {
> case 0:
> - dev_priv->vbt.psr.tp2_tp3_wakeup_time_us = 500;
> + i915->vbt.psr.tp2_tp3_wakeup_time_us = 500;
> break;
> case 1:
> - dev_priv->vbt.psr.tp2_tp3_wakeup_time_us = 100;
> + i915->vbt.psr.tp2_tp3_wakeup_time_us = 100;
> break;
> case 3:
> - dev_priv->vbt.psr.tp2_tp3_wakeup_time_us = 0;
> + i915->vbt.psr.tp2_tp3_wakeup_time_us = 0;
> break;
> default:
> DRM_DEBUG_KMS("VBT tp2_tp3 wakeup time value %d is outside range[0-3], defaulting to max value 2500us\n",
> psr_table->tp2_tp3_wakeup_time);
> /* fallthrough */
> case 2:
> - dev_priv->vbt.psr.tp2_tp3_wakeup_time_us = 2500;
> + i915->vbt.psr.tp2_tp3_wakeup_time_us = 2500;
> break;
> }
> } else {
> - dev_priv->vbt.psr.tp1_wakeup_time_us = psr_table->tp1_wakeup_time * 100;
> - dev_priv->vbt.psr.tp2_tp3_wakeup_time_us = psr_table->tp2_tp3_wakeup_time * 100;
> + i915->vbt.psr.tp1_wakeup_time_us = psr_table->tp1_wakeup_time * 100;
> + i915->vbt.psr.tp2_tp3_wakeup_time_us = psr_table->tp2_tp3_wakeup_time * 100;
> }
>
> if (bdb->version >= 226) {
> @@ -782,74 +782,74 @@ parse_psr(struct drm_i915_private *dev_priv, const struct bdb_header *bdb)
> wakeup_time = 2500;
> break;
> }
> - dev_priv->vbt.psr.psr2_tp2_tp3_wakeup_time_us = wakeup_time;
> + i915->vbt.psr.psr2_tp2_tp3_wakeup_time_us = wakeup_time;
> } else {
> /* Reusing PSR1 wakeup time for PSR2 in older VBTs */
> - dev_priv->vbt.psr.psr2_tp2_tp3_wakeup_time_us = dev_priv->vbt.psr.tp2_tp3_wakeup_time_us;
> + i915->vbt.psr.psr2_tp2_tp3_wakeup_time_us = i915->vbt.psr.tp2_tp3_wakeup_time_us;
> }
> }
>
> -static void parse_dsi_backlight_ports(struct drm_i915_private *dev_priv,
> +static void parse_dsi_backlight_ports(struct drm_i915_private *i915,
> u16 version, enum port port)
> {
> - if (!dev_priv->vbt.dsi.config->dual_link || version < 197) {
> - dev_priv->vbt.dsi.bl_ports = BIT(port);
> - if (dev_priv->vbt.dsi.config->cabc_supported)
> - dev_priv->vbt.dsi.cabc_ports = BIT(port);
> + if (!i915->vbt.dsi.config->dual_link || version < 197) {
> + i915->vbt.dsi.bl_ports = BIT(port);
> + if (i915->vbt.dsi.config->cabc_supported)
> + i915->vbt.dsi.cabc_ports = BIT(port);
>
> return;
> }
>
> - switch (dev_priv->vbt.dsi.config->dl_dcs_backlight_ports) {
> + switch (i915->vbt.dsi.config->dl_dcs_backlight_ports) {
> case DL_DCS_PORT_A:
> - dev_priv->vbt.dsi.bl_ports = BIT(PORT_A);
> + i915->vbt.dsi.bl_ports = BIT(PORT_A);
> break;
> case DL_DCS_PORT_C:
> - dev_priv->vbt.dsi.bl_ports = BIT(PORT_C);
> + i915->vbt.dsi.bl_ports = BIT(PORT_C);
> break;
> default:
> case DL_DCS_PORT_A_AND_C:
> - dev_priv->vbt.dsi.bl_ports = BIT(PORT_A) | BIT(PORT_C);
> + i915->vbt.dsi.bl_ports = BIT(PORT_A) | BIT(PORT_C);
> break;
> }
>
> - if (!dev_priv->vbt.dsi.config->cabc_supported)
> + if (!i915->vbt.dsi.config->cabc_supported)
> return;
>
> - switch (dev_priv->vbt.dsi.config->dl_dcs_cabc_ports) {
> + switch (i915->vbt.dsi.config->dl_dcs_cabc_ports) {
> case DL_DCS_PORT_A:
> - dev_priv->vbt.dsi.cabc_ports = BIT(PORT_A);
> + i915->vbt.dsi.cabc_ports = BIT(PORT_A);
> break;
> case DL_DCS_PORT_C:
> - dev_priv->vbt.dsi.cabc_ports = BIT(PORT_C);
> + i915->vbt.dsi.cabc_ports = BIT(PORT_C);
> break;
> default:
> case DL_DCS_PORT_A_AND_C:
> - dev_priv->vbt.dsi.cabc_ports =
> + i915->vbt.dsi.cabc_ports =
> BIT(PORT_A) | BIT(PORT_C);
> break;
> }
> }
>
> static void
> -parse_mipi_config(struct drm_i915_private *dev_priv,
> +parse_mipi_config(struct drm_i915_private *i915,
> const struct bdb_header *bdb)
> {
> const struct bdb_mipi_config *start;
> const struct mipi_config *config;
> const struct mipi_pps_data *pps;
> - int panel_type = dev_priv->vbt.panel_type;
> + int panel_type = i915->vbt.panel_type;
> enum port port;
>
> /* parse MIPI blocks only if LFP type is MIPI */
> - if (!intel_bios_is_dsi_present(dev_priv, &port))
> + if (!intel_bios_is_dsi_present(i915, &port))
> return;
>
> /* Initialize this to undefined indicating no generic MIPI support */
> - dev_priv->vbt.dsi.panel_id = MIPI_DSI_UNDEFINED_PANEL_ID;
> + i915->vbt.dsi.panel_id = MIPI_DSI_UNDEFINED_PANEL_ID;
>
> /* Block #40 is already parsed and panel_fixed_mode is
> - * stored in dev_priv->lfp_lvds_vbt_mode
> + * stored in i915->lfp_lvds_vbt_mode
> * resuse this when needed
> */
>
> @@ -873,17 +873,17 @@ parse_mipi_config(struct drm_i915_private *dev_priv,
> pps = &start->pps[panel_type];
>
> /* store as of now full data. Trim when we realise all is not needed */
> - dev_priv->vbt.dsi.config = kmemdup(config, sizeof(struct mipi_config), GFP_KERNEL);
> - if (!dev_priv->vbt.dsi.config)
> + i915->vbt.dsi.config = kmemdup(config, sizeof(struct mipi_config), GFP_KERNEL);
> + if (!i915->vbt.dsi.config)
> return;
>
> - dev_priv->vbt.dsi.pps = kmemdup(pps, sizeof(struct mipi_pps_data), GFP_KERNEL);
> - if (!dev_priv->vbt.dsi.pps) {
> - kfree(dev_priv->vbt.dsi.config);
> + i915->vbt.dsi.pps = kmemdup(pps, sizeof(struct mipi_pps_data), GFP_KERNEL);
> + if (!i915->vbt.dsi.pps) {
> + kfree(i915->vbt.dsi.config);
> return;
> }
>
> - parse_dsi_backlight_ports(dev_priv, bdb->version, port);
> + parse_dsi_backlight_ports(i915, bdb->version, port);
>
> /* FIXME is the 90 vs. 270 correct? */
> switch (config->rotation) {
> @@ -892,25 +892,25 @@ parse_mipi_config(struct drm_i915_private *dev_priv,
> * Most (all?) VBTs claim 0 degrees despite having
> * an upside down panel, thus we do not trust this.
> */
> - dev_priv->vbt.dsi.orientation =
> + i915->vbt.dsi.orientation =
> DRM_MODE_PANEL_ORIENTATION_UNKNOWN;
> break;
> case ENABLE_ROTATION_90:
> - dev_priv->vbt.dsi.orientation =
> + i915->vbt.dsi.orientation =
> DRM_MODE_PANEL_ORIENTATION_RIGHT_UP;
> break;
> case ENABLE_ROTATION_180:
> - dev_priv->vbt.dsi.orientation =
> + i915->vbt.dsi.orientation =
> DRM_MODE_PANEL_ORIENTATION_BOTTOM_UP;
> break;
> case ENABLE_ROTATION_270:
> - dev_priv->vbt.dsi.orientation =
> + i915->vbt.dsi.orientation =
> DRM_MODE_PANEL_ORIENTATION_LEFT_UP;
> break;
> }
>
> /* We have mandatory mipi config blocks. Initialize as generic panel */
> - dev_priv->vbt.dsi.panel_id = MIPI_DSI_GENERIC_PANEL_ID;
> + i915->vbt.dsi.panel_id = MIPI_DSI_GENERIC_PANEL_ID;
> }
>
> /* Find the sequence block and size for the given panel. */
> @@ -1073,12 +1073,12 @@ static int goto_next_sequence_v3(const u8 *data, int index, int total)
> * Get len of pre-fixed deassert fragment from a v1 init OTP sequence,
> * skip all delay + gpio operands and stop at the first DSI packet op.
> */
> -static int get_init_otp_deassert_fragment_len(struct drm_i915_private *dev_priv)
> +static int get_init_otp_deassert_fragment_len(struct drm_i915_private *i915)
> {
> - const u8 *data = dev_priv->vbt.dsi.sequence[MIPI_SEQ_INIT_OTP];
> + const u8 *data = i915->vbt.dsi.sequence[MIPI_SEQ_INIT_OTP];
> int index, len;
>
> - if (WARN_ON(!data || dev_priv->vbt.dsi.seq_version != 1))
> + if (WARN_ON(!data || i915->vbt.dsi.seq_version != 1))
> return 0;
>
> /* index = 1 to skip sequence byte */
> @@ -1106,54 +1106,54 @@ static int get_init_otp_deassert_fragment_len(struct drm_i915_private *dev_priv)
> * these devices we split the init OTP sequence into a deassert sequence and
> * the actual init OTP part.
> */
> -static void fixup_mipi_sequences(struct drm_i915_private *dev_priv)
> +static void fixup_mipi_sequences(struct drm_i915_private *i915)
> {
> u8 *init_otp;
> int len;
>
> /* Limit this to VLV for now. */
> - if (!IS_VALLEYVIEW(dev_priv))
> + if (!IS_VALLEYVIEW(i915))
> return;
>
> /* Limit this to v1 vid-mode sequences */
> - if (dev_priv->vbt.dsi.config->is_cmd_mode ||
> - dev_priv->vbt.dsi.seq_version != 1)
> + if (i915->vbt.dsi.config->is_cmd_mode ||
> + i915->vbt.dsi.seq_version != 1)
> return;
>
> /* Only do this if there are otp and assert seqs and no deassert seq */
> - if (!dev_priv->vbt.dsi.sequence[MIPI_SEQ_INIT_OTP] ||
> - !dev_priv->vbt.dsi.sequence[MIPI_SEQ_ASSERT_RESET] ||
> - dev_priv->vbt.dsi.sequence[MIPI_SEQ_DEASSERT_RESET])
> + if (!i915->vbt.dsi.sequence[MIPI_SEQ_INIT_OTP] ||
> + !i915->vbt.dsi.sequence[MIPI_SEQ_ASSERT_RESET] ||
> + i915->vbt.dsi.sequence[MIPI_SEQ_DEASSERT_RESET])
> return;
>
> /* The deassert-sequence ends at the first DSI packet */
> - len = get_init_otp_deassert_fragment_len(dev_priv);
> + len = get_init_otp_deassert_fragment_len(i915);
> if (!len)
> return;
>
> DRM_DEBUG_KMS("Using init OTP fragment to deassert reset\n");
>
> /* Copy the fragment, update seq byte and terminate it */
> - init_otp = (u8 *)dev_priv->vbt.dsi.sequence[MIPI_SEQ_INIT_OTP];
> - dev_priv->vbt.dsi.deassert_seq = kmemdup(init_otp, len + 1, GFP_KERNEL);
> - if (!dev_priv->vbt.dsi.deassert_seq)
> + init_otp = (u8 *)i915->vbt.dsi.sequence[MIPI_SEQ_INIT_OTP];
> + i915->vbt.dsi.deassert_seq = kmemdup(init_otp, len + 1, GFP_KERNEL);
> + if (!i915->vbt.dsi.deassert_seq)
> return;
> - dev_priv->vbt.dsi.deassert_seq[0] = MIPI_SEQ_DEASSERT_RESET;
> - dev_priv->vbt.dsi.deassert_seq[len] = MIPI_SEQ_ELEM_END;
> + i915->vbt.dsi.deassert_seq[0] = MIPI_SEQ_DEASSERT_RESET;
> + i915->vbt.dsi.deassert_seq[len] = MIPI_SEQ_ELEM_END;
> /* Use the copy for deassert */
> - dev_priv->vbt.dsi.sequence[MIPI_SEQ_DEASSERT_RESET] =
> - dev_priv->vbt.dsi.deassert_seq;
> + i915->vbt.dsi.sequence[MIPI_SEQ_DEASSERT_RESET] =
> + i915->vbt.dsi.deassert_seq;
> /* Replace the last byte of the fragment with init OTP seq byte */
> init_otp[len - 1] = MIPI_SEQ_INIT_OTP;
> /* And make MIPI_MIPI_SEQ_INIT_OTP point to it */
> - dev_priv->vbt.dsi.sequence[MIPI_SEQ_INIT_OTP] = init_otp + len - 1;
> + i915->vbt.dsi.sequence[MIPI_SEQ_INIT_OTP] = init_otp + len - 1;
> }
>
> static void
> -parse_mipi_sequence(struct drm_i915_private *dev_priv,
> +parse_mipi_sequence(struct drm_i915_private *i915,
> const struct bdb_header *bdb)
> {
> - int panel_type = dev_priv->vbt.panel_type;
> + int panel_type = i915->vbt.panel_type;
> const struct bdb_mipi_sequence *sequence;
> const u8 *seq_data;
> u32 seq_size;
> @@ -1161,7 +1161,7 @@ parse_mipi_sequence(struct drm_i915_private *dev_priv,
> int index = 0;
>
> /* Only our generic panel driver uses the sequence block. */
> - if (dev_priv->vbt.dsi.panel_id != MIPI_DSI_GENERIC_PANEL_ID)
> + if (i915->vbt.dsi.panel_id != MIPI_DSI_GENERIC_PANEL_ID)
> return;
>
> sequence = find_section(bdb, BDB_MIPI_SEQUENCE);
> @@ -1202,7 +1202,7 @@ parse_mipi_sequence(struct drm_i915_private *dev_priv,
> if (seq_id == MIPI_SEQ_TEAR_ON || seq_id == MIPI_SEQ_TEAR_OFF)
> DRM_DEBUG_KMS("Unsupported sequence %u\n", seq_id);
>
> - dev_priv->vbt.dsi.sequence[seq_id] = data + index;
> + i915->vbt.dsi.sequence[seq_id] = data + index;
>
> if (sequence->version >= 3)
> index = goto_next_sequence_v3(data, index, seq_size);
> @@ -1214,18 +1214,18 @@ parse_mipi_sequence(struct drm_i915_private *dev_priv,
> }
> }
>
> - dev_priv->vbt.dsi.data = data;
> - dev_priv->vbt.dsi.size = seq_size;
> - dev_priv->vbt.dsi.seq_version = sequence->version;
> + i915->vbt.dsi.data = data;
> + i915->vbt.dsi.size = seq_size;
> + i915->vbt.dsi.seq_version = sequence->version;
>
> - fixup_mipi_sequences(dev_priv);
> + fixup_mipi_sequences(i915);
>
> DRM_DEBUG_DRIVER("MIPI related VBT parsing complete\n");
> return;
>
> err:
> kfree(data);
> - memset(dev_priv->vbt.dsi.sequence, 0, sizeof(dev_priv->vbt.dsi.sequence));
> + memset(i915->vbt.dsi.sequence, 0, sizeof(i915->vbt.dsi.sequence));
> }
>
> static u8 translate_iboost(u8 val)
> @@ -1254,16 +1254,16 @@ static enum port get_port_by_ddc_pin(struct drm_i915_private *i915, u8 ddc_pin)
> return PORT_NONE;
> }
>
> -static void sanitize_ddc_pin(struct drm_i915_private *dev_priv,
> +static void sanitize_ddc_pin(struct drm_i915_private *i915,
> enum port port)
> {
> - struct ddi_vbt_port_info *info = &dev_priv->vbt.ddi_port_info[port];
> + struct ddi_vbt_port_info *info = &i915->vbt.ddi_port_info[port];
> enum port p;
>
> if (!info->alternate_ddc_pin)
> return;
>
> - p = get_port_by_ddc_pin(dev_priv, info->alternate_ddc_pin);
> + p = get_port_by_ddc_pin(i915, info->alternate_ddc_pin);
> if (p != PORT_NONE) {
> DRM_DEBUG_KMS("port %c trying to use the same DDC pin (0x%x) as port %c, "
> "disabling port %c DVI/HDMI support\n",
> @@ -1300,16 +1300,16 @@ static enum port get_port_by_aux_ch(struct drm_i915_private *i915, u8 aux_ch)
> return PORT_NONE;
> }
>
> -static void sanitize_aux_ch(struct drm_i915_private *dev_priv,
> +static void sanitize_aux_ch(struct drm_i915_private *i915,
> enum port port)
> {
> - struct ddi_vbt_port_info *info = &dev_priv->vbt.ddi_port_info[port];
> + struct ddi_vbt_port_info *info = &i915->vbt.ddi_port_info[port];
> enum port p;
>
> if (!info->alternate_aux_channel)
> return;
>
> - p = get_port_by_aux_ch(dev_priv, info->alternate_aux_channel);
> + p = get_port_by_aux_ch(i915, info->alternate_aux_channel);
> if (p != PORT_NONE) {
> DRM_DEBUG_KMS("port %c trying to use the same AUX CH (0x%x) as port %c, "
> "disabling port %c DP support\n",
> @@ -1347,15 +1347,15 @@ static const u8 icp_ddc_pin_map[] = {
> [ICL_DDC_BUS_PORT_4] = GMBUS_PIN_12_TC4_ICP,
> };
>
> -static u8 map_ddc_pin(struct drm_i915_private *dev_priv, u8 vbt_pin)
> +static u8 map_ddc_pin(struct drm_i915_private *i915, u8 vbt_pin)
> {
> const u8 *ddc_pin_map;
> int n_entries;
>
> - if (HAS_PCH_ICP(dev_priv)) {
> + if (HAS_PCH_ICP(i915)) {
> ddc_pin_map = icp_ddc_pin_map;
> n_entries = ARRAY_SIZE(icp_ddc_pin_map);
> - } else if (HAS_PCH_CNP(dev_priv)) {
> + } else if (HAS_PCH_CNP(i915)) {
> ddc_pin_map = cnp_ddc_pin_map;
> n_entries = ARRAY_SIZE(cnp_ddc_pin_map);
> } else {
> @@ -1401,7 +1401,7 @@ static enum port dvo_port_to_port(u8 dvo_port)
> return PORT_NONE;
> }
>
> -static void parse_ddi_port(struct drm_i915_private *dev_priv,
> +static void parse_ddi_port(struct drm_i915_private *i915,
> const struct child_device_config *child,
> u8 bdb_version)
> {
> @@ -1413,7 +1413,7 @@ static void parse_ddi_port(struct drm_i915_private *dev_priv,
> if (port == PORT_NONE)
> return;
>
> - info = &dev_priv->vbt.ddi_port_info[port];
> + info = &i915->vbt.ddi_port_info[port];
>
> if (info->child) {
> DRM_DEBUG_KMS("More than one child device for port %c in VBT, using the first.\n",
> @@ -1447,7 +1447,7 @@ static void parse_ddi_port(struct drm_i915_private *dev_priv,
>
> DRM_DEBUG_KMS("Port %c VBT info: CRT:%d DVI:%d HDMI:%d DP:%d eDP:%d LSPCON:%d USB-Type-C:%d TBT:%d\n",
> port_name(port), is_crt, is_dvi, is_hdmi, is_dp, is_edp,
> - HAS_LSPCON(dev_priv) && child->lspcon,
> + HAS_LSPCON(i915) && child->lspcon,
> info->supports_typec_usb, info->supports_tbt);
>
> if (is_edp && is_dvi)
> @@ -1469,10 +1469,10 @@ static void parse_ddi_port(struct drm_i915_private *dev_priv,
> if (is_dvi) {
> u8 ddc_pin;
>
> - ddc_pin = map_ddc_pin(dev_priv, child->ddc_pin);
> - if (intel_gmbus_is_valid_pin(dev_priv, ddc_pin)) {
> + ddc_pin = map_ddc_pin(i915, child->ddc_pin);
> + if (intel_gmbus_is_valid_pin(i915, ddc_pin)) {
> info->alternate_ddc_pin = ddc_pin;
> - sanitize_ddc_pin(dev_priv, port);
> + sanitize_ddc_pin(i915, port);
> } else {
> DRM_DEBUG_KMS("Port %c has invalid DDC pin %d, "
> "sticking to defaults\n",
> @@ -1483,7 +1483,7 @@ static void parse_ddi_port(struct drm_i915_private *dev_priv,
> if (is_dp) {
> info->alternate_aux_channel = child->aux_channel;
>
> - sanitize_aux_ch(dev_priv, port);
> + sanitize_aux_ch(i915, port);
> }
>
> if (bdb_version >= 158) {
> @@ -1553,26 +1553,26 @@ static void parse_ddi_port(struct drm_i915_private *dev_priv,
> info->child = child;
> }
>
> -static void parse_ddi_ports(struct drm_i915_private *dev_priv, u8 bdb_version)
> +static void parse_ddi_ports(struct drm_i915_private *i915, u8 bdb_version)
> {
> const struct child_device_config *child;
> int i;
>
> - if (!HAS_DDI(dev_priv) && !IS_CHERRYVIEW(dev_priv))
> + if (!HAS_DDI(i915) && !IS_CHERRYVIEW(i915))
> return;
>
> if (bdb_version < 155)
> return;
>
> - for (i = 0; i < dev_priv->vbt.child_dev_num; i++) {
> - child = dev_priv->vbt.child_dev + i;
> + for (i = 0; i < i915->vbt.child_dev_num; i++) {
> + child = i915->vbt.child_dev + i;
>
> - parse_ddi_port(dev_priv, child, bdb_version);
> + parse_ddi_port(i915, child, bdb_version);
> }
> }
>
> static void
> -parse_general_definitions(struct drm_i915_private *dev_priv,
> +parse_general_definitions(struct drm_i915_private *i915,
> const struct bdb_header *bdb)
> {
> const struct bdb_general_definitions *defs;
> @@ -1597,8 +1597,8 @@ parse_general_definitions(struct drm_i915_private *dev_priv,
>
> bus_pin = defs->crt_ddc_gmbus_pin;
> DRM_DEBUG_KMS("crt_ddc_bus_pin: %d\n", bus_pin);
> - if (intel_gmbus_is_valid_pin(dev_priv, bus_pin))
> - dev_priv->vbt.crt_ddc_pin = bus_pin;
> + if (intel_gmbus_is_valid_pin(i915, bus_pin))
> + i915->vbt.crt_ddc_pin = bus_pin;
>
> if (bdb->version < 106) {
> expected_size = 22;
> @@ -1645,13 +1645,13 @@ parse_general_definitions(struct drm_i915_private *dev_priv,
> DRM_DEBUG_KMS("no child dev is parsed from VBT\n");
> return;
> }
> - dev_priv->vbt.child_dev = kcalloc(count, sizeof(*child), GFP_KERNEL);
> - if (!dev_priv->vbt.child_dev) {
> + i915->vbt.child_dev = kcalloc(count, sizeof(*child), GFP_KERNEL);
> + if (!i915->vbt.child_dev) {
> DRM_DEBUG_KMS("No memory space for child device\n");
> return;
> }
>
> - dev_priv->vbt.child_dev_num = count;
> + i915->vbt.child_dev_num = count;
> count = 0;
> for (i = 0; i < child_device_num; i++) {
> child = child_device_ptr(defs, i);
> @@ -1663,7 +1663,7 @@ parse_general_definitions(struct drm_i915_private *dev_priv,
> * (child_dev_size) of the child device. Accessing the data must
> * depend on VBT version.
> */
> - memcpy(dev_priv->vbt.child_dev + count, child,
> + memcpy(i915->vbt.child_dev + count, child,
> min_t(size_t, defs->child_dev_size, sizeof(*child)));
> count++;
> }
> @@ -1671,41 +1671,41 @@ parse_general_definitions(struct drm_i915_private *dev_priv,
>
> /* Common defaults which may be overridden by VBT. */
> static void
> -init_vbt_defaults(struct drm_i915_private *dev_priv)
> +init_vbt_defaults(struct drm_i915_private *i915)
> {
> enum port port;
>
> - dev_priv->vbt.crt_ddc_pin = GMBUS_PIN_VGADDC;
> + i915->vbt.crt_ddc_pin = GMBUS_PIN_VGADDC;
>
> /* Default to having backlight */
> - dev_priv->vbt.backlight.present = true;
> + i915->vbt.backlight.present = true;
>
> /* LFP panel data */
> - dev_priv->vbt.lvds_dither = 1;
> + i915->vbt.lvds_dither = 1;
>
> /* SDVO panel data */
> - dev_priv->vbt.sdvo_lvds_vbt_mode = NULL;
> + i915->vbt.sdvo_lvds_vbt_mode = NULL;
>
> /* general features */
> - dev_priv->vbt.int_tv_support = 1;
> - dev_priv->vbt.int_crt_support = 1;
> + i915->vbt.int_tv_support = 1;
> + i915->vbt.int_crt_support = 1;
>
> /* driver features */
> - dev_priv->vbt.int_lvds_support = 1;
> + i915->vbt.int_lvds_support = 1;
>
> /* Default to using SSC */
> - dev_priv->vbt.lvds_use_ssc = 1;
> + i915->vbt.lvds_use_ssc = 1;
> /*
> * Core/SandyBridge/IvyBridge use alternative (120MHz) reference
> * clock for LVDS.
> */
> - dev_priv->vbt.lvds_ssc_freq = intel_bios_ssc_frequency(dev_priv,
> - !HAS_PCH_SPLIT(dev_priv));
> - DRM_DEBUG_KMS("Set default to SSC at %d kHz\n", dev_priv->vbt.lvds_ssc_freq);
> + i915->vbt.lvds_ssc_freq = intel_bios_ssc_frequency(i915,
> + !HAS_PCH_SPLIT(i915));
> + DRM_DEBUG_KMS("Set default to SSC at %d kHz\n", i915->vbt.lvds_ssc_freq);
>
> for (port = PORT_A; port < I915_MAX_PORTS; port++) {
> struct ddi_vbt_port_info *info =
> - &dev_priv->vbt.ddi_port_info[port];
> + &i915->vbt.ddi_port_info[port];
>
> info->hdmi_level_shift = HDMI_LEVEL_SHIFT_UNKNOWN;
> }
> @@ -1713,19 +1713,19 @@ init_vbt_defaults(struct drm_i915_private *dev_priv)
>
> /* Defaults to initialize only if there is no VBT. */
> static void
> -init_vbt_missing_defaults(struct drm_i915_private *dev_priv)
> +init_vbt_missing_defaults(struct drm_i915_private *i915)
> {
> enum port port;
>
> for (port = PORT_A; port < I915_MAX_PORTS; port++) {
> struct ddi_vbt_port_info *info =
> - &dev_priv->vbt.ddi_port_info[port];
> + &i915->vbt.ddi_port_info[port];
>
> /*
> * VBT has the TypeC mode (native,TBT/USB) and we don't want
> * to detect it.
> */
> - if (intel_port_is_tc(dev_priv, port))
> + if (intel_port_is_tc(i915, port))
> continue;
>
> info->supports_dvi = (port != PORT_A && port != PORT_E);
> @@ -1811,25 +1811,25 @@ static const struct vbt_header *find_vbt(void __iomem *bios, size_t size)
>
> /**
> * intel_bios_init - find VBT and initialize settings from the BIOS
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> *
> * Parse and initialize settings from the Video BIOS Tables (VBT). If the VBT
> * was not found in ACPI OpRegion, try to find it in PCI ROM first. Also
> * initialize some defaults if the VBT is not present at all.
> */
> -void intel_bios_init(struct drm_i915_private *dev_priv)
> +void intel_bios_init(struct drm_i915_private *i915)
> {
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> - const struct vbt_header *vbt = dev_priv->opregion.vbt;
> + struct pci_dev *pdev = i915->drm.pdev;
> + const struct vbt_header *vbt = i915->opregion.vbt;
> const struct bdb_header *bdb;
> u8 __iomem *bios = NULL;
>
> - if (!HAS_DISPLAY(dev_priv)) {
> + if (!HAS_DISPLAY(i915)) {
> DRM_DEBUG_KMS("Skipping VBT init due to disabled display.\n");
> return;
> }
>
> - init_vbt_defaults(dev_priv);
> + init_vbt_defaults(i915);
>
> /* If the OpRegion does not have VBT, look in PCI ROM. */
> if (!vbt) {
> @@ -1852,25 +1852,25 @@ void intel_bios_init(struct drm_i915_private *dev_priv)
> (int)sizeof(vbt->signature), vbt->signature, bdb->version);
>
> /* Grab useful general definitions */
> - parse_general_features(dev_priv, bdb);
> - parse_general_definitions(dev_priv, bdb);
> - parse_lfp_panel_data(dev_priv, bdb);
> - parse_lfp_backlight(dev_priv, bdb);
> - parse_sdvo_panel_data(dev_priv, bdb);
> - parse_driver_features(dev_priv, bdb);
> - parse_edp(dev_priv, bdb);
> - parse_psr(dev_priv, bdb);
> - parse_mipi_config(dev_priv, bdb);
> - parse_mipi_sequence(dev_priv, bdb);
> + parse_general_features(i915, bdb);
> + parse_general_definitions(i915, bdb);
> + parse_lfp_panel_data(i915, bdb);
> + parse_lfp_backlight(i915, bdb);
> + parse_sdvo_panel_data(i915, bdb);
> + parse_driver_features(i915, bdb);
> + parse_edp(i915, bdb);
> + parse_psr(i915, bdb);
> + parse_mipi_config(i915, bdb);
> + parse_mipi_sequence(i915, bdb);
>
> /* Further processing on pre-parsed data */
> - parse_sdvo_device_mapping(dev_priv, bdb->version);
> - parse_ddi_ports(dev_priv, bdb->version);
> + parse_sdvo_device_mapping(i915, bdb->version);
> + parse_ddi_ports(i915, bdb->version);
>
> out:
> if (!vbt) {
> DRM_INFO("Failed to find VBIOS tables (VBT)\n");
> - init_vbt_missing_defaults(dev_priv);
> + init_vbt_missing_defaults(i915);
> }
>
> if (bios)
> @@ -1879,47 +1879,47 @@ void intel_bios_init(struct drm_i915_private *dev_priv)
>
> /**
> * intel_bios_cleanup - Free any resources allocated by intel_bios_init()
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> */
> -void intel_bios_cleanup(struct drm_i915_private *dev_priv)
> +void intel_bios_cleanup(struct drm_i915_private *i915)
> {
> - kfree(dev_priv->vbt.child_dev);
> - dev_priv->vbt.child_dev = NULL;
> - dev_priv->vbt.child_dev_num = 0;
> - kfree(dev_priv->vbt.sdvo_lvds_vbt_mode);
> - dev_priv->vbt.sdvo_lvds_vbt_mode = NULL;
> - kfree(dev_priv->vbt.lfp_lvds_vbt_mode);
> - dev_priv->vbt.lfp_lvds_vbt_mode = NULL;
> - kfree(dev_priv->vbt.dsi.data);
> - dev_priv->vbt.dsi.data = NULL;
> - kfree(dev_priv->vbt.dsi.pps);
> - dev_priv->vbt.dsi.pps = NULL;
> - kfree(dev_priv->vbt.dsi.config);
> - dev_priv->vbt.dsi.config = NULL;
> - kfree(dev_priv->vbt.dsi.deassert_seq);
> - dev_priv->vbt.dsi.deassert_seq = NULL;
> + kfree(i915->vbt.child_dev);
> + i915->vbt.child_dev = NULL;
> + i915->vbt.child_dev_num = 0;
> + kfree(i915->vbt.sdvo_lvds_vbt_mode);
> + i915->vbt.sdvo_lvds_vbt_mode = NULL;
> + kfree(i915->vbt.lfp_lvds_vbt_mode);
> + i915->vbt.lfp_lvds_vbt_mode = NULL;
> + kfree(i915->vbt.dsi.data);
> + i915->vbt.dsi.data = NULL;
> + kfree(i915->vbt.dsi.pps);
> + i915->vbt.dsi.pps = NULL;
> + kfree(i915->vbt.dsi.config);
> + i915->vbt.dsi.config = NULL;
> + kfree(i915->vbt.dsi.deassert_seq);
> + i915->vbt.dsi.deassert_seq = NULL;
> }
>
> /**
> * intel_bios_is_tv_present - is integrated TV present in VBT
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> *
> * Return true if TV is present. If no child devices were parsed from VBT,
> * assume TV is present.
> */
> -bool intel_bios_is_tv_present(struct drm_i915_private *dev_priv)
> +bool intel_bios_is_tv_present(struct drm_i915_private *i915)
> {
> const struct child_device_config *child;
> int i;
>
> - if (!dev_priv->vbt.int_tv_support)
> + if (!i915->vbt.int_tv_support)
> return false;
>
> - if (!dev_priv->vbt.child_dev_num)
> + if (!i915->vbt.child_dev_num)
> return true;
>
> - for (i = 0; i < dev_priv->vbt.child_dev_num; i++) {
> - child = dev_priv->vbt.child_dev + i;
> + for (i = 0; i < i915->vbt.child_dev_num; i++) {
> + child = i915->vbt.child_dev + i;
> /*
> * If the device type is not TV, continue.
> */
> @@ -1943,22 +1943,22 @@ bool intel_bios_is_tv_present(struct drm_i915_private *dev_priv)
>
> /**
> * intel_bios_is_lvds_present - is LVDS present in VBT
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> * @i2c_pin: i2c pin for LVDS if present
> *
> * Return true if LVDS is present. If no child devices were parsed from VBT,
> * assume LVDS is present.
> */
> -bool intel_bios_is_lvds_present(struct drm_i915_private *dev_priv, u8 *i2c_pin)
> +bool intel_bios_is_lvds_present(struct drm_i915_private *i915, u8 *i2c_pin)
> {
> const struct child_device_config *child;
> int i;
>
> - if (!dev_priv->vbt.child_dev_num)
> + if (!i915->vbt.child_dev_num)
> return true;
>
> - for (i = 0; i < dev_priv->vbt.child_dev_num; i++) {
> - child = dev_priv->vbt.child_dev + i;
> + for (i = 0; i < i915->vbt.child_dev_num; i++) {
> + child = i915->vbt.child_dev + i;
>
> /* If the device type is not LFP, continue.
> * We have to check both the new identifiers as well as the
> @@ -1968,7 +1968,7 @@ bool intel_bios_is_lvds_present(struct drm_i915_private *dev_priv, u8 *i2c_pin)
> child->device_type != DEVICE_TYPE_LFP)
> continue;
>
> - if (intel_gmbus_is_valid_pin(dev_priv, child->i2c_pin))
> + if (intel_gmbus_is_valid_pin(i915, child->i2c_pin))
> *i2c_pin = child->i2c_pin;
>
> /* However, we cannot trust the BIOS writers to populate
> @@ -1984,7 +1984,7 @@ bool intel_bios_is_lvds_present(struct drm_i915_private *dev_priv, u8 *i2c_pin)
> * additional data. Trust that if the VBT was written into
> * the OpRegion then they have validated the LVDS's existence.
> */
> - if (dev_priv->opregion.vbt)
> + if (i915->opregion.vbt)
> return true;
> }
>
> @@ -1993,12 +1993,12 @@ bool intel_bios_is_lvds_present(struct drm_i915_private *dev_priv, u8 *i2c_pin)
>
> /**
> * intel_bios_is_port_present - is the specified digital port present
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> * @port: port to check
> *
> * Return true if the device in %port is present.
> */
> -bool intel_bios_is_port_present(struct drm_i915_private *dev_priv, enum port port)
> +bool intel_bios_is_port_present(struct drm_i915_private *i915, enum port port)
> {
> const struct child_device_config *child;
> static const struct {
> @@ -2012,9 +2012,9 @@ bool intel_bios_is_port_present(struct drm_i915_private *dev_priv, enum port por
> };
> int i;
>
> - if (HAS_DDI(dev_priv)) {
> + if (HAS_DDI(i915)) {
> const struct ddi_vbt_port_info *port_info =
> - &dev_priv->vbt.ddi_port_info[port];
> + &i915->vbt.ddi_port_info[port];
>
> return port_info->supports_dp ||
> port_info->supports_dvi ||
> @@ -2025,11 +2025,11 @@ bool intel_bios_is_port_present(struct drm_i915_private *dev_priv, enum port por
> if (WARN_ON(port == PORT_A) || port >= ARRAY_SIZE(port_mapping))
> return false;
>
> - if (!dev_priv->vbt.child_dev_num)
> + if (!i915->vbt.child_dev_num)
> return false;
>
> - for (i = 0; i < dev_priv->vbt.child_dev_num; i++) {
> - child = dev_priv->vbt.child_dev + i;
> + for (i = 0; i < i915->vbt.child_dev_num; i++) {
> + child = i915->vbt.child_dev + i;
>
> if ((child->dvo_port == port_mapping[port].dp ||
> child->dvo_port == port_mapping[port].hdmi) &&
> @@ -2043,12 +2043,12 @@ bool intel_bios_is_port_present(struct drm_i915_private *dev_priv, enum port por
>
> /**
> * intel_bios_is_port_edp - is the device in given port eDP
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> * @port: port to check
> *
> * Return true if the device in %port is eDP.
> */
> -bool intel_bios_is_port_edp(struct drm_i915_private *dev_priv, enum port port)
> +bool intel_bios_is_port_edp(struct drm_i915_private *i915, enum port port)
> {
> const struct child_device_config *child;
> static const short port_mapping[] = {
> @@ -2060,14 +2060,14 @@ bool intel_bios_is_port_edp(struct drm_i915_private *dev_priv, enum port port)
> };
> int i;
>
> - if (HAS_DDI(dev_priv))
> - return dev_priv->vbt.ddi_port_info[port].supports_edp;
> + if (HAS_DDI(i915))
> + return i915->vbt.ddi_port_info[port].supports_edp;
>
> - if (!dev_priv->vbt.child_dev_num)
> + if (!i915->vbt.child_dev_num)
> return false;
>
> - for (i = 0; i < dev_priv->vbt.child_dev_num; i++) {
> - child = dev_priv->vbt.child_dev + i;
> + for (i = 0; i < i915->vbt.child_dev_num; i++) {
> + child = i915->vbt.child_dev + i;
>
> if (child->dvo_port == port_mapping[port] &&
> (child->device_type & DEVICE_TYPE_eDP_BITS) ==
> @@ -2113,14 +2113,14 @@ static bool child_dev_is_dp_dual_mode(const struct child_device_config *child,
> return false;
> }
>
> -bool intel_bios_is_port_dp_dual_mode(struct drm_i915_private *dev_priv,
> +bool intel_bios_is_port_dp_dual_mode(struct drm_i915_private *i915,
> enum port port)
> {
> const struct child_device_config *child;
> int i;
>
> - for (i = 0; i < dev_priv->vbt.child_dev_num; i++) {
> - child = dev_priv->vbt.child_dev + i;
> + for (i = 0; i < i915->vbt.child_dev_num; i++) {
> + child = i915->vbt.child_dev + i;
>
> if (child_dev_is_dp_dual_mode(child, port))
> return true;
> @@ -2131,20 +2131,20 @@ bool intel_bios_is_port_dp_dual_mode(struct drm_i915_private *dev_priv,
>
> /**
> * intel_bios_is_dsi_present - is DSI present in VBT
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> * @port: port for DSI if present
> *
> * Return true if DSI is present, and return the port in %port.
> */
> -bool intel_bios_is_dsi_present(struct drm_i915_private *dev_priv,
> +bool intel_bios_is_dsi_present(struct drm_i915_private *i915,
> enum port *port)
> {
> const struct child_device_config *child;
> u8 dvo_port;
> int i;
>
> - for (i = 0; i < dev_priv->vbt.child_dev_num; i++) {
> - child = dev_priv->vbt.child_dev + i;
> + for (i = 0; i < i915->vbt.child_dev_num; i++) {
> + child = i915->vbt.child_dev + i;
>
> if (!(child->device_type & DEVICE_TYPE_MIPI_OUTPUT))
> continue;
> @@ -2152,8 +2152,8 @@ bool intel_bios_is_dsi_present(struct drm_i915_private *dev_priv,
> dvo_port = child->dvo_port;
>
> if (dvo_port == DVO_PORT_MIPIA ||
> - (dvo_port == DVO_PORT_MIPIB && INTEL_GEN(dev_priv) >= 11) ||
> - (dvo_port == DVO_PORT_MIPIC && INTEL_GEN(dev_priv) < 11)) {
> + (dvo_port == DVO_PORT_MIPIB && INTEL_GEN(i915) >= 11) ||
> + (dvo_port == DVO_PORT_MIPIC && INTEL_GEN(i915) < 11)) {
> if (port)
> *port = dvo_port - DVO_PORT_MIPIA;
> return true;
> @@ -2205,11 +2205,11 @@ intel_bios_is_lspcon_present(const struct drm_i915_private *i915,
> return HAS_LSPCON(i915) && child && child->lspcon;
> }
>
> -enum aux_ch intel_bios_port_aux_ch(struct drm_i915_private *dev_priv,
> +enum aux_ch intel_bios_port_aux_ch(struct drm_i915_private *i915,
> enum port port)
> {
> const struct ddi_vbt_port_info *info =
> - &dev_priv->vbt.ddi_port_info[port];
> + &i915->vbt.ddi_port_info[port];
> enum aux_ch aux_ch;
>
> if (!info->alternate_aux_channel) {
> diff --git a/drivers/gpu/drm/i915/intel_bios.h b/drivers/gpu/drm/i915/intel_bios.h
> index 4e42cfaf61a7..d292bd1790ae 100644
> --- a/drivers/gpu/drm/i915/intel_bios.h
> +++ b/drivers/gpu/drm/i915/intel_bios.h
> @@ -226,19 +226,19 @@ struct mipi_pps_data {
> u16 panel_power_cycle_delay;
> } __packed;
>
> -void intel_bios_init(struct drm_i915_private *dev_priv);
> -void intel_bios_cleanup(struct drm_i915_private *dev_priv);
> +void intel_bios_init(struct drm_i915_private *i915);
> +void intel_bios_cleanup(struct drm_i915_private *i915);
> bool intel_bios_is_valid_vbt(const void *buf, size_t size);
> -bool intel_bios_is_tv_present(struct drm_i915_private *dev_priv);
> -bool intel_bios_is_lvds_present(struct drm_i915_private *dev_priv, u8 *i2c_pin);
> -bool intel_bios_is_port_present(struct drm_i915_private *dev_priv, enum port port);
> -bool intel_bios_is_port_edp(struct drm_i915_private *dev_priv, enum port port);
> -bool intel_bios_is_port_dp_dual_mode(struct drm_i915_private *dev_priv, enum port port);
> -bool intel_bios_is_dsi_present(struct drm_i915_private *dev_priv, enum port *port);
> +bool intel_bios_is_tv_present(struct drm_i915_private *i915);
> +bool intel_bios_is_lvds_present(struct drm_i915_private *i915, u8 *i2c_pin);
> +bool intel_bios_is_port_present(struct drm_i915_private *i915, enum port port);
> +bool intel_bios_is_port_edp(struct drm_i915_private *i915, enum port port);
> +bool intel_bios_is_port_dp_dual_mode(struct drm_i915_private *i915, enum port port);
> +bool intel_bios_is_dsi_present(struct drm_i915_private *i915, enum port *port);
> bool intel_bios_is_port_hpd_inverted(const struct drm_i915_private *i915,
> enum port port);
> bool intel_bios_is_lspcon_present(const struct drm_i915_private *i915,
> enum port port);
> -enum aux_ch intel_bios_port_aux_ch(struct drm_i915_private *dev_priv, enum port port);
> +enum aux_ch intel_bios_port_aux_ch(struct drm_i915_private *i915, enum port port);
>
> #endif /* _INTEL_BIOS_H_ */
> diff --git a/drivers/gpu/drm/i915/intel_bw.c b/drivers/gpu/drm/i915/intel_bw.c
> index 753ac3165061..1aca44a258d3 100644
> --- a/drivers/gpu/drm/i915/intel_bw.c
> +++ b/drivers/gpu/drm/i915/intel_bw.c
> @@ -22,13 +22,13 @@ struct intel_qgv_info {
> enum intel_dram_type dram_type;
> };
>
> -static int icl_pcode_read_mem_global_info(struct drm_i915_private *dev_priv,
> +static int icl_pcode_read_mem_global_info(struct drm_i915_private *i915,
> struct intel_qgv_info *qi)
> {
> u32 val = 0;
> int ret;
>
> - ret = sandybridge_pcode_read(dev_priv,
> + ret = sandybridge_pcode_read(i915,
> ICL_PCODE_MEM_SUBSYSYSTEM_INFO |
> ICL_PCODE_MEM_SS_READ_GLOBAL_INFO,
> &val, NULL);
> @@ -61,14 +61,14 @@ static int icl_pcode_read_mem_global_info(struct drm_i915_private *dev_priv,
> return 0;
> }
>
> -static int icl_pcode_read_qgv_point_info(struct drm_i915_private *dev_priv,
> +static int icl_pcode_read_qgv_point_info(struct drm_i915_private *i915,
> struct intel_qgv_point *sp,
> int point)
> {
> u32 val = 0, val2;
> int ret;
>
> - ret = sandybridge_pcode_read(dev_priv,
> + ret = sandybridge_pcode_read(i915,
> ICL_PCODE_MEM_SUBSYSYSTEM_INFO |
> ICL_PCODE_MEM_SS_READ_QGV_POINT_INFO(point),
> &val, &val2);
> @@ -87,12 +87,12 @@ static int icl_pcode_read_qgv_point_info(struct drm_i915_private *dev_priv,
> return 0;
> }
>
> -static int icl_get_qgv_points(struct drm_i915_private *dev_priv,
> +static int icl_get_qgv_points(struct drm_i915_private *i915,
> struct intel_qgv_info *qi)
> {
> int i, ret;
>
> - ret = icl_pcode_read_mem_global_info(dev_priv, qi);
> + ret = icl_pcode_read_mem_global_info(i915, qi);
> if (ret)
> return ret;
>
> @@ -102,7 +102,7 @@ static int icl_get_qgv_points(struct drm_i915_private *dev_priv,
> for (i = 0; i < qi->num_points; i++) {
> struct intel_qgv_point *sp = &qi->points[i];
>
> - ret = icl_pcode_read_qgv_point_info(dev_priv, sp, i);
> + ret = icl_pcode_read_qgv_point_info(i915, sp, i);
> if (ret)
> return ret;
>
> @@ -142,7 +142,7 @@ static const struct intel_sa_info icl_sa_info = {
> .displayrtids = 128,
> };
>
> -static int icl_get_bw_info(struct drm_i915_private *dev_priv)
> +static int icl_get_bw_info(struct drm_i915_private *i915)
> {
> struct intel_qgv_info qi = {};
> const struct intel_sa_info *sa = &icl_sa_info;
> @@ -154,7 +154,7 @@ static int icl_get_bw_info(struct drm_i915_private *dev_priv)
> int maxdebw;
> int i, ret;
>
> - ret = icl_get_qgv_points(dev_priv, &qi);
> + ret = icl_get_qgv_points(i915, &qi);
> if (ret) {
> DRM_DEBUG_KMS("Failed to get memory subsystem information, ignoring bandwidth limits");
> return ret;
> @@ -170,8 +170,8 @@ static int icl_get_bw_info(struct drm_i915_private *dev_priv)
> icl_calc_bw(dclk_max, 16, 1) * 6 / 10); /* 60% */
> ipqdepth = min(ipqdepthpch, sa->displayrtids / num_channels);
>
> - for (i = 0; i < ARRAY_SIZE(dev_priv->max_bw); i++) {
> - struct intel_bw_info *bi = &dev_priv->max_bw[i];
> + for (i = 0; i < ARRAY_SIZE(i915->max_bw); i++) {
> + struct intel_bw_info *bi = &i915->max_bw[i];
> int clpchgroup;
> int j;
>
> @@ -206,18 +206,18 @@ static int icl_get_bw_info(struct drm_i915_private *dev_priv)
> return 0;
> }
>
> -static unsigned int icl_max_bw(struct drm_i915_private *dev_priv,
> +static unsigned int icl_max_bw(struct drm_i915_private *i915,
> int num_planes, int qgv_point)
> {
> int i;
>
> /* Did we initialize the bw limits successfully? */
> - if (dev_priv->max_bw[0].num_planes == 0)
> + if (i915->max_bw[0].num_planes == 0)
> return UINT_MAX;
>
> - for (i = 0; i < ARRAY_SIZE(dev_priv->max_bw); i++) {
> + for (i = 0; i < ARRAY_SIZE(i915->max_bw); i++) {
> const struct intel_bw_info *bi =
> - &dev_priv->max_bw[i];
> + &i915->max_bw[i];
>
> if (num_planes >= bi->num_planes)
> return bi->deratedbw[qgv_point];
> @@ -226,24 +226,24 @@ static unsigned int icl_max_bw(struct drm_i915_private *dev_priv,
> return 0;
> }
>
> -void intel_bw_init_hw(struct drm_i915_private *dev_priv)
> +void intel_bw_init_hw(struct drm_i915_private *i915)
> {
> - if (IS_GEN(dev_priv, 11))
> - icl_get_bw_info(dev_priv);
> + if (IS_GEN(i915, 11))
> + icl_get_bw_info(i915);
> }
>
> -static unsigned int intel_max_data_rate(struct drm_i915_private *dev_priv,
> +static unsigned int intel_max_data_rate(struct drm_i915_private *i915,
> int num_planes)
> {
> - if (IS_GEN(dev_priv, 11))
> + if (IS_GEN(i915, 11))
> /*
> * FIXME with SAGV disabled maybe we can assume
> * point 1 will always be used? Seems to match
> * the behaviour observed in the wild.
> */
> - return min3(icl_max_bw(dev_priv, num_planes, 0),
> - icl_max_bw(dev_priv, num_planes, 1),
> - icl_max_bw(dev_priv, num_planes, 2));
> + return min3(icl_max_bw(i915, num_planes, 0),
> + icl_max_bw(i915, num_planes, 1),
> + icl_max_bw(i915, num_planes, 2));
> else
> return UINT_MAX;
> }
> @@ -293,25 +293,25 @@ void intel_bw_crtc_update(struct intel_bw_state *bw_state,
> bw_state->num_active_planes[crtc->pipe]);
> }
>
> -static unsigned int intel_bw_num_active_planes(struct drm_i915_private *dev_priv,
> +static unsigned int intel_bw_num_active_planes(struct drm_i915_private *i915,
> const struct intel_bw_state *bw_state)
> {
> unsigned int num_active_planes = 0;
> enum pipe pipe;
>
> - for_each_pipe(dev_priv, pipe)
> + for_each_pipe(i915, pipe)
> num_active_planes += bw_state->num_active_planes[pipe];
>
> return num_active_planes;
> }
>
> -static unsigned int intel_bw_data_rate(struct drm_i915_private *dev_priv,
> +static unsigned int intel_bw_data_rate(struct drm_i915_private *i915,
> const struct intel_bw_state *bw_state)
> {
> unsigned int data_rate = 0;
> enum pipe pipe;
>
> - for_each_pipe(dev_priv, pipe)
> + for_each_pipe(i915, pipe)
> data_rate += bw_state->data_rate[pipe];
>
> return data_rate;
> @@ -319,7 +319,7 @@ static unsigned int intel_bw_data_rate(struct drm_i915_private *dev_priv,
>
> int intel_bw_atomic_check(struct intel_atomic_state *state)
> {
> - struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> + struct drm_i915_private *i915 = to_i915(state->base.dev);
> struct intel_crtc_state *new_crtc_state, *old_crtc_state;
> struct intel_bw_state *bw_state = NULL;
> unsigned int data_rate, max_data_rate;
> @@ -328,7 +328,7 @@ int intel_bw_atomic_check(struct intel_atomic_state *state)
> int i;
>
> /* FIXME earlier gens need some checks too */
> - if (INTEL_GEN(dev_priv) < 11)
> + if (INTEL_GEN(i915) < 11)
> return 0;
>
> for_each_oldnew_intel_crtc_in_state(state, crtc, old_crtc_state,
> @@ -366,10 +366,10 @@ int intel_bw_atomic_check(struct intel_atomic_state *state)
> if (!bw_state)
> return 0;
>
> - data_rate = intel_bw_data_rate(dev_priv, bw_state);
> - num_active_planes = intel_bw_num_active_planes(dev_priv, bw_state);
> + data_rate = intel_bw_data_rate(i915, bw_state);
> + num_active_planes = intel_bw_num_active_planes(i915, bw_state);
>
> - max_data_rate = intel_max_data_rate(dev_priv, num_active_planes);
> + max_data_rate = intel_max_data_rate(i915, num_active_planes);
>
> data_rate = DIV_ROUND_UP(data_rate, 1000);
>
> @@ -406,7 +406,7 @@ static const struct drm_private_state_funcs intel_bw_funcs = {
> .atomic_destroy_state = intel_bw_destroy_state,
> };
>
> -int intel_bw_init(struct drm_i915_private *dev_priv)
> +int intel_bw_init(struct drm_i915_private *i915)
> {
> struct intel_bw_state *state;
>
> @@ -414,7 +414,7 @@ int intel_bw_init(struct drm_i915_private *dev_priv)
> if (!state)
> return -ENOMEM;
>
> - drm_atomic_private_obj_init(&dev_priv->drm, &dev_priv->bw_obj,
> + drm_atomic_private_obj_init(&i915->drm, &i915->bw_obj,
> &state->base, &intel_bw_funcs);
>
> return 0;
> diff --git a/drivers/gpu/drm/i915/intel_bw.h b/drivers/gpu/drm/i915/intel_bw.h
> index e9d9c6d63bc3..396e5029f9a4 100644
> --- a/drivers/gpu/drm/i915/intel_bw.h
> +++ b/drivers/gpu/drm/i915/intel_bw.h
> @@ -27,19 +27,19 @@ struct intel_bw_state {
> static inline struct intel_bw_state *
> intel_atomic_get_bw_state(struct intel_atomic_state *state)
> {
> - struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> + struct drm_i915_private *i915 = to_i915(state->base.dev);
> struct drm_private_state *bw_state;
>
> bw_state = drm_atomic_get_private_obj_state(&state->base,
> - &dev_priv->bw_obj);
> + &i915->bw_obj);
> if (IS_ERR(bw_state))
> return ERR_CAST(bw_state);
>
> return to_intel_bw_state(bw_state);
> }
>
> -void intel_bw_init_hw(struct drm_i915_private *dev_priv);
> -int intel_bw_init(struct drm_i915_private *dev_priv);
> +void intel_bw_init_hw(struct drm_i915_private *i915);
> +int intel_bw_init(struct drm_i915_private *i915);
> int intel_bw_atomic_check(struct intel_atomic_state *state);
> void intel_bw_crtc_update(struct intel_bw_state *bw_state,
> const struct intel_crtc_state *crtc_state);
> diff --git a/drivers/gpu/drm/i915/intel_cdclk.c b/drivers/gpu/drm/i915/intel_cdclk.c
> index 8993ab283562..02c617180c4b 100644
> --- a/drivers/gpu/drm/i915/intel_cdclk.c
> +++ b/drivers/gpu/drm/i915/intel_cdclk.c
> @@ -53,46 +53,46 @@
> * dividers can be programmed correctly.
> */
>
> -static void fixed_133mhz_get_cdclk(struct drm_i915_private *dev_priv,
> +static void fixed_133mhz_get_cdclk(struct drm_i915_private *i915,
> struct intel_cdclk_state *cdclk_state)
> {
> cdclk_state->cdclk = 133333;
> }
>
> -static void fixed_200mhz_get_cdclk(struct drm_i915_private *dev_priv,
> +static void fixed_200mhz_get_cdclk(struct drm_i915_private *i915,
> struct intel_cdclk_state *cdclk_state)
> {
> cdclk_state->cdclk = 200000;
> }
>
> -static void fixed_266mhz_get_cdclk(struct drm_i915_private *dev_priv,
> +static void fixed_266mhz_get_cdclk(struct drm_i915_private *i915,
> struct intel_cdclk_state *cdclk_state)
> {
> cdclk_state->cdclk = 266667;
> }
>
> -static void fixed_333mhz_get_cdclk(struct drm_i915_private *dev_priv,
> +static void fixed_333mhz_get_cdclk(struct drm_i915_private *i915,
> struct intel_cdclk_state *cdclk_state)
> {
> cdclk_state->cdclk = 333333;
> }
>
> -static void fixed_400mhz_get_cdclk(struct drm_i915_private *dev_priv,
> +static void fixed_400mhz_get_cdclk(struct drm_i915_private *i915,
> struct intel_cdclk_state *cdclk_state)
> {
> cdclk_state->cdclk = 400000;
> }
>
> -static void fixed_450mhz_get_cdclk(struct drm_i915_private *dev_priv,
> +static void fixed_450mhz_get_cdclk(struct drm_i915_private *i915,
> struct intel_cdclk_state *cdclk_state)
> {
> cdclk_state->cdclk = 450000;
> }
>
> -static void i85x_get_cdclk(struct drm_i915_private *dev_priv,
> +static void i85x_get_cdclk(struct drm_i915_private *i915,
> struct intel_cdclk_state *cdclk_state)
> {
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct pci_dev *pdev = i915->drm.pdev;
> u16 hpllcc = 0;
>
> /*
> @@ -131,10 +131,10 @@ static void i85x_get_cdclk(struct drm_i915_private *dev_priv,
> }
> }
>
> -static void i915gm_get_cdclk(struct drm_i915_private *dev_priv,
> +static void i915gm_get_cdclk(struct drm_i915_private *i915,
> struct intel_cdclk_state *cdclk_state)
> {
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct pci_dev *pdev = i915->drm.pdev;
> u16 gcfgc = 0;
>
> pci_read_config_word(pdev, GCFGC, &gcfgc);
> @@ -155,10 +155,10 @@ static void i915gm_get_cdclk(struct drm_i915_private *dev_priv,
> }
> }
>
> -static void i945gm_get_cdclk(struct drm_i915_private *dev_priv,
> +static void i945gm_get_cdclk(struct drm_i915_private *i915,
> struct intel_cdclk_state *cdclk_state)
> {
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct pci_dev *pdev = i915->drm.pdev;
> u16 gcfgc = 0;
>
> pci_read_config_word(pdev, GCFGC, &gcfgc);
> @@ -179,7 +179,7 @@ static void i945gm_get_cdclk(struct drm_i915_private *dev_priv,
> }
> }
>
> -static unsigned int intel_hpll_vco(struct drm_i915_private *dev_priv)
> +static unsigned int intel_hpll_vco(struct drm_i915_private *i915)
> {
> static const unsigned int blb_vco[8] = {
> [0] = 3200000,
> @@ -223,20 +223,20 @@ static unsigned int intel_hpll_vco(struct drm_i915_private *dev_priv)
> u8 tmp = 0;
>
> /* FIXME other chipsets? */
> - if (IS_GM45(dev_priv))
> + if (IS_GM45(i915))
> vco_table = ctg_vco;
> - else if (IS_G45(dev_priv))
> + else if (IS_G45(i915))
> vco_table = elk_vco;
> - else if (IS_I965GM(dev_priv))
> + else if (IS_I965GM(i915))
> vco_table = cl_vco;
> - else if (IS_PINEVIEW(dev_priv))
> + else if (IS_PINEVIEW(i915))
> vco_table = pnv_vco;
> - else if (IS_G33(dev_priv))
> + else if (IS_G33(i915))
> vco_table = blb_vco;
> else
> return 0;
>
> - tmp = I915_READ(IS_PINEVIEW(dev_priv) || IS_MOBILE(dev_priv) ?
> + tmp = I915_READ(IS_PINEVIEW(i915) || IS_MOBILE(i915) ?
> HPLLVCO_MOBILE : HPLLVCO);
>
> vco = vco_table[tmp & 0x7];
> @@ -248,10 +248,10 @@ static unsigned int intel_hpll_vco(struct drm_i915_private *dev_priv)
> return vco;
> }
>
> -static void g33_get_cdclk(struct drm_i915_private *dev_priv,
> +static void g33_get_cdclk(struct drm_i915_private *i915,
> struct intel_cdclk_state *cdclk_state)
> {
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct pci_dev *pdev = i915->drm.pdev;
> static const u8 div_3200[] = { 12, 10, 8, 7, 5, 16 };
> static const u8 div_4000[] = { 14, 12, 10, 8, 6, 20 };
> static const u8 div_4800[] = { 20, 14, 12, 10, 8, 24 };
> @@ -260,7 +260,7 @@ static void g33_get_cdclk(struct drm_i915_private *dev_priv,
> unsigned int cdclk_sel;
> u16 tmp = 0;
>
> - cdclk_state->vco = intel_hpll_vco(dev_priv);
> + cdclk_state->vco = intel_hpll_vco(i915);
>
> pci_read_config_word(pdev, GCFGC, &tmp);
>
> @@ -296,10 +296,10 @@ static void g33_get_cdclk(struct drm_i915_private *dev_priv,
> cdclk_state->cdclk = 190476;
> }
>
> -static void pnv_get_cdclk(struct drm_i915_private *dev_priv,
> +static void pnv_get_cdclk(struct drm_i915_private *i915,
> struct intel_cdclk_state *cdclk_state)
> {
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct pci_dev *pdev = i915->drm.pdev;
> u16 gcfgc = 0;
>
> pci_read_config_word(pdev, GCFGC, &gcfgc);
> @@ -329,10 +329,10 @@ static void pnv_get_cdclk(struct drm_i915_private *dev_priv,
> }
> }
>
> -static void i965gm_get_cdclk(struct drm_i915_private *dev_priv,
> +static void i965gm_get_cdclk(struct drm_i915_private *i915,
> struct intel_cdclk_state *cdclk_state)
> {
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct pci_dev *pdev = i915->drm.pdev;
> static const u8 div_3200[] = { 16, 10, 8 };
> static const u8 div_4000[] = { 20, 12, 10 };
> static const u8 div_5333[] = { 24, 16, 14 };
> @@ -340,7 +340,7 @@ static void i965gm_get_cdclk(struct drm_i915_private *dev_priv,
> unsigned int cdclk_sel;
> u16 tmp = 0;
>
> - cdclk_state->vco = intel_hpll_vco(dev_priv);
> + cdclk_state->vco = intel_hpll_vco(i915);
>
> pci_read_config_word(pdev, GCFGC, &tmp);
>
> @@ -373,14 +373,14 @@ static void i965gm_get_cdclk(struct drm_i915_private *dev_priv,
> cdclk_state->cdclk = 200000;
> }
>
> -static void gm45_get_cdclk(struct drm_i915_private *dev_priv,
> +static void gm45_get_cdclk(struct drm_i915_private *i915,
> struct intel_cdclk_state *cdclk_state)
> {
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct pci_dev *pdev = i915->drm.pdev;
> unsigned int cdclk_sel;
> u16 tmp = 0;
>
> - cdclk_state->vco = intel_hpll_vco(dev_priv);
> + cdclk_state->vco = intel_hpll_vco(i915);
>
> pci_read_config_word(pdev, GCFGC, &tmp);
>
> @@ -403,7 +403,7 @@ static void gm45_get_cdclk(struct drm_i915_private *dev_priv,
> }
> }
>
> -static void hsw_get_cdclk(struct drm_i915_private *dev_priv,
> +static void hsw_get_cdclk(struct drm_i915_private *i915,
> struct intel_cdclk_state *cdclk_state)
> {
> u32 lcpll = I915_READ(LCPLL_CTL);
> @@ -415,15 +415,15 @@ static void hsw_get_cdclk(struct drm_i915_private *dev_priv,
> cdclk_state->cdclk = 450000;
> else if (freq == LCPLL_CLK_FREQ_450)
> cdclk_state->cdclk = 450000;
> - else if (IS_HSW_ULT(dev_priv))
> + else if (IS_HSW_ULT(i915))
> cdclk_state->cdclk = 337500;
> else
> cdclk_state->cdclk = 540000;
> }
>
> -static int vlv_calc_cdclk(struct drm_i915_private *dev_priv, int min_cdclk)
> +static int vlv_calc_cdclk(struct drm_i915_private *i915, int min_cdclk)
> {
> - int freq_320 = (dev_priv->hpll_freq << 1) % 320000 != 0 ?
> + int freq_320 = (i915->hpll_freq << 1) % 320000 != 0 ?
> 333333 : 320000;
>
> /*
> @@ -431,7 +431,7 @@ static int vlv_calc_cdclk(struct drm_i915_private *dev_priv, int min_cdclk)
> * Not sure what's wrong. For now use 200MHz only when all pipes
> * are off.
> */
> - if (IS_VALLEYVIEW(dev_priv) && min_cdclk > freq_320)
> + if (IS_VALLEYVIEW(i915) && min_cdclk > freq_320)
> return 400000;
> else if (min_cdclk > 266667)
> return freq_320;
> @@ -441,9 +441,9 @@ static int vlv_calc_cdclk(struct drm_i915_private *dev_priv, int min_cdclk)
> return 200000;
> }
>
> -static u8 vlv_calc_voltage_level(struct drm_i915_private *dev_priv, int cdclk)
> +static u8 vlv_calc_voltage_level(struct drm_i915_private *i915, int cdclk)
> {
> - if (IS_VALLEYVIEW(dev_priv)) {
> + if (IS_VALLEYVIEW(i915)) {
> if (cdclk >= 320000) /* jump to highest voltage for 400MHz too */
> return 2;
> else if (cdclk >= 266667)
> @@ -456,29 +456,29 @@ static u8 vlv_calc_voltage_level(struct drm_i915_private *dev_priv, int cdclk)
> * hardware has shown that we just need to write the desired
> * CCK divider into the Punit register.
> */
> - return DIV_ROUND_CLOSEST(dev_priv->hpll_freq << 1, cdclk) - 1;
> + return DIV_ROUND_CLOSEST(i915->hpll_freq << 1, cdclk) - 1;
> }
> }
>
> -static void vlv_get_cdclk(struct drm_i915_private *dev_priv,
> +static void vlv_get_cdclk(struct drm_i915_private *i915,
> struct intel_cdclk_state *cdclk_state)
> {
> u32 val;
>
> - vlv_iosf_sb_get(dev_priv,
> + vlv_iosf_sb_get(i915,
> BIT(VLV_IOSF_SB_CCK) | BIT(VLV_IOSF_SB_PUNIT));
>
> - cdclk_state->vco = vlv_get_hpll_vco(dev_priv);
> - cdclk_state->cdclk = vlv_get_cck_clock(dev_priv, "cdclk",
> + cdclk_state->vco = vlv_get_hpll_vco(i915);
> + cdclk_state->cdclk = vlv_get_cck_clock(i915, "cdclk",
> CCK_DISPLAY_CLOCK_CONTROL,
> cdclk_state->vco);
>
> - val = vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM);
> + val = vlv_punit_read(i915, PUNIT_REG_DSPSSPM);
>
> - vlv_iosf_sb_put(dev_priv,
> + vlv_iosf_sb_put(i915,
> BIT(VLV_IOSF_SB_CCK) | BIT(VLV_IOSF_SB_PUNIT));
>
> - if (IS_VALLEYVIEW(dev_priv))
> + if (IS_VALLEYVIEW(i915))
> cdclk_state->voltage_level = (val & DSPFREQGUAR_MASK) >>
> DSPFREQGUAR_SHIFT;
> else
> @@ -486,18 +486,18 @@ static void vlv_get_cdclk(struct drm_i915_private *dev_priv,
> DSPFREQGUAR_SHIFT_CHV;
> }
>
> -static void vlv_program_pfi_credits(struct drm_i915_private *dev_priv)
> +static void vlv_program_pfi_credits(struct drm_i915_private *i915)
> {
> unsigned int credits, default_credits;
>
> - if (IS_CHERRYVIEW(dev_priv))
> + if (IS_CHERRYVIEW(i915))
> default_credits = PFI_CREDIT(12);
> else
> default_credits = PFI_CREDIT(8);
>
> - if (dev_priv->cdclk.hw.cdclk >= dev_priv->czclk_freq) {
> + if (i915->cdclk.hw.cdclk >= i915->czclk_freq) {
> /* CHV suggested value is 31 or 63 */
> - if (IS_CHERRYVIEW(dev_priv))
> + if (IS_CHERRYVIEW(i915))
> credits = PFI_CREDIT_63;
> else
> credits = PFI_CREDIT(15);
> @@ -522,7 +522,7 @@ static void vlv_program_pfi_credits(struct drm_i915_private *dev_priv)
> WARN_ON(I915_READ(GCI_CONTROL) & PFI_CREDIT_RESEND);
> }
>
> -static void vlv_set_cdclk(struct drm_i915_private *dev_priv,
> +static void vlv_set_cdclk(struct drm_i915_private *i915,
> const struct intel_cdclk_state *cdclk_state,
> enum pipe pipe)
> {
> @@ -548,18 +548,18 @@ static void vlv_set_cdclk(struct drm_i915_private *dev_priv,
> * a system suspend. So grab the PIPE-A domain, which covers
> * the HW blocks needed for the following programming.
> */
> - wakeref = intel_display_power_get(dev_priv, POWER_DOMAIN_PIPE_A);
> + wakeref = intel_display_power_get(i915, POWER_DOMAIN_PIPE_A);
>
> - vlv_iosf_sb_get(dev_priv,
> + vlv_iosf_sb_get(i915,
> BIT(VLV_IOSF_SB_CCK) |
> BIT(VLV_IOSF_SB_BUNIT) |
> BIT(VLV_IOSF_SB_PUNIT));
>
> - val = vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM);
> + val = vlv_punit_read(i915, PUNIT_REG_DSPSSPM);
> val &= ~DSPFREQGUAR_MASK;
> val |= (cmd << DSPFREQGUAR_SHIFT);
> - vlv_punit_write(dev_priv, PUNIT_REG_DSPSSPM, val);
> - if (wait_for((vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM) &
> + vlv_punit_write(i915, PUNIT_REG_DSPSSPM, val);
> + if (wait_for((vlv_punit_read(i915, PUNIT_REG_DSPSSPM) &
> DSPFREQSTAT_MASK) == (cmd << DSPFREQSTAT_SHIFT),
> 50)) {
> DRM_ERROR("timed out waiting for CDclk change\n");
> @@ -568,23 +568,23 @@ static void vlv_set_cdclk(struct drm_i915_private *dev_priv,
> if (cdclk == 400000) {
> u32 divider;
>
> - divider = DIV_ROUND_CLOSEST(dev_priv->hpll_freq << 1,
> + divider = DIV_ROUND_CLOSEST(i915->hpll_freq << 1,
> cdclk) - 1;
>
> /* adjust cdclk divider */
> - val = vlv_cck_read(dev_priv, CCK_DISPLAY_CLOCK_CONTROL);
> + val = vlv_cck_read(i915, CCK_DISPLAY_CLOCK_CONTROL);
> val &= ~CCK_FREQUENCY_VALUES;
> val |= divider;
> - vlv_cck_write(dev_priv, CCK_DISPLAY_CLOCK_CONTROL, val);
> + vlv_cck_write(i915, CCK_DISPLAY_CLOCK_CONTROL, val);
>
> - if (wait_for((vlv_cck_read(dev_priv, CCK_DISPLAY_CLOCK_CONTROL) &
> + if (wait_for((vlv_cck_read(i915, CCK_DISPLAY_CLOCK_CONTROL) &
> CCK_FREQUENCY_STATUS) == (divider << CCK_FREQUENCY_STATUS_SHIFT),
> 50))
> DRM_ERROR("timed out waiting for CDclk change\n");
> }
>
> /* adjust self-refresh exit latency value */
> - val = vlv_bunit_read(dev_priv, BUNIT_REG_BISOC);
> + val = vlv_bunit_read(i915, BUNIT_REG_BISOC);
> val &= ~0x7f;
>
> /*
> @@ -595,21 +595,21 @@ static void vlv_set_cdclk(struct drm_i915_private *dev_priv,
> val |= 4500 / 250; /* 4.5 usec */
> else
> val |= 3000 / 250; /* 3.0 usec */
> - vlv_bunit_write(dev_priv, BUNIT_REG_BISOC, val);
> + vlv_bunit_write(i915, BUNIT_REG_BISOC, val);
>
> - vlv_iosf_sb_put(dev_priv,
> + vlv_iosf_sb_put(i915,
> BIT(VLV_IOSF_SB_CCK) |
> BIT(VLV_IOSF_SB_BUNIT) |
> BIT(VLV_IOSF_SB_PUNIT));
>
> - intel_update_cdclk(dev_priv);
> + intel_update_cdclk(i915);
>
> - vlv_program_pfi_credits(dev_priv);
> + vlv_program_pfi_credits(i915);
>
> - intel_display_power_put(dev_priv, POWER_DOMAIN_PIPE_A, wakeref);
> + intel_display_power_put(i915, POWER_DOMAIN_PIPE_A, wakeref);
> }
>
> -static void chv_set_cdclk(struct drm_i915_private *dev_priv,
> +static void chv_set_cdclk(struct drm_i915_private *i915,
> const struct intel_cdclk_state *cdclk_state,
> enum pipe pipe)
> {
> @@ -634,26 +634,26 @@ static void chv_set_cdclk(struct drm_i915_private *dev_priv,
> * a system suspend. So grab the PIPE-A domain, which covers
> * the HW blocks needed for the following programming.
> */
> - wakeref = intel_display_power_get(dev_priv, POWER_DOMAIN_PIPE_A);
> + wakeref = intel_display_power_get(i915, POWER_DOMAIN_PIPE_A);
>
> - vlv_punit_get(dev_priv);
> - val = vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM);
> + vlv_punit_get(i915);
> + val = vlv_punit_read(i915, PUNIT_REG_DSPSSPM);
> val &= ~DSPFREQGUAR_MASK_CHV;
> val |= (cmd << DSPFREQGUAR_SHIFT_CHV);
> - vlv_punit_write(dev_priv, PUNIT_REG_DSPSSPM, val);
> - if (wait_for((vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM) &
> + vlv_punit_write(i915, PUNIT_REG_DSPSSPM, val);
> + if (wait_for((vlv_punit_read(i915, PUNIT_REG_DSPSSPM) &
> DSPFREQSTAT_MASK_CHV) == (cmd << DSPFREQSTAT_SHIFT_CHV),
> 50)) {
> DRM_ERROR("timed out waiting for CDclk change\n");
> }
>
> - vlv_punit_put(dev_priv);
> + vlv_punit_put(i915);
>
> - intel_update_cdclk(dev_priv);
> + intel_update_cdclk(i915);
>
> - vlv_program_pfi_credits(dev_priv);
> + vlv_program_pfi_credits(i915);
>
> - intel_display_power_put(dev_priv, POWER_DOMAIN_PIPE_A, wakeref);
> + intel_display_power_put(i915, POWER_DOMAIN_PIPE_A, wakeref);
> }
>
> static int bdw_calc_cdclk(int min_cdclk)
> @@ -683,7 +683,7 @@ static u8 bdw_calc_voltage_level(int cdclk)
> }
> }
>
> -static void bdw_get_cdclk(struct drm_i915_private *dev_priv,
> +static void bdw_get_cdclk(struct drm_i915_private *i915,
> struct intel_cdclk_state *cdclk_state)
> {
> u32 lcpll = I915_READ(LCPLL_CTL);
> @@ -710,7 +710,7 @@ static void bdw_get_cdclk(struct drm_i915_private *dev_priv,
> bdw_calc_voltage_level(cdclk_state->cdclk);
> }
>
> -static void bdw_set_cdclk(struct drm_i915_private *dev_priv,
> +static void bdw_set_cdclk(struct drm_i915_private *i915,
> const struct intel_cdclk_state *cdclk_state,
> enum pipe pipe)
> {
> @@ -726,7 +726,7 @@ static void bdw_set_cdclk(struct drm_i915_private *dev_priv,
> "trying to change cdclk frequency with cdclk not enabled\n"))
> return;
>
> - ret = sandybridge_pcode_write(dev_priv,
> + ret = sandybridge_pcode_write(i915,
> BDW_PCODE_DISPLAY_FREQ_CHANGE_REQ, 0x0);
> if (ret) {
> DRM_ERROR("failed to inform pcode about cdclk change\n");
> @@ -776,12 +776,12 @@ static void bdw_set_cdclk(struct drm_i915_private *dev_priv,
> LCPLL_CD_SOURCE_FCLK_DONE) == 0, 1))
> DRM_ERROR("Switching back to LCPLL failed\n");
>
> - sandybridge_pcode_write(dev_priv, HSW_PCODE_DE_WRITE_FREQ_REQ,
> + sandybridge_pcode_write(i915, HSW_PCODE_DE_WRITE_FREQ_REQ,
> cdclk_state->voltage_level);
>
> I915_WRITE(CDCLK_FREQ, DIV_ROUND_CLOSEST(cdclk, 1000) - 1);
>
> - intel_update_cdclk(dev_priv);
> + intel_update_cdclk(i915);
> }
>
> static int skl_calc_cdclk(int min_cdclk, int vco)
> @@ -819,7 +819,7 @@ static u8 skl_calc_voltage_level(int cdclk)
> return 0;
> }
>
> -static void skl_dpll0_update(struct drm_i915_private *dev_priv,
> +static void skl_dpll0_update(struct drm_i915_private *i915,
> struct intel_cdclk_state *cdclk_state)
> {
> u32 val;
> @@ -859,12 +859,12 @@ static void skl_dpll0_update(struct drm_i915_private *dev_priv,
> }
> }
>
> -static void skl_get_cdclk(struct drm_i915_private *dev_priv,
> +static void skl_get_cdclk(struct drm_i915_private *i915,
> struct intel_cdclk_state *cdclk_state)
> {
> u32 cdctl;
>
> - skl_dpll0_update(dev_priv, cdclk_state);
> + skl_dpll0_update(i915, cdclk_state);
>
> cdclk_state->cdclk = cdclk_state->bypass = cdclk_state->ref;
>
> @@ -926,18 +926,18 @@ static int skl_cdclk_decimal(int cdclk)
> return DIV_ROUND_CLOSEST(cdclk - 1000, 500);
> }
>
> -static void skl_set_preferred_cdclk_vco(struct drm_i915_private *dev_priv,
> +static void skl_set_preferred_cdclk_vco(struct drm_i915_private *i915,
> int vco)
> {
> - bool changed = dev_priv->skl_preferred_vco_freq != vco;
> + bool changed = i915->skl_preferred_vco_freq != vco;
>
> - dev_priv->skl_preferred_vco_freq = vco;
> + i915->skl_preferred_vco_freq = vco;
>
> if (changed)
> - intel_update_max_cdclk(dev_priv);
> + intel_update_max_cdclk(i915);
> }
>
> -static void skl_dpll0_enable(struct drm_i915_private *dev_priv, int vco)
> +static void skl_dpll0_enable(struct drm_i915_private *i915, int vco)
> {
> u32 val;
>
> @@ -969,29 +969,29 @@ static void skl_dpll0_enable(struct drm_i915_private *dev_priv, int vco)
>
> I915_WRITE(LCPLL1_CTL, I915_READ(LCPLL1_CTL) | LCPLL_PLL_ENABLE);
>
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> LCPLL1_CTL, LCPLL_PLL_LOCK, LCPLL_PLL_LOCK,
> 5))
> DRM_ERROR("DPLL0 not locked\n");
>
> - dev_priv->cdclk.hw.vco = vco;
> + i915->cdclk.hw.vco = vco;
>
> /* We'll want to keep using the current vco from now on. */
> - skl_set_preferred_cdclk_vco(dev_priv, vco);
> + skl_set_preferred_cdclk_vco(i915, vco);
> }
>
> -static void skl_dpll0_disable(struct drm_i915_private *dev_priv)
> +static void skl_dpll0_disable(struct drm_i915_private *i915)
> {
> I915_WRITE(LCPLL1_CTL, I915_READ(LCPLL1_CTL) & ~LCPLL_PLL_ENABLE);
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> LCPLL1_CTL, LCPLL_PLL_LOCK, 0,
> 1))
> DRM_ERROR("Couldn't disable DPLL0\n");
>
> - dev_priv->cdclk.hw.vco = 0;
> + i915->cdclk.hw.vco = 0;
> }
>
> -static void skl_set_cdclk(struct drm_i915_private *dev_priv,
> +static void skl_set_cdclk(struct drm_i915_private *i915,
> const struct intel_cdclk_state *cdclk_state,
> enum pipe pipe)
> {
> @@ -1008,9 +1008,9 @@ static void skl_set_cdclk(struct drm_i915_private *dev_priv,
> * use the corresponding VCO freq as that always leads to using the
> * minimum 308MHz CDCLK.
> */
> - WARN_ON_ONCE(IS_SKYLAKE(dev_priv) && vco == 8640000);
> + WARN_ON_ONCE(IS_SKYLAKE(i915) && vco == 8640000);
>
> - ret = skl_pcode_request(dev_priv, SKL_PCODE_CDCLK_CONTROL,
> + ret = skl_pcode_request(i915, SKL_PCODE_CDCLK_CONTROL,
> SKL_CDCLK_PREPARE_FOR_CHANGE,
> SKL_CDCLK_READY_FOR_CHANGE,
> SKL_CDCLK_READY_FOR_CHANGE, 3);
> @@ -1023,7 +1023,7 @@ static void skl_set_cdclk(struct drm_i915_private *dev_priv,
> /* Choose frequency for this cdclk */
> switch (cdclk) {
> default:
> - WARN_ON(cdclk != dev_priv->cdclk.hw.bypass);
> + WARN_ON(cdclk != i915->cdclk.hw.bypass);
> WARN_ON(vco != 0);
> /* fall through */
> case 308571:
> @@ -1043,13 +1043,13 @@ static void skl_set_cdclk(struct drm_i915_private *dev_priv,
> break;
> }
>
> - if (dev_priv->cdclk.hw.vco != 0 &&
> - dev_priv->cdclk.hw.vco != vco)
> - skl_dpll0_disable(dev_priv);
> + if (i915->cdclk.hw.vco != 0 &&
> + i915->cdclk.hw.vco != vco)
> + skl_dpll0_disable(i915);
>
> cdclk_ctl = I915_READ(CDCLK_CTL);
>
> - if (dev_priv->cdclk.hw.vco != vco) {
> + if (i915->cdclk.hw.vco != vco) {
> /* Wa Display #1183: skl,kbl,cfl */
> cdclk_ctl &= ~(CDCLK_FREQ_SEL_MASK | CDCLK_FREQ_DECIMAL_MASK);
> cdclk_ctl |= freq_select | skl_cdclk_decimal(cdclk);
> @@ -1061,8 +1061,8 @@ static void skl_set_cdclk(struct drm_i915_private *dev_priv,
> I915_WRITE(CDCLK_CTL, cdclk_ctl);
> POSTING_READ(CDCLK_CTL);
>
> - if (dev_priv->cdclk.hw.vco != vco)
> - skl_dpll0_enable(dev_priv, vco);
> + if (i915->cdclk.hw.vco != vco)
> + skl_dpll0_enable(i915, vco);
>
> /* Wa Display #1183: skl,kbl,cfl */
> cdclk_ctl &= ~(CDCLK_FREQ_SEL_MASK | CDCLK_FREQ_DECIMAL_MASK);
> @@ -1077,13 +1077,13 @@ static void skl_set_cdclk(struct drm_i915_private *dev_priv,
> POSTING_READ(CDCLK_CTL);
>
> /* inform PCU of the change */
> - sandybridge_pcode_write(dev_priv, SKL_PCODE_CDCLK_CONTROL,
> + sandybridge_pcode_write(i915, SKL_PCODE_CDCLK_CONTROL,
> cdclk_state->voltage_level);
>
> - intel_update_cdclk(dev_priv);
> + intel_update_cdclk(i915);
> }
>
> -static void skl_sanitize_cdclk(struct drm_i915_private *dev_priv)
> +static void skl_sanitize_cdclk(struct drm_i915_private *i915)
> {
> u32 cdctl, expected;
>
> @@ -1095,12 +1095,12 @@ static void skl_sanitize_cdclk(struct drm_i915_private *dev_priv)
> if ((I915_READ(SWF_ILK(0x18)) & 0x00FFFFFF) == 0)
> goto sanitize;
>
> - intel_update_cdclk(dev_priv);
> - intel_dump_cdclk_state(&dev_priv->cdclk.hw, "Current CDCLK");
> + intel_update_cdclk(i915);
> + intel_dump_cdclk_state(&i915->cdclk.hw, "Current CDCLK");
>
> /* Is PLL enabled and locked ? */
> - if (dev_priv->cdclk.hw.vco == 0 ||
> - dev_priv->cdclk.hw.cdclk == dev_priv->cdclk.hw.bypass)
> + if (i915->cdclk.hw.vco == 0 ||
> + i915->cdclk.hw.cdclk == i915->cdclk.hw.bypass)
> goto sanitize;
>
> /* DPLL okay; verify the cdclock
> @@ -1111,7 +1111,7 @@ static void skl_sanitize_cdclk(struct drm_i915_private *dev_priv)
> */
> cdctl = I915_READ(CDCLK_CTL);
> expected = (cdctl & CDCLK_FREQ_SEL_MASK) |
> - skl_cdclk_decimal(dev_priv->cdclk.hw.cdclk);
> + skl_cdclk_decimal(i915->cdclk.hw.cdclk);
> if (cdctl == expected)
> /* All well; nothing to sanitize */
> return;
> @@ -1120,49 +1120,49 @@ static void skl_sanitize_cdclk(struct drm_i915_private *dev_priv)
> DRM_DEBUG_KMS("Sanitizing cdclk programmed by pre-os\n");
>
> /* force cdclk programming */
> - dev_priv->cdclk.hw.cdclk = 0;
> + i915->cdclk.hw.cdclk = 0;
> /* force full PLL disable + enable */
> - dev_priv->cdclk.hw.vco = -1;
> + i915->cdclk.hw.vco = -1;
> }
>
> -static void skl_init_cdclk(struct drm_i915_private *dev_priv)
> +static void skl_init_cdclk(struct drm_i915_private *i915)
> {
> struct intel_cdclk_state cdclk_state;
>
> - skl_sanitize_cdclk(dev_priv);
> + skl_sanitize_cdclk(i915);
>
> - if (dev_priv->cdclk.hw.cdclk != 0 &&
> - dev_priv->cdclk.hw.vco != 0) {
> + if (i915->cdclk.hw.cdclk != 0 &&
> + i915->cdclk.hw.vco != 0) {
> /*
> * Use the current vco as our initial
> * guess as to what the preferred vco is.
> */
> - if (dev_priv->skl_preferred_vco_freq == 0)
> - skl_set_preferred_cdclk_vco(dev_priv,
> - dev_priv->cdclk.hw.vco);
> + if (i915->skl_preferred_vco_freq == 0)
> + skl_set_preferred_cdclk_vco(i915,
> + i915->cdclk.hw.vco);
> return;
> }
>
> - cdclk_state = dev_priv->cdclk.hw;
> + cdclk_state = i915->cdclk.hw;
>
> - cdclk_state.vco = dev_priv->skl_preferred_vco_freq;
> + cdclk_state.vco = i915->skl_preferred_vco_freq;
> if (cdclk_state.vco == 0)
> cdclk_state.vco = 8100000;
> cdclk_state.cdclk = skl_calc_cdclk(0, cdclk_state.vco);
> cdclk_state.voltage_level = skl_calc_voltage_level(cdclk_state.cdclk);
>
> - skl_set_cdclk(dev_priv, &cdclk_state, INVALID_PIPE);
> + skl_set_cdclk(i915, &cdclk_state, INVALID_PIPE);
> }
>
> -static void skl_uninit_cdclk(struct drm_i915_private *dev_priv)
> +static void skl_uninit_cdclk(struct drm_i915_private *i915)
> {
> - struct intel_cdclk_state cdclk_state = dev_priv->cdclk.hw;
> + struct intel_cdclk_state cdclk_state = i915->cdclk.hw;
>
> cdclk_state.cdclk = cdclk_state.bypass;
> cdclk_state.vco = 0;
> cdclk_state.voltage_level = skl_calc_voltage_level(cdclk_state.cdclk);
>
> - skl_set_cdclk(dev_priv, &cdclk_state, INVALID_PIPE);
> + skl_set_cdclk(i915, &cdclk_state, INVALID_PIPE);
> }
>
> static int bxt_calc_cdclk(int min_cdclk)
> @@ -1194,11 +1194,11 @@ static u8 bxt_calc_voltage_level(int cdclk)
> return DIV_ROUND_UP(cdclk, 25000);
> }
>
> -static int bxt_de_pll_vco(struct drm_i915_private *dev_priv, int cdclk)
> +static int bxt_de_pll_vco(struct drm_i915_private *i915, int cdclk)
> {
> int ratio;
>
> - if (cdclk == dev_priv->cdclk.hw.bypass)
> + if (cdclk == i915->cdclk.hw.bypass)
> return 0;
>
> switch (cdclk) {
> @@ -1216,14 +1216,14 @@ static int bxt_de_pll_vco(struct drm_i915_private *dev_priv, int cdclk)
> break;
> }
>
> - return dev_priv->cdclk.hw.ref * ratio;
> + return i915->cdclk.hw.ref * ratio;
> }
>
> -static int glk_de_pll_vco(struct drm_i915_private *dev_priv, int cdclk)
> +static int glk_de_pll_vco(struct drm_i915_private *i915, int cdclk)
> {
> int ratio;
>
> - if (cdclk == dev_priv->cdclk.hw.bypass)
> + if (cdclk == i915->cdclk.hw.bypass)
> return 0;
>
> switch (cdclk) {
> @@ -1237,10 +1237,10 @@ static int glk_de_pll_vco(struct drm_i915_private *dev_priv, int cdclk)
> break;
> }
>
> - return dev_priv->cdclk.hw.ref * ratio;
> + return i915->cdclk.hw.ref * ratio;
> }
>
> -static void bxt_de_pll_update(struct drm_i915_private *dev_priv,
> +static void bxt_de_pll_update(struct drm_i915_private *i915,
> struct intel_cdclk_state *cdclk_state)
> {
> u32 val;
> @@ -1259,13 +1259,13 @@ static void bxt_de_pll_update(struct drm_i915_private *dev_priv,
> cdclk_state->vco = (val & BXT_DE_PLL_RATIO_MASK) * cdclk_state->ref;
> }
>
> -static void bxt_get_cdclk(struct drm_i915_private *dev_priv,
> +static void bxt_get_cdclk(struct drm_i915_private *i915,
> struct intel_cdclk_state *cdclk_state)
> {
> u32 divider;
> int div;
>
> - bxt_de_pll_update(dev_priv, cdclk_state);
> + bxt_de_pll_update(i915, cdclk_state);
>
> cdclk_state->cdclk = cdclk_state->bypass = cdclk_state->ref;
>
> @@ -1279,7 +1279,7 @@ static void bxt_get_cdclk(struct drm_i915_private *dev_priv,
> div = 2;
> break;
> case BXT_CDCLK_CD2X_DIV_SEL_1_5:
> - WARN(IS_GEMINILAKE(dev_priv), "Unsupported divider\n");
> + WARN(IS_GEMINILAKE(i915), "Unsupported divider\n");
> div = 3;
> break;
> case BXT_CDCLK_CD2X_DIV_SEL_2:
> @@ -1304,22 +1304,22 @@ static void bxt_get_cdclk(struct drm_i915_private *dev_priv,
> bxt_calc_voltage_level(cdclk_state->cdclk);
> }
>
> -static void bxt_de_pll_disable(struct drm_i915_private *dev_priv)
> +static void bxt_de_pll_disable(struct drm_i915_private *i915)
> {
> I915_WRITE(BXT_DE_PLL_ENABLE, 0);
>
> /* Timeout 200us */
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> BXT_DE_PLL_ENABLE, BXT_DE_PLL_LOCK, 0,
> 1))
> DRM_ERROR("timeout waiting for DE PLL unlock\n");
>
> - dev_priv->cdclk.hw.vco = 0;
> + i915->cdclk.hw.vco = 0;
> }
>
> -static void bxt_de_pll_enable(struct drm_i915_private *dev_priv, int vco)
> +static void bxt_de_pll_enable(struct drm_i915_private *i915, int vco)
> {
> - int ratio = DIV_ROUND_CLOSEST(vco, dev_priv->cdclk.hw.ref);
> + int ratio = DIV_ROUND_CLOSEST(vco, i915->cdclk.hw.ref);
> u32 val;
>
> val = I915_READ(BXT_DE_PLL_CTL);
> @@ -1330,17 +1330,17 @@ static void bxt_de_pll_enable(struct drm_i915_private *dev_priv, int vco)
> I915_WRITE(BXT_DE_PLL_ENABLE, BXT_DE_PLL_PLL_ENABLE);
>
> /* Timeout 200us */
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> BXT_DE_PLL_ENABLE,
> BXT_DE_PLL_LOCK,
> BXT_DE_PLL_LOCK,
> 1))
> DRM_ERROR("timeout waiting for DE PLL lock\n");
>
> - dev_priv->cdclk.hw.vco = vco;
> + i915->cdclk.hw.vco = vco;
> }
>
> -static void bxt_set_cdclk(struct drm_i915_private *dev_priv,
> +static void bxt_set_cdclk(struct drm_i915_private *i915,
> const struct intel_cdclk_state *cdclk_state,
> enum pipe pipe)
> {
> @@ -1352,14 +1352,14 @@ static void bxt_set_cdclk(struct drm_i915_private *dev_priv,
> /* cdclk = vco / 2 / div{1,1.5,2,4} */
> switch (DIV_ROUND_CLOSEST(vco, cdclk)) {
> default:
> - WARN_ON(cdclk != dev_priv->cdclk.hw.bypass);
> + WARN_ON(cdclk != i915->cdclk.hw.bypass);
> WARN_ON(vco != 0);
> /* fall through */
> case 2:
> divider = BXT_CDCLK_CD2X_DIV_SEL_1;
> break;
> case 3:
> - WARN(IS_GEMINILAKE(dev_priv), "Unsupported divider\n");
> + WARN(IS_GEMINILAKE(i915), "Unsupported divider\n");
> divider = BXT_CDCLK_CD2X_DIV_SEL_1_5;
> break;
> case 4:
> @@ -1375,7 +1375,7 @@ static void bxt_set_cdclk(struct drm_i915_private *dev_priv,
> * requires us to wait up to 150usec, but that leads to timeouts;
> * the 2ms used here is based on experiment.
> */
> - ret = sandybridge_pcode_write_timeout(dev_priv,
> + ret = sandybridge_pcode_write_timeout(i915,
> HSW_PCODE_DE_WRITE_FREQ_REQ,
> 0x80000000, 150, 2);
> if (ret) {
> @@ -1384,12 +1384,12 @@ static void bxt_set_cdclk(struct drm_i915_private *dev_priv,
> return;
> }
>
> - if (dev_priv->cdclk.hw.vco != 0 &&
> - dev_priv->cdclk.hw.vco != vco)
> - bxt_de_pll_disable(dev_priv);
> + if (i915->cdclk.hw.vco != 0 &&
> + i915->cdclk.hw.vco != vco)
> + bxt_de_pll_disable(i915);
>
> - if (dev_priv->cdclk.hw.vco != vco)
> - bxt_de_pll_enable(dev_priv, vco);
> + if (i915->cdclk.hw.vco != vco)
> + bxt_de_pll_enable(i915, vco);
>
> val = divider | skl_cdclk_decimal(cdclk);
> if (pipe == INVALID_PIPE)
> @@ -1405,7 +1405,7 @@ static void bxt_set_cdclk(struct drm_i915_private *dev_priv,
> I915_WRITE(CDCLK_CTL, val);
>
> if (pipe != INVALID_PIPE)
> - intel_wait_for_vblank(dev_priv, pipe);
> + intel_wait_for_vblank(i915, pipe);
>
> /*
> * The timeout isn't specified, the 2ms used here is based on
> @@ -1413,7 +1413,7 @@ static void bxt_set_cdclk(struct drm_i915_private *dev_priv,
> * FIXME: Waiting for the request completion could be delayed until
> * the next PCODE request based on BSpec.
> */
> - ret = sandybridge_pcode_write_timeout(dev_priv,
> + ret = sandybridge_pcode_write_timeout(i915,
> HSW_PCODE_DE_WRITE_FREQ_REQ,
> cdclk_state->voltage_level, 150, 2);
> if (ret) {
> @@ -1422,18 +1422,18 @@ static void bxt_set_cdclk(struct drm_i915_private *dev_priv,
> return;
> }
>
> - intel_update_cdclk(dev_priv);
> + intel_update_cdclk(i915);
> }
>
> -static void bxt_sanitize_cdclk(struct drm_i915_private *dev_priv)
> +static void bxt_sanitize_cdclk(struct drm_i915_private *i915)
> {
> u32 cdctl, expected;
>
> - intel_update_cdclk(dev_priv);
> - intel_dump_cdclk_state(&dev_priv->cdclk.hw, "Current CDCLK");
> + intel_update_cdclk(i915);
> + intel_dump_cdclk_state(&i915->cdclk.hw, "Current CDCLK");
>
> - if (dev_priv->cdclk.hw.vco == 0 ||
> - dev_priv->cdclk.hw.cdclk == dev_priv->cdclk.hw.bypass)
> + if (i915->cdclk.hw.vco == 0 ||
> + i915->cdclk.hw.cdclk == i915->cdclk.hw.bypass)
> goto sanitize;
>
> /* DPLL okay; verify the cdclock
> @@ -1451,12 +1451,12 @@ static void bxt_sanitize_cdclk(struct drm_i915_private *dev_priv)
> cdctl &= ~BXT_CDCLK_CD2X_PIPE_NONE;
>
> expected = (cdctl & BXT_CDCLK_CD2X_DIV_SEL_MASK) |
> - skl_cdclk_decimal(dev_priv->cdclk.hw.cdclk);
> + skl_cdclk_decimal(i915->cdclk.hw.cdclk);
> /*
> * Disable SSA Precharge when CD clock frequency < 500 MHz,
> * enable otherwise.
> */
> - if (dev_priv->cdclk.hw.cdclk >= 500000)
> + if (i915->cdclk.hw.cdclk >= 500000)
> expected |= BXT_CDCLK_SSA_PRECHARGE_ENABLE;
>
> if (cdctl == expected)
> @@ -1467,50 +1467,50 @@ static void bxt_sanitize_cdclk(struct drm_i915_private *dev_priv)
> DRM_DEBUG_KMS("Sanitizing cdclk programmed by pre-os\n");
>
> /* force cdclk programming */
> - dev_priv->cdclk.hw.cdclk = 0;
> + i915->cdclk.hw.cdclk = 0;
>
> /* force full PLL disable + enable */
> - dev_priv->cdclk.hw.vco = -1;
> + i915->cdclk.hw.vco = -1;
> }
>
> -static void bxt_init_cdclk(struct drm_i915_private *dev_priv)
> +static void bxt_init_cdclk(struct drm_i915_private *i915)
> {
> struct intel_cdclk_state cdclk_state;
>
> - bxt_sanitize_cdclk(dev_priv);
> + bxt_sanitize_cdclk(i915);
>
> - if (dev_priv->cdclk.hw.cdclk != 0 &&
> - dev_priv->cdclk.hw.vco != 0)
> + if (i915->cdclk.hw.cdclk != 0 &&
> + i915->cdclk.hw.vco != 0)
> return;
>
> - cdclk_state = dev_priv->cdclk.hw;
> + cdclk_state = i915->cdclk.hw;
>
> /*
> * FIXME:
> * - The initial CDCLK needs to be read from VBT.
> * Need to make this change after VBT has changes for BXT.
> */
> - if (IS_GEMINILAKE(dev_priv)) {
> + if (IS_GEMINILAKE(i915)) {
> cdclk_state.cdclk = glk_calc_cdclk(0);
> - cdclk_state.vco = glk_de_pll_vco(dev_priv, cdclk_state.cdclk);
> + cdclk_state.vco = glk_de_pll_vco(i915, cdclk_state.cdclk);
> } else {
> cdclk_state.cdclk = bxt_calc_cdclk(0);
> - cdclk_state.vco = bxt_de_pll_vco(dev_priv, cdclk_state.cdclk);
> + cdclk_state.vco = bxt_de_pll_vco(i915, cdclk_state.cdclk);
> }
> cdclk_state.voltage_level = bxt_calc_voltage_level(cdclk_state.cdclk);
>
> - bxt_set_cdclk(dev_priv, &cdclk_state, INVALID_PIPE);
> + bxt_set_cdclk(i915, &cdclk_state, INVALID_PIPE);
> }
>
> -static void bxt_uninit_cdclk(struct drm_i915_private *dev_priv)
> +static void bxt_uninit_cdclk(struct drm_i915_private *i915)
> {
> - struct intel_cdclk_state cdclk_state = dev_priv->cdclk.hw;
> + struct intel_cdclk_state cdclk_state = i915->cdclk.hw;
>
> cdclk_state.cdclk = cdclk_state.bypass;
> cdclk_state.vco = 0;
> cdclk_state.voltage_level = bxt_calc_voltage_level(cdclk_state.cdclk);
>
> - bxt_set_cdclk(dev_priv, &cdclk_state, INVALID_PIPE);
> + bxt_set_cdclk(i915, &cdclk_state, INVALID_PIPE);
> }
>
> static int cnl_calc_cdclk(int min_cdclk)
> @@ -1533,7 +1533,7 @@ static u8 cnl_calc_voltage_level(int cdclk)
> return 0;
> }
>
> -static void cnl_cdclk_pll_update(struct drm_i915_private *dev_priv,
> +static void cnl_cdclk_pll_update(struct drm_i915_private *i915,
> struct intel_cdclk_state *cdclk_state)
> {
> u32 val;
> @@ -1555,13 +1555,13 @@ static void cnl_cdclk_pll_update(struct drm_i915_private *dev_priv,
> cdclk_state->vco = (val & CNL_CDCLK_PLL_RATIO_MASK) * cdclk_state->ref;
> }
>
> -static void cnl_get_cdclk(struct drm_i915_private *dev_priv,
> +static void cnl_get_cdclk(struct drm_i915_private *i915,
> struct intel_cdclk_state *cdclk_state)
> {
> u32 divider;
> int div;
>
> - cnl_cdclk_pll_update(dev_priv, cdclk_state);
> + cnl_cdclk_pll_update(i915, cdclk_state);
>
> cdclk_state->cdclk = cdclk_state->bypass = cdclk_state->ref;
>
> @@ -1593,7 +1593,7 @@ static void cnl_get_cdclk(struct drm_i915_private *dev_priv,
> cnl_calc_voltage_level(cdclk_state->cdclk);
> }
>
> -static void cnl_cdclk_pll_disable(struct drm_i915_private *dev_priv)
> +static void cnl_cdclk_pll_disable(struct drm_i915_private *i915)
> {
> u32 val;
>
> @@ -1605,12 +1605,12 @@ static void cnl_cdclk_pll_disable(struct drm_i915_private *dev_priv)
> if (wait_for((I915_READ(BXT_DE_PLL_ENABLE) & BXT_DE_PLL_LOCK) == 0, 1))
> DRM_ERROR("timeout waiting for CDCLK PLL unlock\n");
>
> - dev_priv->cdclk.hw.vco = 0;
> + i915->cdclk.hw.vco = 0;
> }
>
> -static void cnl_cdclk_pll_enable(struct drm_i915_private *dev_priv, int vco)
> +static void cnl_cdclk_pll_enable(struct drm_i915_private *i915, int vco)
> {
> - int ratio = DIV_ROUND_CLOSEST(vco, dev_priv->cdclk.hw.ref);
> + int ratio = DIV_ROUND_CLOSEST(vco, i915->cdclk.hw.ref);
> u32 val;
>
> val = CNL_CDCLK_PLL_RATIO(ratio);
> @@ -1623,10 +1623,10 @@ static void cnl_cdclk_pll_enable(struct drm_i915_private *dev_priv, int vco)
> if (wait_for((I915_READ(BXT_DE_PLL_ENABLE) & BXT_DE_PLL_LOCK) != 0, 1))
> DRM_ERROR("timeout waiting for CDCLK PLL lock\n");
>
> - dev_priv->cdclk.hw.vco = vco;
> + i915->cdclk.hw.vco = vco;
> }
>
> -static void cnl_set_cdclk(struct drm_i915_private *dev_priv,
> +static void cnl_set_cdclk(struct drm_i915_private *i915,
> const struct intel_cdclk_state *cdclk_state,
> enum pipe pipe)
> {
> @@ -1635,7 +1635,7 @@ static void cnl_set_cdclk(struct drm_i915_private *dev_priv,
> u32 val, divider;
> int ret;
>
> - ret = skl_pcode_request(dev_priv, SKL_PCODE_CDCLK_CONTROL,
> + ret = skl_pcode_request(i915, SKL_PCODE_CDCLK_CONTROL,
> SKL_CDCLK_PREPARE_FOR_CHANGE,
> SKL_CDCLK_READY_FOR_CHANGE,
> SKL_CDCLK_READY_FOR_CHANGE, 3);
> @@ -1648,7 +1648,7 @@ static void cnl_set_cdclk(struct drm_i915_private *dev_priv,
> /* cdclk = vco / 2 / div{1,2} */
> switch (DIV_ROUND_CLOSEST(vco, cdclk)) {
> default:
> - WARN_ON(cdclk != dev_priv->cdclk.hw.bypass);
> + WARN_ON(cdclk != i915->cdclk.hw.bypass);
> WARN_ON(vco != 0);
> /* fall through */
> case 2:
> @@ -1659,12 +1659,12 @@ static void cnl_set_cdclk(struct drm_i915_private *dev_priv,
> break;
> }
>
> - if (dev_priv->cdclk.hw.vco != 0 &&
> - dev_priv->cdclk.hw.vco != vco)
> - cnl_cdclk_pll_disable(dev_priv);
> + if (i915->cdclk.hw.vco != 0 &&
> + i915->cdclk.hw.vco != vco)
> + cnl_cdclk_pll_disable(i915);
>
> - if (dev_priv->cdclk.hw.vco != vco)
> - cnl_cdclk_pll_enable(dev_priv, vco);
> + if (i915->cdclk.hw.vco != vco)
> + cnl_cdclk_pll_enable(i915, vco);
>
> val = divider | skl_cdclk_decimal(cdclk);
> if (pipe == INVALID_PIPE)
> @@ -1674,26 +1674,26 @@ static void cnl_set_cdclk(struct drm_i915_private *dev_priv,
> I915_WRITE(CDCLK_CTL, val);
>
> if (pipe != INVALID_PIPE)
> - intel_wait_for_vblank(dev_priv, pipe);
> + intel_wait_for_vblank(i915, pipe);
>
> /* inform PCU of the change */
> - sandybridge_pcode_write(dev_priv, SKL_PCODE_CDCLK_CONTROL,
> + sandybridge_pcode_write(i915, SKL_PCODE_CDCLK_CONTROL,
> cdclk_state->voltage_level);
>
> - intel_update_cdclk(dev_priv);
> + intel_update_cdclk(i915);
>
> /*
> * Can't read out the voltage level :(
> * Let's just assume everything is as expected.
> */
> - dev_priv->cdclk.hw.voltage_level = cdclk_state->voltage_level;
> + i915->cdclk.hw.voltage_level = cdclk_state->voltage_level;
> }
>
> -static int cnl_cdclk_pll_vco(struct drm_i915_private *dev_priv, int cdclk)
> +static int cnl_cdclk_pll_vco(struct drm_i915_private *i915, int cdclk)
> {
> int ratio;
>
> - if (cdclk == dev_priv->cdclk.hw.bypass)
> + if (cdclk == i915->cdclk.hw.bypass)
> return 0;
>
> switch (cdclk) {
> @@ -1702,25 +1702,25 @@ static int cnl_cdclk_pll_vco(struct drm_i915_private *dev_priv, int cdclk)
> /* fall through */
> case 168000:
> case 336000:
> - ratio = dev_priv->cdclk.hw.ref == 19200 ? 35 : 28;
> + ratio = i915->cdclk.hw.ref == 19200 ? 35 : 28;
> break;
> case 528000:
> - ratio = dev_priv->cdclk.hw.ref == 19200 ? 55 : 44;
> + ratio = i915->cdclk.hw.ref == 19200 ? 55 : 44;
> break;
> }
>
> - return dev_priv->cdclk.hw.ref * ratio;
> + return i915->cdclk.hw.ref * ratio;
> }
>
> -static void cnl_sanitize_cdclk(struct drm_i915_private *dev_priv)
> +static void cnl_sanitize_cdclk(struct drm_i915_private *i915)
> {
> u32 cdctl, expected;
>
> - intel_update_cdclk(dev_priv);
> - intel_dump_cdclk_state(&dev_priv->cdclk.hw, "Current CDCLK");
> + intel_update_cdclk(i915);
> + intel_dump_cdclk_state(&i915->cdclk.hw, "Current CDCLK");
>
> - if (dev_priv->cdclk.hw.vco == 0 ||
> - dev_priv->cdclk.hw.cdclk == dev_priv->cdclk.hw.bypass)
> + if (i915->cdclk.hw.vco == 0 ||
> + i915->cdclk.hw.cdclk == i915->cdclk.hw.bypass)
> goto sanitize;
>
> /* DPLL okay; verify the cdclock
> @@ -1738,7 +1738,7 @@ static void cnl_sanitize_cdclk(struct drm_i915_private *dev_priv)
> cdctl &= ~BXT_CDCLK_CD2X_PIPE_NONE;
>
> expected = (cdctl & BXT_CDCLK_CD2X_DIV_SEL_MASK) |
> - skl_cdclk_decimal(dev_priv->cdclk.hw.cdclk);
> + skl_cdclk_decimal(i915->cdclk.hw.cdclk);
>
> if (cdctl == expected)
> /* All well; nothing to sanitize */
> @@ -1748,10 +1748,10 @@ static void cnl_sanitize_cdclk(struct drm_i915_private *dev_priv)
> DRM_DEBUG_KMS("Sanitizing cdclk programmed by pre-os\n");
>
> /* force cdclk programming */
> - dev_priv->cdclk.hw.cdclk = 0;
> + i915->cdclk.hw.cdclk = 0;
>
> /* force full PLL disable + enable */
> - dev_priv->cdclk.hw.vco = -1;
> + i915->cdclk.hw.vco = -1;
> }
>
> static int icl_calc_cdclk(int min_cdclk, unsigned int ref)
> @@ -1781,11 +1781,11 @@ static int icl_calc_cdclk(int min_cdclk, unsigned int ref)
> return ranges[0];
> }
>
> -static int icl_calc_cdclk_pll_vco(struct drm_i915_private *dev_priv, int cdclk)
> +static int icl_calc_cdclk_pll_vco(struct drm_i915_private *i915, int cdclk)
> {
> int ratio;
>
> - if (cdclk == dev_priv->cdclk.hw.bypass)
> + if (cdclk == i915->cdclk.hw.bypass)
> return 0;
>
> switch (cdclk) {
> @@ -1795,21 +1795,21 @@ static int icl_calc_cdclk_pll_vco(struct drm_i915_private *dev_priv, int cdclk)
> case 307200:
> case 556800:
> case 652800:
> - WARN_ON(dev_priv->cdclk.hw.ref != 19200 &&
> - dev_priv->cdclk.hw.ref != 38400);
> + WARN_ON(i915->cdclk.hw.ref != 19200 &&
> + i915->cdclk.hw.ref != 38400);
> break;
> case 312000:
> case 552000:
> case 648000:
> - WARN_ON(dev_priv->cdclk.hw.ref != 24000);
> + WARN_ON(i915->cdclk.hw.ref != 24000);
> }
>
> - ratio = cdclk / (dev_priv->cdclk.hw.ref / 2);
> + ratio = cdclk / (i915->cdclk.hw.ref / 2);
>
> - return dev_priv->cdclk.hw.ref * ratio;
> + return i915->cdclk.hw.ref * ratio;
> }
>
> -static void icl_set_cdclk(struct drm_i915_private *dev_priv,
> +static void icl_set_cdclk(struct drm_i915_private *i915,
> const struct intel_cdclk_state *cdclk_state,
> enum pipe pipe)
> {
> @@ -1817,7 +1817,7 @@ static void icl_set_cdclk(struct drm_i915_private *dev_priv,
> unsigned int vco = cdclk_state->vco;
> int ret;
>
> - ret = skl_pcode_request(dev_priv, SKL_PCODE_CDCLK_CONTROL,
> + ret = skl_pcode_request(i915, SKL_PCODE_CDCLK_CONTROL,
> SKL_CDCLK_PREPARE_FOR_CHANGE,
> SKL_CDCLK_READY_FOR_CHANGE,
> SKL_CDCLK_READY_FOR_CHANGE, 3);
> @@ -1827,12 +1827,12 @@ static void icl_set_cdclk(struct drm_i915_private *dev_priv,
> return;
> }
>
> - if (dev_priv->cdclk.hw.vco != 0 &&
> - dev_priv->cdclk.hw.vco != vco)
> - cnl_cdclk_pll_disable(dev_priv);
> + if (i915->cdclk.hw.vco != 0 &&
> + i915->cdclk.hw.vco != vco)
> + cnl_cdclk_pll_disable(i915);
>
> - if (dev_priv->cdclk.hw.vco != vco)
> - cnl_cdclk_pll_enable(dev_priv, vco);
> + if (i915->cdclk.hw.vco != vco)
> + cnl_cdclk_pll_enable(i915, vco);
>
> /*
> * On ICL CD2X_DIV can only be 1, so we'll never end up changing the
> @@ -1842,16 +1842,16 @@ static void icl_set_cdclk(struct drm_i915_private *dev_priv,
> I915_WRITE(CDCLK_CTL, ICL_CDCLK_CD2X_PIPE_NONE |
> skl_cdclk_decimal(cdclk));
>
> - sandybridge_pcode_write(dev_priv, SKL_PCODE_CDCLK_CONTROL,
> + sandybridge_pcode_write(i915, SKL_PCODE_CDCLK_CONTROL,
> cdclk_state->voltage_level);
>
> - intel_update_cdclk(dev_priv);
> + intel_update_cdclk(i915);
>
> /*
> * Can't read out the voltage level :(
> * Let's just assume everything is as expected.
> */
> - dev_priv->cdclk.hw.voltage_level = cdclk_state->voltage_level;
> + i915->cdclk.hw.voltage_level = cdclk_state->voltage_level;
> }
>
> static u8 icl_calc_voltage_level(int cdclk)
> @@ -1864,7 +1864,7 @@ static u8 icl_calc_voltage_level(int cdclk)
> return 0;
> }
>
> -static void icl_get_cdclk(struct drm_i915_private *dev_priv,
> +static void icl_get_cdclk(struct drm_i915_private *i915,
> struct intel_cdclk_state *cdclk_state)
> {
> u32 val;
> @@ -1915,17 +1915,17 @@ static void icl_get_cdclk(struct drm_i915_private *dev_priv,
> icl_calc_voltage_level(cdclk_state->cdclk);
> }
>
> -static void icl_init_cdclk(struct drm_i915_private *dev_priv)
> +static void icl_init_cdclk(struct drm_i915_private *i915)
> {
> struct intel_cdclk_state sanitized_state;
> u32 val;
>
> - /* This sets dev_priv->cdclk.hw. */
> - intel_update_cdclk(dev_priv);
> - intel_dump_cdclk_state(&dev_priv->cdclk.hw, "Current CDCLK");
> + /* This sets i915->cdclk.hw. */
> + intel_update_cdclk(i915);
> + intel_dump_cdclk_state(&i915->cdclk.hw, "Current CDCLK");
>
> /* This means CDCLK disabled. */
> - if (dev_priv->cdclk.hw.cdclk == dev_priv->cdclk.hw.bypass)
> + if (i915->cdclk.hw.cdclk == i915->cdclk.hw.bypass)
> goto sanitize;
>
> val = I915_READ(CDCLK_CTL);
> @@ -1934,7 +1934,7 @@ static void icl_init_cdclk(struct drm_i915_private *dev_priv)
> goto sanitize;
>
> if ((val & CDCLK_FREQ_DECIMAL_MASK) !=
> - skl_cdclk_decimal(dev_priv->cdclk.hw.cdclk))
> + skl_cdclk_decimal(i915->cdclk.hw.cdclk))
> goto sanitize;
>
> return;
> @@ -1942,62 +1942,62 @@ static void icl_init_cdclk(struct drm_i915_private *dev_priv)
> sanitize:
> DRM_DEBUG_KMS("Sanitizing cdclk programmed by pre-os\n");
>
> - sanitized_state.ref = dev_priv->cdclk.hw.ref;
> + sanitized_state.ref = i915->cdclk.hw.ref;
> sanitized_state.cdclk = icl_calc_cdclk(0, sanitized_state.ref);
> - sanitized_state.vco = icl_calc_cdclk_pll_vco(dev_priv,
> + sanitized_state.vco = icl_calc_cdclk_pll_vco(i915,
> sanitized_state.cdclk);
> sanitized_state.voltage_level =
> icl_calc_voltage_level(sanitized_state.cdclk);
>
> - icl_set_cdclk(dev_priv, &sanitized_state, INVALID_PIPE);
> + icl_set_cdclk(i915, &sanitized_state, INVALID_PIPE);
> }
>
> -static void icl_uninit_cdclk(struct drm_i915_private *dev_priv)
> +static void icl_uninit_cdclk(struct drm_i915_private *i915)
> {
> - struct intel_cdclk_state cdclk_state = dev_priv->cdclk.hw;
> + struct intel_cdclk_state cdclk_state = i915->cdclk.hw;
>
> cdclk_state.cdclk = cdclk_state.bypass;
> cdclk_state.vco = 0;
> cdclk_state.voltage_level = icl_calc_voltage_level(cdclk_state.cdclk);
>
> - icl_set_cdclk(dev_priv, &cdclk_state, INVALID_PIPE);
> + icl_set_cdclk(i915, &cdclk_state, INVALID_PIPE);
> }
>
> -static void cnl_init_cdclk(struct drm_i915_private *dev_priv)
> +static void cnl_init_cdclk(struct drm_i915_private *i915)
> {
> struct intel_cdclk_state cdclk_state;
>
> - cnl_sanitize_cdclk(dev_priv);
> + cnl_sanitize_cdclk(i915);
>
> - if (dev_priv->cdclk.hw.cdclk != 0 &&
> - dev_priv->cdclk.hw.vco != 0)
> + if (i915->cdclk.hw.cdclk != 0 &&
> + i915->cdclk.hw.vco != 0)
> return;
>
> - cdclk_state = dev_priv->cdclk.hw;
> + cdclk_state = i915->cdclk.hw;
>
> cdclk_state.cdclk = cnl_calc_cdclk(0);
> - cdclk_state.vco = cnl_cdclk_pll_vco(dev_priv, cdclk_state.cdclk);
> + cdclk_state.vco = cnl_cdclk_pll_vco(i915, cdclk_state.cdclk);
> cdclk_state.voltage_level = cnl_calc_voltage_level(cdclk_state.cdclk);
>
> - cnl_set_cdclk(dev_priv, &cdclk_state, INVALID_PIPE);
> + cnl_set_cdclk(i915, &cdclk_state, INVALID_PIPE);
> }
>
> -static void cnl_uninit_cdclk(struct drm_i915_private *dev_priv)
> +static void cnl_uninit_cdclk(struct drm_i915_private *i915)
> {
> - struct intel_cdclk_state cdclk_state = dev_priv->cdclk.hw;
> + struct intel_cdclk_state cdclk_state = i915->cdclk.hw;
>
> cdclk_state.cdclk = cdclk_state.bypass;
> cdclk_state.vco = 0;
> cdclk_state.voltage_level = cnl_calc_voltage_level(cdclk_state.cdclk);
>
> - cnl_set_cdclk(dev_priv, &cdclk_state, INVALID_PIPE);
> + cnl_set_cdclk(i915, &cdclk_state, INVALID_PIPE);
> }
>
> /**
> * intel_cdclk_init - Initialize CDCLK
> * @i915: i915 device
> *
> - * Initialize CDCLK. This consists mainly of initializing dev_priv->cdclk.hw and
> + * Initialize CDCLK. This consists mainly of initializing i915->cdclk.hw and
> * sanitizing the state of the hardware if needed. This is generally done only
> * during the display core initialization sequence, after which the DMC will
> * take care of turning CDCLK off/on as needed.
> @@ -2051,19 +2051,19 @@ bool intel_cdclk_needs_modeset(const struct intel_cdclk_state *a,
>
> /**
> * intel_cdclk_needs_cd2x_update - Determine if two CDCLK states require a cd2x divider update
> - * @dev_priv: Not a CDCLK state, it's the drm_i915_private!
> + * @i915: Not a CDCLK state, it's the drm_i915_private!
> * @a: first CDCLK state
> * @b: second CDCLK state
> *
> * Returns:
> * True if the CDCLK states require just a cd2x divider update, false if not.
> */
> -bool intel_cdclk_needs_cd2x_update(struct drm_i915_private *dev_priv,
> +bool intel_cdclk_needs_cd2x_update(struct drm_i915_private *i915,
> const struct intel_cdclk_state *a,
> const struct intel_cdclk_state *b)
> {
> /* Older hw doesn't have the capability */
> - if (INTEL_GEN(dev_priv) < 10 && !IS_GEN9_LP(dev_priv))
> + if (INTEL_GEN(i915) < 10 && !IS_GEN9_LP(i915))
> return false;
>
> return a->cdclk != b->cdclk &&
> @@ -2100,10 +2100,10 @@ bool intel_cdclk_changed(const struct intel_cdclk_state *a,
> */
> void intel_cdclk_swap_state(struct intel_atomic_state *state)
> {
> - struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> + struct drm_i915_private *i915 = to_i915(state->base.dev);
>
> - swap(state->cdclk.logical, dev_priv->cdclk.logical);
> - swap(state->cdclk.actual, dev_priv->cdclk.actual);
> + swap(state->cdclk.logical, i915->cdclk.logical);
> + swap(state->cdclk.actual, i915->cdclk.actual);
> }
>
> void intel_dump_cdclk_state(const struct intel_cdclk_state *cdclk_state,
> @@ -2117,37 +2117,37 @@ void intel_dump_cdclk_state(const struct intel_cdclk_state *cdclk_state,
>
> /**
> * intel_set_cdclk - Push the CDCLK state to the hardware
> - * @dev_priv: i915 device
> + * @i915: i915 device
> * @cdclk_state: new CDCLK state
> * @pipe: pipe with which to synchronize the update
> *
> * Program the hardware based on the passed in CDCLK state,
> * if necessary.
> */
> -static void intel_set_cdclk(struct drm_i915_private *dev_priv,
> +static void intel_set_cdclk(struct drm_i915_private *i915,
> const struct intel_cdclk_state *cdclk_state,
> enum pipe pipe)
> {
> - if (!intel_cdclk_changed(&dev_priv->cdclk.hw, cdclk_state))
> + if (!intel_cdclk_changed(&i915->cdclk.hw, cdclk_state))
> return;
>
> - if (WARN_ON_ONCE(!dev_priv->display.set_cdclk))
> + if (WARN_ON_ONCE(!i915->display.set_cdclk))
> return;
>
> intel_dump_cdclk_state(cdclk_state, "Changing CDCLK to");
>
> - dev_priv->display.set_cdclk(dev_priv, cdclk_state, pipe);
> + i915->display.set_cdclk(i915, cdclk_state, pipe);
>
> - if (WARN(intel_cdclk_changed(&dev_priv->cdclk.hw, cdclk_state),
> + if (WARN(intel_cdclk_changed(&i915->cdclk.hw, cdclk_state),
> "cdclk state doesn't match!\n")) {
> - intel_dump_cdclk_state(&dev_priv->cdclk.hw, "[hw state]");
> + intel_dump_cdclk_state(&i915->cdclk.hw, "[hw state]");
> intel_dump_cdclk_state(cdclk_state, "[sw state]");
> }
> }
>
> /**
> * intel_set_cdclk_pre_plane_update - Push the CDCLK state to the hardware
> - * @dev_priv: i915 device
> + * @i915: i915 device
> * @old_state: old CDCLK state
> * @new_state: new CDCLK state
> * @pipe: pipe with which to synchronize the update
> @@ -2156,18 +2156,18 @@ static void intel_set_cdclk(struct drm_i915_private *dev_priv,
> * in CDCLK state, if necessary.
> */
> void
> -intel_set_cdclk_pre_plane_update(struct drm_i915_private *dev_priv,
> +intel_set_cdclk_pre_plane_update(struct drm_i915_private *i915,
> const struct intel_cdclk_state *old_state,
> const struct intel_cdclk_state *new_state,
> enum pipe pipe)
> {
> if (pipe == INVALID_PIPE || old_state->cdclk <= new_state->cdclk)
> - intel_set_cdclk(dev_priv, new_state, pipe);
> + intel_set_cdclk(i915, new_state, pipe);
> }
>
> /**
> * intel_set_cdclk_post_plane_update - Push the CDCLK state to the hardware
> - * @dev_priv: i915 device
> + * @i915: i915 device
> * @old_state: old CDCLK state
> * @new_state: new CDCLK state
> * @pipe: pipe with which to synchronize the update
> @@ -2176,24 +2176,24 @@ intel_set_cdclk_pre_plane_update(struct drm_i915_private *dev_priv,
> * in CDCLK state, if necessary.
> */
> void
> -intel_set_cdclk_post_plane_update(struct drm_i915_private *dev_priv,
> +intel_set_cdclk_post_plane_update(struct drm_i915_private *i915,
> const struct intel_cdclk_state *old_state,
> const struct intel_cdclk_state *new_state,
> enum pipe pipe)
> {
> if (pipe != INVALID_PIPE && old_state->cdclk > new_state->cdclk)
> - intel_set_cdclk(dev_priv, new_state, pipe);
> + intel_set_cdclk(i915, new_state, pipe);
> }
>
> -static int intel_pixel_rate_to_cdclk(struct drm_i915_private *dev_priv,
> +static int intel_pixel_rate_to_cdclk(struct drm_i915_private *i915,
> int pixel_rate)
> {
> - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> return DIV_ROUND_UP(pixel_rate, 2);
> - else if (IS_GEN(dev_priv, 9) ||
> - IS_BROADWELL(dev_priv) || IS_HASWELL(dev_priv))
> + else if (IS_GEN(i915, 9) ||
> + IS_BROADWELL(i915) || IS_HASWELL(i915))
> return pixel_rate;
> - else if (IS_CHERRYVIEW(dev_priv))
> + else if (IS_CHERRYVIEW(i915))
> return DIV_ROUND_UP(pixel_rate * 100, 95);
> else
> return DIV_ROUND_UP(pixel_rate * 100, 90);
> @@ -2201,17 +2201,17 @@ static int intel_pixel_rate_to_cdclk(struct drm_i915_private *dev_priv,
>
> int intel_crtc_compute_min_cdclk(const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> to_i915(crtc_state->base.crtc->dev);
> int min_cdclk;
>
> if (!crtc_state->base.enable)
> return 0;
>
> - min_cdclk = intel_pixel_rate_to_cdclk(dev_priv, crtc_state->pixel_rate);
> + min_cdclk = intel_pixel_rate_to_cdclk(i915, crtc_state->pixel_rate);
>
> /* pixel rate mustn't exceed 95% of cdclk with IPS on BDW */
> - if (IS_BROADWELL(dev_priv) && hsw_crtc_state_ips_capable(crtc_state))
> + if (IS_BROADWELL(i915) && hsw_crtc_state_ips_capable(crtc_state))
> min_cdclk = DIV_ROUND_UP(min_cdclk * 100, 95);
>
> /* BSpec says "Do not use DisplayPort with CDCLK less than 432 MHz,
> @@ -2223,10 +2223,10 @@ int intel_crtc_compute_min_cdclk(const struct intel_crtc_state *crtc_state)
> crtc_state->has_audio &&
> crtc_state->port_clock >= 540000 &&
> crtc_state->lane_count == 4) {
> - if (IS_CANNONLAKE(dev_priv) || IS_GEMINILAKE(dev_priv)) {
> + if (IS_CANNONLAKE(i915) || IS_GEMINILAKE(i915)) {
> /* Display WA #1145: glk,cnl */
> min_cdclk = max(316800, min_cdclk);
> - } else if (IS_GEN(dev_priv, 9) || IS_BROADWELL(dev_priv)) {
> + } else if (IS_GEN(i915, 9) || IS_BROADWELL(i915)) {
> /* Display WA #1144: skl,bxt */
> min_cdclk = max(432000, min_cdclk);
> }
> @@ -2236,7 +2236,7 @@ int intel_crtc_compute_min_cdclk(const struct intel_crtc_state *crtc_state)
> * According to BSpec, "The CD clock frequency must be at least twice
> * the frequency of the Azalia BCLK." and BCLK is 96 MHz by default.
> */
> - if (crtc_state->has_audio && INTEL_GEN(dev_priv) >= 9)
> + if (crtc_state->has_audio && INTEL_GEN(i915) >= 9)
> min_cdclk = max(2 * 96000, min_cdclk);
>
> /*
> @@ -2244,7 +2244,7 @@ int intel_crtc_compute_min_cdclk(const struct intel_crtc_state *crtc_state)
> * than 320000KHz.
> */
> if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_DSI) &&
> - IS_VALLEYVIEW(dev_priv))
> + IS_VALLEYVIEW(i915))
> min_cdclk = max(320000, min_cdclk);
>
> /*
> @@ -2253,12 +2253,12 @@ int intel_crtc_compute_min_cdclk(const struct intel_crtc_state *crtc_state)
> * correct for DSI PLL and DE PLL.
> */
> if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_DSI) &&
> - IS_GEMINILAKE(dev_priv))
> + IS_GEMINILAKE(i915))
> min_cdclk = max(158400, min_cdclk);
>
> - if (min_cdclk > dev_priv->max_cdclk_freq) {
> + if (min_cdclk > i915->max_cdclk_freq) {
> DRM_DEBUG_KMS("required cdclk (%d kHz) exceeds max (%d kHz)\n",
> - min_cdclk, dev_priv->max_cdclk_freq);
> + min_cdclk, i915->max_cdclk_freq);
> return -EINVAL;
> }
>
> @@ -2267,13 +2267,13 @@ int intel_crtc_compute_min_cdclk(const struct intel_crtc_state *crtc_state)
>
> static int intel_compute_min_cdclk(struct intel_atomic_state *state)
> {
> - struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> + struct drm_i915_private *i915 = to_i915(state->base.dev);
> struct intel_crtc *crtc;
> struct intel_crtc_state *crtc_state;
> int min_cdclk, i;
> enum pipe pipe;
>
> - memcpy(state->min_cdclk, dev_priv->min_cdclk,
> + memcpy(state->min_cdclk, i915->min_cdclk,
> sizeof(state->min_cdclk));
>
> for_each_new_intel_crtc_in_state(state, crtc, crtc_state, i) {
> @@ -2285,7 +2285,7 @@ static int intel_compute_min_cdclk(struct intel_atomic_state *state)
> }
>
> min_cdclk = state->cdclk.force_min_cdclk;
> - for_each_pipe(dev_priv, pipe)
> + for_each_pipe(i915, pipe)
> min_cdclk = max(state->min_cdclk[pipe], min_cdclk);
>
> return min_cdclk;
> @@ -2302,14 +2302,14 @@ static int intel_compute_min_cdclk(struct intel_atomic_state *state)
> */
> static u8 cnl_compute_min_voltage_level(struct intel_atomic_state *state)
> {
> - struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> + struct drm_i915_private *i915 = to_i915(state->base.dev);
> struct intel_crtc *crtc;
> struct intel_crtc_state *crtc_state;
> u8 min_voltage_level;
> int i;
> enum pipe pipe;
>
> - memcpy(state->min_voltage_level, dev_priv->min_voltage_level,
> + memcpy(state->min_voltage_level, i915->min_voltage_level,
> sizeof(state->min_voltage_level));
>
> for_each_new_intel_crtc_in_state(state, crtc, crtc_state, i) {
> @@ -2321,7 +2321,7 @@ static u8 cnl_compute_min_voltage_level(struct intel_atomic_state *state)
> }
>
> min_voltage_level = 0;
> - for_each_pipe(dev_priv, pipe)
> + for_each_pipe(i915, pipe)
> min_voltage_level = max(state->min_voltage_level[pipe],
> min_voltage_level);
>
> @@ -2330,25 +2330,25 @@ static u8 cnl_compute_min_voltage_level(struct intel_atomic_state *state)
>
> static int vlv_modeset_calc_cdclk(struct intel_atomic_state *state)
> {
> - struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> + struct drm_i915_private *i915 = to_i915(state->base.dev);
> int min_cdclk, cdclk;
>
> min_cdclk = intel_compute_min_cdclk(state);
> if (min_cdclk < 0)
> return min_cdclk;
>
> - cdclk = vlv_calc_cdclk(dev_priv, min_cdclk);
> + cdclk = vlv_calc_cdclk(i915, min_cdclk);
>
> state->cdclk.logical.cdclk = cdclk;
> state->cdclk.logical.voltage_level =
> - vlv_calc_voltage_level(dev_priv, cdclk);
> + vlv_calc_voltage_level(i915, cdclk);
>
> if (!state->active_crtcs) {
> - cdclk = vlv_calc_cdclk(dev_priv, state->cdclk.force_min_cdclk);
> + cdclk = vlv_calc_cdclk(i915, state->cdclk.force_min_cdclk);
>
> state->cdclk.actual.cdclk = cdclk;
> state->cdclk.actual.voltage_level =
> - vlv_calc_voltage_level(dev_priv, cdclk);
> + vlv_calc_voltage_level(i915, cdclk);
> } else {
> state->cdclk.actual = state->cdclk.logical;
> }
> @@ -2389,14 +2389,14 @@ static int bdw_modeset_calc_cdclk(struct intel_atomic_state *state)
>
> static int skl_dpll0_vco(struct intel_atomic_state *state)
> {
> - struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> + struct drm_i915_private *i915 = to_i915(state->base.dev);
> struct intel_crtc *crtc;
> struct intel_crtc_state *crtc_state;
> int vco, i;
>
> vco = state->cdclk.logical.vco;
> if (!vco)
> - vco = dev_priv->skl_preferred_vco_freq;
> + vco = i915->skl_preferred_vco_freq;
>
> for_each_new_intel_crtc_in_state(state, crtc, crtc_state, i) {
> if (!crtc_state->base.enable)
> @@ -2460,19 +2460,19 @@ static int skl_modeset_calc_cdclk(struct intel_atomic_state *state)
>
> static int bxt_modeset_calc_cdclk(struct intel_atomic_state *state)
> {
> - struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> + struct drm_i915_private *i915 = to_i915(state->base.dev);
> int min_cdclk, cdclk, vco;
>
> min_cdclk = intel_compute_min_cdclk(state);
> if (min_cdclk < 0)
> return min_cdclk;
>
> - if (IS_GEMINILAKE(dev_priv)) {
> + if (IS_GEMINILAKE(i915)) {
> cdclk = glk_calc_cdclk(min_cdclk);
> - vco = glk_de_pll_vco(dev_priv, cdclk);
> + vco = glk_de_pll_vco(i915, cdclk);
> } else {
> cdclk = bxt_calc_cdclk(min_cdclk);
> - vco = bxt_de_pll_vco(dev_priv, cdclk);
> + vco = bxt_de_pll_vco(i915, cdclk);
> }
>
> state->cdclk.logical.vco = vco;
> @@ -2481,12 +2481,12 @@ static int bxt_modeset_calc_cdclk(struct intel_atomic_state *state)
> bxt_calc_voltage_level(cdclk);
>
> if (!state->active_crtcs) {
> - if (IS_GEMINILAKE(dev_priv)) {
> + if (IS_GEMINILAKE(i915)) {
> cdclk = glk_calc_cdclk(state->cdclk.force_min_cdclk);
> - vco = glk_de_pll_vco(dev_priv, cdclk);
> + vco = glk_de_pll_vco(i915, cdclk);
> } else {
> cdclk = bxt_calc_cdclk(state->cdclk.force_min_cdclk);
> - vco = bxt_de_pll_vco(dev_priv, cdclk);
> + vco = bxt_de_pll_vco(i915, cdclk);
> }
>
> state->cdclk.actual.vco = vco;
> @@ -2502,7 +2502,7 @@ static int bxt_modeset_calc_cdclk(struct intel_atomic_state *state)
>
> static int cnl_modeset_calc_cdclk(struct intel_atomic_state *state)
> {
> - struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> + struct drm_i915_private *i915 = to_i915(state->base.dev);
> int min_cdclk, cdclk, vco;
>
> min_cdclk = intel_compute_min_cdclk(state);
> @@ -2510,7 +2510,7 @@ static int cnl_modeset_calc_cdclk(struct intel_atomic_state *state)
> return min_cdclk;
>
> cdclk = cnl_calc_cdclk(min_cdclk);
> - vco = cnl_cdclk_pll_vco(dev_priv, cdclk);
> + vco = cnl_cdclk_pll_vco(i915, cdclk);
>
> state->cdclk.logical.vco = vco;
> state->cdclk.logical.cdclk = cdclk;
> @@ -2520,7 +2520,7 @@ static int cnl_modeset_calc_cdclk(struct intel_atomic_state *state)
>
> if (!state->active_crtcs) {
> cdclk = cnl_calc_cdclk(state->cdclk.force_min_cdclk);
> - vco = cnl_cdclk_pll_vco(dev_priv, cdclk);
> + vco = cnl_cdclk_pll_vco(i915, cdclk);
>
> state->cdclk.actual.vco = vco;
> state->cdclk.actual.cdclk = cdclk;
> @@ -2535,7 +2535,7 @@ static int cnl_modeset_calc_cdclk(struct intel_atomic_state *state)
>
> static int icl_modeset_calc_cdclk(struct intel_atomic_state *state)
> {
> - struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> + struct drm_i915_private *i915 = to_i915(state->base.dev);
> unsigned int ref = state->cdclk.logical.ref;
> int min_cdclk, cdclk, vco;
>
> @@ -2544,7 +2544,7 @@ static int icl_modeset_calc_cdclk(struct intel_atomic_state *state)
> return min_cdclk;
>
> cdclk = icl_calc_cdclk(min_cdclk, ref);
> - vco = icl_calc_cdclk_pll_vco(dev_priv, cdclk);
> + vco = icl_calc_cdclk_pll_vco(i915, cdclk);
>
> state->cdclk.logical.vco = vco;
> state->cdclk.logical.cdclk = cdclk;
> @@ -2554,7 +2554,7 @@ static int icl_modeset_calc_cdclk(struct intel_atomic_state *state)
>
> if (!state->active_crtcs) {
> cdclk = icl_calc_cdclk(state->cdclk.force_min_cdclk, ref);
> - vco = icl_calc_cdclk_pll_vco(dev_priv, cdclk);
> + vco = icl_calc_cdclk_pll_vco(i915, cdclk);
>
> state->cdclk.actual.vco = vco;
> state->cdclk.actual.cdclk = cdclk;
> @@ -2567,18 +2567,18 @@ static int icl_modeset_calc_cdclk(struct intel_atomic_state *state)
> return 0;
> }
>
> -static int intel_compute_max_dotclk(struct drm_i915_private *dev_priv)
> +static int intel_compute_max_dotclk(struct drm_i915_private *i915)
> {
> - int max_cdclk_freq = dev_priv->max_cdclk_freq;
> + int max_cdclk_freq = i915->max_cdclk_freq;
>
> - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> return 2 * max_cdclk_freq;
> - else if (IS_GEN(dev_priv, 9) ||
> - IS_BROADWELL(dev_priv) || IS_HASWELL(dev_priv))
> + else if (IS_GEN(i915, 9) ||
> + IS_BROADWELL(i915) || IS_HASWELL(i915))
> return max_cdclk_freq;
> - else if (IS_CHERRYVIEW(dev_priv))
> + else if (IS_CHERRYVIEW(i915))
> return max_cdclk_freq*95/100;
> - else if (INTEL_GEN(dev_priv) < 4)
> + else if (INTEL_GEN(i915) < 4)
> return 2*max_cdclk_freq*90/100;
> else
> return max_cdclk_freq*90/100;
> @@ -2586,26 +2586,26 @@ static int intel_compute_max_dotclk(struct drm_i915_private *dev_priv)
>
> /**
> * intel_update_max_cdclk - Determine the maximum support CDCLK frequency
> - * @dev_priv: i915 device
> + * @i915: i915 device
> *
> * Determine the maximum CDCLK frequency the platform supports, and also
> * derive the maximum dot clock frequency the maximum CDCLK frequency
> * allows.
> */
> -void intel_update_max_cdclk(struct drm_i915_private *dev_priv)
> +void intel_update_max_cdclk(struct drm_i915_private *i915)
> {
> - if (INTEL_GEN(dev_priv) >= 11) {
> - if (dev_priv->cdclk.hw.ref == 24000)
> - dev_priv->max_cdclk_freq = 648000;
> + if (INTEL_GEN(i915) >= 11) {
> + if (i915->cdclk.hw.ref == 24000)
> + i915->max_cdclk_freq = 648000;
> else
> - dev_priv->max_cdclk_freq = 652800;
> - } else if (IS_CANNONLAKE(dev_priv)) {
> - dev_priv->max_cdclk_freq = 528000;
> - } else if (IS_GEN9_BC(dev_priv)) {
> + i915->max_cdclk_freq = 652800;
> + } else if (IS_CANNONLAKE(i915)) {
> + i915->max_cdclk_freq = 528000;
> + } else if (IS_GEN9_BC(i915)) {
> u32 limit = I915_READ(SKL_DFSM) & SKL_DFSM_CDCLK_LIMIT_MASK;
> int max_cdclk, vco;
>
> - vco = dev_priv->skl_preferred_vco_freq;
> + vco = i915->skl_preferred_vco_freq;
> WARN_ON(vco != 8100000 && vco != 8640000);
>
> /*
> @@ -2622,12 +2622,12 @@ void intel_update_max_cdclk(struct drm_i915_private *dev_priv)
> else
> max_cdclk = 308571;
>
> - dev_priv->max_cdclk_freq = skl_calc_cdclk(max_cdclk, vco);
> - } else if (IS_GEMINILAKE(dev_priv)) {
> - dev_priv->max_cdclk_freq = 316800;
> - } else if (IS_BROXTON(dev_priv)) {
> - dev_priv->max_cdclk_freq = 624000;
> - } else if (IS_BROADWELL(dev_priv)) {
> + i915->max_cdclk_freq = skl_calc_cdclk(max_cdclk, vco);
> + } else if (IS_GEMINILAKE(i915)) {
> + i915->max_cdclk_freq = 316800;
> + } else if (IS_BROXTON(i915)) {
> + i915->max_cdclk_freq = 624000;
> + } else if (IS_BROADWELL(i915)) {
> /*
> * FIXME with extra cooling we can allow
> * 540 MHz for ULX and 675 Mhz for ULT.
> @@ -2635,40 +2635,40 @@ void intel_update_max_cdclk(struct drm_i915_private *dev_priv)
> * available? PCI ID, VTB, something else?
> */
> if (I915_READ(FUSE_STRAP) & HSW_CDCLK_LIMIT)
> - dev_priv->max_cdclk_freq = 450000;
> - else if (IS_BDW_ULX(dev_priv))
> - dev_priv->max_cdclk_freq = 450000;
> - else if (IS_BDW_ULT(dev_priv))
> - dev_priv->max_cdclk_freq = 540000;
> + i915->max_cdclk_freq = 450000;
> + else if (IS_BDW_ULX(i915))
> + i915->max_cdclk_freq = 450000;
> + else if (IS_BDW_ULT(i915))
> + i915->max_cdclk_freq = 540000;
> else
> - dev_priv->max_cdclk_freq = 675000;
> - } else if (IS_CHERRYVIEW(dev_priv)) {
> - dev_priv->max_cdclk_freq = 320000;
> - } else if (IS_VALLEYVIEW(dev_priv)) {
> - dev_priv->max_cdclk_freq = 400000;
> + i915->max_cdclk_freq = 675000;
> + } else if (IS_CHERRYVIEW(i915)) {
> + i915->max_cdclk_freq = 320000;
> + } else if (IS_VALLEYVIEW(i915)) {
> + i915->max_cdclk_freq = 400000;
> } else {
> /* otherwise assume cdclk is fixed */
> - dev_priv->max_cdclk_freq = dev_priv->cdclk.hw.cdclk;
> + i915->max_cdclk_freq = i915->cdclk.hw.cdclk;
> }
>
> - dev_priv->max_dotclk_freq = intel_compute_max_dotclk(dev_priv);
> + i915->max_dotclk_freq = intel_compute_max_dotclk(i915);
>
> DRM_DEBUG_DRIVER("Max CD clock rate: %d kHz\n",
> - dev_priv->max_cdclk_freq);
> + i915->max_cdclk_freq);
>
> DRM_DEBUG_DRIVER("Max dotclock rate: %d kHz\n",
> - dev_priv->max_dotclk_freq);
> + i915->max_dotclk_freq);
> }
>
> /**
> * intel_update_cdclk - Determine the current CDCLK frequency
> - * @dev_priv: i915 device
> + * @i915: i915 device
> *
> * Determine the current CDCLK frequency.
> */
> -void intel_update_cdclk(struct drm_i915_private *dev_priv)
> +void intel_update_cdclk(struct drm_i915_private *i915)
> {
> - dev_priv->display.get_cdclk(dev_priv, &dev_priv->cdclk.hw);
> + i915->display.get_cdclk(i915, &i915->cdclk.hw);
>
> /*
> * 9:0 CMBUS [sic] CDCLK frequency (cdfreq):
> @@ -2676,12 +2676,12 @@ void intel_update_cdclk(struct drm_i915_private *dev_priv)
> * of cdclk that generates 4MHz reference clock freq which is used to
> * generate GMBus clock. This will vary with the cdclk freq.
> */
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> I915_WRITE(GMBUSFREQ_VLV,
> - DIV_ROUND_UP(dev_priv->cdclk.hw.cdclk, 1000));
> + DIV_ROUND_UP(i915->cdclk.hw.cdclk, 1000));
> }
>
> -static int cnp_rawclk(struct drm_i915_private *dev_priv)
> +static int cnp_rawclk(struct drm_i915_private *i915)
> {
> u32 rawclk;
> int divider, fraction;
> @@ -2702,7 +2702,7 @@ static int cnp_rawclk(struct drm_i915_private *dev_priv)
>
> rawclk |= CNP_RAWCLK_DEN(DIV_ROUND_CLOSEST(numerator * 1000,
> fraction) - 1);
> - if (INTEL_PCH_TYPE(dev_priv) >= PCH_ICP)
> + if (INTEL_PCH_TYPE(i915) >= PCH_ICP)
> rawclk |= ICP_RAWCLK_NUM(numerator);
> }
>
> @@ -2710,19 +2710,19 @@ static int cnp_rawclk(struct drm_i915_private *dev_priv)
> return divider + fraction;
> }
>
> -static int pch_rawclk(struct drm_i915_private *dev_priv)
> +static int pch_rawclk(struct drm_i915_private *i915)
> {
> return (I915_READ(PCH_RAWCLK_FREQ) & RAWCLK_FREQ_MASK) * 1000;
> }
>
> -static int vlv_hrawclk(struct drm_i915_private *dev_priv)
> +static int vlv_hrawclk(struct drm_i915_private *i915)
> {
> /* RAWCLK_FREQ_VLV register updated from power well code */
> - return vlv_get_cck_clock_hpll(dev_priv, "hrawclk",
> + return vlv_get_cck_clock_hpll(i915, "hrawclk",
> CCK_DISPLAY_REF_CLOCK_CONTROL);
> }
>
> -static int g4x_hrawclk(struct drm_i915_private *dev_priv)
> +static int g4x_hrawclk(struct drm_i915_private *i915)
> {
> u32 clkcfg;
>
> @@ -2750,104 +2750,104 @@ static int g4x_hrawclk(struct drm_i915_private *dev_priv)
>
> /**
> * intel_update_rawclk - Determine the current RAWCLK frequency
> - * @dev_priv: i915 device
> + * @i915: i915 device
> *
> * Determine the current RAWCLK frequency. RAWCLK is a fixed
> * frequency clock so this needs to done only once.
> */
> -void intel_update_rawclk(struct drm_i915_private *dev_priv)
> -{
> - if (INTEL_PCH_TYPE(dev_priv) >= PCH_CNP)
> - dev_priv->rawclk_freq = cnp_rawclk(dev_priv);
> - else if (HAS_PCH_SPLIT(dev_priv))
> - dev_priv->rawclk_freq = pch_rawclk(dev_priv);
> - else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> - dev_priv->rawclk_freq = vlv_hrawclk(dev_priv);
> - else if (IS_G4X(dev_priv) || IS_PINEVIEW(dev_priv))
> - dev_priv->rawclk_freq = g4x_hrawclk(dev_priv);
> +void intel_update_rawclk(struct drm_i915_private *i915)
> +{
> + if (INTEL_PCH_TYPE(i915) >= PCH_CNP)
> + i915->rawclk_freq = cnp_rawclk(i915);
> + else if (HAS_PCH_SPLIT(i915))
> + i915->rawclk_freq = pch_rawclk(i915);
> + else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> + i915->rawclk_freq = vlv_hrawclk(i915);
> + else if (IS_G4X(i915) || IS_PINEVIEW(i915))
> + i915->rawclk_freq = g4x_hrawclk(i915);
> else
> /* no rawclk on other platforms, or no need to know it */
> return;
>
> - DRM_DEBUG_DRIVER("rawclk rate: %d kHz\n", dev_priv->rawclk_freq);
> + DRM_DEBUG_DRIVER("rawclk rate: %d kHz\n", i915->rawclk_freq);
> }
>
> /**
> * intel_init_cdclk_hooks - Initialize CDCLK related modesetting hooks
> - * @dev_priv: i915 device
> + * @i915: i915 device
> */
> -void intel_init_cdclk_hooks(struct drm_i915_private *dev_priv)
> -{
> - if (INTEL_GEN(dev_priv) >= 11) {
> - dev_priv->display.set_cdclk = icl_set_cdclk;
> - dev_priv->display.modeset_calc_cdclk = icl_modeset_calc_cdclk;
> - } else if (IS_CANNONLAKE(dev_priv)) {
> - dev_priv->display.set_cdclk = cnl_set_cdclk;
> - dev_priv->display.modeset_calc_cdclk = cnl_modeset_calc_cdclk;
> - } else if (IS_GEN9_LP(dev_priv)) {
> - dev_priv->display.set_cdclk = bxt_set_cdclk;
> - dev_priv->display.modeset_calc_cdclk = bxt_modeset_calc_cdclk;
> - } else if (IS_GEN9_BC(dev_priv)) {
> - dev_priv->display.set_cdclk = skl_set_cdclk;
> - dev_priv->display.modeset_calc_cdclk = skl_modeset_calc_cdclk;
> - } else if (IS_BROADWELL(dev_priv)) {
> - dev_priv->display.set_cdclk = bdw_set_cdclk;
> - dev_priv->display.modeset_calc_cdclk = bdw_modeset_calc_cdclk;
> - } else if (IS_CHERRYVIEW(dev_priv)) {
> - dev_priv->display.set_cdclk = chv_set_cdclk;
> - dev_priv->display.modeset_calc_cdclk = vlv_modeset_calc_cdclk;
> - } else if (IS_VALLEYVIEW(dev_priv)) {
> - dev_priv->display.set_cdclk = vlv_set_cdclk;
> - dev_priv->display.modeset_calc_cdclk = vlv_modeset_calc_cdclk;
> +void intel_init_cdclk_hooks(struct drm_i915_private *i915)
> +{
> + if (INTEL_GEN(i915) >= 11) {
> + i915->display.set_cdclk = icl_set_cdclk;
> + i915->display.modeset_calc_cdclk = icl_modeset_calc_cdclk;
> + } else if (IS_CANNONLAKE(i915)) {
> + i915->display.set_cdclk = cnl_set_cdclk;
> + i915->display.modeset_calc_cdclk = cnl_modeset_calc_cdclk;
> + } else if (IS_GEN9_LP(i915)) {
> + i915->display.set_cdclk = bxt_set_cdclk;
> + i915->display.modeset_calc_cdclk = bxt_modeset_calc_cdclk;
> + } else if (IS_GEN9_BC(i915)) {
> + i915->display.set_cdclk = skl_set_cdclk;
> + i915->display.modeset_calc_cdclk = skl_modeset_calc_cdclk;
> + } else if (IS_BROADWELL(i915)) {
> + i915->display.set_cdclk = bdw_set_cdclk;
> + i915->display.modeset_calc_cdclk = bdw_modeset_calc_cdclk;
> + } else if (IS_CHERRYVIEW(i915)) {
> + i915->display.set_cdclk = chv_set_cdclk;
> + i915->display.modeset_calc_cdclk = vlv_modeset_calc_cdclk;
> + } else if (IS_VALLEYVIEW(i915)) {
> + i915->display.set_cdclk = vlv_set_cdclk;
> + i915->display.modeset_calc_cdclk = vlv_modeset_calc_cdclk;
> }
>
> - if (INTEL_GEN(dev_priv) >= 11)
> - dev_priv->display.get_cdclk = icl_get_cdclk;
> - else if (IS_CANNONLAKE(dev_priv))
> - dev_priv->display.get_cdclk = cnl_get_cdclk;
> - else if (IS_GEN9_LP(dev_priv))
> - dev_priv->display.get_cdclk = bxt_get_cdclk;
> - else if (IS_GEN9_BC(dev_priv))
> - dev_priv->display.get_cdclk = skl_get_cdclk;
> - else if (IS_BROADWELL(dev_priv))
> - dev_priv->display.get_cdclk = bdw_get_cdclk;
> - else if (IS_HASWELL(dev_priv))
> - dev_priv->display.get_cdclk = hsw_get_cdclk;
> - else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> - dev_priv->display.get_cdclk = vlv_get_cdclk;
> - else if (IS_GEN(dev_priv, 6) || IS_IVYBRIDGE(dev_priv))
> - dev_priv->display.get_cdclk = fixed_400mhz_get_cdclk;
> - else if (IS_GEN(dev_priv, 5))
> - dev_priv->display.get_cdclk = fixed_450mhz_get_cdclk;
> - else if (IS_GM45(dev_priv))
> - dev_priv->display.get_cdclk = gm45_get_cdclk;
> - else if (IS_G45(dev_priv))
> - dev_priv->display.get_cdclk = g33_get_cdclk;
> - else if (IS_I965GM(dev_priv))
> - dev_priv->display.get_cdclk = i965gm_get_cdclk;
> - else if (IS_I965G(dev_priv))
> - dev_priv->display.get_cdclk = fixed_400mhz_get_cdclk;
> - else if (IS_PINEVIEW(dev_priv))
> - dev_priv->display.get_cdclk = pnv_get_cdclk;
> - else if (IS_G33(dev_priv))
> - dev_priv->display.get_cdclk = g33_get_cdclk;
> - else if (IS_I945GM(dev_priv))
> - dev_priv->display.get_cdclk = i945gm_get_cdclk;
> - else if (IS_I945G(dev_priv))
> - dev_priv->display.get_cdclk = fixed_400mhz_get_cdclk;
> - else if (IS_I915GM(dev_priv))
> - dev_priv->display.get_cdclk = i915gm_get_cdclk;
> - else if (IS_I915G(dev_priv))
> - dev_priv->display.get_cdclk = fixed_333mhz_get_cdclk;
> - else if (IS_I865G(dev_priv))
> - dev_priv->display.get_cdclk = fixed_266mhz_get_cdclk;
> - else if (IS_I85X(dev_priv))
> - dev_priv->display.get_cdclk = i85x_get_cdclk;
> - else if (IS_I845G(dev_priv))
> - dev_priv->display.get_cdclk = fixed_200mhz_get_cdclk;
> + if (INTEL_GEN(i915) >= 11)
> + i915->display.get_cdclk = icl_get_cdclk;
> + else if (IS_CANNONLAKE(i915))
> + i915->display.get_cdclk = cnl_get_cdclk;
> + else if (IS_GEN9_LP(i915))
> + i915->display.get_cdclk = bxt_get_cdclk;
> + else if (IS_GEN9_BC(i915))
> + i915->display.get_cdclk = skl_get_cdclk;
> + else if (IS_BROADWELL(i915))
> + i915->display.get_cdclk = bdw_get_cdclk;
> + else if (IS_HASWELL(i915))
> + i915->display.get_cdclk = hsw_get_cdclk;
> + else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> + i915->display.get_cdclk = vlv_get_cdclk;
> + else if (IS_GEN(i915, 6) || IS_IVYBRIDGE(i915))
> + i915->display.get_cdclk = fixed_400mhz_get_cdclk;
> + else if (IS_GEN(i915, 5))
> + i915->display.get_cdclk = fixed_450mhz_get_cdclk;
> + else if (IS_GM45(i915))
> + i915->display.get_cdclk = gm45_get_cdclk;
> + else if (IS_G45(i915))
> + i915->display.get_cdclk = g33_get_cdclk;
> + else if (IS_I965GM(i915))
> + i915->display.get_cdclk = i965gm_get_cdclk;
> + else if (IS_I965G(i915))
> + i915->display.get_cdclk = fixed_400mhz_get_cdclk;
> + else if (IS_PINEVIEW(i915))
> + i915->display.get_cdclk = pnv_get_cdclk;
> + else if (IS_G33(i915))
> + i915->display.get_cdclk = g33_get_cdclk;
> + else if (IS_I945GM(i915))
> + i915->display.get_cdclk = i945gm_get_cdclk;
> + else if (IS_I945G(i915))
> + i915->display.get_cdclk = fixed_400mhz_get_cdclk;
> + else if (IS_I915GM(i915))
> + i915->display.get_cdclk = i915gm_get_cdclk;
> + else if (IS_I915G(i915))
> + i915->display.get_cdclk = fixed_333mhz_get_cdclk;
> + else if (IS_I865G(i915))
> + i915->display.get_cdclk = fixed_266mhz_get_cdclk;
> + else if (IS_I85X(i915))
> + i915->display.get_cdclk = i85x_get_cdclk;
> + else if (IS_I845G(i915))
> + i915->display.get_cdclk = fixed_200mhz_get_cdclk;
> else { /* 830 */
> - WARN(!IS_I830(dev_priv),
> + WARN(!IS_I830(i915),
> "Unknown platform. Assuming 133 MHz CDCLK\n");
> - dev_priv->display.get_cdclk = fixed_133mhz_get_cdclk;
> + i915->display.get_cdclk = fixed_133mhz_get_cdclk;
> }
> }
> diff --git a/drivers/gpu/drm/i915/intel_cdclk.h b/drivers/gpu/drm/i915/intel_cdclk.h
> index 4d6f7f5f8930..4e17102af66c 100644
> --- a/drivers/gpu/drm/i915/intel_cdclk.h
> +++ b/drivers/gpu/drm/i915/intel_cdclk.h
> @@ -18,11 +18,11 @@ struct intel_crtc_state;
> int intel_crtc_compute_min_cdclk(const struct intel_crtc_state *crtc_state);
> void intel_cdclk_init(struct drm_i915_private *i915);
> void intel_cdclk_uninit(struct drm_i915_private *i915);
> -void intel_init_cdclk_hooks(struct drm_i915_private *dev_priv);
> -void intel_update_max_cdclk(struct drm_i915_private *dev_priv);
> -void intel_update_cdclk(struct drm_i915_private *dev_priv);
> -void intel_update_rawclk(struct drm_i915_private *dev_priv);
> -bool intel_cdclk_needs_cd2x_update(struct drm_i915_private *dev_priv,
> +void intel_init_cdclk_hooks(struct drm_i915_private *i915);
> +void intel_update_max_cdclk(struct drm_i915_private *i915);
> +void intel_update_cdclk(struct drm_i915_private *i915);
> +void intel_update_rawclk(struct drm_i915_private *i915);
> +bool intel_cdclk_needs_cd2x_update(struct drm_i915_private *i915,
> const struct intel_cdclk_state *a,
> const struct intel_cdclk_state *b);
> bool intel_cdclk_needs_modeset(const struct intel_cdclk_state *a,
> @@ -31,12 +31,12 @@ bool intel_cdclk_changed(const struct intel_cdclk_state *a,
> const struct intel_cdclk_state *b);
> void intel_cdclk_swap_state(struct intel_atomic_state *state);
> void
> -intel_set_cdclk_pre_plane_update(struct drm_i915_private *dev_priv,
> +intel_set_cdclk_pre_plane_update(struct drm_i915_private *i915,
> const struct intel_cdclk_state *old_state,
> const struct intel_cdclk_state *new_state,
> enum pipe pipe);
> void
> -intel_set_cdclk_post_plane_update(struct drm_i915_private *dev_priv,
> +intel_set_cdclk_post_plane_update(struct drm_i915_private *i915,
> const struct intel_cdclk_state *old_state,
> const struct intel_cdclk_state *new_state,
> enum pipe pipe);
> diff --git a/drivers/gpu/drm/i915/intel_color.c b/drivers/gpu/drm/i915/intel_color.c
> index 0b8cf3e8c963..efc67e4a0ba5 100644
> --- a/drivers/gpu/drm/i915/intel_color.c
> +++ b/drivers/gpu/drm/i915/intel_color.c
> @@ -137,7 +137,7 @@ static void ilk_update_pipe_csc(struct intel_crtc *crtc,
> const u16 coeff[9],
> const u16 postoff[3])
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
>
> I915_WRITE(PIPE_CSC_PREOFF_HI(pipe), preoff[0]);
> @@ -153,7 +153,7 @@ static void ilk_update_pipe_csc(struct intel_crtc *crtc,
> I915_WRITE(PIPE_CSC_COEFF_RV_GV(pipe), coeff[6] << 16 | coeff[7]);
> I915_WRITE(PIPE_CSC_COEFF_BV(pipe), coeff[8] << 16);
>
> - if (INTEL_GEN(dev_priv) >= 7) {
> + if (INTEL_GEN(i915) >= 7) {
> I915_WRITE(PIPE_CSC_POSTOFF_HI(pipe), postoff[0]);
> I915_WRITE(PIPE_CSC_POSTOFF_ME(pipe), postoff[1]);
> I915_WRITE(PIPE_CSC_POSTOFF_LO(pipe), postoff[2]);
> @@ -165,7 +165,7 @@ static void icl_update_output_csc(struct intel_crtc *crtc,
> const u16 coeff[9],
> const u16 postoff[3])
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
>
> I915_WRITE(PIPE_CSC_OUTPUT_PREOFF_HI(pipe), preoff[0]);
> @@ -188,15 +188,15 @@ static void icl_update_output_csc(struct intel_crtc *crtc,
>
> static bool ilk_csc_limited_range(const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
>
> /*
> * FIXME if there's a gamma LUT after the CSC, we should
> * do the range compression using the gamma LUT instead.
> */
> return crtc_state->limited_color_range &&
> - (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv) ||
> - IS_GEN_RANGE(dev_priv, 9, 10));
> + (IS_HASWELL(i915) || IS_BROADWELL(i915) ||
> + IS_GEN_RANGE(i915, 9, 10));
> }
>
> static void ilk_csc_convert_ctm(const struct intel_crtc_state *crtc_state,
> @@ -254,7 +254,7 @@ static void ilk_csc_convert_ctm(const struct intel_crtc_state *crtc_state,
> static void ilk_load_csc_matrix(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> bool limited_color_range = ilk_csc_limited_range(crtc_state);
>
> if (crtc_state->base.ctm) {
> @@ -280,7 +280,7 @@ static void ilk_load_csc_matrix(const struct intel_crtc_state *crtc_state)
> * LUT is needed but CSC is not we need to load an
> * identity matrix.
> */
> - WARN_ON(!IS_CANNONLAKE(dev_priv) && !IS_GEMINILAKE(dev_priv));
> + WARN_ON(!IS_CANNONLAKE(i915) && !IS_GEMINILAKE(i915));
>
> ilk_update_pipe_csc(crtc, ilk_csc_off_zero,
> ilk_csc_coeff_identity,
> @@ -293,7 +293,7 @@ static void ilk_load_csc_matrix(const struct intel_crtc_state *crtc_state)
> static void icl_load_csc_matrix(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
>
> if (crtc_state->base.ctm) {
> u16 coeff[9];
> @@ -322,7 +322,7 @@ static void icl_load_csc_matrix(const struct intel_crtc_state *crtc_state)
> static void cherryview_load_csc_matrix(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
>
> if (crtc_state->base.ctm) {
> @@ -388,15 +388,15 @@ static void i9xx_load_luts_internal(const struct intel_crtc_state *crtc_state,
> const struct drm_property_blob *blob)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
> int i;
>
> - if (HAS_GMCH(dev_priv)) {
> + if (HAS_GMCH(i915)) {
> if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_DSI))
> - assert_dsi_pll_enabled(dev_priv);
> + assert_dsi_pll_enabled(i915);
> else
> - assert_pll_enabled(dev_priv, pipe);
> + assert_pll_enabled(i915, pipe);
> }
>
> if (blob) {
> @@ -408,7 +408,7 @@ static void i9xx_load_luts_internal(const struct intel_crtc_state *crtc_state,
> (drm_color_lut_extract(lut[i].green, 8) << 8) |
> drm_color_lut_extract(lut[i].blue, 8);
>
> - if (HAS_GMCH(dev_priv))
> + if (HAS_GMCH(i915))
> I915_WRITE(PALETTE(pipe, i), word);
> else
> I915_WRITE(LGC_PALETTE(pipe, i), word);
> @@ -424,7 +424,7 @@ static void i9xx_load_luts(const struct intel_crtc_state *crtc_state)
> static void i9xx_color_commit(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
> u32 val;
>
> @@ -437,7 +437,7 @@ static void i9xx_color_commit(const struct intel_crtc_state *crtc_state)
> static void ilk_color_commit(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
> u32 val;
>
> @@ -452,7 +452,7 @@ static void ilk_color_commit(const struct intel_crtc_state *crtc_state)
> static void hsw_color_commit(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
>
> I915_WRITE(GAMMA_MODE(crtc->pipe), crtc_state->gamma_mode);
>
> @@ -462,7 +462,7 @@ static void hsw_color_commit(const struct intel_crtc_state *crtc_state)
> static void skl_color_commit(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
> u32 val = 0;
>
> @@ -479,7 +479,7 @@ static void skl_color_commit(const struct intel_crtc_state *crtc_state)
>
> I915_WRITE(GAMMA_MODE(crtc->pipe), crtc_state->gamma_mode);
>
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> icl_load_csc_matrix(crtc_state);
> else
> ilk_load_csc_matrix(crtc_state);
> @@ -488,7 +488,7 @@ static void skl_color_commit(const struct intel_crtc_state *crtc_state)
> static void i965_load_lut_10p6(struct intel_crtc *crtc,
> const struct drm_property_blob *blob)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> const struct drm_color_lut *lut = blob->data;
> int i, lut_size = drm_color_lut_size(blob);
> enum pipe pipe = crtc->pipe;
> @@ -519,7 +519,7 @@ static void i965_load_luts(const struct intel_crtc_state *crtc_state)
> static void ilk_load_lut_10(struct intel_crtc *crtc,
> const struct drm_property_blob *blob)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> const struct drm_color_lut *lut = blob->data;
> int i, lut_size = drm_color_lut_size(blob);
> enum pipe pipe = crtc->pipe;
> @@ -556,7 +556,7 @@ static void ivb_load_lut_10(struct intel_crtc *crtc,
> const struct drm_property_blob *blob,
> u32 prec_index)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> int hw_lut_size = ivb_lut_10_size(prec_index);
> const struct drm_color_lut *lut = blob->data;
> int i, lut_size = drm_color_lut_size(blob);
> @@ -583,7 +583,7 @@ static void bdw_load_lut_10(struct intel_crtc *crtc,
> const struct drm_property_blob *blob,
> u32 prec_index)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> int hw_lut_size = ivb_lut_10_size(prec_index);
> const struct drm_color_lut *lut = blob->data;
> int i, lut_size = drm_color_lut_size(blob);
> @@ -609,7 +609,7 @@ static void bdw_load_lut_10(struct intel_crtc *crtc,
>
> static void ivb_load_lut_10_max(struct intel_crtc *crtc)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
>
> /* Program the max register to clamp values > 1.0. */
> @@ -622,7 +622,7 @@ static void ivb_load_lut_10_max(struct intel_crtc *crtc)
> * ToDo: Extend the ABI to be able to program values
> * from 3.0 to 7.0
> */
> - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) {
> + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) {
> I915_WRITE(PREC_PAL_EXT2_GC_MAX(pipe, 0), 1 << 16);
> I915_WRITE(PREC_PAL_EXT2_GC_MAX(pipe, 1), 1 << 16);
> I915_WRITE(PREC_PAL_EXT2_GC_MAX(pipe, 2), 1 << 16);
> @@ -678,9 +678,9 @@ static void bdw_load_luts(const struct intel_crtc_state *crtc_state)
> static void glk_load_degamma_lut(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
> - const u32 lut_size = INTEL_INFO(dev_priv)->color.degamma_lut_size;
> + const u32 lut_size = INTEL_INFO(i915)->color.degamma_lut_size;
> const struct drm_color_lut *lut = crtc_state->base.degamma_lut->data;
> u32 i;
>
> @@ -717,9 +717,9 @@ static void glk_load_degamma_lut(const struct intel_crtc_state *crtc_state)
> static void glk_load_degamma_lut_linear(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
> - const u32 lut_size = INTEL_INFO(dev_priv)->color.degamma_lut_size;
> + const u32 lut_size = INTEL_INFO(i915)->color.degamma_lut_size;
> u32 i;
>
> /*
> @@ -798,7 +798,7 @@ static u32 chv_cgm_degamma_udw(const struct drm_color_lut *color)
> static void chv_load_cgm_degamma(struct intel_crtc *crtc,
> const struct drm_property_blob *blob)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> const struct drm_color_lut *lut = blob->data;
> int i, lut_size = drm_color_lut_size(blob);
> enum pipe pipe = crtc->pipe;
> @@ -825,7 +825,7 @@ static u32 chv_cgm_gamma_udw(const struct drm_color_lut *color)
> static void chv_load_cgm_gamma(struct intel_crtc *crtc,
> const struct drm_property_blob *blob)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> const struct drm_color_lut *lut = blob->data;
> int i, lut_size = drm_color_lut_size(blob);
> enum pipe pipe = crtc->pipe;
> @@ -860,37 +860,37 @@ static void chv_load_luts(const struct intel_crtc_state *crtc_state)
>
> void intel_color_load_luts(const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
>
> - dev_priv->display.load_luts(crtc_state);
> + i915->display.load_luts(crtc_state);
> }
>
> void intel_color_commit(const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
>
> - dev_priv->display.color_commit(crtc_state);
> + i915->display.color_commit(crtc_state);
> }
>
> int intel_color_check(struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
>
> - return dev_priv->display.color_check(crtc_state);
> + return i915->display.color_check(crtc_state);
> }
>
> void intel_color_get_config(struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
>
> - if (dev_priv->display.read_luts)
> - dev_priv->display.read_luts(crtc_state);
> + if (i915->display.read_luts)
> + i915->display.read_luts(crtc_state);
> }
>
> static bool need_plane_update(struct intel_plane *plane,
> const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
>
> /*
> * On pre-SKL the pipe gamma enable and pipe csc enable for
> @@ -898,7 +898,7 @@ static bool need_plane_update(struct intel_plane *plane,
> * We have to reconfigure that even if the plane is inactive.
> */
> return crtc_state->active_planes & BIT(plane->id) ||
> - (INTEL_GEN(dev_priv) < 9 &&
> + (INTEL_GEN(i915) < 9 &&
> plane->id == PLANE_PRIMARY);
> }
>
> @@ -906,7 +906,7 @@ static int
> intel_color_add_affected_planes(struct intel_crtc_state *new_crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(new_crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> struct intel_atomic_state *state =
> to_intel_atomic_state(new_crtc_state->base.state);
> const struct intel_crtc_state *old_crtc_state =
> @@ -921,7 +921,7 @@ intel_color_add_affected_planes(struct intel_crtc_state *new_crtc_state)
> new_crtc_state->csc_enable == old_crtc_state->csc_enable)
> return 0;
>
> - for_each_intel_plane_on_crtc(&dev_priv->drm, crtc, plane) {
> + for_each_intel_plane_on_crtc(&i915->drm, crtc, plane) {
> struct intel_plane_state *plane_state;
>
> if (!need_plane_update(plane, new_crtc_state))
> @@ -956,7 +956,7 @@ static int check_lut_size(const struct drm_property_blob *lut, int expected)
>
> static int check_luts(const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> const struct drm_property_blob *gamma_lut = crtc_state->base.gamma_lut;
> const struct drm_property_blob *degamma_lut = crtc_state->base.degamma_lut;
> int gamma_length, degamma_length;
> @@ -972,10 +972,10 @@ static int check_luts(const struct intel_crtc_state *crtc_state)
> return -EINVAL;
> }
>
> - degamma_length = INTEL_INFO(dev_priv)->color.degamma_lut_size;
> - gamma_length = INTEL_INFO(dev_priv)->color.gamma_lut_size;
> - degamma_tests = INTEL_INFO(dev_priv)->color.degamma_lut_tests;
> - gamma_tests = INTEL_INFO(dev_priv)->color.gamma_lut_tests;
> + degamma_length = INTEL_INFO(i915)->color.degamma_lut_size;
> + gamma_length = INTEL_INFO(i915)->color.gamma_lut_size;
> + degamma_tests = INTEL_INFO(i915)->color.degamma_lut_tests;
> + gamma_tests = INTEL_INFO(i915)->color.gamma_lut_tests;
>
> if (check_lut_size(degamma_lut, degamma_length) ||
> check_lut_size(gamma_lut, gamma_length))
> @@ -1255,56 +1255,56 @@ static int icl_color_check(struct intel_crtc_state *crtc_state)
>
> void intel_color_init(struct intel_crtc *crtc)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> - bool has_ctm = INTEL_INFO(dev_priv)->color.degamma_lut_size != 0;
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> + bool has_ctm = INTEL_INFO(i915)->color.degamma_lut_size != 0;
>
> drm_mode_crtc_set_gamma_size(&crtc->base, 256);
>
> - if (HAS_GMCH(dev_priv)) {
> - if (IS_CHERRYVIEW(dev_priv)) {
> - dev_priv->display.color_check = chv_color_check;
> - dev_priv->display.color_commit = i9xx_color_commit;
> - dev_priv->display.load_luts = chv_load_luts;
> - } else if (INTEL_GEN(dev_priv) >= 4) {
> - dev_priv->display.color_check = i9xx_color_check;
> - dev_priv->display.color_commit = i9xx_color_commit;
> - dev_priv->display.load_luts = i965_load_luts;
> + if (HAS_GMCH(i915)) {
> + if (IS_CHERRYVIEW(i915)) {
> + i915->display.color_check = chv_color_check;
> + i915->display.color_commit = i9xx_color_commit;
> + i915->display.load_luts = chv_load_luts;
> + } else if (INTEL_GEN(i915) >= 4) {
> + i915->display.color_check = i9xx_color_check;
> + i915->display.color_commit = i9xx_color_commit;
> + i915->display.load_luts = i965_load_luts;
> } else {
> - dev_priv->display.color_check = i9xx_color_check;
> - dev_priv->display.color_commit = i9xx_color_commit;
> - dev_priv->display.load_luts = i9xx_load_luts;
> + i915->display.color_check = i9xx_color_check;
> + i915->display.color_commit = i9xx_color_commit;
> + i915->display.load_luts = i9xx_load_luts;
> }
> } else {
> - if (INTEL_GEN(dev_priv) >= 11)
> - dev_priv->display.color_check = icl_color_check;
> - else if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> - dev_priv->display.color_check = glk_color_check;
> - else if (INTEL_GEN(dev_priv) >= 7)
> - dev_priv->display.color_check = ivb_color_check;
> + if (INTEL_GEN(i915) >= 11)
> + i915->display.color_check = icl_color_check;
> + else if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> + i915->display.color_check = glk_color_check;
> + else if (INTEL_GEN(i915) >= 7)
> + i915->display.color_check = ivb_color_check;
> else
> - dev_priv->display.color_check = ilk_color_check;
> + i915->display.color_check = ilk_color_check;
>
> - if (INTEL_GEN(dev_priv) >= 9)
> - dev_priv->display.color_commit = skl_color_commit;
> - else if (IS_BROADWELL(dev_priv) || IS_HASWELL(dev_priv))
> - dev_priv->display.color_commit = hsw_color_commit;
> + if (INTEL_GEN(i915) >= 9)
> + i915->display.color_commit = skl_color_commit;
> + else if (IS_BROADWELL(i915) || IS_HASWELL(i915))
> + i915->display.color_commit = hsw_color_commit;
> else
> - dev_priv->display.color_commit = ilk_color_commit;
> -
> - if (INTEL_GEN(dev_priv) >= 11)
> - dev_priv->display.load_luts = icl_load_luts;
> - else if (IS_CANNONLAKE(dev_priv) || IS_GEMINILAKE(dev_priv))
> - dev_priv->display.load_luts = glk_load_luts;
> - else if (INTEL_GEN(dev_priv) >= 8)
> - dev_priv->display.load_luts = bdw_load_luts;
> - else if (INTEL_GEN(dev_priv) >= 7)
> - dev_priv->display.load_luts = ivb_load_luts;
> + i915->display.color_commit = ilk_color_commit;
> +
> + if (INTEL_GEN(i915) >= 11)
> + i915->display.load_luts = icl_load_luts;
> + else if (IS_CANNONLAKE(i915) || IS_GEMINILAKE(i915))
> + i915->display.load_luts = glk_load_luts;
> + else if (INTEL_GEN(i915) >= 8)
> + i915->display.load_luts = bdw_load_luts;
> + else if (INTEL_GEN(i915) >= 7)
> + i915->display.load_luts = ivb_load_luts;
> else
> - dev_priv->display.load_luts = ilk_load_luts;
> + i915->display.load_luts = ilk_load_luts;
> }
>
> drm_crtc_enable_color_mgmt(&crtc->base,
> - INTEL_INFO(dev_priv)->color.degamma_lut_size,
> + INTEL_INFO(i915)->color.degamma_lut_size,
> has_ctm,
> - INTEL_INFO(dev_priv)->color.gamma_lut_size);
> + INTEL_INFO(i915)->color.gamma_lut_size);
> }
> diff --git a/drivers/gpu/drm/i915/intel_combo_phy.c b/drivers/gpu/drm/i915/intel_combo_phy.c
> index 841708da5a56..0e7c18d783ca 100644
> --- a/drivers/gpu/drm/i915/intel_combo_phy.c
> +++ b/drivers/gpu/drm/i915/intel_combo_phy.c
> @@ -44,7 +44,7 @@ static const struct cnl_procmon {
> * on its name.
> */
> static const struct cnl_procmon *
> -cnl_get_procmon_ref_values(struct drm_i915_private *dev_priv, enum port port)
> +cnl_get_procmon_ref_values(struct drm_i915_private *i915, enum port port)
> {
> const struct cnl_procmon *procmon;
> u32 val;
> @@ -74,13 +74,13 @@ cnl_get_procmon_ref_values(struct drm_i915_private *dev_priv, enum port port)
> return procmon;
> }
>
> -static void cnl_set_procmon_ref_values(struct drm_i915_private *dev_priv,
> +static void cnl_set_procmon_ref_values(struct drm_i915_private *i915,
> enum port port)
> {
> const struct cnl_procmon *procmon;
> u32 val;
>
> - procmon = cnl_get_procmon_ref_values(dev_priv, port);
> + procmon = cnl_get_procmon_ref_values(i915, port);
>
> val = I915_READ(ICL_PORT_COMP_DW1(port));
> val &= ~((0xff << 16) | 0xff);
> @@ -91,7 +91,7 @@ static void cnl_set_procmon_ref_values(struct drm_i915_private *dev_priv,
> I915_WRITE(ICL_PORT_COMP_DW10(port), procmon->dw10);
> }
>
> -static bool check_phy_reg(struct drm_i915_private *dev_priv,
> +static bool check_phy_reg(struct drm_i915_private *i915,
> enum port port, i915_reg_t reg, u32 mask,
> u32 expected_val)
> {
> @@ -108,47 +108,47 @@ static bool check_phy_reg(struct drm_i915_private *dev_priv,
> return true;
> }
>
> -static bool cnl_verify_procmon_ref_values(struct drm_i915_private *dev_priv,
> +static bool cnl_verify_procmon_ref_values(struct drm_i915_private *i915,
> enum port port)
> {
> const struct cnl_procmon *procmon;
> bool ret;
>
> - procmon = cnl_get_procmon_ref_values(dev_priv, port);
> + procmon = cnl_get_procmon_ref_values(i915, port);
>
> - ret = check_phy_reg(dev_priv, port, ICL_PORT_COMP_DW1(port),
> + ret = check_phy_reg(i915, port, ICL_PORT_COMP_DW1(port),
> (0xff << 16) | 0xff, procmon->dw1);
> - ret &= check_phy_reg(dev_priv, port, ICL_PORT_COMP_DW9(port),
> + ret &= check_phy_reg(i915, port, ICL_PORT_COMP_DW9(port),
> -1U, procmon->dw9);
> - ret &= check_phy_reg(dev_priv, port, ICL_PORT_COMP_DW10(port),
> + ret &= check_phy_reg(i915, port, ICL_PORT_COMP_DW10(port),
> -1U, procmon->dw10);
>
> return ret;
> }
>
> -static bool cnl_combo_phy_enabled(struct drm_i915_private *dev_priv)
> +static bool cnl_combo_phy_enabled(struct drm_i915_private *i915)
> {
> return !(I915_READ(CHICKEN_MISC_2) & CNL_COMP_PWR_DOWN) &&
> (I915_READ(CNL_PORT_COMP_DW0) & COMP_INIT);
> }
>
> -static bool cnl_combo_phy_verify_state(struct drm_i915_private *dev_priv)
> +static bool cnl_combo_phy_verify_state(struct drm_i915_private *i915)
> {
> enum port port = PORT_A;
> bool ret;
>
> - if (!cnl_combo_phy_enabled(dev_priv))
> + if (!cnl_combo_phy_enabled(i915))
> return false;
>
> - ret = cnl_verify_procmon_ref_values(dev_priv, port);
> + ret = cnl_verify_procmon_ref_values(i915, port);
>
> - ret &= check_phy_reg(dev_priv, port, CNL_PORT_CL1CM_DW5,
> + ret &= check_phy_reg(i915, port, CNL_PORT_CL1CM_DW5,
> CL_POWER_DOWN_ENABLE, CL_POWER_DOWN_ENABLE);
>
> return ret;
> }
>
> -static void cnl_combo_phys_init(struct drm_i915_private *dev_priv)
> +static void cnl_combo_phys_init(struct drm_i915_private *i915)
> {
> u32 val;
>
> @@ -157,7 +157,7 @@ static void cnl_combo_phys_init(struct drm_i915_private *dev_priv)
> I915_WRITE(CHICKEN_MISC_2, val);
>
> /* Dummy PORT_A to get the correct CNL register from the ICL macro */
> - cnl_set_procmon_ref_values(dev_priv, PORT_A);
> + cnl_set_procmon_ref_values(i915, PORT_A);
>
> val = I915_READ(CNL_PORT_COMP_DW0);
> val |= COMP_INIT;
> @@ -168,11 +168,11 @@ static void cnl_combo_phys_init(struct drm_i915_private *dev_priv)
> I915_WRITE(CNL_PORT_CL1CM_DW5, val);
> }
>
> -static void cnl_combo_phys_uninit(struct drm_i915_private *dev_priv)
> +static void cnl_combo_phys_uninit(struct drm_i915_private *i915)
> {
> u32 val;
>
> - if (!cnl_combo_phy_verify_state(dev_priv))
> + if (!cnl_combo_phy_verify_state(i915))
> DRM_WARN("Combo PHY HW state changed unexpectedly.\n");
>
> val = I915_READ(CHICKEN_MISC_2);
> @@ -180,7 +180,7 @@ static void cnl_combo_phys_uninit(struct drm_i915_private *dev_priv)
> I915_WRITE(CHICKEN_MISC_2, val);
> }
>
> -static bool icl_combo_phy_enabled(struct drm_i915_private *dev_priv,
> +static bool icl_combo_phy_enabled(struct drm_i915_private *i915,
> enum port port)
> {
> return !(I915_READ(ICL_PHY_MISC(port)) &
> @@ -188,27 +188,27 @@ static bool icl_combo_phy_enabled(struct drm_i915_private *dev_priv,
> (I915_READ(ICL_PORT_COMP_DW0(port)) & COMP_INIT);
> }
>
> -static bool icl_combo_phy_verify_state(struct drm_i915_private *dev_priv,
> +static bool icl_combo_phy_verify_state(struct drm_i915_private *i915,
> enum port port)
> {
> bool ret;
>
> - if (!icl_combo_phy_enabled(dev_priv, port))
> + if (!icl_combo_phy_enabled(i915, port))
> return false;
>
> - ret = cnl_verify_procmon_ref_values(dev_priv, port);
> + ret = cnl_verify_procmon_ref_values(i915, port);
>
> if (port == PORT_A)
> - ret &= check_phy_reg(dev_priv, port, ICL_PORT_COMP_DW8(port),
> + ret &= check_phy_reg(i915, port, ICL_PORT_COMP_DW8(port),
> IREFGEN, IREFGEN);
>
> - ret &= check_phy_reg(dev_priv, port, ICL_PORT_CL_DW5(port),
> + ret &= check_phy_reg(i915, port, ICL_PORT_CL_DW5(port),
> CL_POWER_DOWN_ENABLE, CL_POWER_DOWN_ENABLE);
>
> return ret;
> }
>
> -void intel_combo_phy_power_up_lanes(struct drm_i915_private *dev_priv,
> +void intel_combo_phy_power_up_lanes(struct drm_i915_private *i915,
> enum port port, bool is_dsi,
> int lane_count, bool lane_reversal)
> {
> @@ -260,14 +260,14 @@ void intel_combo_phy_power_up_lanes(struct drm_i915_private *dev_priv,
> I915_WRITE(ICL_PORT_CL_DW10(port), val);
> }
>
> -static void icl_combo_phys_init(struct drm_i915_private *dev_priv)
> +static void icl_combo_phys_init(struct drm_i915_private *i915)
> {
> enum port port;
>
> - for_each_combo_port(dev_priv, port) {
> + for_each_combo_port(i915, port) {
> u32 val;
>
> - if (icl_combo_phy_verify_state(dev_priv, port)) {
> + if (icl_combo_phy_verify_state(i915, port)) {
> DRM_DEBUG_DRIVER("Port %c combo PHY already enabled, won't reprogram it.\n",
> port_name(port));
> continue;
> @@ -277,7 +277,7 @@ static void icl_combo_phys_init(struct drm_i915_private *dev_priv)
> val &= ~ICL_PHY_MISC_DE_IO_COMP_PWR_DOWN;
> I915_WRITE(ICL_PHY_MISC(port), val);
>
> - cnl_set_procmon_ref_values(dev_priv, port);
> + cnl_set_procmon_ref_values(i915, port);
>
> if (port == PORT_A) {
> val = I915_READ(ICL_PORT_COMP_DW8(port));
> @@ -295,15 +295,15 @@ static void icl_combo_phys_init(struct drm_i915_private *dev_priv)
> }
> }
>
> -static void icl_combo_phys_uninit(struct drm_i915_private *dev_priv)
> +static void icl_combo_phys_uninit(struct drm_i915_private *i915)
> {
> enum port port;
>
> - for_each_combo_port_reverse(dev_priv, port) {
> + for_each_combo_port_reverse(i915, port) {
> u32 val;
>
> if (port == PORT_A &&
> - !icl_combo_phy_verify_state(dev_priv, port))
> + !icl_combo_phy_verify_state(i915, port))
> DRM_WARN("Port %c combo PHY HW state changed unexpectedly\n",
> port_name(port));
>
> diff --git a/drivers/gpu/drm/i915/intel_combo_phy.h b/drivers/gpu/drm/i915/intel_combo_phy.h
> index e6e195a83b19..a08286cfcecf 100644
> --- a/drivers/gpu/drm/i915/intel_combo_phy.h
> +++ b/drivers/gpu/drm/i915/intel_combo_phy.h
> @@ -11,9 +11,9 @@
>
> struct drm_i915_private;
>
> -void intel_combo_phy_init(struct drm_i915_private *dev_priv);
> -void intel_combo_phy_uninit(struct drm_i915_private *dev_priv);
> -void intel_combo_phy_power_up_lanes(struct drm_i915_private *dev_priv,
> +void intel_combo_phy_init(struct drm_i915_private *i915);
> +void intel_combo_phy_uninit(struct drm_i915_private *i915);
> +void intel_combo_phy_power_up_lanes(struct drm_i915_private *i915,
> enum port port, bool is_dsi,
> int lane_count, bool lane_reversal);
>
> diff --git a/drivers/gpu/drm/i915/intel_connector.c b/drivers/gpu/drm/i915/intel_connector.c
> index 073b6c3ab7cc..2912804d5a0d 100644
> --- a/drivers/gpu/drm/i915/intel_connector.c
> +++ b/drivers/gpu/drm/i915/intel_connector.c
> @@ -219,10 +219,10 @@ void
> intel_attach_force_audio_property(struct drm_connector *connector)
> {
> struct drm_device *dev = connector->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct drm_property *prop;
>
> - prop = dev_priv->force_audio_property;
> + prop = i915->force_audio_property;
> if (prop == NULL) {
> prop = drm_property_create_enum(dev, 0,
> "audio",
> @@ -231,7 +231,7 @@ intel_attach_force_audio_property(struct drm_connector *connector)
> if (prop == NULL)
> return;
>
> - dev_priv->force_audio_property = prop;
> + i915->force_audio_property = prop;
> }
> drm_object_attach_property(&connector->base, prop, 0);
> }
> @@ -246,10 +246,10 @@ void
> intel_attach_broadcast_rgb_property(struct drm_connector *connector)
> {
> struct drm_device *dev = connector->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct drm_property *prop;
>
> - prop = dev_priv->broadcast_rgb_property;
> + prop = i915->broadcast_rgb_property;
> if (prop == NULL) {
> prop = drm_property_create_enum(dev, DRM_MODE_PROP_ENUM,
> "Broadcast RGB",
> @@ -258,7 +258,7 @@ intel_attach_broadcast_rgb_property(struct drm_connector *connector)
> if (prop == NULL)
> return;
>
> - dev_priv->broadcast_rgb_property = prop;
> + i915->broadcast_rgb_property = prop;
> }
>
> drm_object_attach_property(&connector->base, prop, 0);
> diff --git a/drivers/gpu/drm/i915/intel_crt.c b/drivers/gpu/drm/i915/intel_crt.c
> index 3fcf2f84bcce..2a794e0bf457 100644
> --- a/drivers/gpu/drm/i915/intel_crt.c
> +++ b/drivers/gpu/drm/i915/intel_crt.c
> @@ -70,7 +70,7 @@ static struct intel_crt *intel_attached_crt(struct drm_connector *connector)
> return intel_encoder_to_crt(intel_attached_encoder(connector));
> }
>
> -bool intel_crt_port_enabled(struct drm_i915_private *dev_priv,
> +bool intel_crt_port_enabled(struct drm_i915_private *i915,
> i915_reg_t adpa_reg, enum pipe *pipe)
> {
> u32 val;
> @@ -78,7 +78,7 @@ bool intel_crt_port_enabled(struct drm_i915_private *dev_priv,
> val = I915_READ(adpa_reg);
>
> /* asserts want to know the pipe even if the port is disabled */
> - if (HAS_PCH_CPT(dev_priv))
> + if (HAS_PCH_CPT(i915))
> *pipe = (val & ADPA_PIPE_SEL_MASK_CPT) >> ADPA_PIPE_SEL_SHIFT_CPT;
> else
> *pipe = (val & ADPA_PIPE_SEL_MASK) >> ADPA_PIPE_SEL_SHIFT;
> @@ -89,26 +89,26 @@ bool intel_crt_port_enabled(struct drm_i915_private *dev_priv,
> static bool intel_crt_get_hw_state(struct intel_encoder *encoder,
> enum pipe *pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crt *crt = intel_encoder_to_crt(encoder);
> intel_wakeref_t wakeref;
> bool ret;
>
> - wakeref = intel_display_power_get_if_enabled(dev_priv,
> + wakeref = intel_display_power_get_if_enabled(i915,
> encoder->power_domain);
> if (!wakeref)
> return false;
>
> - ret = intel_crt_port_enabled(dev_priv, crt->adpa_reg, pipe);
> + ret = intel_crt_port_enabled(i915, crt->adpa_reg, pipe);
>
> - intel_display_power_put(dev_priv, encoder->power_domain, wakeref);
> + intel_display_power_put(i915, encoder->power_domain, wakeref);
>
> return ret;
> }
>
> static unsigned int intel_crt_get_flags(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crt *crt = intel_encoder_to_crt(encoder);
> u32 tmp, flags = 0;
>
> @@ -140,7 +140,7 @@ static void intel_crt_get_config(struct intel_encoder *encoder,
> static void hsw_crt_get_config(struct intel_encoder *encoder,
> struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
>
> intel_ddi_get_config(encoder, pipe_config);
>
> @@ -150,7 +150,7 @@ static void hsw_crt_get_config(struct intel_encoder *encoder,
> DRM_MODE_FLAG_NVSYNC);
> pipe_config->base.adjusted_mode.flags |= intel_crt_get_flags(encoder);
>
> - pipe_config->base.adjusted_mode.crtc_clock = lpt_get_iclkip(dev_priv);
> + pipe_config->base.adjusted_mode.crtc_clock = lpt_get_iclkip(i915);
> }
>
> /* Note: The caller is required to filter out dpms modes not supported by the
> @@ -159,13 +159,13 @@ static void intel_crt_set_dpms(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state,
> int mode)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crt *crt = intel_encoder_to_crt(encoder);
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> const struct drm_display_mode *adjusted_mode = &crtc_state->base.adjusted_mode;
> u32 adpa;
>
> - if (INTEL_GEN(dev_priv) >= 5)
> + if (INTEL_GEN(i915) >= 5)
> adpa = ADPA_HOTPLUG_BITS;
> else
> adpa = 0;
> @@ -176,14 +176,14 @@ static void intel_crt_set_dpms(struct intel_encoder *encoder,
> adpa |= ADPA_VSYNC_ACTIVE_HIGH;
>
> /* For CPT allow 3 pipe config, for others just use A or B */
> - if (HAS_PCH_LPT(dev_priv))
> + if (HAS_PCH_LPT(i915))
> ; /* Those bits don't exist here */
> - else if (HAS_PCH_CPT(dev_priv))
> + else if (HAS_PCH_CPT(i915))
> adpa |= ADPA_PIPE_SEL_CPT(crtc->pipe);
> else
> adpa |= ADPA_PIPE_SEL(crtc->pipe);
>
> - if (!HAS_PCH_SPLIT(dev_priv))
> + if (!HAS_PCH_SPLIT(i915))
> I915_WRITE(BCLRPAT(crtc->pipe), 0);
>
> switch (mode) {
> @@ -228,57 +228,57 @@ static void hsw_disable_crt(struct intel_encoder *encoder,
> const struct intel_crtc_state *old_crtc_state,
> const struct drm_connector_state *old_conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
>
> WARN_ON(!old_crtc_state->has_pch_encoder);
>
> - intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, false);
> + intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, false);
> }
>
> static void hsw_post_disable_crt(struct intel_encoder *encoder,
> const struct intel_crtc_state *old_crtc_state,
> const struct drm_connector_state *old_conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
>
> intel_ddi_disable_pipe_clock(old_crtc_state);
>
> pch_post_disable_crt(encoder, old_crtc_state, old_conn_state);
>
> - lpt_disable_pch_transcoder(dev_priv);
> - lpt_disable_iclkip(dev_priv);
> + lpt_disable_pch_transcoder(i915);
> + lpt_disable_iclkip(i915);
>
> intel_ddi_fdi_post_disable(encoder, old_crtc_state, old_conn_state);
>
> WARN_ON(!old_crtc_state->has_pch_encoder);
>
> - intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, true);
> + intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, true);
> }
>
> static void hsw_pre_pll_enable_crt(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state,
> const struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
>
> WARN_ON(!crtc_state->has_pch_encoder);
>
> - intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, false);
> + intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, false);
> }
>
> static void hsw_pre_enable_crt(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state,
> const struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> enum pipe pipe = crtc->pipe;
>
> WARN_ON(!crtc_state->has_pch_encoder);
>
> - intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, false);
> + intel_set_cpu_fifo_underrun_reporting(i915, pipe, false);
>
> - dev_priv->display.fdi_link_train(crtc, crtc_state);
> + i915->display.fdi_link_train(crtc, crtc_state);
>
> intel_ddi_enable_pipe_clock(crtc_state);
> }
> @@ -287,7 +287,7 @@ static void hsw_enable_crt(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state,
> const struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> enum pipe pipe = crtc->pipe;
>
> @@ -295,10 +295,10 @@ static void hsw_enable_crt(struct intel_encoder *encoder,
>
> intel_crt_set_dpms(encoder, crtc_state, DRM_MODE_DPMS_ON);
>
> - intel_wait_for_vblank(dev_priv, pipe);
> - intel_wait_for_vblank(dev_priv, pipe);
> - intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, true);
> - intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, true);
> + intel_wait_for_vblank(i915, pipe);
> + intel_wait_for_vblank(i915, pipe);
> + intel_set_cpu_fifo_underrun_reporting(i915, pipe, true);
> + intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, true);
> }
>
> static void intel_enable_crt(struct intel_encoder *encoder,
> @@ -313,8 +313,8 @@ intel_crt_mode_valid(struct drm_connector *connector,
> struct drm_display_mode *mode)
> {
> struct drm_device *dev = connector->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - int max_dotclk = dev_priv->max_dotclk_freq;
> + struct drm_i915_private *i915 = to_i915(dev);
> + int max_dotclk = i915->max_dotclk_freq;
> int max_clock;
>
> if (mode->flags & DRM_MODE_FLAG_DBLSCAN)
> @@ -323,15 +323,15 @@ intel_crt_mode_valid(struct drm_connector *connector,
> if (mode->clock < 25000)
> return MODE_CLOCK_LOW;
>
> - if (HAS_PCH_LPT(dev_priv))
> + if (HAS_PCH_LPT(i915))
> max_clock = 180000;
> - else if (IS_VALLEYVIEW(dev_priv))
> + else if (IS_VALLEYVIEW(i915))
> /*
> * 270 MHz due to current DPLL limits,
> * DAC limit supposedly 355 MHz.
> */
> max_clock = 270000;
> - else if (IS_GEN_RANGE(dev_priv, 3, 4))
> + else if (IS_GEN_RANGE(i915, 3, 4))
> max_clock = 400000;
> else
> max_clock = 350000;
> @@ -342,7 +342,7 @@ intel_crt_mode_valid(struct drm_connector *connector,
> return MODE_CLOCK_HIGH;
>
> /* The FDI receiver on LPT only supports 8bpc and only has 2 lanes. */
> - if (HAS_PCH_LPT(dev_priv) &&
> + if (HAS_PCH_LPT(i915) &&
> (ironlake_get_lanes_required(mode->clock, 270000, 24) > 2))
> return MODE_CLOCK_HIGH;
>
> @@ -388,7 +388,7 @@ static int hsw_crt_compute_config(struct intel_encoder *encoder,
> struct intel_crtc_state *pipe_config,
> struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct drm_display_mode *adjusted_mode =
> &pipe_config->base.adjusted_mode;
>
> @@ -404,7 +404,7 @@ static int hsw_crt_compute_config(struct intel_encoder *encoder,
> pipe_config->output_format = INTEL_OUTPUT_FORMAT_RGB;
>
> /* LPT FDI RX only supports 8bpc. */
> - if (HAS_PCH_LPT(dev_priv)) {
> + if (HAS_PCH_LPT(i915)) {
> if (pipe_config->bw_constrained && pipe_config->pipe_bpp < 24) {
> DRM_DEBUG_KMS("LPT only supports 24bpp\n");
> return -EINVAL;
> @@ -423,13 +423,13 @@ static bool intel_ironlake_crt_detect_hotplug(struct drm_connector *connector)
> {
> struct drm_device *dev = connector->dev;
> struct intel_crt *crt = intel_attached_crt(connector);
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> u32 adpa;
> bool ret;
>
> /* The first time through, trigger an explicit detection cycle */
> if (crt->force_hotplug_required) {
> - bool turn_off_dac = HAS_PCH_SPLIT(dev_priv);
> + bool turn_off_dac = HAS_PCH_SPLIT(i915);
> u32 save_adpa;
>
> crt->force_hotplug_required = 0;
> @@ -443,7 +443,7 @@ static bool intel_ironlake_crt_detect_hotplug(struct drm_connector *connector)
>
> I915_WRITE(crt->adpa_reg, adpa);
>
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> crt->adpa_reg,
> ADPA_CRT_HOTPLUG_FORCE_TRIGGER, 0,
> 1000))
> @@ -470,7 +470,7 @@ static bool valleyview_crt_detect_hotplug(struct drm_connector *connector)
> {
> struct drm_device *dev = connector->dev;
> struct intel_crt *crt = intel_attached_crt(connector);
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> bool reenable_hpd;
> u32 adpa;
> bool ret;
> @@ -488,7 +488,7 @@ static bool valleyview_crt_detect_hotplug(struct drm_connector *connector)
> *
> * Just disable HPD interrupts here to prevent this
> */
> - reenable_hpd = intel_hpd_disable(dev_priv, crt->base.hpd_pin);
> + reenable_hpd = intel_hpd_disable(i915, crt->base.hpd_pin);
>
> save_adpa = adpa = I915_READ(crt->adpa_reg);
> DRM_DEBUG_KMS("trigger hotplug detect cycle: adpa=0x%x\n", adpa);
> @@ -497,7 +497,7 @@ static bool valleyview_crt_detect_hotplug(struct drm_connector *connector)
>
> I915_WRITE(crt->adpa_reg, adpa);
>
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> crt->adpa_reg,
> ADPA_CRT_HOTPLUG_FORCE_TRIGGER, 0,
> 1000)) {
> @@ -515,7 +515,7 @@ static bool valleyview_crt_detect_hotplug(struct drm_connector *connector)
> DRM_DEBUG_KMS("valleyview hotplug adpa=0x%x, result %d\n", adpa, ret);
>
> if (reenable_hpd)
> - intel_hpd_enable(dev_priv, crt->base.hpd_pin);
> + intel_hpd_enable(i915, crt->base.hpd_pin);
>
> return ret;
> }
> @@ -523,15 +523,15 @@ static bool valleyview_crt_detect_hotplug(struct drm_connector *connector)
> static bool intel_crt_detect_hotplug(struct drm_connector *connector)
> {
> struct drm_device *dev = connector->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> u32 stat;
> bool ret = false;
> int i, tries = 0;
>
> - if (HAS_PCH_SPLIT(dev_priv))
> + if (HAS_PCH_SPLIT(i915))
> return intel_ironlake_crt_detect_hotplug(connector);
>
> - if (IS_VALLEYVIEW(dev_priv))
> + if (IS_VALLEYVIEW(i915))
> return valleyview_crt_detect_hotplug(connector);
>
> /*
> @@ -539,18 +539,18 @@ static bool intel_crt_detect_hotplug(struct drm_connector *connector)
> * to get a reliable result.
> */
>
> - if (IS_G45(dev_priv))
> + if (IS_G45(i915))
> tries = 2;
> else
> tries = 1;
>
> for (i = 0; i < tries ; i++) {
> /* turn on the FORCE_DETECT */
> - i915_hotplug_interrupt_update(dev_priv,
> + i915_hotplug_interrupt_update(i915,
> CRT_HOTPLUG_FORCE_DETECT,
> CRT_HOTPLUG_FORCE_DETECT);
> /* wait for FORCE_DETECT to go off */
> - if (intel_wait_for_register(&dev_priv->uncore, PORT_HOTPLUG_EN,
> + if (intel_wait_for_register(&i915->uncore, PORT_HOTPLUG_EN,
> CRT_HOTPLUG_FORCE_DETECT, 0,
> 1000))
> DRM_DEBUG_KMS("timed out waiting for FORCE_DETECT to go off");
> @@ -563,7 +563,7 @@ static bool intel_crt_detect_hotplug(struct drm_connector *connector)
> /* clear the interrupt we just generated, if any */
> I915_WRITE(PORT_HOTPLUG_STAT, CRT_HOTPLUG_INT_STATUS);
>
> - i915_hotplug_interrupt_update(dev_priv, CRT_HOTPLUG_FORCE_DETECT, 0);
> + i915_hotplug_interrupt_update(i915, CRT_HOTPLUG_FORCE_DETECT, 0);
>
> return ret;
> }
> @@ -605,14 +605,14 @@ static int intel_crt_ddc_get_modes(struct drm_connector *connector,
> static bool intel_crt_detect_ddc(struct drm_connector *connector)
> {
> struct intel_crt *crt = intel_attached_crt(connector);
> - struct drm_i915_private *dev_priv = to_i915(crt->base.base.dev);
> + struct drm_i915_private *i915 = to_i915(crt->base.base.dev);
> struct edid *edid;
> struct i2c_adapter *i2c;
> bool ret = false;
>
> BUG_ON(crt->base.type != INTEL_OUTPUT_ANALOG);
>
> - i2c = intel_gmbus_get_adapter(dev_priv, dev_priv->vbt.crt_ddc_pin);
> + i2c = intel_gmbus_get_adapter(i915, i915->vbt.crt_ddc_pin);
> edid = intel_crt_get_edid(connector, i2c);
>
> if (edid) {
> @@ -642,8 +642,8 @@ static enum drm_connector_status
> intel_crt_load_detect(struct intel_crt *crt, u32 pipe)
> {
> struct drm_device *dev = crt->base.base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct intel_uncore *uncore = &i915->uncore;
> u32 save_bclrpat;
> u32 save_vtotal;
> u32 vtotal, vactive;
> @@ -677,7 +677,7 @@ intel_crt_load_detect(struct intel_crt *crt, u32 pipe)
> /* Set the border color to purple. */
> intel_uncore_write(uncore, bclrpat_reg, 0x500050);
>
> - if (!IS_GEN(dev_priv, 2)) {
> + if (!IS_GEN(i915, 2)) {
> u32 pipeconf = intel_uncore_read(uncore, pipeconf_reg);
> intel_uncore_write(uncore,
> pipeconf_reg,
> @@ -685,7 +685,7 @@ intel_crt_load_detect(struct intel_crt *crt, u32 pipe)
> intel_uncore_posting_read(uncore, pipeconf_reg);
> /* Wait for next Vblank to substitue
> * border color for Color info */
> - intel_wait_for_vblank(dev_priv, pipe);
> + intel_wait_for_vblank(i915, pipe);
> st00 = intel_uncore_read8(uncore, _VGA_MSR_WRITE);
> status = ((st00 & (1 << 4)) != 0) ?
> connector_status_connected :
> @@ -789,7 +789,7 @@ intel_crt_detect(struct drm_connector *connector,
> struct drm_modeset_acquire_ctx *ctx,
> bool force)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->dev);
> + struct drm_i915_private *i915 = to_i915(connector->dev);
> struct intel_crt *crt = intel_attached_crt(connector);
> struct intel_encoder *intel_encoder = &crt->base;
> intel_wakeref_t wakeref;
> @@ -801,7 +801,7 @@ intel_crt_detect(struct drm_connector *connector,
> force);
>
> if (i915_modparams.load_detect_test) {
> - wakeref = intel_display_power_get(dev_priv,
> + wakeref = intel_display_power_get(i915,
> intel_encoder->power_domain);
> goto load_detect;
> }
> @@ -810,10 +810,10 @@ intel_crt_detect(struct drm_connector *connector,
> if (dmi_check_system(intel_spurious_crt_detect))
> return connector_status_disconnected;
>
> - wakeref = intel_display_power_get(dev_priv,
> + wakeref = intel_display_power_get(i915,
> intel_encoder->power_domain);
>
> - if (I915_HAS_HOTPLUG(dev_priv)) {
> + if (I915_HAS_HOTPLUG(i915)) {
> /* We can not rely on the HPD pin always being correctly wired
> * up, for example many KVM do not pass it through, and so
> * only trust an assertion that the monitor is connected.
> @@ -835,7 +835,7 @@ intel_crt_detect(struct drm_connector *connector,
> * broken monitor (without edid) to work behind a broken kvm (that fails
> * to have the right resistors for HP detection) needs to fix this up.
> * For now just bail out. */
> - if (I915_HAS_HOTPLUG(dev_priv)) {
> + if (I915_HAS_HOTPLUG(i915)) {
> status = connector_status_disconnected;
> goto out;
> }
> @@ -851,7 +851,7 @@ intel_crt_detect(struct drm_connector *connector,
> if (ret > 0) {
> if (intel_crt_detect_ddc(connector))
> status = connector_status_connected;
> - else if (INTEL_GEN(dev_priv) < 4)
> + else if (INTEL_GEN(i915) < 4)
> status = intel_crt_load_detect(crt,
> to_intel_crtc(connector->state->crtc)->pipe);
> else if (i915_modparams.load_detect_test)
> @@ -866,44 +866,44 @@ intel_crt_detect(struct drm_connector *connector,
> }
>
> out:
> - intel_display_power_put(dev_priv, intel_encoder->power_domain, wakeref);
> + intel_display_power_put(i915, intel_encoder->power_domain, wakeref);
> return status;
> }
>
> static int intel_crt_get_modes(struct drm_connector *connector)
> {
> struct drm_device *dev = connector->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_crt *crt = intel_attached_crt(connector);
> struct intel_encoder *intel_encoder = &crt->base;
> intel_wakeref_t wakeref;
> struct i2c_adapter *i2c;
> int ret;
>
> - wakeref = intel_display_power_get(dev_priv,
> + wakeref = intel_display_power_get(i915,
> intel_encoder->power_domain);
>
> - i2c = intel_gmbus_get_adapter(dev_priv, dev_priv->vbt.crt_ddc_pin);
> + i2c = intel_gmbus_get_adapter(i915, i915->vbt.crt_ddc_pin);
> ret = intel_crt_ddc_get_modes(connector, i2c);
> - if (ret || !IS_G4X(dev_priv))
> + if (ret || !IS_G4X(i915))
> goto out;
>
> /* Try to probe digital port for output in DVI-I -> VGA mode. */
> - i2c = intel_gmbus_get_adapter(dev_priv, GMBUS_PIN_DPB);
> + i2c = intel_gmbus_get_adapter(i915, GMBUS_PIN_DPB);
> ret = intel_crt_ddc_get_modes(connector, i2c);
>
> out:
> - intel_display_power_put(dev_priv, intel_encoder->power_domain, wakeref);
> + intel_display_power_put(i915, intel_encoder->power_domain, wakeref);
>
> return ret;
> }
>
> void intel_crt_reset(struct drm_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->dev);
> + struct drm_i915_private *i915 = to_i915(encoder->dev);
> struct intel_crt *crt = intel_encoder_to_crt(to_intel_encoder(encoder));
>
> - if (INTEL_GEN(dev_priv) >= 5) {
> + if (INTEL_GEN(i915) >= 5) {
> u32 adpa;
>
> adpa = I915_READ(crt->adpa_reg);
> @@ -942,7 +942,7 @@ static const struct drm_encoder_funcs intel_crt_enc_funcs = {
> .destroy = intel_encoder_destroy,
> };
>
> -void intel_crt_init(struct drm_i915_private *dev_priv)
> +void intel_crt_init(struct drm_i915_private *i915)
> {
> struct drm_connector *connector;
> struct intel_crt *crt;
> @@ -950,9 +950,9 @@ void intel_crt_init(struct drm_i915_private *dev_priv)
> i915_reg_t adpa_reg;
> u32 adpa;
>
> - if (HAS_PCH_SPLIT(dev_priv))
> + if (HAS_PCH_SPLIT(i915))
> adpa_reg = PCH_ADPA;
> - else if (IS_VALLEYVIEW(dev_priv))
> + else if (IS_VALLEYVIEW(i915))
> adpa_reg = VLV_ADPA;
> else
> adpa_reg = ADPA;
> @@ -986,22 +986,22 @@ void intel_crt_init(struct drm_i915_private *dev_priv)
>
> connector = &intel_connector->base;
> crt->connector = intel_connector;
> - drm_connector_init(&dev_priv->drm, &intel_connector->base,
> + drm_connector_init(&i915->drm, &intel_connector->base,
> &intel_crt_connector_funcs, DRM_MODE_CONNECTOR_VGA);
>
> - drm_encoder_init(&dev_priv->drm, &crt->base.base, &intel_crt_enc_funcs,
> + drm_encoder_init(&i915->drm, &crt->base.base, &intel_crt_enc_funcs,
> DRM_MODE_ENCODER_DAC, "CRT");
>
> intel_connector_attach_encoder(intel_connector, &crt->base);
>
> crt->base.type = INTEL_OUTPUT_ANALOG;
> crt->base.cloneable = (1 << INTEL_OUTPUT_DVO) | (1 << INTEL_OUTPUT_HDMI);
> - if (IS_I830(dev_priv))
> + if (IS_I830(i915))
> crt->base.crtc_mask = (1 << 0);
> else
> crt->base.crtc_mask = (1 << 0) | (1 << 1) | (1 << 2);
>
> - if (IS_GEN(dev_priv, 2))
> + if (IS_GEN(i915, 2))
> connector->interlace_allowed = 0;
> else
> connector->interlace_allowed = 1;
> @@ -1011,13 +1011,13 @@ void intel_crt_init(struct drm_i915_private *dev_priv)
>
> crt->base.power_domain = POWER_DOMAIN_PORT_CRT;
>
> - if (I915_HAS_HOTPLUG(dev_priv) &&
> + if (I915_HAS_HOTPLUG(i915) &&
> !dmi_check_system(intel_spurious_crt_detect)) {
> crt->base.hpd_pin = HPD_CRT;
> crt->base.hotplug = intel_encoder_hotplug;
> }
>
> - if (HAS_DDI(dev_priv)) {
> + if (HAS_DDI(i915)) {
> crt->base.port = PORT_E;
> crt->base.get_config = hsw_crt_get_config;
> crt->base.get_hw_state = intel_ddi_get_hw_state;
> @@ -1028,7 +1028,7 @@ void intel_crt_init(struct drm_i915_private *dev_priv)
> crt->base.disable = hsw_disable_crt;
> crt->base.post_disable = hsw_post_disable_crt;
> } else {
> - if (HAS_PCH_SPLIT(dev_priv)) {
> + if (HAS_PCH_SPLIT(i915)) {
> crt->base.compute_config = pch_crt_compute_config;
> crt->base.disable = pch_disable_crt;
> crt->base.post_disable = pch_post_disable_crt;
> @@ -1045,7 +1045,7 @@ void intel_crt_init(struct drm_i915_private *dev_priv)
>
> drm_connector_helper_add(connector, &intel_crt_connector_helper_funcs);
>
> - if (!I915_HAS_HOTPLUG(dev_priv))
> + if (!I915_HAS_HOTPLUG(i915))
> intel_connector->polled = DRM_CONNECTOR_POLL_CONNECT;
>
> /*
> @@ -1058,11 +1058,11 @@ void intel_crt_init(struct drm_i915_private *dev_priv)
> * polarity and link reversal bits or not, instead of relying on the
> * BIOS.
> */
> - if (HAS_PCH_LPT(dev_priv)) {
> + if (HAS_PCH_LPT(i915)) {
> u32 fdi_config = FDI_RX_POLARITY_REVERSED_LPT |
> FDI_RX_LINK_REVERSAL_OVERRIDE;
>
> - dev_priv->fdi_rx_config = I915_READ(FDI_RX_CTL(PIPE_A)) & fdi_config;
> + i915->fdi_rx_config = I915_READ(FDI_RX_CTL(PIPE_A)) & fdi_config;
> }
>
> intel_crt_reset(&crt->base.base);
> diff --git a/drivers/gpu/drm/i915/intel_crt.h b/drivers/gpu/drm/i915/intel_crt.h
> index 1b3fba359efc..f771fbcf8b27 100644
> --- a/drivers/gpu/drm/i915/intel_crt.h
> +++ b/drivers/gpu/drm/i915/intel_crt.h
> @@ -13,9 +13,9 @@ struct drm_encoder;
> struct drm_i915_private;
> struct drm_i915_private;
>
> -bool intel_crt_port_enabled(struct drm_i915_private *dev_priv,
> +bool intel_crt_port_enabled(struct drm_i915_private *i915,
> i915_reg_t adpa_reg, enum pipe *pipe);
> -void intel_crt_init(struct drm_i915_private *dev_priv);
> +void intel_crt_init(struct drm_i915_private *i915);
> void intel_crt_reset(struct drm_encoder *encoder);
>
> #endif /* __INTEL_CRT_H__ */
> diff --git a/drivers/gpu/drm/i915/intel_csr.c b/drivers/gpu/drm/i915/intel_csr.c
> index bf0eebd385b9..29ea3cb6bbaf 100644
> --- a/drivers/gpu/drm/i915/intel_csr.c
> +++ b/drivers/gpu/drm/i915/intel_csr.c
> @@ -206,18 +206,18 @@ static const struct stepping_info icl_stepping_info[] = {
> static const struct stepping_info no_stepping_info = { '*', '*' };
>
> static const struct stepping_info *
> -intel_get_stepping_info(struct drm_i915_private *dev_priv)
> +intel_get_stepping_info(struct drm_i915_private *i915)
> {
> const struct stepping_info *si;
> unsigned int size;
>
> - if (IS_ICELAKE(dev_priv)) {
> + if (IS_ICELAKE(i915)) {
> size = ARRAY_SIZE(icl_stepping_info);
> si = icl_stepping_info;
> - } else if (IS_SKYLAKE(dev_priv)) {
> + } else if (IS_SKYLAKE(i915)) {
> size = ARRAY_SIZE(skl_stepping_info);
> si = skl_stepping_info;
> - } else if (IS_BROXTON(dev_priv)) {
> + } else if (IS_BROXTON(i915)) {
> size = ARRAY_SIZE(bxt_stepping_info);
> si = bxt_stepping_info;
> } else {
> @@ -225,19 +225,19 @@ intel_get_stepping_info(struct drm_i915_private *dev_priv)
> si = NULL;
> }
>
> - if (INTEL_REVID(dev_priv) < size)
> - return si + INTEL_REVID(dev_priv);
> + if (INTEL_REVID(i915) < size)
> + return si + INTEL_REVID(i915);
>
> return &no_stepping_info;
> }
>
> -static void gen9_set_dc_state_debugmask(struct drm_i915_private *dev_priv)
> +static void gen9_set_dc_state_debugmask(struct drm_i915_private *i915)
> {
> u32 val, mask;
>
> mask = DC_STATE_DEBUG_MASK_MEMORY_UP;
>
> - if (IS_GEN9_LP(dev_priv))
> + if (IS_GEN9_LP(i915))
> mask |= DC_STATE_DEBUG_MASK_CORES;
>
> /* The below bit doesn't need to be cleared ever afterwards */
> @@ -251,29 +251,29 @@ static void gen9_set_dc_state_debugmask(struct drm_i915_private *dev_priv)
>
> /**
> * intel_csr_load_program() - write the firmware from memory to register.
> - * @dev_priv: i915 drm device.
> + * @i915: i915 drm device.
> *
> * CSR firmware is read from a .bin file and kept in internal memory one time.
> * Everytime display comes back from low power state this function is called to
> * copy the firmware from internal memory to registers.
> */
> -void intel_csr_load_program(struct drm_i915_private *dev_priv)
> +void intel_csr_load_program(struct drm_i915_private *i915)
> {
> - u32 *payload = dev_priv->csr.dmc_payload;
> + u32 *payload = i915->csr.dmc_payload;
> u32 i, fw_size;
>
> - if (!HAS_CSR(dev_priv)) {
> + if (!HAS_CSR(i915)) {
> DRM_ERROR("No CSR support available for this platform\n");
> return;
> }
>
> - if (!dev_priv->csr.dmc_payload) {
> + if (!i915->csr.dmc_payload) {
> DRM_ERROR("Tried to program CSR with empty payload\n");
> return;
> }
>
> - fw_size = dev_priv->csr.dmc_fw_size;
> - assert_rpm_wakelock_held(dev_priv);
> + fw_size = i915->csr.dmc_fw_size;
> + assert_rpm_wakelock_held(i915);
>
> preempt_disable();
>
> @@ -282,24 +282,24 @@ void intel_csr_load_program(struct drm_i915_private *dev_priv)
>
> preempt_enable();
>
> - for (i = 0; i < dev_priv->csr.mmio_count; i++) {
> - I915_WRITE(dev_priv->csr.mmioaddr[i],
> - dev_priv->csr.mmiodata[i]);
> + for (i = 0; i < i915->csr.mmio_count; i++) {
> + I915_WRITE(i915->csr.mmioaddr[i],
> + i915->csr.mmiodata[i]);
> }
>
> - dev_priv->csr.dc_state = 0;
> + i915->csr.dc_state = 0;
>
> - gen9_set_dc_state_debugmask(dev_priv);
> + gen9_set_dc_state_debugmask(i915);
> }
>
> -static u32 *parse_csr_fw(struct drm_i915_private *dev_priv,
> +static u32 *parse_csr_fw(struct drm_i915_private *i915,
> const struct firmware *fw)
> {
> struct intel_css_header *css_header;
> struct intel_package_header *package_header;
> struct intel_dmc_header *dmc_header;
> - struct intel_csr *csr = &dev_priv->csr;
> - const struct stepping_info *si = intel_get_stepping_info(dev_priv);
> + struct intel_csr *csr = &i915->csr;
> + const struct stepping_info *si = intel_get_stepping_info(i915);
> u32 dmc_offset = CSR_DEFAULT_FW_OFFSET, readcount = 0, nbytes;
> u32 i;
> u32 *dmc_payload;
> @@ -430,48 +430,48 @@ static u32 *parse_csr_fw(struct drm_i915_private *dev_priv,
> return NULL;
> }
>
> -static void intel_csr_runtime_pm_get(struct drm_i915_private *dev_priv)
> +static void intel_csr_runtime_pm_get(struct drm_i915_private *i915)
> {
> - WARN_ON(dev_priv->csr.wakeref);
> - dev_priv->csr.wakeref =
> - intel_display_power_get(dev_priv, POWER_DOMAIN_INIT);
> + WARN_ON(i915->csr.wakeref);
> + i915->csr.wakeref =
> + intel_display_power_get(i915, POWER_DOMAIN_INIT);
> }
>
> -static void intel_csr_runtime_pm_put(struct drm_i915_private *dev_priv)
> +static void intel_csr_runtime_pm_put(struct drm_i915_private *i915)
> {
> intel_wakeref_t wakeref __maybe_unused =
> - fetch_and_zero(&dev_priv->csr.wakeref);
> + fetch_and_zero(&i915->csr.wakeref);
>
> - intel_display_power_put(dev_priv, POWER_DOMAIN_INIT, wakeref);
> + intel_display_power_put(i915, POWER_DOMAIN_INIT, wakeref);
> }
>
> static void csr_load_work_fn(struct work_struct *work)
> {
> - struct drm_i915_private *dev_priv;
> + struct drm_i915_private *i915;
> struct intel_csr *csr;
> const struct firmware *fw = NULL;
>
> - dev_priv = container_of(work, typeof(*dev_priv), csr.work);
> - csr = &dev_priv->csr;
> + i915 = container_of(work, typeof(*i915), csr.work);
> + csr = &i915->csr;
>
> - request_firmware(&fw, dev_priv->csr.fw_path, &dev_priv->drm.pdev->dev);
> + request_firmware(&fw, i915->csr.fw_path, &i915->drm.pdev->dev);
> if (fw)
> - dev_priv->csr.dmc_payload = parse_csr_fw(dev_priv, fw);
> + i915->csr.dmc_payload = parse_csr_fw(i915, fw);
>
> - if (dev_priv->csr.dmc_payload) {
> - intel_csr_load_program(dev_priv);
> - intel_csr_runtime_pm_put(dev_priv);
> + if (i915->csr.dmc_payload) {
> + intel_csr_load_program(i915);
> + intel_csr_runtime_pm_put(i915);
>
> DRM_INFO("Finished loading DMC firmware %s (v%u.%u)\n",
> - dev_priv->csr.fw_path,
> + i915->csr.fw_path,
> CSR_VERSION_MAJOR(csr->version),
> CSR_VERSION_MINOR(csr->version));
> } else {
> - dev_notice(dev_priv->drm.dev,
> + dev_notice(i915->drm.dev,
> "Failed to load DMC firmware %s."
> " Disabling runtime power management.\n",
> csr->fw_path);
> - dev_notice(dev_priv->drm.dev, "DMC firmware homepage: %s",
> + dev_notice(i915->drm.dev, "DMC firmware homepage: %s",
> INTEL_UC_FIRMWARE_URL);
> }
>
> @@ -480,18 +480,18 @@ static void csr_load_work_fn(struct work_struct *work)
>
> /**
> * intel_csr_ucode_init() - initialize the firmware loading.
> - * @dev_priv: i915 drm device.
> + * @i915: i915 drm device.
> *
> * This function is called at the time of loading the display driver to read
> * firmware from a .bin file and copied into a internal memory.
> */
> -void intel_csr_ucode_init(struct drm_i915_private *dev_priv)
> +void intel_csr_ucode_init(struct drm_i915_private *i915)
> {
> - struct intel_csr *csr = &dev_priv->csr;
> + struct intel_csr *csr = &i915->csr;
>
> - INIT_WORK(&dev_priv->csr.work, csr_load_work_fn);
> + INIT_WORK(&i915->csr.work, csr_load_work_fn);
>
> - if (!HAS_CSR(dev_priv))
> + if (!HAS_CSR(i915))
> return;
>
> /*
> @@ -502,32 +502,32 @@ void intel_csr_ucode_init(struct drm_i915_private *dev_priv)
> * suspend as runtime suspend *requires* a working CSR for whatever
> * reason.
> */
> - intel_csr_runtime_pm_get(dev_priv);
> + intel_csr_runtime_pm_get(i915);
>
> - if (INTEL_GEN(dev_priv) >= 12) {
> + if (INTEL_GEN(i915) >= 12) {
> /* Allow to load fw via parameter using the last known size */
> csr->max_fw_size = GEN12_CSR_MAX_FW_SIZE;
> - } else if (IS_GEN(dev_priv, 11)) {
> + } else if (IS_GEN(i915, 11)) {
> csr->fw_path = ICL_CSR_PATH;
> csr->required_version = ICL_CSR_VERSION_REQUIRED;
> csr->max_fw_size = ICL_CSR_MAX_FW_SIZE;
> - } else if (IS_CANNONLAKE(dev_priv)) {
> + } else if (IS_CANNONLAKE(i915)) {
> csr->fw_path = CNL_CSR_PATH;
> csr->required_version = CNL_CSR_VERSION_REQUIRED;
> csr->max_fw_size = CNL_CSR_MAX_FW_SIZE;
> - } else if (IS_GEMINILAKE(dev_priv)) {
> + } else if (IS_GEMINILAKE(i915)) {
> csr->fw_path = GLK_CSR_PATH;
> csr->required_version = GLK_CSR_VERSION_REQUIRED;
> csr->max_fw_size = GLK_CSR_MAX_FW_SIZE;
> - } else if (IS_KABYLAKE(dev_priv) || IS_COFFEELAKE(dev_priv)) {
> + } else if (IS_KABYLAKE(i915) || IS_COFFEELAKE(i915)) {
> csr->fw_path = KBL_CSR_PATH;
> csr->required_version = KBL_CSR_VERSION_REQUIRED;
> csr->max_fw_size = KBL_CSR_MAX_FW_SIZE;
> - } else if (IS_SKYLAKE(dev_priv)) {
> + } else if (IS_SKYLAKE(i915)) {
> csr->fw_path = SKL_CSR_PATH;
> csr->required_version = SKL_CSR_VERSION_REQUIRED;
> csr->max_fw_size = SKL_CSR_MAX_FW_SIZE;
> - } else if (IS_BROXTON(dev_priv)) {
> + } else if (IS_BROXTON(i915)) {
> csr->fw_path = BXT_CSR_PATH;
> csr->required_version = BXT_CSR_VERSION_REQUIRED;
> csr->max_fw_size = BXT_CSR_MAX_FW_SIZE;
> @@ -551,63 +551,63 @@ void intel_csr_ucode_init(struct drm_i915_private *dev_priv)
> }
>
> DRM_DEBUG_KMS("Loading %s\n", csr->fw_path);
> - schedule_work(&dev_priv->csr.work);
> + schedule_work(&i915->csr.work);
> }
>
> /**
> * intel_csr_ucode_suspend() - prepare CSR firmware before system suspend
> - * @dev_priv: i915 drm device
> + * @i915: i915 drm device
> *
> * Prepare the DMC firmware before entering system suspend. This includes
> * flushing pending work items and releasing any resources acquired during
> * init.
> */
> -void intel_csr_ucode_suspend(struct drm_i915_private *dev_priv)
> +void intel_csr_ucode_suspend(struct drm_i915_private *i915)
> {
> - if (!HAS_CSR(dev_priv))
> + if (!HAS_CSR(i915))
> return;
>
> - flush_work(&dev_priv->csr.work);
> + flush_work(&i915->csr.work);
>
> /* Drop the reference held in case DMC isn't loaded. */
> - if (!dev_priv->csr.dmc_payload)
> - intel_csr_runtime_pm_put(dev_priv);
> + if (!i915->csr.dmc_payload)
> + intel_csr_runtime_pm_put(i915);
> }
>
> /**
> * intel_csr_ucode_resume() - init CSR firmware during system resume
> - * @dev_priv: i915 drm device
> + * @i915: i915 drm device
> *
> * Reinitialize the DMC firmware during system resume, reacquiring any
> * resources released in intel_csr_ucode_suspend().
> */
> -void intel_csr_ucode_resume(struct drm_i915_private *dev_priv)
> +void intel_csr_ucode_resume(struct drm_i915_private *i915)
> {
> - if (!HAS_CSR(dev_priv))
> + if (!HAS_CSR(i915))
> return;
>
> /*
> * Reacquire the reference to keep RPM disabled in case DMC isn't
> * loaded.
> */
> - if (!dev_priv->csr.dmc_payload)
> - intel_csr_runtime_pm_get(dev_priv);
> + if (!i915->csr.dmc_payload)
> + intel_csr_runtime_pm_get(i915);
> }
>
> /**
> * intel_csr_ucode_fini() - unload the CSR firmware.
> - * @dev_priv: i915 drm device.
> + * @i915: i915 drm device.
> *
> * Firmmware unloading includes freeing the internal memory and reset the
> * firmware loading status.
> */
> -void intel_csr_ucode_fini(struct drm_i915_private *dev_priv)
> +void intel_csr_ucode_fini(struct drm_i915_private *i915)
> {
> - if (!HAS_CSR(dev_priv))
> + if (!HAS_CSR(i915))
> return;
>
> - intel_csr_ucode_suspend(dev_priv);
> - WARN_ON(dev_priv->csr.wakeref);
> + intel_csr_ucode_suspend(i915);
> + WARN_ON(i915->csr.wakeref);
>
> - kfree(dev_priv->csr.dmc_payload);
> + kfree(i915->csr.dmc_payload);
> }
> diff --git a/drivers/gpu/drm/i915/intel_ddi.c b/drivers/gpu/drm/i915/intel_ddi.c
> index 7925a176f900..eb61731f1abc 100644
> --- a/drivers/gpu/drm/i915/intel_ddi.c
> +++ b/drivers/gpu/drm/i915/intel_ddi.c
> @@ -586,9 +586,9 @@ static const struct icl_mg_phy_ddi_buf_trans icl_mg_phy_ddi_translations[] = {
> };
>
> static const struct ddi_buf_trans *
> -bdw_get_buf_trans_edp(struct drm_i915_private *dev_priv, int *n_entries)
> +bdw_get_buf_trans_edp(struct drm_i915_private *i915, int *n_entries)
> {
> - if (dev_priv->vbt.edp.low_vswing) {
> + if (i915->vbt.edp.low_vswing) {
> *n_entries = ARRAY_SIZE(bdw_ddi_translations_edp);
> return bdw_ddi_translations_edp;
> } else {
> @@ -598,12 +598,12 @@ bdw_get_buf_trans_edp(struct drm_i915_private *dev_priv, int *n_entries)
> }
>
> static const struct ddi_buf_trans *
> -skl_get_buf_trans_dp(struct drm_i915_private *dev_priv, int *n_entries)
> +skl_get_buf_trans_dp(struct drm_i915_private *i915, int *n_entries)
> {
> - if (IS_SKL_ULX(dev_priv)) {
> + if (IS_SKL_ULX(i915)) {
> *n_entries = ARRAY_SIZE(skl_y_ddi_translations_dp);
> return skl_y_ddi_translations_dp;
> - } else if (IS_SKL_ULT(dev_priv)) {
> + } else if (IS_SKL_ULT(i915)) {
> *n_entries = ARRAY_SIZE(skl_u_ddi_translations_dp);
> return skl_u_ddi_translations_dp;
> } else {
> @@ -613,12 +613,12 @@ skl_get_buf_trans_dp(struct drm_i915_private *dev_priv, int *n_entries)
> }
>
> static const struct ddi_buf_trans *
> -kbl_get_buf_trans_dp(struct drm_i915_private *dev_priv, int *n_entries)
> +kbl_get_buf_trans_dp(struct drm_i915_private *i915, int *n_entries)
> {
> - if (IS_KBL_ULX(dev_priv) || IS_CFL_ULX(dev_priv)) {
> + if (IS_KBL_ULX(i915) || IS_CFL_ULX(i915)) {
> *n_entries = ARRAY_SIZE(kbl_y_ddi_translations_dp);
> return kbl_y_ddi_translations_dp;
> - } else if (IS_KBL_ULT(dev_priv) || IS_CFL_ULT(dev_priv)) {
> + } else if (IS_KBL_ULT(i915) || IS_CFL_ULT(i915)) {
> *n_entries = ARRAY_SIZE(kbl_u_ddi_translations_dp);
> return kbl_u_ddi_translations_dp;
> } else {
> @@ -628,15 +628,15 @@ kbl_get_buf_trans_dp(struct drm_i915_private *dev_priv, int *n_entries)
> }
>
> static const struct ddi_buf_trans *
> -skl_get_buf_trans_edp(struct drm_i915_private *dev_priv, int *n_entries)
> +skl_get_buf_trans_edp(struct drm_i915_private *i915, int *n_entries)
> {
> - if (dev_priv->vbt.edp.low_vswing) {
> - if (IS_SKL_ULX(dev_priv) || IS_KBL_ULX(dev_priv) ||
> - IS_CFL_ULX(dev_priv)) {
> + if (i915->vbt.edp.low_vswing) {
> + if (IS_SKL_ULX(i915) || IS_KBL_ULX(i915) ||
> + IS_CFL_ULX(i915)) {
> *n_entries = ARRAY_SIZE(skl_y_ddi_translations_edp);
> return skl_y_ddi_translations_edp;
> - } else if (IS_SKL_ULT(dev_priv) || IS_KBL_ULT(dev_priv) ||
> - IS_CFL_ULT(dev_priv)) {
> + } else if (IS_SKL_ULT(i915) || IS_KBL_ULT(i915) ||
> + IS_CFL_ULT(i915)) {
> *n_entries = ARRAY_SIZE(skl_u_ddi_translations_edp);
> return skl_u_ddi_translations_edp;
> } else {
> @@ -645,17 +645,17 @@ skl_get_buf_trans_edp(struct drm_i915_private *dev_priv, int *n_entries)
> }
> }
>
> - if (IS_KABYLAKE(dev_priv) || IS_COFFEELAKE(dev_priv))
> - return kbl_get_buf_trans_dp(dev_priv, n_entries);
> + if (IS_KABYLAKE(i915) || IS_COFFEELAKE(i915))
> + return kbl_get_buf_trans_dp(i915, n_entries);
> else
> - return skl_get_buf_trans_dp(dev_priv, n_entries);
> + return skl_get_buf_trans_dp(i915, n_entries);
> }
>
> static const struct ddi_buf_trans *
> -skl_get_buf_trans_hdmi(struct drm_i915_private *dev_priv, int *n_entries)
> +skl_get_buf_trans_hdmi(struct drm_i915_private *i915, int *n_entries)
> {
> - if (IS_SKL_ULX(dev_priv) || IS_KBL_ULX(dev_priv) ||
> - IS_CFL_ULX(dev_priv)) {
> + if (IS_SKL_ULX(i915) || IS_KBL_ULX(i915) ||
> + IS_CFL_ULX(i915)) {
> *n_entries = ARRAY_SIZE(skl_y_ddi_translations_hdmi);
> return skl_y_ddi_translations_hdmi;
> } else {
> @@ -674,23 +674,23 @@ static int skl_buf_trans_num_entries(enum port port, int n_entries)
> }
>
> static const struct ddi_buf_trans *
> -intel_ddi_get_buf_trans_dp(struct drm_i915_private *dev_priv,
> +intel_ddi_get_buf_trans_dp(struct drm_i915_private *i915,
> enum port port, int *n_entries)
> {
> - if (IS_KABYLAKE(dev_priv) || IS_COFFEELAKE(dev_priv)) {
> + if (IS_KABYLAKE(i915) || IS_COFFEELAKE(i915)) {
> const struct ddi_buf_trans *ddi_translations =
> - kbl_get_buf_trans_dp(dev_priv, n_entries);
> + kbl_get_buf_trans_dp(i915, n_entries);
> *n_entries = skl_buf_trans_num_entries(port, *n_entries);
> return ddi_translations;
> - } else if (IS_SKYLAKE(dev_priv)) {
> + } else if (IS_SKYLAKE(i915)) {
> const struct ddi_buf_trans *ddi_translations =
> - skl_get_buf_trans_dp(dev_priv, n_entries);
> + skl_get_buf_trans_dp(i915, n_entries);
> *n_entries = skl_buf_trans_num_entries(port, *n_entries);
> return ddi_translations;
> - } else if (IS_BROADWELL(dev_priv)) {
> + } else if (IS_BROADWELL(i915)) {
> *n_entries = ARRAY_SIZE(bdw_ddi_translations_dp);
> return bdw_ddi_translations_dp;
> - } else if (IS_HASWELL(dev_priv)) {
> + } else if (IS_HASWELL(i915)) {
> *n_entries = ARRAY_SIZE(hsw_ddi_translations_dp);
> return hsw_ddi_translations_dp;
> }
> @@ -700,17 +700,17 @@ intel_ddi_get_buf_trans_dp(struct drm_i915_private *dev_priv,
> }
>
> static const struct ddi_buf_trans *
> -intel_ddi_get_buf_trans_edp(struct drm_i915_private *dev_priv,
> +intel_ddi_get_buf_trans_edp(struct drm_i915_private *i915,
> enum port port, int *n_entries)
> {
> - if (IS_GEN9_BC(dev_priv)) {
> + if (IS_GEN9_BC(i915)) {
> const struct ddi_buf_trans *ddi_translations =
> - skl_get_buf_trans_edp(dev_priv, n_entries);
> + skl_get_buf_trans_edp(i915, n_entries);
> *n_entries = skl_buf_trans_num_entries(port, *n_entries);
> return ddi_translations;
> - } else if (IS_BROADWELL(dev_priv)) {
> - return bdw_get_buf_trans_edp(dev_priv, n_entries);
> - } else if (IS_HASWELL(dev_priv)) {
> + } else if (IS_BROADWELL(i915)) {
> + return bdw_get_buf_trans_edp(i915, n_entries);
> + } else if (IS_HASWELL(i915)) {
> *n_entries = ARRAY_SIZE(hsw_ddi_translations_dp);
> return hsw_ddi_translations_dp;
> }
> @@ -720,13 +720,13 @@ intel_ddi_get_buf_trans_edp(struct drm_i915_private *dev_priv,
> }
>
> static const struct ddi_buf_trans *
> -intel_ddi_get_buf_trans_fdi(struct drm_i915_private *dev_priv,
> +intel_ddi_get_buf_trans_fdi(struct drm_i915_private *i915,
> int *n_entries)
> {
> - if (IS_BROADWELL(dev_priv)) {
> + if (IS_BROADWELL(i915)) {
> *n_entries = ARRAY_SIZE(bdw_ddi_translations_fdi);
> return bdw_ddi_translations_fdi;
> - } else if (IS_HASWELL(dev_priv)) {
> + } else if (IS_HASWELL(i915)) {
> *n_entries = ARRAY_SIZE(hsw_ddi_translations_fdi);
> return hsw_ddi_translations_fdi;
> }
> @@ -736,15 +736,15 @@ intel_ddi_get_buf_trans_fdi(struct drm_i915_private *dev_priv,
> }
>
> static const struct ddi_buf_trans *
> -intel_ddi_get_buf_trans_hdmi(struct drm_i915_private *dev_priv,
> +intel_ddi_get_buf_trans_hdmi(struct drm_i915_private *i915,
> int *n_entries)
> {
> - if (IS_GEN9_BC(dev_priv)) {
> - return skl_get_buf_trans_hdmi(dev_priv, n_entries);
> - } else if (IS_BROADWELL(dev_priv)) {
> + if (IS_GEN9_BC(i915)) {
> + return skl_get_buf_trans_hdmi(i915, n_entries);
> + } else if (IS_BROADWELL(i915)) {
> *n_entries = ARRAY_SIZE(bdw_ddi_translations_hdmi);
> return bdw_ddi_translations_hdmi;
> - } else if (IS_HASWELL(dev_priv)) {
> + } else if (IS_HASWELL(i915)) {
> *n_entries = ARRAY_SIZE(hsw_ddi_translations_hdmi);
> return hsw_ddi_translations_hdmi;
> }
> @@ -754,32 +754,32 @@ intel_ddi_get_buf_trans_hdmi(struct drm_i915_private *dev_priv,
> }
>
> static const struct bxt_ddi_buf_trans *
> -bxt_get_buf_trans_dp(struct drm_i915_private *dev_priv, int *n_entries)
> +bxt_get_buf_trans_dp(struct drm_i915_private *i915, int *n_entries)
> {
> *n_entries = ARRAY_SIZE(bxt_ddi_translations_dp);
> return bxt_ddi_translations_dp;
> }
>
> static const struct bxt_ddi_buf_trans *
> -bxt_get_buf_trans_edp(struct drm_i915_private *dev_priv, int *n_entries)
> +bxt_get_buf_trans_edp(struct drm_i915_private *i915, int *n_entries)
> {
> - if (dev_priv->vbt.edp.low_vswing) {
> + if (i915->vbt.edp.low_vswing) {
> *n_entries = ARRAY_SIZE(bxt_ddi_translations_edp);
> return bxt_ddi_translations_edp;
> }
>
> - return bxt_get_buf_trans_dp(dev_priv, n_entries);
> + return bxt_get_buf_trans_dp(i915, n_entries);
> }
>
> static const struct bxt_ddi_buf_trans *
> -bxt_get_buf_trans_hdmi(struct drm_i915_private *dev_priv, int *n_entries)
> +bxt_get_buf_trans_hdmi(struct drm_i915_private *i915, int *n_entries)
> {
> *n_entries = ARRAY_SIZE(bxt_ddi_translations_hdmi);
> return bxt_ddi_translations_hdmi;
> }
>
> static const struct cnl_ddi_buf_trans *
> -cnl_get_buf_trans_hdmi(struct drm_i915_private *dev_priv, int *n_entries)
> +cnl_get_buf_trans_hdmi(struct drm_i915_private *i915, int *n_entries)
> {
> u32 voltage = I915_READ(CNL_PORT_COMP_DW3) & VOLTAGE_INFO_MASK;
>
> @@ -800,7 +800,7 @@ cnl_get_buf_trans_hdmi(struct drm_i915_private *dev_priv, int *n_entries)
> }
>
> static const struct cnl_ddi_buf_trans *
> -cnl_get_buf_trans_dp(struct drm_i915_private *dev_priv, int *n_entries)
> +cnl_get_buf_trans_dp(struct drm_i915_private *i915, int *n_entries)
> {
> u32 voltage = I915_READ(CNL_PORT_COMP_DW3) & VOLTAGE_INFO_MASK;
>
> @@ -821,11 +821,11 @@ cnl_get_buf_trans_dp(struct drm_i915_private *dev_priv, int *n_entries)
> }
>
> static const struct cnl_ddi_buf_trans *
> -cnl_get_buf_trans_edp(struct drm_i915_private *dev_priv, int *n_entries)
> +cnl_get_buf_trans_edp(struct drm_i915_private *i915, int *n_entries)
> {
> u32 voltage = I915_READ(CNL_PORT_COMP_DW3) & VOLTAGE_INFO_MASK;
>
> - if (dev_priv->vbt.edp.low_vswing) {
> + if (i915->vbt.edp.low_vswing) {
> if (voltage == VOLTAGE_INFO_0_85V) {
> *n_entries = ARRAY_SIZE(cnl_ddi_translations_edp_0_85V);
> return cnl_ddi_translations_edp_0_85V;
> @@ -841,12 +841,12 @@ cnl_get_buf_trans_edp(struct drm_i915_private *dev_priv, int *n_entries)
> }
> return NULL;
> } else {
> - return cnl_get_buf_trans_dp(dev_priv, n_entries);
> + return cnl_get_buf_trans_dp(i915, n_entries);
> }
> }
>
> static const struct cnl_ddi_buf_trans *
> -icl_get_combo_buf_trans(struct drm_i915_private *dev_priv, enum port port,
> +icl_get_combo_buf_trans(struct drm_i915_private *i915, enum port port,
> int type, int rate, int *n_entries)
> {
> if (type == INTEL_OUTPUT_HDMI) {
> @@ -855,7 +855,7 @@ icl_get_combo_buf_trans(struct drm_i915_private *dev_priv, enum port port,
> } else if (rate > 540000 && type == INTEL_OUTPUT_EDP) {
> *n_entries = ARRAY_SIZE(icl_combo_phy_ddi_translations_edp_hbr3);
> return icl_combo_phy_ddi_translations_edp_hbr3;
> - } else if (type == INTEL_OUTPUT_EDP && dev_priv->vbt.edp.low_vswing) {
> + } else if (type == INTEL_OUTPUT_EDP && i915->vbt.edp.low_vswing) {
> *n_entries = ARRAY_SIZE(icl_combo_phy_ddi_translations_edp_hbr2);
> return icl_combo_phy_ddi_translations_edp_hbr2;
> }
> @@ -864,33 +864,33 @@ icl_get_combo_buf_trans(struct drm_i915_private *dev_priv, enum port port,
> return icl_combo_phy_ddi_translations_dp_hbr2;
> }
>
> -static int intel_ddi_hdmi_level(struct drm_i915_private *dev_priv, enum port port)
> +static int intel_ddi_hdmi_level(struct drm_i915_private *i915, enum port port)
> {
> int n_entries, level, default_entry;
>
> - level = dev_priv->vbt.ddi_port_info[port].hdmi_level_shift;
> + level = i915->vbt.ddi_port_info[port].hdmi_level_shift;
>
> - if (INTEL_GEN(dev_priv) >= 11) {
> - if (intel_port_is_combophy(dev_priv, port))
> - icl_get_combo_buf_trans(dev_priv, port, INTEL_OUTPUT_HDMI,
> + if (INTEL_GEN(i915) >= 11) {
> + if (intel_port_is_combophy(i915, port))
> + icl_get_combo_buf_trans(i915, port, INTEL_OUTPUT_HDMI,
> 0, &n_entries);
> else
> n_entries = ARRAY_SIZE(icl_mg_phy_ddi_translations);
> default_entry = n_entries - 1;
> - } else if (IS_CANNONLAKE(dev_priv)) {
> - cnl_get_buf_trans_hdmi(dev_priv, &n_entries);
> + } else if (IS_CANNONLAKE(i915)) {
> + cnl_get_buf_trans_hdmi(i915, &n_entries);
> default_entry = n_entries - 1;
> - } else if (IS_GEN9_LP(dev_priv)) {
> - bxt_get_buf_trans_hdmi(dev_priv, &n_entries);
> + } else if (IS_GEN9_LP(i915)) {
> + bxt_get_buf_trans_hdmi(i915, &n_entries);
> default_entry = n_entries - 1;
> - } else if (IS_GEN9_BC(dev_priv)) {
> - intel_ddi_get_buf_trans_hdmi(dev_priv, &n_entries);
> + } else if (IS_GEN9_BC(i915)) {
> + intel_ddi_get_buf_trans_hdmi(i915, &n_entries);
> default_entry = 8;
> - } else if (IS_BROADWELL(dev_priv)) {
> - intel_ddi_get_buf_trans_hdmi(dev_priv, &n_entries);
> + } else if (IS_BROADWELL(i915)) {
> + intel_ddi_get_buf_trans_hdmi(i915, &n_entries);
> default_entry = 7;
> - } else if (IS_HASWELL(dev_priv)) {
> - intel_ddi_get_buf_trans_hdmi(dev_priv, &n_entries);
> + } else if (IS_HASWELL(i915)) {
> + intel_ddi_get_buf_trans_hdmi(i915, &n_entries);
> default_entry = 6;
> } else {
> WARN(1, "ddi translation table missing\n");
> @@ -917,25 +917,25 @@ static int intel_ddi_hdmi_level(struct drm_i915_private *dev_priv, enum port por
> static void intel_prepare_dp_ddi_buffers(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> u32 iboost_bit = 0;
> int i, n_entries;
> enum port port = encoder->port;
> const struct ddi_buf_trans *ddi_translations;
>
> if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_ANALOG))
> - ddi_translations = intel_ddi_get_buf_trans_fdi(dev_priv,
> + ddi_translations = intel_ddi_get_buf_trans_fdi(i915,
> &n_entries);
> else if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_EDP))
> - ddi_translations = intel_ddi_get_buf_trans_edp(dev_priv, port,
> + ddi_translations = intel_ddi_get_buf_trans_edp(i915, port,
> &n_entries);
> else
> - ddi_translations = intel_ddi_get_buf_trans_dp(dev_priv, port,
> + ddi_translations = intel_ddi_get_buf_trans_dp(i915, port,
> &n_entries);
>
> /* If we're boosting the current, set bit 31 of trans1 */
> - if (IS_GEN9_BC(dev_priv) &&
> - dev_priv->vbt.ddi_port_info[port].dp_boost_level)
> + if (IS_GEN9_BC(i915) &&
> + i915->vbt.ddi_port_info[port].dp_boost_level)
> iboost_bit = DDI_BUF_BALANCE_LEG_ENABLE;
>
> for (i = 0; i < n_entries; i++) {
> @@ -954,13 +954,13 @@ static void intel_prepare_dp_ddi_buffers(struct intel_encoder *encoder,
> static void intel_prepare_hdmi_ddi_buffers(struct intel_encoder *encoder,
> int level)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> u32 iboost_bit = 0;
> int n_entries;
> enum port port = encoder->port;
> const struct ddi_buf_trans *ddi_translations;
>
> - ddi_translations = intel_ddi_get_buf_trans_hdmi(dev_priv, &n_entries);
> + ddi_translations = intel_ddi_get_buf_trans_hdmi(i915, &n_entries);
>
> if (WARN_ON_ONCE(!ddi_translations))
> return;
> @@ -968,8 +968,8 @@ static void intel_prepare_hdmi_ddi_buffers(struct intel_encoder *encoder,
> level = n_entries - 1;
>
> /* If we're boosting the current, set bit 31 of trans1 */
> - if (IS_GEN9_BC(dev_priv) &&
> - dev_priv->vbt.ddi_port_info[port].hdmi_boost_level)
> + if (IS_GEN9_BC(i915) &&
> + i915->vbt.ddi_port_info[port].hdmi_boost_level)
> iboost_bit = DDI_BUF_BALANCE_LEG_ENABLE;
>
> /* Entry 9 is for HDMI: */
> @@ -979,7 +979,7 @@ static void intel_prepare_hdmi_ddi_buffers(struct intel_encoder *encoder,
> ddi_translations[level].trans2);
> }
>
> -static void intel_wait_ddi_buf_idle(struct drm_i915_private *dev_priv,
> +static void intel_wait_ddi_buf_idle(struct drm_i915_private *i915,
> enum port port)
> {
> i915_reg_t reg = DDI_BUF_CTL(port);
> @@ -1064,7 +1064,7 @@ void hsw_fdi_link_train(struct intel_crtc *crtc,
> const struct intel_crtc_state *crtc_state)
> {
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_encoder *encoder;
> u32 temp, i, rx_ctl_val, ddi_pll_sel;
>
> @@ -1085,7 +1085,7 @@ void hsw_fdi_link_train(struct intel_crtc *crtc,
> FDI_RX_TP1_TO_TP2_48 | FDI_RX_FDI_DELAY_90);
>
> /* Enable the PCH Receiver FDI PLL */
> - rx_ctl_val = dev_priv->fdi_rx_config | FDI_RX_ENHANCE_FRAME_ENABLE |
> + rx_ctl_val = i915->fdi_rx_config | FDI_RX_ENHANCE_FRAME_ENABLE |
> FDI_RX_PLL_ENABLE |
> FDI_DP_PORT_WIDTH(crtc_state->fdi_lanes);
> I915_WRITE(FDI_RX_CTL(PIPE_A), rx_ctl_val);
> @@ -1174,7 +1174,7 @@ void hsw_fdi_link_train(struct intel_crtc *crtc,
> I915_WRITE(DP_TP_CTL(PORT_E), temp);
> POSTING_READ(DP_TP_CTL(PORT_E));
>
> - intel_wait_ddi_buf_idle(dev_priv, PORT_E);
> + intel_wait_ddi_buf_idle(i915, PORT_E);
>
> /* Reset FDI_RX_MISC pwrdn lanes */
> temp = I915_READ(FDI_RX_MISC(PIPE_A));
> @@ -1223,7 +1223,7 @@ intel_ddi_get_crtc_encoder(struct intel_crtc *crtc)
> return ret;
> }
>
> -static int hsw_ddi_calc_wrpll_link(struct drm_i915_private *dev_priv,
> +static int hsw_ddi_calc_wrpll_link(struct drm_i915_private *i915,
> i915_reg_t reg)
> {
> int refclk;
> @@ -1238,7 +1238,7 @@ static int hsw_ddi_calc_wrpll_link(struct drm_i915_private *dev_priv,
> * non-SSC for non-ULT HSW. Check FUSE_STRAP3
> * for the non-SSC reference frequency.
> */
> - if (IS_HASWELL(dev_priv) && !IS_HSW_ULT(dev_priv)) {
> + if (IS_HASWELL(i915) && !IS_HSW_ULT(i915)) {
> if (I915_READ(FUSE_STRAP3) & HSW_REF_CLK_SELECT)
> refclk = 24;
> else
> @@ -1325,7 +1325,7 @@ static int skl_calc_wrpll_link(const struct intel_dpll_hw_state *pll_state)
> return dco_freq / (p0 * p1 * p2 * 5);
> }
>
> -int cnl_calc_wrpll_link(struct drm_i915_private *dev_priv,
> +int cnl_calc_wrpll_link(struct drm_i915_private *i915,
> struct intel_dpll_hw_state *pll_state)
> {
> u32 p0, p1, p2, dco_freq, ref_clock;
> @@ -1367,7 +1367,7 @@ int cnl_calc_wrpll_link(struct drm_i915_private *dev_priv,
> break;
> }
>
> - ref_clock = cnl_hdmi_pll_ref_clock(dev_priv);
> + ref_clock = cnl_hdmi_pll_ref_clock(i915);
>
> dco_freq = (pll_state->cfgcr0 & DPLL_CFGCR0_DCO_INTEGER_MASK)
> * ref_clock;
> @@ -1381,7 +1381,7 @@ int cnl_calc_wrpll_link(struct drm_i915_private *dev_priv,
> return dco_freq / (p0 * p1 * p2 * 5);
> }
>
> -static int icl_calc_tbt_pll_link(struct drm_i915_private *dev_priv,
> +static int icl_calc_tbt_pll_link(struct drm_i915_private *i915,
> enum port port)
> {
> u32 val = I915_READ(DDI_CLK_SEL(port)) & DDI_CLK_SEL_MASK;
> @@ -1403,13 +1403,13 @@ static int icl_calc_tbt_pll_link(struct drm_i915_private *dev_priv,
> }
> }
>
> -static int icl_calc_mg_pll_link(struct drm_i915_private *dev_priv,
> +static int icl_calc_mg_pll_link(struct drm_i915_private *i915,
> const struct intel_dpll_hw_state *pll_state)
> {
> u32 m1, m2_int, m2_frac, div1, div2, ref_clock;
> u64 tmp;
>
> - ref_clock = dev_priv->cdclk.hw.ref;
> + ref_clock = i915->cdclk.hw.ref;
>
> m1 = pll_state->mg_pll_div1 & MG_PLL_DIV1_FBPREDIV_MASK;
> m2_int = pll_state->mg_pll_div0 & MG_PLL_DIV0_FBDIV_INT_MASK;
> @@ -1483,21 +1483,21 @@ static void ddi_dotclock_get(struct intel_crtc_state *pipe_config)
> static void icl_ddi_clock_get(struct intel_encoder *encoder,
> struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dpll_hw_state *pll_state = &pipe_config->dpll_hw_state;
> enum port port = encoder->port;
> int link_clock;
>
> - if (intel_port_is_combophy(dev_priv, port)) {
> - link_clock = cnl_calc_wrpll_link(dev_priv, pll_state);
> + if (intel_port_is_combophy(i915, port)) {
> + link_clock = cnl_calc_wrpll_link(i915, pll_state);
> } else {
> - enum intel_dpll_id pll_id = intel_get_shared_dpll_id(dev_priv,
> + enum intel_dpll_id pll_id = intel_get_shared_dpll_id(i915,
> pipe_config->shared_dpll);
>
> if (pll_id == DPLL_ID_ICL_TBTPLL)
> - link_clock = icl_calc_tbt_pll_link(dev_priv, port);
> + link_clock = icl_calc_tbt_pll_link(i915, port);
> else
> - link_clock = icl_calc_mg_pll_link(dev_priv, pll_state);
> + link_clock = icl_calc_mg_pll_link(i915, pll_state);
> }
>
> pipe_config->port_clock = link_clock;
> @@ -1508,12 +1508,12 @@ static void icl_ddi_clock_get(struct intel_encoder *encoder,
> static void cnl_ddi_clock_get(struct intel_encoder *encoder,
> struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dpll_hw_state *pll_state = &pipe_config->dpll_hw_state;
> int link_clock;
>
> if (pll_state->cfgcr0 & DPLL_CFGCR0_HDMI_MODE) {
> - link_clock = cnl_calc_wrpll_link(dev_priv, pll_state);
> + link_clock = cnl_calc_wrpll_link(i915, pll_state);
> } else {
> link_clock = pll_state->cfgcr0 & DPLL_CFGCR0_LINK_RATE_MASK;
>
> @@ -1604,7 +1604,7 @@ static void skl_ddi_clock_get(struct intel_encoder *encoder,
> static void hsw_ddi_clock_get(struct intel_encoder *encoder,
> struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> int link_clock = 0;
> u32 val, pll;
>
> @@ -1620,10 +1620,10 @@ static void hsw_ddi_clock_get(struct intel_encoder *encoder,
> link_clock = 270000;
> break;
> case PORT_CLK_SEL_WRPLL1:
> - link_clock = hsw_ddi_calc_wrpll_link(dev_priv, WRPLL_CTL(0));
> + link_clock = hsw_ddi_calc_wrpll_link(i915, WRPLL_CTL(0));
> break;
> case PORT_CLK_SEL_WRPLL2:
> - link_clock = hsw_ddi_calc_wrpll_link(dev_priv, WRPLL_CTL(1));
> + link_clock = hsw_ddi_calc_wrpll_link(i915, WRPLL_CTL(1));
> break;
> case PORT_CLK_SEL_SPLL:
> pll = I915_READ(SPLL_CTL) & SPLL_FREQ_MASK;
> @@ -1675,24 +1675,24 @@ static void bxt_ddi_clock_get(struct intel_encoder *encoder,
> static void intel_ddi_clock_get(struct intel_encoder *encoder,
> struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
>
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> icl_ddi_clock_get(encoder, pipe_config);
> - else if (IS_CANNONLAKE(dev_priv))
> + else if (IS_CANNONLAKE(i915))
> cnl_ddi_clock_get(encoder, pipe_config);
> - else if (IS_GEN9_LP(dev_priv))
> + else if (IS_GEN9_LP(i915))
> bxt_ddi_clock_get(encoder, pipe_config);
> - else if (IS_GEN9_BC(dev_priv))
> + else if (IS_GEN9_BC(i915))
> skl_ddi_clock_get(encoder, pipe_config);
> - else if (INTEL_GEN(dev_priv) <= 8)
> + else if (INTEL_GEN(i915) <= 8)
> hsw_ddi_clock_get(encoder, pipe_config);
> }
>
> void intel_ddi_set_pipe_settings(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> u32 temp;
>
> @@ -1746,7 +1746,7 @@ void intel_ddi_set_vc_payload_alloc(const struct intel_crtc_state *crtc_state,
> bool state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> u32 temp;
>
> @@ -1762,7 +1762,7 @@ void intel_ddi_enable_transcoder_func(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> struct intel_encoder *encoder = intel_ddi_get_crtc_encoder(crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
> enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> enum port port = encoder->port;
> @@ -1845,7 +1845,7 @@ void intel_ddi_enable_transcoder_func(const struct intel_crtc_state *crtc_state)
> void intel_ddi_disable_transcoder_func(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> i915_reg_t reg = TRANS_DDI_FUNC_CTL(cpu_transcoder);
> u32 val = I915_READ(reg);
> @@ -1854,7 +1854,7 @@ void intel_ddi_disable_transcoder_func(const struct intel_crtc_state *crtc_state
> val |= TRANS_DDI_PORT_NONE;
> I915_WRITE(reg, val);
>
> - if (dev_priv->quirks & QUIRK_INCREASE_DDI_DISABLED_TIME &&
> + if (i915->quirks & QUIRK_INCREASE_DDI_DISABLED_TIME &&
> intel_crtc_has_type(crtc_state, INTEL_OUTPUT_HDMI)) {
> DRM_DEBUG_KMS("Quirk Increase DDI disabled time\n");
> /* Quirk time at 100ms for reliable operation */
> @@ -1866,13 +1866,13 @@ int intel_ddi_toggle_hdcp_signalling(struct intel_encoder *intel_encoder,
> bool enable)
> {
> struct drm_device *dev = intel_encoder->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> intel_wakeref_t wakeref;
> enum pipe pipe = 0;
> int ret = 0;
> u32 tmp;
>
> - wakeref = intel_display_power_get_if_enabled(dev_priv,
> + wakeref = intel_display_power_get_if_enabled(i915,
> intel_encoder->power_domain);
> if (WARN_ON(!wakeref))
> return -ENXIO;
> @@ -1889,14 +1889,14 @@ int intel_ddi_toggle_hdcp_signalling(struct intel_encoder *intel_encoder,
> tmp &= ~TRANS_DDI_HDCP_SIGNALLING;
> I915_WRITE(TRANS_DDI_FUNC_CTL(pipe), tmp);
> out:
> - intel_display_power_put(dev_priv, intel_encoder->power_domain, wakeref);
> + intel_display_power_put(i915, intel_encoder->power_domain, wakeref);
> return ret;
> }
>
> bool intel_ddi_connector_get_hw_state(struct intel_connector *intel_connector)
> {
> struct drm_device *dev = intel_connector->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_encoder *encoder = intel_connector->encoder;
> int type = intel_connector->base.connector_type;
> enum port port = encoder->port;
> @@ -1906,7 +1906,7 @@ bool intel_ddi_connector_get_hw_state(struct intel_connector *intel_connector)
> u32 tmp;
> bool ret;
>
> - wakeref = intel_display_power_get_if_enabled(dev_priv,
> + wakeref = intel_display_power_get_if_enabled(i915,
> encoder->power_domain);
> if (!wakeref)
> return false;
> @@ -1916,7 +1916,7 @@ bool intel_ddi_connector_get_hw_state(struct intel_connector *intel_connector)
> goto out;
> }
>
> - if (HAS_TRANSCODER_EDP(dev_priv) && port == PORT_A)
> + if (HAS_TRANSCODER_EDP(i915) && port == PORT_A)
> cpu_transcoder = TRANSCODER_EDP;
> else
> cpu_transcoder = (enum transcoder) pipe;
> @@ -1950,7 +1950,7 @@ bool intel_ddi_connector_get_hw_state(struct intel_connector *intel_connector)
> }
>
> out:
> - intel_display_power_put(dev_priv, encoder->power_domain, wakeref);
> + intel_display_power_put(i915, encoder->power_domain, wakeref);
>
> return ret;
> }
> @@ -1959,7 +1959,7 @@ static void intel_ddi_get_encoder_pipes(struct intel_encoder *encoder,
> u8 *pipe_mask, bool *is_dp_mst)
> {
> struct drm_device *dev = encoder->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> enum port port = encoder->port;
> intel_wakeref_t wakeref;
> enum pipe p;
> @@ -1969,7 +1969,7 @@ static void intel_ddi_get_encoder_pipes(struct intel_encoder *encoder,
> *pipe_mask = 0;
> *is_dp_mst = false;
>
> - wakeref = intel_display_power_get_if_enabled(dev_priv,
> + wakeref = intel_display_power_get_if_enabled(i915,
> encoder->power_domain);
> if (!wakeref)
> return;
> @@ -1978,7 +1978,7 @@ static void intel_ddi_get_encoder_pipes(struct intel_encoder *encoder,
> if (!(tmp & DDI_BUF_CTL_ENABLE))
> goto out;
>
> - if (HAS_TRANSCODER_EDP(dev_priv) && port == PORT_A) {
> + if (HAS_TRANSCODER_EDP(i915) && port == PORT_A) {
> tmp = I915_READ(TRANS_DDI_FUNC_CTL(TRANSCODER_EDP));
>
> switch (tmp & TRANS_DDI_EDP_INPUT_MASK) {
> @@ -2001,7 +2001,7 @@ static void intel_ddi_get_encoder_pipes(struct intel_encoder *encoder,
> }
>
> mst_pipe_mask = 0;
> - for_each_pipe(dev_priv, p) {
> + for_each_pipe(i915, p) {
> enum transcoder cpu_transcoder = (enum transcoder)p;
>
> tmp = I915_READ(TRANS_DDI_FUNC_CTL(cpu_transcoder));
> @@ -2033,7 +2033,7 @@ static void intel_ddi_get_encoder_pipes(struct intel_encoder *encoder,
> *is_dp_mst = mst_pipe_mask;
>
> out:
> - if (*pipe_mask && IS_GEN9_LP(dev_priv)) {
> + if (*pipe_mask && IS_GEN9_LP(i915)) {
> tmp = I915_READ(BXT_PHY_CTL(port));
> if ((tmp & (BXT_PHY_CMNLANE_POWERDOWN_ACK |
> BXT_PHY_LANE_POWERDOWN_ACK |
> @@ -2042,7 +2042,7 @@ static void intel_ddi_get_encoder_pipes(struct intel_encoder *encoder,
> "(PHY_CTL %08x)\n", port_name(port), tmp);
> }
>
> - intel_display_power_put(dev_priv, encoder->power_domain, wakeref);
> + intel_display_power_put(i915, encoder->power_domain, wakeref);
> }
>
> bool intel_ddi_get_hw_state(struct intel_encoder *encoder,
> @@ -2083,7 +2083,7 @@ intel_ddi_main_link_aux_domain(struct intel_digital_port *dig_port)
> static void intel_ddi_get_power_domains(struct intel_encoder *encoder,
> struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_digital_port *dig_port;
>
> /*
> @@ -2095,29 +2095,29 @@ static void intel_ddi_get_power_domains(struct intel_encoder *encoder,
> return;
>
> dig_port = enc_to_dig_port(&encoder->base);
> - intel_display_power_get(dev_priv, dig_port->ddi_io_power_domain);
> + intel_display_power_get(i915, dig_port->ddi_io_power_domain);
>
> /*
> * AUX power is only needed for (e)DP mode, and for HDMI mode on TC
> * ports.
> */
> if (intel_crtc_has_dp_encoder(crtc_state) ||
> - intel_port_is_tc(dev_priv, encoder->port))
> - intel_display_power_get(dev_priv,
> + intel_port_is_tc(i915, encoder->port))
> + intel_display_power_get(i915,
> intel_ddi_main_link_aux_domain(dig_port));
>
> /*
> * VDSC power is needed when DSC is enabled
> */
> if (crtc_state->dsc_params.compression_enable)
> - intel_display_power_get(dev_priv,
> + intel_display_power_get(i915,
> intel_dsc_power_domain(crtc_state));
> }
>
> void intel_ddi_enable_pipe_clock(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> struct intel_encoder *encoder = intel_ddi_get_crtc_encoder(crtc);
> enum port port = encoder->port;
> enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> @@ -2129,7 +2129,7 @@ void intel_ddi_enable_pipe_clock(const struct intel_crtc_state *crtc_state)
>
> void intel_ddi_disable_pipe_clock(const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
>
> if (cpu_transcoder != TRANSCODER_EDP)
> @@ -2137,7 +2137,7 @@ void intel_ddi_disable_pipe_clock(const struct intel_crtc_state *crtc_state)
> TRANS_CLK_SEL_DISABLED);
> }
>
> -static void _skl_ddi_set_iboost(struct drm_i915_private *dev_priv,
> +static void _skl_ddi_set_iboost(struct drm_i915_private *i915,
> enum port port, u8 iboost)
> {
> u32 tmp;
> @@ -2155,25 +2155,25 @@ static void skl_ddi_set_iboost(struct intel_encoder *encoder,
> int level, enum intel_output_type type)
> {
> struct intel_digital_port *intel_dig_port = enc_to_dig_port(&encoder->base);
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> enum port port = encoder->port;
> u8 iboost;
>
> if (type == INTEL_OUTPUT_HDMI)
> - iboost = dev_priv->vbt.ddi_port_info[port].hdmi_boost_level;
> + iboost = i915->vbt.ddi_port_info[port].hdmi_boost_level;
> else
> - iboost = dev_priv->vbt.ddi_port_info[port].dp_boost_level;
> + iboost = i915->vbt.ddi_port_info[port].dp_boost_level;
>
> if (iboost == 0) {
> const struct ddi_buf_trans *ddi_translations;
> int n_entries;
>
> if (type == INTEL_OUTPUT_HDMI)
> - ddi_translations = intel_ddi_get_buf_trans_hdmi(dev_priv, &n_entries);
> + ddi_translations = intel_ddi_get_buf_trans_hdmi(i915, &n_entries);
> else if (type == INTEL_OUTPUT_EDP)
> - ddi_translations = intel_ddi_get_buf_trans_edp(dev_priv, port, &n_entries);
> + ddi_translations = intel_ddi_get_buf_trans_edp(i915, port, &n_entries);
> else
> - ddi_translations = intel_ddi_get_buf_trans_dp(dev_priv, port, &n_entries);
> + ddi_translations = intel_ddi_get_buf_trans_dp(i915, port, &n_entries);
>
> if (WARN_ON_ONCE(!ddi_translations))
> return;
> @@ -2189,33 +2189,33 @@ static void skl_ddi_set_iboost(struct intel_encoder *encoder,
> return;
> }
>
> - _skl_ddi_set_iboost(dev_priv, port, iboost);
> + _skl_ddi_set_iboost(i915, port, iboost);
>
> if (port == PORT_A && intel_dig_port->max_lanes == 4)
> - _skl_ddi_set_iboost(dev_priv, PORT_E, iboost);
> + _skl_ddi_set_iboost(i915, PORT_E, iboost);
> }
>
> static void bxt_ddi_vswing_sequence(struct intel_encoder *encoder,
> int level, enum intel_output_type type)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> const struct bxt_ddi_buf_trans *ddi_translations;
> enum port port = encoder->port;
> int n_entries;
>
> if (type == INTEL_OUTPUT_HDMI)
> - ddi_translations = bxt_get_buf_trans_hdmi(dev_priv, &n_entries);
> + ddi_translations = bxt_get_buf_trans_hdmi(i915, &n_entries);
> else if (type == INTEL_OUTPUT_EDP)
> - ddi_translations = bxt_get_buf_trans_edp(dev_priv, &n_entries);
> + ddi_translations = bxt_get_buf_trans_edp(i915, &n_entries);
> else
> - ddi_translations = bxt_get_buf_trans_dp(dev_priv, &n_entries);
> + ddi_translations = bxt_get_buf_trans_dp(i915, &n_entries);
>
> if (WARN_ON_ONCE(!ddi_translations))
> return;
> if (WARN_ON_ONCE(level >= n_entries))
> level = n_entries - 1;
>
> - bxt_ddi_phy_set_signal_level(dev_priv, port,
> + bxt_ddi_phy_set_signal_level(i915, port,
> ddi_translations[level].margin,
> ddi_translations[level].scale,
> ddi_translations[level].enable,
> @@ -2224,32 +2224,32 @@ static void bxt_ddi_vswing_sequence(struct intel_encoder *encoder,
>
> u8 intel_ddi_dp_voltage_max(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> enum port port = encoder->port;
> int n_entries;
>
> - if (INTEL_GEN(dev_priv) >= 11) {
> - if (intel_port_is_combophy(dev_priv, port))
> - icl_get_combo_buf_trans(dev_priv, port, encoder->type,
> + if (INTEL_GEN(i915) >= 11) {
> + if (intel_port_is_combophy(i915, port))
> + icl_get_combo_buf_trans(i915, port, encoder->type,
> intel_dp->link_rate, &n_entries);
> else
> n_entries = ARRAY_SIZE(icl_mg_phy_ddi_translations);
> - } else if (IS_CANNONLAKE(dev_priv)) {
> + } else if (IS_CANNONLAKE(i915)) {
> if (encoder->type == INTEL_OUTPUT_EDP)
> - cnl_get_buf_trans_edp(dev_priv, &n_entries);
> + cnl_get_buf_trans_edp(i915, &n_entries);
> else
> - cnl_get_buf_trans_dp(dev_priv, &n_entries);
> - } else if (IS_GEN9_LP(dev_priv)) {
> + cnl_get_buf_trans_dp(i915, &n_entries);
> + } else if (IS_GEN9_LP(i915)) {
> if (encoder->type == INTEL_OUTPUT_EDP)
> - bxt_get_buf_trans_edp(dev_priv, &n_entries);
> + bxt_get_buf_trans_edp(i915, &n_entries);
> else
> - bxt_get_buf_trans_dp(dev_priv, &n_entries);
> + bxt_get_buf_trans_dp(i915, &n_entries);
> } else {
> if (encoder->type == INTEL_OUTPUT_EDP)
> - intel_ddi_get_buf_trans_edp(dev_priv, port, &n_entries);
> + intel_ddi_get_buf_trans_edp(i915, port, &n_entries);
> else
> - intel_ddi_get_buf_trans_dp(dev_priv, port, &n_entries);
> + intel_ddi_get_buf_trans_dp(i915, port, &n_entries);
> }
>
> if (WARN_ON(n_entries < 1))
> @@ -2284,18 +2284,18 @@ u8 intel_ddi_dp_pre_emphasis_max(struct intel_encoder *encoder, u8 voltage_swing
> static void cnl_ddi_vswing_program(struct intel_encoder *encoder,
> int level, enum intel_output_type type)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> const struct cnl_ddi_buf_trans *ddi_translations;
> enum port port = encoder->port;
> int n_entries, ln;
> u32 val;
>
> if (type == INTEL_OUTPUT_HDMI)
> - ddi_translations = cnl_get_buf_trans_hdmi(dev_priv, &n_entries);
> + ddi_translations = cnl_get_buf_trans_hdmi(i915, &n_entries);
> else if (type == INTEL_OUTPUT_EDP)
> - ddi_translations = cnl_get_buf_trans_edp(dev_priv, &n_entries);
> + ddi_translations = cnl_get_buf_trans_edp(i915, &n_entries);
> else
> - ddi_translations = cnl_get_buf_trans_dp(dev_priv, &n_entries);
> + ddi_translations = cnl_get_buf_trans_dp(i915, &n_entries);
>
> if (WARN_ON_ONCE(!ddi_translations))
> return;
> @@ -2348,7 +2348,7 @@ static void cnl_ddi_vswing_program(struct intel_encoder *encoder,
> static void cnl_ddi_vswing_sequence(struct intel_encoder *encoder,
> int level, enum intel_output_type type)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> enum port port = encoder->port;
> int width, rate, ln;
> u32 val;
> @@ -2412,7 +2412,7 @@ static void cnl_ddi_vswing_sequence(struct intel_encoder *encoder,
> I915_WRITE(CNL_PORT_TX_DW5_GRP(port), val);
> }
>
> -static void icl_ddi_combo_vswing_program(struct drm_i915_private *dev_priv,
> +static void icl_ddi_combo_vswing_program(struct drm_i915_private *i915,
> u32 level, enum port port, int type,
> int rate)
> {
> @@ -2420,7 +2420,7 @@ static void icl_ddi_combo_vswing_program(struct drm_i915_private *dev_priv,
> u32 n_entries, val;
> int ln;
>
> - ddi_translations = icl_get_combo_buf_trans(dev_priv, port, type,
> + ddi_translations = icl_get_combo_buf_trans(i915, port, type,
> rate, &n_entries);
> if (!ddi_translations)
> return;
> @@ -2472,7 +2472,7 @@ static void icl_combo_phy_ddi_vswing_sequence(struct intel_encoder *encoder,
> u32 level,
> enum intel_output_type type)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> enum port port = encoder->port;
> int width = 0;
> int rate = 0;
> @@ -2530,7 +2530,7 @@ static void icl_combo_phy_ddi_vswing_sequence(struct intel_encoder *encoder,
> I915_WRITE(ICL_PORT_TX_DW5_GRP(port), val);
>
> /* 5. Program swing and de-emphasis */
> - icl_ddi_combo_vswing_program(dev_priv, level, port, type, rate);
> + icl_ddi_combo_vswing_program(i915, level, port, type, rate);
>
> /* 6. Set training enable to trigger update */
> val = I915_READ(ICL_PORT_TX_DW5_LN0(port));
> @@ -2542,7 +2542,7 @@ static void icl_mg_phy_ddi_vswing_sequence(struct intel_encoder *encoder,
> int link_clock,
> u32 level)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> enum port port = encoder->port;
> const struct icl_mg_phy_ddi_buf_trans *ddi_translations;
> u32 n_entries, val;
> @@ -2662,10 +2662,10 @@ static void icl_ddi_vswing_sequence(struct intel_encoder *encoder,
> u32 level,
> enum intel_output_type type)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> enum port port = encoder->port;
>
> - if (intel_port_is_combophy(dev_priv, port))
> + if (intel_port_is_combophy(i915, port))
> icl_combo_phy_ddi_vswing_sequence(encoder, level, type);
> else
> icl_mg_phy_ddi_vswing_sequence(encoder, link_clock, level);
> @@ -2698,14 +2698,14 @@ static u32 intel_ddi_dp_level(struct intel_dp *intel_dp)
> u32 bxt_signal_levels(struct intel_dp *intel_dp)
> {
> struct intel_digital_port *dport = dp_to_dig_port(intel_dp);
> - struct drm_i915_private *dev_priv = to_i915(dport->base.base.dev);
> + struct drm_i915_private *i915 = to_i915(dport->base.base.dev);
> struct intel_encoder *encoder = &dport->base;
> int level = intel_ddi_dp_level(intel_dp);
>
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> icl_ddi_vswing_sequence(encoder, intel_dp->link_rate,
> level, encoder->type);
> - else if (IS_CANNONLAKE(dev_priv))
> + else if (IS_CANNONLAKE(i915))
> cnl_ddi_vswing_sequence(encoder, level, encoder->type);
> else
> bxt_ddi_vswing_sequence(encoder, level, encoder->type);
> @@ -2716,24 +2716,24 @@ u32 bxt_signal_levels(struct intel_dp *intel_dp)
> u32 ddi_signal_levels(struct intel_dp *intel_dp)
> {
> struct intel_digital_port *dport = dp_to_dig_port(intel_dp);
> - struct drm_i915_private *dev_priv = to_i915(dport->base.base.dev);
> + struct drm_i915_private *i915 = to_i915(dport->base.base.dev);
> struct intel_encoder *encoder = &dport->base;
> int level = intel_ddi_dp_level(intel_dp);
>
> - if (IS_GEN9_BC(dev_priv))
> + if (IS_GEN9_BC(i915))
> skl_ddi_set_iboost(encoder, level, encoder->type);
>
> return DDI_BUF_TRANS_SELECT(level);
> }
>
> static inline
> -u32 icl_dpclka_cfgcr0_clk_off(struct drm_i915_private *dev_priv,
> +u32 icl_dpclka_cfgcr0_clk_off(struct drm_i915_private *i915,
> enum port port)
> {
> - if (intel_port_is_combophy(dev_priv, port)) {
> + if (intel_port_is_combophy(i915, port)) {
> return ICL_DPCLKA_CFGCR0_DDI_CLK_OFF(port);
> - } else if (intel_port_is_tc(dev_priv, port)) {
> - enum tc_port tc_port = intel_port_to_tc(dev_priv, port);
> + } else if (intel_port_is_tc(i915, port)) {
> + enum tc_port tc_port = intel_port_to_tc(i915, port);
>
> return ICL_DPCLKA_CFGCR0_TC_CLK_OFF(tc_port);
> }
> @@ -2744,47 +2744,47 @@ u32 icl_dpclka_cfgcr0_clk_off(struct drm_i915_private *dev_priv,
> static void icl_map_plls_to_ports(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_shared_dpll *pll = crtc_state->shared_dpll;
> enum port port = encoder->port;
> u32 val;
>
> - mutex_lock(&dev_priv->dpll_lock);
> + mutex_lock(&i915->dpll_lock);
>
> val = I915_READ(DPCLKA_CFGCR0_ICL);
> - WARN_ON((val & icl_dpclka_cfgcr0_clk_off(dev_priv, port)) == 0);
> + WARN_ON((val & icl_dpclka_cfgcr0_clk_off(i915, port)) == 0);
>
> - if (intel_port_is_combophy(dev_priv, port)) {
> + if (intel_port_is_combophy(i915, port)) {
> val &= ~DPCLKA_CFGCR0_DDI_CLK_SEL_MASK(port);
> val |= DPCLKA_CFGCR0_DDI_CLK_SEL(pll->info->id, port);
> I915_WRITE(DPCLKA_CFGCR0_ICL, val);
> POSTING_READ(DPCLKA_CFGCR0_ICL);
> }
>
> - val &= ~icl_dpclka_cfgcr0_clk_off(dev_priv, port);
> + val &= ~icl_dpclka_cfgcr0_clk_off(i915, port);
> I915_WRITE(DPCLKA_CFGCR0_ICL, val);
>
> - mutex_unlock(&dev_priv->dpll_lock);
> + mutex_unlock(&i915->dpll_lock);
> }
>
> static void icl_unmap_plls_to_ports(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> enum port port = encoder->port;
> u32 val;
>
> - mutex_lock(&dev_priv->dpll_lock);
> + mutex_lock(&i915->dpll_lock);
>
> val = I915_READ(DPCLKA_CFGCR0_ICL);
> - val |= icl_dpclka_cfgcr0_clk_off(dev_priv, port);
> + val |= icl_dpclka_cfgcr0_clk_off(i915, port);
> I915_WRITE(DPCLKA_CFGCR0_ICL, val);
>
> - mutex_unlock(&dev_priv->dpll_lock);
> + mutex_unlock(&i915->dpll_lock);
> }
>
> void icl_sanitize_encoder_pll_mapping(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> u32 val;
> enum port port;
> u32 port_mask;
> @@ -2821,7 +2821,7 @@ void icl_sanitize_encoder_pll_mapping(struct intel_encoder *encoder)
> * Sanity check that we haven't incorrectly registered another
> * encoder using any of the ports of this DSI encoder.
> */
> - for_each_intel_encoder(&dev_priv->drm, other_encoder) {
> + for_each_intel_encoder(&i915->drm, other_encoder) {
> if (other_encoder == encoder)
> continue;
>
> @@ -2838,7 +2838,7 @@ void icl_sanitize_encoder_pll_mapping(struct intel_encoder *encoder)
> val = I915_READ(DPCLKA_CFGCR0_ICL);
> for_each_port_masked(port, port_mask) {
> bool ddi_clk_ungated = !(val &
> - icl_dpclka_cfgcr0_clk_off(dev_priv,
> + icl_dpclka_cfgcr0_clk_off(i915,
> port));
>
> if (ddi_clk_needed == ddi_clk_ungated)
> @@ -2853,7 +2853,7 @@ void icl_sanitize_encoder_pll_mapping(struct intel_encoder *encoder)
>
> DRM_NOTE("Port %c is disabled/in DSI mode with an ungated DDI clock, gate it\n",
> port_name(port));
> - val |= icl_dpclka_cfgcr0_clk_off(dev_priv, port);
> + val |= icl_dpclka_cfgcr0_clk_off(i915, port);
> I915_WRITE(DPCLKA_CFGCR0_ICL, val);
> }
> }
> @@ -2861,7 +2861,7 @@ void icl_sanitize_encoder_pll_mapping(struct intel_encoder *encoder)
> static void intel_ddi_clk_select(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> enum port port = encoder->port;
> u32 val;
> const struct intel_shared_dpll *pll = crtc_state->shared_dpll;
> @@ -2869,13 +2869,13 @@ static void intel_ddi_clk_select(struct intel_encoder *encoder,
> if (WARN_ON(!pll))
> return;
>
> - mutex_lock(&dev_priv->dpll_lock);
> + mutex_lock(&i915->dpll_lock);
>
> - if (INTEL_GEN(dev_priv) >= 11) {
> - if (!intel_port_is_combophy(dev_priv, port))
> + if (INTEL_GEN(i915) >= 11) {
> + if (!intel_port_is_combophy(i915, port))
> I915_WRITE(DDI_CLK_SEL(port),
> icl_pll_to_ddi_clk_sel(encoder, crtc_state));
> - } else if (IS_CANNONLAKE(dev_priv)) {
> + } else if (IS_CANNONLAKE(i915)) {
> /* Configure DPCLKA_CFGCR0 to map the DPLL to the DDI. */
> val = I915_READ(DPCLKA_CFGCR0);
> val &= ~DPCLKA_CFGCR0_DDI_CLK_SEL_MASK(port);
> @@ -2890,7 +2890,7 @@ static void intel_ddi_clk_select(struct intel_encoder *encoder,
> val = I915_READ(DPCLKA_CFGCR0);
> val &= ~DPCLKA_CFGCR0_DDI_CLK_OFF(port);
> I915_WRITE(DPCLKA_CFGCR0, val);
> - } else if (IS_GEN9_BC(dev_priv)) {
> + } else if (IS_GEN9_BC(i915)) {
> /* DDI -> PLL mapping */
> val = I915_READ(DPLL_CTRL2);
>
> @@ -2901,37 +2901,37 @@ static void intel_ddi_clk_select(struct intel_encoder *encoder,
>
> I915_WRITE(DPLL_CTRL2, val);
>
> - } else if (INTEL_GEN(dev_priv) < 9) {
> + } else if (INTEL_GEN(i915) < 9) {
> I915_WRITE(PORT_CLK_SEL(port), hsw_pll_to_ddi_pll_sel(pll));
> }
>
> - mutex_unlock(&dev_priv->dpll_lock);
> + mutex_unlock(&i915->dpll_lock);
> }
>
> static void intel_ddi_clk_disable(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> enum port port = encoder->port;
>
> - if (INTEL_GEN(dev_priv) >= 11) {
> - if (!intel_port_is_combophy(dev_priv, port))
> + if (INTEL_GEN(i915) >= 11) {
> + if (!intel_port_is_combophy(i915, port))
> I915_WRITE(DDI_CLK_SEL(port), DDI_CLK_SEL_NONE);
> - } else if (IS_CANNONLAKE(dev_priv)) {
> + } else if (IS_CANNONLAKE(i915)) {
> I915_WRITE(DPCLKA_CFGCR0, I915_READ(DPCLKA_CFGCR0) |
> DPCLKA_CFGCR0_DDI_CLK_OFF(port));
> - } else if (IS_GEN9_BC(dev_priv)) {
> + } else if (IS_GEN9_BC(i915)) {
> I915_WRITE(DPLL_CTRL2, I915_READ(DPLL_CTRL2) |
> DPLL_CTRL2_DDI_CLK_OFF(port));
> - } else if (INTEL_GEN(dev_priv) < 9) {
> + } else if (INTEL_GEN(i915) < 9) {
> I915_WRITE(PORT_CLK_SEL(port), PORT_CLK_SEL_NONE);
> }
> }
>
> static void icl_enable_phy_clock_gating(struct intel_digital_port *dig_port)
> {
> - struct drm_i915_private *dev_priv = to_i915(dig_port->base.base.dev);
> + struct drm_i915_private *i915 = to_i915(dig_port->base.base.dev);
> enum port port = dig_port->base.port;
> - enum tc_port tc_port = intel_port_to_tc(dev_priv, port);
> + enum tc_port tc_port = intel_port_to_tc(i915, port);
> u32 val;
> int ln;
>
> @@ -2961,9 +2961,9 @@ static void icl_enable_phy_clock_gating(struct intel_digital_port *dig_port)
>
> static void icl_disable_phy_clock_gating(struct intel_digital_port *dig_port)
> {
> - struct drm_i915_private *dev_priv = to_i915(dig_port->base.base.dev);
> + struct drm_i915_private *i915 = to_i915(dig_port->base.base.dev);
> enum port port = dig_port->base.port;
> - enum tc_port tc_port = intel_port_to_tc(dev_priv, port);
> + enum tc_port tc_port = intel_port_to_tc(i915, port);
> u32 val;
> int ln;
>
> @@ -2993,9 +2993,9 @@ static void icl_disable_phy_clock_gating(struct intel_digital_port *dig_port)
>
> static void icl_program_mg_dp_mode(struct intel_digital_port *intel_dig_port)
> {
> - struct drm_i915_private *dev_priv = to_i915(intel_dig_port->base.base.dev);
> + struct drm_i915_private *i915 = to_i915(intel_dig_port->base.base.dev);
> enum port port = intel_dig_port->base.port;
> - enum tc_port tc_port = intel_port_to_tc(dev_priv, port);
> + enum tc_port tc_port = intel_port_to_tc(i915, port);
> u32 ln0, ln1, lane_info;
>
> if (tc_port == PORT_TC_NONE || intel_dig_port->tc_type == TC_PORT_TBT)
> @@ -3069,7 +3069,7 @@ static void intel_dp_sink_set_fec_ready(struct intel_dp *intel_dp,
> static void intel_ddi_enable_fec(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> enum port port = encoder->port;
> u32 val;
>
> @@ -3080,7 +3080,7 @@ static void intel_ddi_enable_fec(struct intel_encoder *encoder,
> val |= DP_TP_CTL_FEC_ENABLE;
> I915_WRITE(DP_TP_CTL(port), val);
>
> - if (intel_wait_for_register(&dev_priv->uncore, DP_TP_STATUS(port),
> + if (intel_wait_for_register(&i915->uncore, DP_TP_STATUS(port),
> DP_TP_STATUS_FEC_ENABLE_LIVE,
> DP_TP_STATUS_FEC_ENABLE_LIVE,
> 1))
> @@ -3090,7 +3090,7 @@ static void intel_ddi_enable_fec(struct intel_encoder *encoder,
> static void intel_ddi_disable_fec_state(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> enum port port = encoder->port;
> u32 val;
>
> @@ -3108,7 +3108,7 @@ static void intel_ddi_pre_enable_dp(struct intel_encoder *encoder,
> const struct drm_connector_state *conn_state)
> {
> struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> enum port port = encoder->port;
> struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base);
> bool is_mst = intel_crtc_has_type(crtc_state, INTEL_OUTPUT_DP_MST);
> @@ -3123,26 +3123,26 @@ static void intel_ddi_pre_enable_dp(struct intel_encoder *encoder,
>
> intel_ddi_clk_select(encoder, crtc_state);
>
> - intel_display_power_get(dev_priv, dig_port->ddi_io_power_domain);
> + intel_display_power_get(i915, dig_port->ddi_io_power_domain);
>
> icl_program_mg_dp_mode(dig_port);
> icl_disable_phy_clock_gating(dig_port);
>
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> icl_ddi_vswing_sequence(encoder, crtc_state->port_clock,
> level, encoder->type);
> - else if (IS_CANNONLAKE(dev_priv))
> + else if (IS_CANNONLAKE(i915))
> cnl_ddi_vswing_sequence(encoder, level, encoder->type);
> - else if (IS_GEN9_LP(dev_priv))
> + else if (IS_GEN9_LP(i915))
> bxt_ddi_vswing_sequence(encoder, level, encoder->type);
> else
> intel_prepare_dp_ddi_buffers(encoder, crtc_state);
>
> - if (intel_port_is_combophy(dev_priv, port)) {
> + if (intel_port_is_combophy(i915, port)) {
> bool lane_reversal =
> dig_port->saved_port_bits & DDI_BUF_PORT_REVERSAL;
>
> - intel_combo_phy_power_up_lanes(dev_priv, port, false,
> + intel_combo_phy_power_up_lanes(i915, port, false,
> crtc_state->lane_count,
> lane_reversal);
> }
> @@ -3154,7 +3154,7 @@ static void intel_ddi_pre_enable_dp(struct intel_encoder *encoder,
> true);
> intel_dp_sink_set_fec_ready(intel_dp, crtc_state);
> intel_dp_start_link_train(intel_dp);
> - if (port != PORT_A || INTEL_GEN(dev_priv) >= 9)
> + if (port != PORT_A || INTEL_GEN(i915) >= 9)
> intel_dp_stop_link_train(intel_dp);
>
> intel_ddi_enable_fec(encoder, crtc_state);
> @@ -3173,32 +3173,32 @@ static void intel_ddi_pre_enable_hdmi(struct intel_encoder *encoder,
> {
> struct intel_digital_port *intel_dig_port = enc_to_dig_port(&encoder->base);
> struct intel_hdmi *intel_hdmi = &intel_dig_port->hdmi;
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> enum port port = encoder->port;
> - int level = intel_ddi_hdmi_level(dev_priv, port);
> + int level = intel_ddi_hdmi_level(i915, port);
> struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base);
>
> intel_dp_dual_mode_set_tmds_output(intel_hdmi, true);
> intel_ddi_clk_select(encoder, crtc_state);
>
> - intel_display_power_get(dev_priv, dig_port->ddi_io_power_domain);
> + intel_display_power_get(i915, dig_port->ddi_io_power_domain);
>
> icl_program_mg_dp_mode(dig_port);
> icl_disable_phy_clock_gating(dig_port);
>
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> icl_ddi_vswing_sequence(encoder, crtc_state->port_clock,
> level, INTEL_OUTPUT_HDMI);
> - else if (IS_CANNONLAKE(dev_priv))
> + else if (IS_CANNONLAKE(i915))
> cnl_ddi_vswing_sequence(encoder, level, INTEL_OUTPUT_HDMI);
> - else if (IS_GEN9_LP(dev_priv))
> + else if (IS_GEN9_LP(i915))
> bxt_ddi_vswing_sequence(encoder, level, INTEL_OUTPUT_HDMI);
> else
> intel_prepare_hdmi_ddi_buffers(encoder, level);
>
> icl_enable_phy_clock_gating(dig_port);
>
> - if (IS_GEN9_BC(dev_priv))
> + if (IS_GEN9_BC(i915))
> skl_ddi_set_iboost(encoder, level, INTEL_OUTPUT_HDMI);
>
> intel_ddi_enable_pipe_clock(crtc_state);
> @@ -3213,7 +3213,7 @@ static void intel_ddi_pre_enable(struct intel_encoder *encoder,
> const struct drm_connector_state *conn_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
>
> /*
> @@ -3231,10 +3231,10 @@ static void intel_ddi_pre_enable(struct intel_encoder *encoder,
>
> WARN_ON(crtc_state->has_pch_encoder);
>
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> icl_map_plls_to_ports(encoder, crtc_state);
>
> - intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, true);
> + intel_set_cpu_fifo_underrun_reporting(i915, pipe, true);
>
> if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_HDMI)) {
> intel_ddi_pre_enable_hdmi(encoder, crtc_state, conn_state);
> @@ -3257,7 +3257,7 @@ static void intel_ddi_pre_enable(struct intel_encoder *encoder,
> static void intel_disable_ddi_buf(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> enum port port = encoder->port;
> bool wait = false;
> u32 val;
> @@ -3278,14 +3278,14 @@ static void intel_disable_ddi_buf(struct intel_encoder *encoder,
> intel_ddi_disable_fec_state(encoder, crtc_state);
>
> if (wait)
> - intel_wait_ddi_buf_idle(dev_priv, port);
> + intel_wait_ddi_buf_idle(i915, port);
> }
>
> static void intel_ddi_post_disable_dp(struct intel_encoder *encoder,
> const struct intel_crtc_state *old_crtc_state,
> const struct drm_connector_state *old_conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base);
> struct intel_dp *intel_dp = &dig_port->dp;
> bool is_mst = intel_crtc_has_type(old_crtc_state,
> @@ -3305,7 +3305,7 @@ static void intel_ddi_post_disable_dp(struct intel_encoder *encoder,
> intel_edp_panel_vdd_on(intel_dp);
> intel_edp_panel_off(intel_dp);
>
> - intel_display_power_put_unchecked(dev_priv,
> + intel_display_power_put_unchecked(i915,
> dig_port->ddi_io_power_domain);
>
> intel_ddi_clk_disable(encoder);
> @@ -3315,7 +3315,7 @@ static void intel_ddi_post_disable_hdmi(struct intel_encoder *encoder,
> const struct intel_crtc_state *old_crtc_state,
> const struct drm_connector_state *old_conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base);
> struct intel_hdmi *intel_hdmi = &dig_port->hdmi;
>
> @@ -3326,7 +3326,7 @@ static void intel_ddi_post_disable_hdmi(struct intel_encoder *encoder,
>
> intel_disable_ddi_buf(encoder, old_crtc_state);
>
> - intel_display_power_put_unchecked(dev_priv,
> + intel_display_power_put_unchecked(i915,
> dig_port->ddi_io_power_domain);
>
> intel_ddi_clk_disable(encoder);
> @@ -3338,7 +3338,7 @@ static void intel_ddi_post_disable(struct intel_encoder *encoder,
> const struct intel_crtc_state *old_crtc_state,
> const struct drm_connector_state *old_conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
>
> /*
> * When called from DP MST code:
> @@ -3360,7 +3360,7 @@ static void intel_ddi_post_disable(struct intel_encoder *encoder,
> intel_ddi_post_disable_dp(encoder,
> old_crtc_state, old_conn_state);
>
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> icl_unmap_plls_to_ports(encoder);
> }
>
> @@ -3368,7 +3368,7 @@ void intel_ddi_fdi_post_disable(struct intel_encoder *encoder,
> const struct intel_crtc_state *old_crtc_state,
> const struct drm_connector_state *old_conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> u32 val;
>
> /*
> @@ -3402,11 +3402,11 @@ static void intel_enable_ddi_dp(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state,
> const struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> enum port port = encoder->port;
>
> - if (port == PORT_A && INTEL_GEN(dev_priv) < 9)
> + if (port == PORT_A && INTEL_GEN(i915) < 9)
> intel_dp_stop_link_train(intel_dp);
>
> intel_edp_backlight_on(crtc_state, conn_state);
> @@ -3419,7 +3419,7 @@ static void intel_enable_ddi_dp(struct intel_encoder *encoder,
> }
>
> static i915_reg_t
> -gen9_chicken_trans_reg_by_port(struct drm_i915_private *dev_priv,
> +gen9_chicken_trans_reg_by_port(struct drm_i915_private *i915,
> enum port port)
> {
> static const i915_reg_t regs[] = {
> @@ -3430,7 +3430,7 @@ gen9_chicken_trans_reg_by_port(struct drm_i915_private *dev_priv,
> [PORT_E] = CHICKEN_TRANS_A,
> };
>
> - WARN_ON(INTEL_GEN(dev_priv) < 9);
> + WARN_ON(INTEL_GEN(i915) < 9);
>
> if (WARN_ON(port < PORT_A || port > PORT_E))
> port = PORT_A;
> @@ -3442,7 +3442,7 @@ static void intel_enable_ddi_hdmi(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state,
> const struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base);
> struct drm_connector *connector = conn_state->connector;
> enum port port = encoder->port;
> @@ -3454,14 +3454,14 @@ static void intel_enable_ddi_hdmi(struct intel_encoder *encoder,
> connector->base.id, connector->name);
>
> /* Display WA #1143: skl,kbl,cfl */
> - if (IS_GEN9_BC(dev_priv)) {
> + if (IS_GEN9_BC(i915)) {
> /*
> * For some reason these chicken bits have been
> * stuffed into a transcoder register, event though
> * the bits affect a specific DDI port rather than
> * a specific transcoder.
> */
> - i915_reg_t reg = gen9_chicken_trans_reg_by_port(dev_priv, port);
> + i915_reg_t reg = gen9_chicken_trans_reg_by_port(i915, port);
> u32 val;
>
> val = I915_READ(reg);
> @@ -3595,9 +3595,9 @@ static void intel_ddi_set_fia_lane_count(struct intel_encoder *encoder,
> const struct intel_crtc_state *pipe_config,
> enum port port)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base);
> - enum tc_port tc_port = intel_port_to_tc(dev_priv, port);
> + enum tc_port tc_port = intel_port_to_tc(i915, port);
> u32 val = I915_READ(PORT_TX_DFLEXDPMLE1);
> bool lane_reversal = dig_port->saved_port_bits & DDI_BUF_PORT_REVERSAL;
>
> @@ -3625,16 +3625,16 @@ intel_ddi_pre_pll_enable(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state,
> const struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base);
> enum port port = encoder->port;
>
> if (intel_crtc_has_dp_encoder(crtc_state) ||
> - intel_port_is_tc(dev_priv, encoder->port))
> - intel_display_power_get(dev_priv,
> + intel_port_is_tc(i915, encoder->port))
> + intel_display_power_get(i915,
> intel_ddi_main_link_aux_domain(dig_port));
>
> - if (IS_GEN9_LP(dev_priv))
> + if (IS_GEN9_LP(i915))
> bxt_ddi_phy_set_lane_optim_mask(encoder,
> crtc_state->lane_lat_optim_mask);
>
> @@ -3654,19 +3654,19 @@ intel_ddi_post_pll_disable(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state,
> const struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base);
>
> if (intel_crtc_has_dp_encoder(crtc_state) ||
> - intel_port_is_tc(dev_priv, encoder->port))
> - intel_display_power_put_unchecked(dev_priv,
> + intel_port_is_tc(i915, encoder->port))
> + intel_display_power_put_unchecked(i915,
> intel_ddi_main_link_aux_domain(dig_port));
> }
>
> static void intel_ddi_prepare_link_retrain(struct intel_dp *intel_dp)
> {
> struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp);
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> to_i915(intel_dig_port->base.base.dev);
> enum port port = intel_dig_port->base.port;
> u32 val;
> @@ -3687,7 +3687,7 @@ static void intel_ddi_prepare_link_retrain(struct intel_dp *intel_dp)
> POSTING_READ(DP_TP_CTL(port));
>
> if (wait)
> - intel_wait_ddi_buf_idle(dev_priv, port);
> + intel_wait_ddi_buf_idle(i915, port);
> }
>
> val = DP_TP_CTL_ENABLE |
> @@ -3709,32 +3709,32 @@ static void intel_ddi_prepare_link_retrain(struct intel_dp *intel_dp)
> udelay(600);
> }
>
> -static bool intel_ddi_is_audio_enabled(struct drm_i915_private *dev_priv,
> +static bool intel_ddi_is_audio_enabled(struct drm_i915_private *i915,
> enum transcoder cpu_transcoder)
> {
> if (cpu_transcoder == TRANSCODER_EDP)
> return false;
>
> - if (!intel_display_power_is_enabled(dev_priv, POWER_DOMAIN_AUDIO))
> + if (!intel_display_power_is_enabled(i915, POWER_DOMAIN_AUDIO))
> return false;
>
> return I915_READ(HSW_AUD_PIN_ELD_CP_VLD) &
> AUDIO_OUTPUT_ENABLE(cpu_transcoder);
> }
>
> -void intel_ddi_compute_min_voltage_level(struct drm_i915_private *dev_priv,
> +void intel_ddi_compute_min_voltage_level(struct drm_i915_private *i915,
> struct intel_crtc_state *crtc_state)
> {
> - if (INTEL_GEN(dev_priv) >= 11 && crtc_state->port_clock > 594000)
> + if (INTEL_GEN(i915) >= 11 && crtc_state->port_clock > 594000)
> crtc_state->min_voltage_level = 1;
> - else if (IS_CANNONLAKE(dev_priv) && crtc_state->port_clock > 594000)
> + else if (IS_CANNONLAKE(i915) && crtc_state->port_clock > 594000)
> crtc_state->min_voltage_level = 2;
> }
>
> void intel_ddi_get_config(struct intel_encoder *encoder,
> struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(pipe_config->base.crtc);
> enum transcoder cpu_transcoder = pipe_config->cpu_transcoder;
> struct intel_digital_port *intel_dig_port;
> @@ -3816,10 +3816,10 @@ void intel_ddi_get_config(struct intel_encoder *encoder,
> }
>
> pipe_config->has_audio =
> - intel_ddi_is_audio_enabled(dev_priv, cpu_transcoder);
> + intel_ddi_is_audio_enabled(i915, cpu_transcoder);
>
> - if (encoder->type == INTEL_OUTPUT_EDP && dev_priv->vbt.edp.bpp &&
> - pipe_config->pipe_bpp > dev_priv->vbt.edp.bpp) {
> + if (encoder->type == INTEL_OUTPUT_EDP && i915->vbt.edp.bpp &&
> + pipe_config->pipe_bpp > i915->vbt.edp.bpp) {
> /*
> * This is a big fat ugly hack.
> *
> @@ -3834,17 +3834,17 @@ void intel_ddi_get_config(struct intel_encoder *encoder,
> * load.
> */
> DRM_DEBUG_KMS("pipe has %d bpp for eDP panel, overriding BIOS-provided max %d bpp\n",
> - pipe_config->pipe_bpp, dev_priv->vbt.edp.bpp);
> - dev_priv->vbt.edp.bpp = pipe_config->pipe_bpp;
> + pipe_config->pipe_bpp, i915->vbt.edp.bpp);
> + i915->vbt.edp.bpp = pipe_config->pipe_bpp;
> }
>
> intel_ddi_clock_get(encoder, pipe_config);
>
> - if (IS_GEN9_LP(dev_priv))
> + if (IS_GEN9_LP(i915))
> pipe_config->lane_lat_optim_mask =
> bxt_ddi_phy_get_lane_lat_optim_mask(encoder);
>
> - intel_ddi_compute_min_voltage_level(dev_priv, pipe_config);
> + intel_ddi_compute_min_voltage_level(i915, pipe_config);
>
> intel_hdmi_read_gcp_infoframe(encoder, pipe_config);
>
> @@ -3885,11 +3885,11 @@ static int intel_ddi_compute_config(struct intel_encoder *encoder,
> struct drm_connector_state *conn_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(pipe_config->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> enum port port = encoder->port;
> int ret;
>
> - if (HAS_TRANSCODER_EDP(dev_priv) && port == PORT_A)
> + if (HAS_TRANSCODER_EDP(i915) && port == PORT_A)
> pipe_config->cpu_transcoder = TRANSCODER_EDP;
>
> if (intel_crtc_has_type(pipe_config, INTEL_OUTPUT_HDMI))
> @@ -3899,17 +3899,17 @@ static int intel_ddi_compute_config(struct intel_encoder *encoder,
> if (ret)
> return ret;
>
> - if (IS_HASWELL(dev_priv) && crtc->pipe == PIPE_A &&
> + if (IS_HASWELL(i915) && crtc->pipe == PIPE_A &&
> pipe_config->cpu_transcoder == TRANSCODER_EDP)
> pipe_config->pch_pfit.force_thru =
> pipe_config->pch_pfit.enabled ||
> pipe_config->crc_enabled;
>
> - if (IS_GEN9_LP(dev_priv))
> + if (IS_GEN9_LP(i915))
> pipe_config->lane_lat_optim_mask =
> bxt_ddi_phy_calc_lane_lat_optim_mask(pipe_config->lane_count);
>
> - intel_ddi_compute_min_voltage_level(dev_priv, pipe_config);
> + intel_ddi_compute_min_voltage_level(i915, pipe_config);
>
> return 0;
> }
> @@ -4013,11 +4013,11 @@ static int modeset_pipe(struct drm_crtc *crtc,
> static int intel_hdmi_reset_link(struct intel_encoder *encoder,
> struct drm_modeset_acquire_ctx *ctx)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_hdmi *hdmi = enc_to_intel_hdmi(&encoder->base);
> struct intel_connector *connector = hdmi->attached_connector;
> struct i2c_adapter *adapter =
> - intel_gmbus_get_adapter(dev_priv, hdmi->ddc_bus);
> + intel_gmbus_get_adapter(i915, hdmi->ddc_bus);
> struct drm_connector_state *conn_state;
> struct intel_crtc_state *crtc_state;
> struct intel_crtc *crtc;
> @@ -4027,7 +4027,7 @@ static int intel_hdmi_reset_link(struct intel_encoder *encoder,
> if (!connector || connector->base.status != connector_status_connected)
> return 0;
>
> - ret = drm_modeset_lock(&dev_priv->drm.mode_config.connection_mutex,
> + ret = drm_modeset_lock(&i915->drm.mode_config.connection_mutex,
> ctx);
> if (ret)
> return ret;
> @@ -4131,7 +4131,7 @@ intel_ddi_init_hdmi_connector(struct intel_digital_port *intel_dig_port)
>
> static bool intel_ddi_a_force_4_lanes(struct intel_digital_port *dport)
> {
> - struct drm_i915_private *dev_priv = to_i915(dport->base.base.dev);
> + struct drm_i915_private *i915 = to_i915(dport->base.base.dev);
>
> if (dport->base.port != PORT_A)
> return false;
> @@ -4142,7 +4142,7 @@ static bool intel_ddi_a_force_4_lanes(struct intel_digital_port *dport)
> /* Broxton/Geminilake: Bspec says that DDI_A_4_LANES is the only
> * supported configuration
> */
> - if (IS_GEN9_LP(dev_priv))
> + if (IS_GEN9_LP(i915))
> return true;
>
> /* Cannonlake: Most of SKUs don't support DDI_E, and the only
> @@ -4150,8 +4150,8 @@ static bool intel_ddi_a_force_4_lanes(struct intel_digital_port *dport)
> * DDI_F what makes DDI_E useless. However for this
> * case let's trust VBT info.
> */
> - if (IS_CANNONLAKE(dev_priv) &&
> - !intel_bios_is_port_present(dev_priv, PORT_E))
> + if (IS_CANNONLAKE(i915) &&
> + !intel_bios_is_port_present(i915, PORT_E))
> return true;
>
> return false;
> @@ -4160,11 +4160,11 @@ static bool intel_ddi_a_force_4_lanes(struct intel_digital_port *dport)
> static int
> intel_ddi_max_lanes(struct intel_digital_port *intel_dport)
> {
> - struct drm_i915_private *dev_priv = to_i915(intel_dport->base.base.dev);
> + struct drm_i915_private *i915 = to_i915(intel_dport->base.base.dev);
> enum port port = intel_dport->base.port;
> int max_lanes = 4;
>
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> return max_lanes;
>
> if (port == PORT_A || port == PORT_E) {
> @@ -4189,10 +4189,10 @@ intel_ddi_max_lanes(struct intel_digital_port *intel_dport)
> return max_lanes;
> }
>
> -void intel_ddi_init(struct drm_i915_private *dev_priv, enum port port)
> +void intel_ddi_init(struct drm_i915_private *i915, enum port port)
> {
> struct ddi_vbt_port_info *port_info =
> - &dev_priv->vbt.ddi_port_info[port];
> + &i915->vbt.ddi_port_info[port];
> struct intel_digital_port *intel_dig_port;
> struct intel_encoder *intel_encoder;
> struct drm_encoder *encoder;
> @@ -4202,7 +4202,7 @@ void intel_ddi_init(struct drm_i915_private *dev_priv, enum port port)
> init_hdmi = port_info->supports_dvi || port_info->supports_hdmi;
> init_dp = port_info->supports_dp;
>
> - if (intel_bios_is_lspcon_present(dev_priv, port)) {
> + if (intel_bios_is_lspcon_present(i915, port)) {
> /*
> * Lspcon device needs to be driven with DP connector
> * with special detection sequence. So make sure DP
> @@ -4227,7 +4227,7 @@ void intel_ddi_init(struct drm_i915_private *dev_priv, enum port port)
> intel_encoder = &intel_dig_port->base;
> encoder = &intel_encoder->base;
>
> - drm_encoder_init(&dev_priv->drm, encoder, &intel_ddi_funcs,
> + drm_encoder_init(&i915->drm, encoder, &intel_ddi_funcs,
> DRM_MODE_ENCODER_TMDS, "DDI %c", port_name(port));
>
> intel_encoder->hotplug = intel_ddi_hotplug;
> @@ -4248,10 +4248,10 @@ void intel_ddi_init(struct drm_i915_private *dev_priv, enum port port)
> intel_encoder->power_domain = intel_port_to_power_domain(port);
> intel_encoder->port = port;
> intel_encoder->cloneable = 0;
> - for_each_pipe(dev_priv, pipe)
> + for_each_pipe(i915, pipe)
> intel_encoder->crtc_mask |= BIT(pipe);
>
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> intel_dig_port->saved_port_bits = I915_READ(DDI_BUF_CTL(port)) &
> DDI_BUF_PORT_REVERSAL;
> else
> @@ -4259,9 +4259,9 @@ void intel_ddi_init(struct drm_i915_private *dev_priv, enum port port)
> (DDI_BUF_PORT_REVERSAL | DDI_A_4_LANES);
> intel_dig_port->dp.output_reg = INVALID_MMIO_REG;
> intel_dig_port->max_lanes = intel_ddi_max_lanes(intel_dig_port);
> - intel_dig_port->aux_ch = intel_bios_port_aux_ch(dev_priv, port);
> + intel_dig_port->aux_ch = intel_bios_port_aux_ch(i915, port);
>
> - intel_dig_port->tc_legacy_port = intel_port_is_tc(dev_priv, port) &&
> + intel_dig_port->tc_legacy_port = intel_port_is_tc(i915, port) &&
> !port_info->supports_typec_usb &&
> !port_info->supports_tbt;
>
> @@ -4324,7 +4324,7 @@ void intel_ddi_init(struct drm_i915_private *dev_priv, enum port port)
>
> intel_infoframe_init(intel_dig_port);
>
> - if (intel_port_is_tc(dev_priv, port))
> + if (intel_port_is_tc(i915, port))
> intel_digital_port_connected(intel_encoder);
>
> return;
> diff --git a/drivers/gpu/drm/i915/intel_ddi.h b/drivers/gpu/drm/i915/intel_ddi.h
> index a08365da2643..50dd6fac4e3b 100644
> --- a/drivers/gpu/drm/i915/intel_ddi.h
> +++ b/drivers/gpu/drm/i915/intel_ddi.h
> @@ -24,7 +24,7 @@ void intel_ddi_fdi_post_disable(struct intel_encoder *intel_encoder,
> const struct drm_connector_state *old_conn_state);
> void hsw_fdi_link_train(struct intel_crtc *crtc,
> const struct intel_crtc_state *crtc_state);
> -void intel_ddi_init(struct drm_i915_private *dev_priv, enum port port);
> +void intel_ddi_init(struct drm_i915_private *i915, enum port port);
> bool intel_ddi_get_hw_state(struct intel_encoder *encoder, enum pipe *pipe);
> void intel_ddi_enable_transcoder_func(const struct intel_crtc_state *crtc_state);
> void intel_ddi_disable_transcoder_func(const struct intel_crtc_state *crtc_state);
> @@ -36,7 +36,7 @@ void intel_ddi_get_config(struct intel_encoder *encoder,
> struct intel_crtc_state *pipe_config);
> void intel_ddi_set_vc_payload_alloc(const struct intel_crtc_state *crtc_state,
> bool state);
> -void intel_ddi_compute_min_voltage_level(struct drm_i915_private *dev_priv,
> +void intel_ddi_compute_min_voltage_level(struct drm_i915_private *i915,
> struct intel_crtc_state *crtc_state);
> u32 bxt_signal_levels(struct intel_dp *intel_dp);
> u32 ddi_signal_levels(struct intel_dp *intel_dp);
> @@ -46,7 +46,7 @@ u8 intel_ddi_dp_pre_emphasis_max(struct intel_encoder *encoder,
> int intel_ddi_toggle_hdcp_signalling(struct intel_encoder *intel_encoder,
> bool enable);
> void icl_sanitize_encoder_pll_mapping(struct intel_encoder *encoder);
> -int cnl_calc_wrpll_link(struct drm_i915_private *dev_priv,
> +int cnl_calc_wrpll_link(struct drm_i915_private *i915,
> struct intel_dpll_hw_state *state);
>
> #endif /* __INTEL_DDI_H__ */
> diff --git a/drivers/gpu/drm/i915/intel_device_info.c b/drivers/gpu/drm/i915/intel_device_info.c
> index 7135d8dc32a7..a670881b3f1e 100644
> --- a/drivers/gpu/drm/i915/intel_device_info.c
> +++ b/drivers/gpu/drm/i915/intel_device_info.c
> @@ -182,15 +182,15 @@ static u16 compute_eu_total(const struct sseu_dev_info *sseu)
> return total;
> }
>
> -static void gen11_sseu_info_init(struct drm_i915_private *dev_priv)
> +static void gen11_sseu_info_init(struct drm_i915_private *i915)
> {
> - struct sseu_dev_info *sseu = &RUNTIME_INFO(dev_priv)->sseu;
> + struct sseu_dev_info *sseu = &RUNTIME_INFO(i915)->sseu;
> u8 s_en;
> u32 ss_en, ss_en_mask;
> u8 eu_en;
> int s;
>
> - if (IS_ELKHARTLAKE(dev_priv)) {
> + if (IS_ELKHARTLAKE(i915)) {
> sseu->max_slices = 1;
> sseu->max_subslices = 4;
> sseu->max_eus_per_subslice = 8;
> @@ -227,9 +227,9 @@ static void gen11_sseu_info_init(struct drm_i915_private *dev_priv)
> sseu->has_eu_pg = 1;
> }
>
> -static void gen10_sseu_info_init(struct drm_i915_private *dev_priv)
> +static void gen10_sseu_info_init(struct drm_i915_private *i915)
> {
> - struct sseu_dev_info *sseu = &RUNTIME_INFO(dev_priv)->sseu;
> + struct sseu_dev_info *sseu = &RUNTIME_INFO(i915)->sseu;
> const u32 fuse2 = I915_READ(GEN8_FUSE2);
> int s, ss;
> const int eu_mask = 0xff;
> @@ -305,9 +305,9 @@ static void gen10_sseu_info_init(struct drm_i915_private *dev_priv)
> sseu->has_eu_pg = 1;
> }
>
> -static void cherryview_sseu_info_init(struct drm_i915_private *dev_priv)
> +static void cherryview_sseu_info_init(struct drm_i915_private *i915)
> {
> - struct sseu_dev_info *sseu = &RUNTIME_INFO(dev_priv)->sseu;
> + struct sseu_dev_info *sseu = &RUNTIME_INFO(i915)->sseu;
> u32 fuse;
>
> fuse = I915_READ(CHV_FUSE_GT);
> @@ -359,10 +359,10 @@ static void cherryview_sseu_info_init(struct drm_i915_private *dev_priv)
> sseu->has_eu_pg = (sseu->eu_per_subslice > 2);
> }
>
> -static void gen9_sseu_info_init(struct drm_i915_private *dev_priv)
> +static void gen9_sseu_info_init(struct drm_i915_private *i915)
> {
> - struct intel_device_info *info = mkwrite_device_info(dev_priv);
> - struct sseu_dev_info *sseu = &RUNTIME_INFO(dev_priv)->sseu;
> + struct intel_device_info *info = mkwrite_device_info(i915);
> + struct sseu_dev_info *sseu = &RUNTIME_INFO(i915)->sseu;
> int s, ss;
> u32 fuse2, eu_disable, subslice_mask;
> const u8 eu_mask = 0xff;
> @@ -371,8 +371,8 @@ static void gen9_sseu_info_init(struct drm_i915_private *dev_priv)
> sseu->slice_mask = (fuse2 & GEN8_F2_S_ENA_MASK) >> GEN8_F2_S_ENA_SHIFT;
>
> /* BXT has a single slice and at most 3 subslices. */
> - sseu->max_slices = IS_GEN9_LP(dev_priv) ? 1 : 3;
> - sseu->max_subslices = IS_GEN9_LP(dev_priv) ? 3 : 4;
> + sseu->max_slices = IS_GEN9_LP(i915) ? 1 : 3;
> + sseu->max_subslices = IS_GEN9_LP(i915) ? 3 : 4;
> sseu->max_eus_per_subslice = 8;
>
> /*
> @@ -442,12 +442,12 @@ static void gen9_sseu_info_init(struct drm_i915_private *dev_priv)
> * pair per subslice.
> */
> sseu->has_slice_pg =
> - !IS_GEN9_LP(dev_priv) && hweight8(sseu->slice_mask) > 1;
> + !IS_GEN9_LP(i915) && hweight8(sseu->slice_mask) > 1;
> sseu->has_subslice_pg =
> - IS_GEN9_LP(dev_priv) && intel_sseu_subslice_total(sseu) > 1;
> + IS_GEN9_LP(i915) && intel_sseu_subslice_total(sseu) > 1;
> sseu->has_eu_pg = sseu->eu_per_subslice > 2;
>
> - if (IS_GEN9_LP(dev_priv)) {
> + if (IS_GEN9_LP(i915)) {
> #define IS_SS_DISABLED(ss) (!(sseu->subslice_mask[0] & BIT(ss)))
> info->has_pooled_eu = hweight8(sseu->subslice_mask[0]) == 3;
>
> @@ -464,9 +464,9 @@ static void gen9_sseu_info_init(struct drm_i915_private *dev_priv)
> }
> }
>
> -static void broadwell_sseu_info_init(struct drm_i915_private *dev_priv)
> +static void broadwell_sseu_info_init(struct drm_i915_private *i915)
> {
> - struct sseu_dev_info *sseu = &RUNTIME_INFO(dev_priv)->sseu;
> + struct sseu_dev_info *sseu = &RUNTIME_INFO(i915)->sseu;
> int s, ss;
> u32 fuse2, subslice_mask, eu_disable[3]; /* s_max */
>
> @@ -547,9 +547,9 @@ static void broadwell_sseu_info_init(struct drm_i915_private *dev_priv)
> sseu->has_eu_pg = 0;
> }
>
> -static void haswell_sseu_info_init(struct drm_i915_private *dev_priv)
> +static void haswell_sseu_info_init(struct drm_i915_private *i915)
> {
> - struct sseu_dev_info *sseu = &RUNTIME_INFO(dev_priv)->sseu;
> + struct sseu_dev_info *sseu = &RUNTIME_INFO(i915)->sseu;
> u32 fuse1;
> int s, ss;
>
> @@ -557,9 +557,9 @@ static void haswell_sseu_info_init(struct drm_i915_private *dev_priv)
> * There isn't a register to tell us how many slices/subslices. We
> * work off the PCI-ids here.
> */
> - switch (INTEL_INFO(dev_priv)->gt) {
> + switch (INTEL_INFO(i915)->gt) {
> default:
> - MISSING_CASE(INTEL_INFO(dev_priv)->gt);
> + MISSING_CASE(INTEL_INFO(i915)->gt);
> /* fall through */
> case 1:
> sseu->slice_mask = BIT(0);
> @@ -612,7 +612,7 @@ static void haswell_sseu_info_init(struct drm_i915_private *dev_priv)
> sseu->has_eu_pg = 0;
> }
>
> -static u32 read_reference_ts_freq(struct drm_i915_private *dev_priv)
> +static u32 read_reference_ts_freq(struct drm_i915_private *i915)
> {
> u32 ts_override = I915_READ(GEN9_TIMESTAMP_OVERRIDE);
> u32 base_freq, frac_freq;
> @@ -629,7 +629,7 @@ static u32 read_reference_ts_freq(struct drm_i915_private *dev_priv)
> return base_freq + frac_freq;
> }
>
> -static u32 gen10_get_crystal_clock_freq(struct drm_i915_private *dev_priv,
> +static u32 gen10_get_crystal_clock_freq(struct drm_i915_private *i915,
> u32 rpm_config_reg)
> {
> u32 f19_2_mhz = 19200;
> @@ -649,7 +649,7 @@ static u32 gen10_get_crystal_clock_freq(struct drm_i915_private *dev_priv,
> }
> }
>
> -static u32 gen11_get_crystal_clock_freq(struct drm_i915_private *dev_priv,
> +static u32 gen11_get_crystal_clock_freq(struct drm_i915_private *i915,
> u32 rpm_config_reg)
> {
> u32 f19_2_mhz = 19200;
> @@ -675,21 +675,21 @@ static u32 gen11_get_crystal_clock_freq(struct drm_i915_private *dev_priv,
> }
> }
>
> -static u32 read_timestamp_frequency(struct drm_i915_private *dev_priv)
> +static u32 read_timestamp_frequency(struct drm_i915_private *i915)
> {
> u32 f12_5_mhz = 12500;
> u32 f19_2_mhz = 19200;
> u32 f24_mhz = 24000;
>
> - if (INTEL_GEN(dev_priv) <= 4) {
> + if (INTEL_GEN(i915) <= 4) {
> /* PRMs say:
> *
> * "The value in this register increments once every 16
> * hclks." (through the “Clocking Configuration”
> * (“CLKCFG”) MCHBAR register)
> */
> - return dev_priv->rawclk_freq / 16;
> - } else if (INTEL_GEN(dev_priv) <= 8) {
> + return i915->rawclk_freq / 16;
> + } else if (INTEL_GEN(i915) <= 8) {
> /* PRMs say:
> *
> * "The PCU TSC counts 10ns increments; this timestamp
> @@ -697,14 +697,14 @@ static u32 read_timestamp_frequency(struct drm_i915_private *dev_priv)
> * rolling over every 1.5 hours).
> */
> return f12_5_mhz;
> - } else if (INTEL_GEN(dev_priv) <= 9) {
> + } else if (INTEL_GEN(i915) <= 9) {
> u32 ctc_reg = I915_READ(CTC_MODE);
> u32 freq = 0;
>
> if ((ctc_reg & CTC_SOURCE_PARAMETER_MASK) == CTC_SOURCE_DIVIDE_LOGIC) {
> - freq = read_reference_ts_freq(dev_priv);
> + freq = read_reference_ts_freq(i915);
> } else {
> - freq = IS_GEN9_LP(dev_priv) ? f19_2_mhz : f24_mhz;
> + freq = IS_GEN9_LP(i915) ? f19_2_mhz : f24_mhz;
>
> /* Now figure out how the command stream's timestamp
> * register increments from this frequency (it might
> @@ -715,7 +715,7 @@ static u32 read_timestamp_frequency(struct drm_i915_private *dev_priv)
> }
>
> return freq;
> - } else if (INTEL_GEN(dev_priv) <= 11) {
> + } else if (INTEL_GEN(i915) <= 11) {
> u32 ctc_reg = I915_READ(CTC_MODE);
> u32 freq = 0;
>
> @@ -725,15 +725,15 @@ static u32 read_timestamp_frequency(struct drm_i915_private *dev_priv)
> * tells us which one we should use.
> */
> if ((ctc_reg & CTC_SOURCE_PARAMETER_MASK) == CTC_SOURCE_DIVIDE_LOGIC) {
> - freq = read_reference_ts_freq(dev_priv);
> + freq = read_reference_ts_freq(i915);
> } else {
> u32 rpm_config_reg = I915_READ(RPM_CONFIG0);
>
> - if (INTEL_GEN(dev_priv) <= 10)
> - freq = gen10_get_crystal_clock_freq(dev_priv,
> + if (INTEL_GEN(i915) <= 10)
> + freq = gen10_get_crystal_clock_freq(i915,
> rpm_config_reg);
> else
> - freq = gen11_get_crystal_clock_freq(dev_priv,
> + freq = gen11_get_crystal_clock_freq(i915,
> rpm_config_reg);
>
> /* Now figure out how the command stream's timestamp
> @@ -841,7 +841,7 @@ void intel_device_info_subplatform_init(struct drm_i915_private *i915)
>
> /**
> * intel_device_info_runtime_init - initialize runtime info
> - * @dev_priv: the i915 device
> + * @i915: the i915 device
> *
> * Determine various intel_device_info fields at runtime.
> *
> @@ -855,16 +855,16 @@ void intel_device_info_subplatform_init(struct drm_i915_private *i915)
> * - after the PCH has been detected,
> * - before the first usage of the fields it can tweak.
> */
> -void intel_device_info_runtime_init(struct drm_i915_private *dev_priv)
> +void intel_device_info_runtime_init(struct drm_i915_private *i915)
> {
> - struct intel_device_info *info = mkwrite_device_info(dev_priv);
> - struct intel_runtime_info *runtime = RUNTIME_INFO(dev_priv);
> + struct intel_device_info *info = mkwrite_device_info(i915);
> + struct intel_runtime_info *runtime = RUNTIME_INFO(i915);
> enum pipe pipe;
>
> - if (INTEL_GEN(dev_priv) >= 10) {
> - for_each_pipe(dev_priv, pipe)
> + if (INTEL_GEN(i915) >= 10) {
> + for_each_pipe(i915, pipe)
> runtime->num_scalers[pipe] = 2;
> - } else if (IS_GEN(dev_priv, 9)) {
> + } else if (IS_GEN(i915, 9)) {
> runtime->num_scalers[PIPE_A] = 2;
> runtime->num_scalers[PIPE_B] = 2;
> runtime->num_scalers[PIPE_C] = 1;
> @@ -872,13 +872,13 @@ void intel_device_info_runtime_init(struct drm_i915_private *dev_priv)
>
> BUILD_BUG_ON(BITS_PER_TYPE(intel_engine_mask_t) < I915_NUM_ENGINES);
>
> - if (INTEL_GEN(dev_priv) >= 11)
> - for_each_pipe(dev_priv, pipe)
> + if (INTEL_GEN(i915) >= 11)
> + for_each_pipe(i915, pipe)
> runtime->num_sprites[pipe] = 6;
> - else if (IS_GEN(dev_priv, 10) || IS_GEMINILAKE(dev_priv))
> - for_each_pipe(dev_priv, pipe)
> + else if (IS_GEN(i915, 10) || IS_GEMINILAKE(i915))
> + for_each_pipe(i915, pipe)
> runtime->num_sprites[pipe] = 3;
> - else if (IS_BROXTON(dev_priv)) {
> + else if (IS_BROXTON(i915)) {
> /*
> * Skylake and Broxton currently don't expose the topmost plane as its
> * use is exclusive with the legacy cursor and we only want to expose
> @@ -891,20 +891,20 @@ void intel_device_info_runtime_init(struct drm_i915_private *dev_priv)
> runtime->num_sprites[PIPE_A] = 2;
> runtime->num_sprites[PIPE_B] = 2;
> runtime->num_sprites[PIPE_C] = 1;
> - } else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> - for_each_pipe(dev_priv, pipe)
> + } else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> + for_each_pipe(i915, pipe)
> runtime->num_sprites[pipe] = 2;
> - } else if (INTEL_GEN(dev_priv) >= 5 || IS_G4X(dev_priv)) {
> - for_each_pipe(dev_priv, pipe)
> + } else if (INTEL_GEN(i915) >= 5 || IS_G4X(i915)) {
> + for_each_pipe(i915, pipe)
> runtime->num_sprites[pipe] = 1;
> }
>
> if (i915_modparams.disable_display) {
> DRM_INFO("Display disabled (module parameter)\n");
> info->num_pipes = 0;
> - } else if (HAS_DISPLAY(dev_priv) &&
> - (IS_GEN_RANGE(dev_priv, 7, 8)) &&
> - HAS_PCH_SPLIT(dev_priv)) {
> + } else if (HAS_DISPLAY(i915) &&
> + (IS_GEN_RANGE(i915, 7, 8)) &&
> + HAS_PCH_SPLIT(i915)) {
> u32 fuse_strap = I915_READ(FUSE_STRAP);
> u32 sfuse_strap = I915_READ(SFUSE_STRAP);
>
> @@ -919,7 +919,7 @@ void intel_device_info_runtime_init(struct drm_i915_private *dev_priv)
> */
> if (fuse_strap & ILK_INTERNAL_DISPLAY_DISABLE ||
> sfuse_strap & SFUSE_STRAP_DISPLAY_DISABLED ||
> - (HAS_PCH_CPT(dev_priv) &&
> + (HAS_PCH_CPT(i915) &&
> !(sfuse_strap & SFUSE_STRAP_FUSE_LOCK))) {
> DRM_INFO("Display fused off, disabling\n");
> info->num_pipes = 0;
> @@ -927,7 +927,7 @@ void intel_device_info_runtime_init(struct drm_i915_private *dev_priv)
> DRM_INFO("PipeC fused off\n");
> info->num_pipes -= 1;
> }
> - } else if (HAS_DISPLAY(dev_priv) && INTEL_GEN(dev_priv) >= 9) {
> + } else if (HAS_DISPLAY(i915) && INTEL_GEN(i915) >= 9) {
> u32 dfsm = I915_READ(SKL_DFSM);
> u8 disabled_mask = 0;
> bool invalid;
> @@ -961,26 +961,26 @@ void intel_device_info_runtime_init(struct drm_i915_private *dev_priv)
> }
>
> /* Initialize slice/subslice/EU info */
> - if (IS_HASWELL(dev_priv))
> - haswell_sseu_info_init(dev_priv);
> - else if (IS_CHERRYVIEW(dev_priv))
> - cherryview_sseu_info_init(dev_priv);
> - else if (IS_BROADWELL(dev_priv))
> - broadwell_sseu_info_init(dev_priv);
> - else if (IS_GEN(dev_priv, 9))
> - gen9_sseu_info_init(dev_priv);
> - else if (IS_GEN(dev_priv, 10))
> - gen10_sseu_info_init(dev_priv);
> - else if (INTEL_GEN(dev_priv) >= 11)
> - gen11_sseu_info_init(dev_priv);
> -
> - if (IS_GEN(dev_priv, 6) && intel_vtd_active()) {
> + if (IS_HASWELL(i915))
> + haswell_sseu_info_init(i915);
> + else if (IS_CHERRYVIEW(i915))
> + cherryview_sseu_info_init(i915);
> + else if (IS_BROADWELL(i915))
> + broadwell_sseu_info_init(i915);
> + else if (IS_GEN(i915, 9))
> + gen9_sseu_info_init(i915);
> + else if (IS_GEN(i915, 10))
> + gen10_sseu_info_init(i915);
> + else if (INTEL_GEN(i915) >= 11)
> + gen11_sseu_info_init(i915);
> +
> + if (IS_GEN(i915, 6) && intel_vtd_active()) {
> DRM_INFO("Disabling ppGTT for VT-d support\n");
> info->ppgtt_type = INTEL_PPGTT_NONE;
> }
>
> /* Initialize command stream timestamp frequency */
> - runtime->cs_timestamp_frequency_khz = read_timestamp_frequency(dev_priv);
> + runtime->cs_timestamp_frequency_khz = read_timestamp_frequency(i915);
> }
>
> void intel_driver_caps_print(const struct intel_driver_caps *caps,
> @@ -997,16 +997,16 @@ void intel_driver_caps_print(const struct intel_driver_caps *caps,
> * this point (but later we need to prune the forcewake domains for engines that
> * are indeed fused off).
> */
> -void intel_device_info_init_mmio(struct drm_i915_private *dev_priv)
> +void intel_device_info_init_mmio(struct drm_i915_private *i915)
> {
> - struct intel_device_info *info = mkwrite_device_info(dev_priv);
> + struct intel_device_info *info = mkwrite_device_info(i915);
> unsigned int logical_vdbox = 0;
> unsigned int i;
> u32 media_fuse;
> u16 vdbox_mask;
> u16 vebox_mask;
>
> - if (INTEL_GEN(dev_priv) < 11)
> + if (INTEL_GEN(i915) < 11)
> return;
>
> media_fuse = ~I915_READ(GEN11_GT_VEBOX_VDBOX_DISABLE);
> @@ -1016,7 +1016,7 @@ void intel_device_info_init_mmio(struct drm_i915_private *dev_priv)
> GEN11_GT_VEBOX_DISABLE_SHIFT;
>
> for (i = 0; i < I915_MAX_VCS; i++) {
> - if (!HAS_ENGINE(dev_priv, _VCS(i)))
> + if (!HAS_ENGINE(i915, _VCS(i)))
> continue;
>
> if (!(BIT(i) & vdbox_mask)) {
> @@ -1030,14 +1030,14 @@ void intel_device_info_init_mmio(struct drm_i915_private *dev_priv)
> * hooked up to an SFC (Scaler & Format Converter) unit.
> */
> if (logical_vdbox++ % 2 == 0)
> - RUNTIME_INFO(dev_priv)->vdbox_sfc_access |= BIT(i);
> + RUNTIME_INFO(i915)->vdbox_sfc_access |= BIT(i);
> }
> DRM_DEBUG_DRIVER("vdbox enable: %04x, instances: %04lx\n",
> - vdbox_mask, VDBOX_MASK(dev_priv));
> - GEM_BUG_ON(vdbox_mask != VDBOX_MASK(dev_priv));
> + vdbox_mask, VDBOX_MASK(i915));
> + GEM_BUG_ON(vdbox_mask != VDBOX_MASK(i915));
>
> for (i = 0; i < I915_MAX_VECS; i++) {
> - if (!HAS_ENGINE(dev_priv, _VECS(i)))
> + if (!HAS_ENGINE(i915, _VECS(i)))
> continue;
>
> if (!(BIT(i) & vebox_mask)) {
> @@ -1046,6 +1046,6 @@ void intel_device_info_init_mmio(struct drm_i915_private *dev_priv)
> }
> }
> DRM_DEBUG_DRIVER("vebox enable: %04x, instances: %04lx\n",
> - vebox_mask, VEBOX_MASK(dev_priv));
> - GEM_BUG_ON(vebox_mask != VEBOX_MASK(dev_priv));
> + vebox_mask, VEBOX_MASK(i915));
> + GEM_BUG_ON(vebox_mask != VEBOX_MASK(i915));
> }
> diff --git a/drivers/gpu/drm/i915/intel_device_info.h b/drivers/gpu/drm/i915/intel_device_info.h
> index 3ea953a230b3..93e5cc95e5c0 100644
> --- a/drivers/gpu/drm/i915/intel_device_info.h
> +++ b/drivers/gpu/drm/i915/intel_device_info.h
> @@ -218,8 +218,8 @@ struct intel_driver_caps {
>
> const char *intel_platform_name(enum intel_platform platform);
>
> -void intel_device_info_subplatform_init(struct drm_i915_private *dev_priv);
> -void intel_device_info_runtime_init(struct drm_i915_private *dev_priv);
> +void intel_device_info_subplatform_init(struct drm_i915_private *i915);
> +void intel_device_info_runtime_init(struct drm_i915_private *i915);
> void intel_device_info_dump_flags(const struct intel_device_info *info,
> struct drm_printer *p);
> void intel_device_info_dump_runtime(const struct intel_runtime_info *info,
> @@ -227,7 +227,7 @@ void intel_device_info_dump_runtime(const struct intel_runtime_info *info,
> void intel_device_info_dump_topology(const struct sseu_dev_info *sseu,
> struct drm_printer *p);
>
> -void intel_device_info_init_mmio(struct drm_i915_private *dev_priv);
> +void intel_device_info_init_mmio(struct drm_i915_private *i915);
>
> void intel_driver_caps_print(const struct intel_driver_caps *caps,
> struct drm_printer *p);
> diff --git a/drivers/gpu/drm/i915/intel_display.c b/drivers/gpu/drm/i915/intel_display.c
> index 1b1ddb48ca7a..be26ad9fc194 100644
> --- a/drivers/gpu/drm/i915/intel_display.c
> +++ b/drivers/gpu/drm/i915/intel_display.c
> @@ -156,24 +156,24 @@ struct intel_limit {
> };
>
> /* returns HPLL frequency in kHz */
> -int vlv_get_hpll_vco(struct drm_i915_private *dev_priv)
> +int vlv_get_hpll_vco(struct drm_i915_private *i915)
> {
> int hpll_freq, vco_freq[] = { 800, 1600, 2000, 2400 };
>
> /* Obtain SKU information */
> - hpll_freq = vlv_cck_read(dev_priv, CCK_FUSE_REG) &
> + hpll_freq = vlv_cck_read(i915, CCK_FUSE_REG) &
> CCK_FUSE_HPLL_FREQ_MASK;
>
> return vco_freq[hpll_freq] * 1000;
> }
>
> -int vlv_get_cck_clock(struct drm_i915_private *dev_priv,
> +int vlv_get_cck_clock(struct drm_i915_private *i915,
> const char *name, u32 reg, int ref_freq)
> {
> u32 val;
> int divider;
>
> - val = vlv_cck_read(dev_priv, reg);
> + val = vlv_cck_read(i915, reg);
> divider = val & CCK_FREQUENCY_VALUES;
>
> WARN((val & CCK_FREQUENCY_STATUS) !=
> @@ -183,42 +183,42 @@ int vlv_get_cck_clock(struct drm_i915_private *dev_priv,
> return DIV_ROUND_CLOSEST(ref_freq << 1, divider + 1);
> }
>
> -int vlv_get_cck_clock_hpll(struct drm_i915_private *dev_priv,
> +int vlv_get_cck_clock_hpll(struct drm_i915_private *i915,
> const char *name, u32 reg)
> {
> int hpll;
>
> - vlv_cck_get(dev_priv);
> + vlv_cck_get(i915);
>
> - if (dev_priv->hpll_freq == 0)
> - dev_priv->hpll_freq = vlv_get_hpll_vco(dev_priv);
> + if (i915->hpll_freq == 0)
> + i915->hpll_freq = vlv_get_hpll_vco(i915);
>
> - hpll = vlv_get_cck_clock(dev_priv, name, reg, dev_priv->hpll_freq);
> + hpll = vlv_get_cck_clock(i915, name, reg, i915->hpll_freq);
>
> - vlv_cck_put(dev_priv);
> + vlv_cck_put(i915);
>
> return hpll;
> }
>
> -static void intel_update_czclk(struct drm_i915_private *dev_priv)
> +static void intel_update_czclk(struct drm_i915_private *i915)
> {
> - if (!(IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)))
> + if (!(IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)))
> return;
>
> - dev_priv->czclk_freq = vlv_get_cck_clock_hpll(dev_priv, "czclk",
> + i915->czclk_freq = vlv_get_cck_clock_hpll(i915, "czclk",
> CCK_CZ_CLOCK_CONTROL);
>
> - DRM_DEBUG_DRIVER("CZ clock rate: %d kHz\n", dev_priv->czclk_freq);
> + DRM_DEBUG_DRIVER("CZ clock rate: %d kHz\n", i915->czclk_freq);
> }
>
> static inline u32 /* units of 100MHz */
> -intel_fdi_link_freq(struct drm_i915_private *dev_priv,
> +intel_fdi_link_freq(struct drm_i915_private *i915,
> const struct intel_crtc_state *pipe_config)
> {
> - if (HAS_DDI(dev_priv))
> + if (HAS_DDI(i915))
> return pipe_config->port_clock; /* SPLL */
> else
> - return dev_priv->fdi_pll_freq;
> + return i915->fdi_pll_freq;
> }
>
> static const struct intel_limit intel_limits_i8xx_dac = {
> @@ -488,7 +488,7 @@ static const struct intel_limit intel_limits_bxt = {
>
> /* WA Display #0827: Gen9:all */
> static void
> -skl_wa_827(struct drm_i915_private *dev_priv, int pipe, bool enable)
> +skl_wa_827(struct drm_i915_private *i915, int pipe, bool enable)
> {
> if (enable)
> I915_WRITE(CLKGATE_DIS_PSL(pipe),
> @@ -502,7 +502,7 @@ skl_wa_827(struct drm_i915_private *dev_priv, int pipe, bool enable)
>
> /* Wa_2006604312:icl */
> static void
> -icl_wa_scalerclkgating(struct drm_i915_private *dev_priv, enum pipe pipe,
> +icl_wa_scalerclkgating(struct drm_i915_private *i915, enum pipe pipe,
> bool enable)
> {
> if (enable)
> @@ -588,7 +588,7 @@ int chv_calc_dpll_params(int refclk, struct dpll *clock)
> * Returns whether the given set of divisors are valid for a given refclk with
> * the given connectors.
> */
> -static bool intel_PLL_is_valid(struct drm_i915_private *dev_priv,
> +static bool intel_PLL_is_valid(struct drm_i915_private *i915,
> const struct intel_limit *limit,
> const struct dpll *clock)
> {
> @@ -601,13 +601,13 @@ static bool intel_PLL_is_valid(struct drm_i915_private *dev_priv,
> if (clock->m1 < limit->m1.min || limit->m1.max < clock->m1)
> INTELPllInvalid("m1 out of range\n");
>
> - if (!IS_PINEVIEW(dev_priv) && !IS_VALLEYVIEW(dev_priv) &&
> - !IS_CHERRYVIEW(dev_priv) && !IS_GEN9_LP(dev_priv))
> + if (!IS_PINEVIEW(i915) && !IS_VALLEYVIEW(i915) &&
> + !IS_CHERRYVIEW(i915) && !IS_GEN9_LP(i915))
> if (clock->m1 <= clock->m2)
> INTELPllInvalid("m1 <= m2\n");
>
> - if (!IS_VALLEYVIEW(dev_priv) && !IS_CHERRYVIEW(dev_priv) &&
> - !IS_GEN9_LP(dev_priv)) {
> + if (!IS_VALLEYVIEW(i915) && !IS_CHERRYVIEW(i915) &&
> + !IS_GEN9_LP(i915)) {
> if (clock->p < limit->p.min || limit->p.max < clock->p)
> INTELPllInvalid("p out of range\n");
> if (clock->m < limit->m.min || limit->m.max < clock->m)
> @@ -630,7 +630,7 @@ i9xx_select_p2_div(const struct intel_limit *limit,
> const struct intel_crtc_state *crtc_state,
> int target)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
>
> if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS)) {
> /*
> @@ -638,7 +638,7 @@ i9xx_select_p2_div(const struct intel_limit *limit,
> * We haven't figured out how to reliably set up different
> * single/dual channel state, if we even can.
> */
> - if (intel_is_dual_link_lvds(dev_priv))
> + if (intel_is_dual_link_lvds(i915))
> return limit->p2.p2_fast;
> else
> return limit->p2.p2_slow;
> @@ -1016,22 +1016,22 @@ bool intel_crtc_active(struct intel_crtc *crtc)
> crtc->config->base.adjusted_mode.crtc_clock;
> }
>
> -enum transcoder intel_pipe_to_cpu_transcoder(struct drm_i915_private *dev_priv,
> +enum transcoder intel_pipe_to_cpu_transcoder(struct drm_i915_private *i915,
> enum pipe pipe)
> {
> - struct intel_crtc *crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> + struct intel_crtc *crtc = intel_get_crtc_for_pipe(i915, pipe);
>
> return crtc->config->cpu_transcoder;
> }
>
> -static bool pipe_scanline_is_moving(struct drm_i915_private *dev_priv,
> +static bool pipe_scanline_is_moving(struct drm_i915_private *i915,
> enum pipe pipe)
> {
> i915_reg_t reg = PIPEDSL(pipe);
> u32 line1, line2;
> u32 line_mask;
>
> - if (IS_GEN(dev_priv, 2))
> + if (IS_GEN(i915, 2))
> line_mask = DSL_LINEMASK_GEN2;
> else
> line_mask = DSL_LINEMASK_GEN3;
> @@ -1045,11 +1045,11 @@ static bool pipe_scanline_is_moving(struct drm_i915_private *dev_priv,
>
> static void wait_for_pipe_scanline_moving(struct intel_crtc *crtc, bool state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
>
> /* Wait for the display line to settle/start moving */
> - if (wait_for(pipe_scanline_is_moving(dev_priv, pipe) == state, 100))
> + if (wait_for(pipe_scanline_is_moving(i915, pipe) == state, 100))
> DRM_ERROR("pipe %c scanline %s wait timed out\n",
> pipe_name(pipe), onoff(state));
> }
> @@ -1068,14 +1068,14 @@ static void
> intel_wait_for_pipe_off(const struct intel_crtc_state *old_crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
>
> - if (INTEL_GEN(dev_priv) >= 4) {
> + if (INTEL_GEN(i915) >= 4) {
> enum transcoder cpu_transcoder = old_crtc_state->cpu_transcoder;
> i915_reg_t reg = PIPECONF(cpu_transcoder);
>
> /* Wait for the Pipe State to go off */
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> reg, I965_PIPECONF_ACTIVE, 0,
> 100))
> WARN(1, "pipe_off wait timed out\n");
> @@ -1085,7 +1085,7 @@ intel_wait_for_pipe_off(const struct intel_crtc_state *old_crtc_state)
> }
>
> /* Only for pre-ILK configs */
> -void assert_pll(struct drm_i915_private *dev_priv,
> +void assert_pll(struct drm_i915_private *i915,
> enum pipe pipe, bool state)
> {
> u32 val;
> @@ -1099,14 +1099,14 @@ void assert_pll(struct drm_i915_private *dev_priv,
> }
>
> /* XXX: the dsi pll is shared between MIPI DSI ports */
> -void assert_dsi_pll(struct drm_i915_private *dev_priv, bool state)
> +void assert_dsi_pll(struct drm_i915_private *i915, bool state)
> {
> u32 val;
> bool cur_state;
>
> - vlv_cck_get(dev_priv);
> - val = vlv_cck_read(dev_priv, CCK_REG_DSI_PLL_CONTROL);
> - vlv_cck_put(dev_priv);
> + vlv_cck_get(i915);
> + val = vlv_cck_read(i915, CCK_REG_DSI_PLL_CONTROL);
> + vlv_cck_put(i915);
>
> cur_state = val & DSI_PLL_VCO_EN;
> I915_STATE_WARN(cur_state != state,
> @@ -1114,14 +1114,14 @@ void assert_dsi_pll(struct drm_i915_private *dev_priv, bool state)
> onoff(state), onoff(cur_state));
> }
>
> -static void assert_fdi_tx(struct drm_i915_private *dev_priv,
> +static void assert_fdi_tx(struct drm_i915_private *i915,
> enum pipe pipe, bool state)
> {
> bool cur_state;
> - enum transcoder cpu_transcoder = intel_pipe_to_cpu_transcoder(dev_priv,
> + enum transcoder cpu_transcoder = intel_pipe_to_cpu_transcoder(i915,
> pipe);
>
> - if (HAS_DDI(dev_priv)) {
> + if (HAS_DDI(i915)) {
> /* DDI does not have a specific FDI_TX register */
> u32 val = I915_READ(TRANS_DDI_FUNC_CTL(cpu_transcoder));
> cur_state = !!(val & TRANS_DDI_FUNC_ENABLE);
> @@ -1136,7 +1136,7 @@ static void assert_fdi_tx(struct drm_i915_private *dev_priv,
> #define assert_fdi_tx_enabled(d, p) assert_fdi_tx(d, p, true)
> #define assert_fdi_tx_disabled(d, p) assert_fdi_tx(d, p, false)
>
> -static void assert_fdi_rx(struct drm_i915_private *dev_priv,
> +static void assert_fdi_rx(struct drm_i915_private *i915,
> enum pipe pipe, bool state)
> {
> u32 val;
> @@ -1151,24 +1151,24 @@ static void assert_fdi_rx(struct drm_i915_private *dev_priv,
> #define assert_fdi_rx_enabled(d, p) assert_fdi_rx(d, p, true)
> #define assert_fdi_rx_disabled(d, p) assert_fdi_rx(d, p, false)
>
> -static void assert_fdi_tx_pll_enabled(struct drm_i915_private *dev_priv,
> +static void assert_fdi_tx_pll_enabled(struct drm_i915_private *i915,
> enum pipe pipe)
> {
> u32 val;
>
> /* ILK FDI PLL is always enabled */
> - if (IS_GEN(dev_priv, 5))
> + if (IS_GEN(i915, 5))
> return;
>
> /* On Haswell, DDI ports are responsible for the FDI PLL setup */
> - if (HAS_DDI(dev_priv))
> + if (HAS_DDI(i915))
> return;
>
> val = I915_READ(FDI_TX_CTL(pipe));
> I915_STATE_WARN(!(val & FDI_TX_PLL_ENABLE), "FDI TX PLL assertion failure, should be active but is disabled\n");
> }
>
> -void assert_fdi_rx_pll(struct drm_i915_private *dev_priv,
> +void assert_fdi_rx_pll(struct drm_i915_private *i915,
> enum pipe pipe, bool state)
> {
> u32 val;
> @@ -1181,17 +1181,17 @@ void assert_fdi_rx_pll(struct drm_i915_private *dev_priv,
> onoff(state), onoff(cur_state));
> }
>
> -void assert_panel_unlocked(struct drm_i915_private *dev_priv, enum pipe pipe)
> +void assert_panel_unlocked(struct drm_i915_private *i915, enum pipe pipe)
> {
> i915_reg_t pp_reg;
> u32 val;
> enum pipe panel_pipe = INVALID_PIPE;
> bool locked = true;
>
> - if (WARN_ON(HAS_DDI(dev_priv)))
> + if (WARN_ON(HAS_DDI(i915)))
> return;
>
> - if (HAS_PCH_SPLIT(dev_priv)) {
> + if (HAS_PCH_SPLIT(i915)) {
> u32 port_sel;
>
> pp_reg = PP_CONTROL(0);
> @@ -1199,22 +1199,22 @@ void assert_panel_unlocked(struct drm_i915_private *dev_priv, enum pipe pipe)
>
> switch (port_sel) {
> case PANEL_PORT_SELECT_LVDS:
> - intel_lvds_port_enabled(dev_priv, PCH_LVDS, &panel_pipe);
> + intel_lvds_port_enabled(i915, PCH_LVDS, &panel_pipe);
> break;
> case PANEL_PORT_SELECT_DPA:
> - intel_dp_port_enabled(dev_priv, DP_A, PORT_A, &panel_pipe);
> + intel_dp_port_enabled(i915, DP_A, PORT_A, &panel_pipe);
> break;
> case PANEL_PORT_SELECT_DPC:
> - intel_dp_port_enabled(dev_priv, PCH_DP_C, PORT_C, &panel_pipe);
> + intel_dp_port_enabled(i915, PCH_DP_C, PORT_C, &panel_pipe);
> break;
> case PANEL_PORT_SELECT_DPD:
> - intel_dp_port_enabled(dev_priv, PCH_DP_D, PORT_D, &panel_pipe);
> + intel_dp_port_enabled(i915, PCH_DP_D, PORT_D, &panel_pipe);
> break;
> default:
> MISSING_CASE(port_sel);
> break;
> }
> - } else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> + } else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> /* presumably write lock depends on pipe, not port select */
> pp_reg = PP_CONTROL(pipe);
> panel_pipe = pipe;
> @@ -1225,7 +1225,7 @@ void assert_panel_unlocked(struct drm_i915_private *dev_priv, enum pipe pipe)
> port_sel = I915_READ(PP_ON_DELAYS(0)) & PANEL_PORT_SELECT_MASK;
>
> WARN_ON(port_sel != PANEL_PORT_SELECT_LVDS);
> - intel_lvds_port_enabled(dev_priv, LVDS, &panel_pipe);
> + intel_lvds_port_enabled(i915, LVDS, &panel_pipe);
> }
>
> val = I915_READ(pp_reg);
> @@ -1238,26 +1238,26 @@ void assert_panel_unlocked(struct drm_i915_private *dev_priv, enum pipe pipe)
> pipe_name(pipe));
> }
>
> -void assert_pipe(struct drm_i915_private *dev_priv,
> +void assert_pipe(struct drm_i915_private *i915,
> enum pipe pipe, bool state)
> {
> bool cur_state;
> - enum transcoder cpu_transcoder = intel_pipe_to_cpu_transcoder(dev_priv,
> + enum transcoder cpu_transcoder = intel_pipe_to_cpu_transcoder(i915,
> pipe);
> enum intel_display_power_domain power_domain;
> intel_wakeref_t wakeref;
>
> /* we keep both pipes enabled on 830 */
> - if (IS_I830(dev_priv))
> + if (IS_I830(i915))
> state = true;
>
> power_domain = POWER_DOMAIN_TRANSCODER(cpu_transcoder);
> - wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain);
> + wakeref = intel_display_power_get_if_enabled(i915, power_domain);
> if (wakeref) {
> u32 val = I915_READ(PIPECONF(cpu_transcoder));
> cur_state = !!(val & PIPECONF_ENABLE);
>
> - intel_display_power_put(dev_priv, power_domain, wakeref);
> + intel_display_power_put(i915, power_domain, wakeref);
> } else {
> cur_state = false;
> }
> @@ -1284,10 +1284,10 @@ static void assert_plane(struct intel_plane *plane, bool state)
>
> static void assert_planes_disabled(struct intel_crtc *crtc)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> struct intel_plane *plane;
>
> - for_each_intel_plane_on_crtc(&dev_priv->drm, crtc, plane)
> + for_each_intel_plane_on_crtc(&i915->drm, crtc, plane)
> assert_plane_disabled(plane);
> }
>
> @@ -1297,7 +1297,7 @@ static void assert_vblank_disabled(struct drm_crtc *crtc)
> drm_crtc_vblank_put(crtc);
> }
>
> -void assert_pch_transcoder_disabled(struct drm_i915_private *dev_priv,
> +void assert_pch_transcoder_disabled(struct drm_i915_private *i915,
> enum pipe pipe)
> {
> u32 val;
> @@ -1310,78 +1310,78 @@ void assert_pch_transcoder_disabled(struct drm_i915_private *dev_priv,
> pipe_name(pipe));
> }
>
> -static void assert_pch_dp_disabled(struct drm_i915_private *dev_priv,
> +static void assert_pch_dp_disabled(struct drm_i915_private *i915,
> enum pipe pipe, enum port port,
> i915_reg_t dp_reg)
> {
> enum pipe port_pipe;
> bool state;
>
> - state = intel_dp_port_enabled(dev_priv, dp_reg, port, &port_pipe);
> + state = intel_dp_port_enabled(i915, dp_reg, port, &port_pipe);
>
> I915_STATE_WARN(state && port_pipe == pipe,
> "PCH DP %c enabled on transcoder %c, should be disabled\n",
> port_name(port), pipe_name(pipe));
>
> - I915_STATE_WARN(HAS_PCH_IBX(dev_priv) && !state && port_pipe == PIPE_B,
> + I915_STATE_WARN(HAS_PCH_IBX(i915) && !state && port_pipe == PIPE_B,
> "IBX PCH DP %c still using transcoder B\n",
> port_name(port));
> }
>
> -static void assert_pch_hdmi_disabled(struct drm_i915_private *dev_priv,
> +static void assert_pch_hdmi_disabled(struct drm_i915_private *i915,
> enum pipe pipe, enum port port,
> i915_reg_t hdmi_reg)
> {
> enum pipe port_pipe;
> bool state;
>
> - state = intel_sdvo_port_enabled(dev_priv, hdmi_reg, &port_pipe);
> + state = intel_sdvo_port_enabled(i915, hdmi_reg, &port_pipe);
>
> I915_STATE_WARN(state && port_pipe == pipe,
> "PCH HDMI %c enabled on transcoder %c, should be disabled\n",
> port_name(port), pipe_name(pipe));
>
> - I915_STATE_WARN(HAS_PCH_IBX(dev_priv) && !state && port_pipe == PIPE_B,
> + I915_STATE_WARN(HAS_PCH_IBX(i915) && !state && port_pipe == PIPE_B,
> "IBX PCH HDMI %c still using transcoder B\n",
> port_name(port));
> }
>
> -static void assert_pch_ports_disabled(struct drm_i915_private *dev_priv,
> +static void assert_pch_ports_disabled(struct drm_i915_private *i915,
> enum pipe pipe)
> {
> enum pipe port_pipe;
>
> - assert_pch_dp_disabled(dev_priv, pipe, PORT_B, PCH_DP_B);
> - assert_pch_dp_disabled(dev_priv, pipe, PORT_C, PCH_DP_C);
> - assert_pch_dp_disabled(dev_priv, pipe, PORT_D, PCH_DP_D);
> + assert_pch_dp_disabled(i915, pipe, PORT_B, PCH_DP_B);
> + assert_pch_dp_disabled(i915, pipe, PORT_C, PCH_DP_C);
> + assert_pch_dp_disabled(i915, pipe, PORT_D, PCH_DP_D);
>
> - I915_STATE_WARN(intel_crt_port_enabled(dev_priv, PCH_ADPA, &port_pipe) &&
> + I915_STATE_WARN(intel_crt_port_enabled(i915, PCH_ADPA, &port_pipe) &&
> port_pipe == pipe,
> "PCH VGA enabled on transcoder %c, should be disabled\n",
> pipe_name(pipe));
>
> - I915_STATE_WARN(intel_lvds_port_enabled(dev_priv, PCH_LVDS, &port_pipe) &&
> + I915_STATE_WARN(intel_lvds_port_enabled(i915, PCH_LVDS, &port_pipe) &&
> port_pipe == pipe,
> "PCH LVDS enabled on transcoder %c, should be disabled\n",
> pipe_name(pipe));
>
> /* PCH SDVOB multiplex with HDMIB */
> - assert_pch_hdmi_disabled(dev_priv, pipe, PORT_B, PCH_HDMIB);
> - assert_pch_hdmi_disabled(dev_priv, pipe, PORT_C, PCH_HDMIC);
> - assert_pch_hdmi_disabled(dev_priv, pipe, PORT_D, PCH_HDMID);
> + assert_pch_hdmi_disabled(i915, pipe, PORT_B, PCH_HDMIB);
> + assert_pch_hdmi_disabled(i915, pipe, PORT_C, PCH_HDMIC);
> + assert_pch_hdmi_disabled(i915, pipe, PORT_D, PCH_HDMID);
> }
>
> static void _vlv_enable_pll(struct intel_crtc *crtc,
> const struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
>
> I915_WRITE(DPLL(pipe), pipe_config->dpll_hw_state.dpll);
> POSTING_READ(DPLL(pipe));
> udelay(150);
>
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> DPLL(pipe),
> DPLL_LOCK_VLV,
> DPLL_LOCK_VLV,
> @@ -1392,13 +1392,13 @@ static void _vlv_enable_pll(struct intel_crtc *crtc,
> static void vlv_enable_pll(struct intel_crtc *crtc,
> const struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
>
> - assert_pipe_disabled(dev_priv, pipe);
> + assert_pipe_disabled(i915, pipe);
>
> /* PLL is protected by panel, make sure we can write it */
> - assert_panel_unlocked(dev_priv, pipe);
> + assert_panel_unlocked(i915, pipe);
>
> if (pipe_config->dpll_hw_state.dpll & DPLL_VCO_ENABLE)
> _vlv_enable_pll(crtc, pipe_config);
> @@ -1411,19 +1411,19 @@ static void vlv_enable_pll(struct intel_crtc *crtc,
> static void _chv_enable_pll(struct intel_crtc *crtc,
> const struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
> enum dpio_channel port = vlv_pipe_to_channel(pipe);
> u32 tmp;
>
> - vlv_dpio_get(dev_priv);
> + vlv_dpio_get(i915);
>
> /* Enable back the 10bit clock to display controller */
> - tmp = vlv_dpio_read(dev_priv, pipe, CHV_CMN_DW14(port));
> + tmp = vlv_dpio_read(i915, pipe, CHV_CMN_DW14(port));
> tmp |= DPIO_DCLKP_EN;
> - vlv_dpio_write(dev_priv, pipe, CHV_CMN_DW14(port), tmp);
> + vlv_dpio_write(i915, pipe, CHV_CMN_DW14(port), tmp);
>
> - vlv_dpio_put(dev_priv);
> + vlv_dpio_put(i915);
>
> /*
> * Need to wait > 100ns between dclkp clock enable bit and PLL enable.
> @@ -1434,7 +1434,7 @@ static void _chv_enable_pll(struct intel_crtc *crtc,
> I915_WRITE(DPLL(pipe), pipe_config->dpll_hw_state.dpll);
>
> /* Check PLL is locked */
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> DPLL(pipe), DPLL_LOCK_VLV, DPLL_LOCK_VLV,
> 1))
> DRM_ERROR("PLL %d failed to lock\n", pipe);
> @@ -1443,13 +1443,13 @@ static void _chv_enable_pll(struct intel_crtc *crtc,
> static void chv_enable_pll(struct intel_crtc *crtc,
> const struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
>
> - assert_pipe_disabled(dev_priv, pipe);
> + assert_pipe_disabled(i915, pipe);
>
> /* PLL is protected by panel, make sure we can write it */
> - assert_panel_unlocked(dev_priv, pipe);
> + assert_panel_unlocked(i915, pipe);
>
> if (pipe_config->dpll_hw_state.dpll & DPLL_VCO_ENABLE)
> _chv_enable_pll(crtc, pipe_config);
> @@ -1464,7 +1464,7 @@ static void chv_enable_pll(struct intel_crtc *crtc,
> I915_WRITE(CBR4_VLV, CBR_DPLLBMD_PIPE(pipe));
> I915_WRITE(DPLL_MD(PIPE_B), pipe_config->dpll_hw_state.dpll_md);
> I915_WRITE(CBR4_VLV, 0);
> - dev_priv->chv_dpll_md[pipe] = pipe_config->dpll_hw_state.dpll_md;
> + i915->chv_dpll_md[pipe] = pipe_config->dpll_hw_state.dpll_md;
>
> /*
> * DPLLB VGA mode also seems to cause problems.
> @@ -1477,27 +1477,27 @@ static void chv_enable_pll(struct intel_crtc *crtc,
> }
> }
>
> -static bool i9xx_has_pps(struct drm_i915_private *dev_priv)
> +static bool i9xx_has_pps(struct drm_i915_private *i915)
> {
> - if (IS_I830(dev_priv))
> + if (IS_I830(i915))
> return false;
>
> - return IS_PINEVIEW(dev_priv) || IS_MOBILE(dev_priv);
> + return IS_PINEVIEW(i915) || IS_MOBILE(i915);
> }
>
> static void i9xx_enable_pll(struct intel_crtc *crtc,
> const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> i915_reg_t reg = DPLL(crtc->pipe);
> u32 dpll = crtc_state->dpll_hw_state.dpll;
> int i;
>
> - assert_pipe_disabled(dev_priv, crtc->pipe);
> + assert_pipe_disabled(i915, crtc->pipe);
>
> /* PLL is protected by panel, make sure we can write it */
> - if (i9xx_has_pps(dev_priv))
> - assert_panel_unlocked(dev_priv, crtc->pipe);
> + if (i9xx_has_pps(i915))
> + assert_panel_unlocked(i915, crtc->pipe);
>
> /*
> * Apparently we need to have VGA mode enabled prior to changing
> @@ -1511,7 +1511,7 @@ static void i9xx_enable_pll(struct intel_crtc *crtc,
> POSTING_READ(reg);
> udelay(150);
>
> - if (INTEL_GEN(dev_priv) >= 4) {
> + if (INTEL_GEN(i915) >= 4) {
> I915_WRITE(DPLL_MD(crtc->pipe),
> crtc_state->dpll_hw_state.dpll_md);
> } else {
> @@ -1534,26 +1534,26 @@ static void i9xx_enable_pll(struct intel_crtc *crtc,
> static void i9xx_disable_pll(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
>
> /* Don't disable pipe or pipe PLLs if needed */
> - if (IS_I830(dev_priv))
> + if (IS_I830(i915))
> return;
>
> /* Make sure the pipe isn't still relying on us */
> - assert_pipe_disabled(dev_priv, pipe);
> + assert_pipe_disabled(i915, pipe);
>
> I915_WRITE(DPLL(pipe), DPLL_VGA_MODE_DIS);
> POSTING_READ(DPLL(pipe));
> }
>
> -static void vlv_disable_pll(struct drm_i915_private *dev_priv, enum pipe pipe)
> +static void vlv_disable_pll(struct drm_i915_private *i915, enum pipe pipe)
> {
> u32 val;
>
> /* Make sure the pipe isn't still relying on us */
> - assert_pipe_disabled(dev_priv, pipe);
> + assert_pipe_disabled(i915, pipe);
>
> val = DPLL_INTEGRATED_REF_CLK_VLV |
> DPLL_REF_CLK_ENABLE_VLV | DPLL_VGA_MODE_DIS;
> @@ -1564,13 +1564,13 @@ static void vlv_disable_pll(struct drm_i915_private *dev_priv, enum pipe pipe)
> POSTING_READ(DPLL(pipe));
> }
>
> -static void chv_disable_pll(struct drm_i915_private *dev_priv, enum pipe pipe)
> +static void chv_disable_pll(struct drm_i915_private *i915, enum pipe pipe)
> {
> enum dpio_channel port = vlv_pipe_to_channel(pipe);
> u32 val;
>
> /* Make sure the pipe isn't still relying on us */
> - assert_pipe_disabled(dev_priv, pipe);
> + assert_pipe_disabled(i915, pipe);
>
> val = DPLL_SSC_REF_CLK_CHV |
> DPLL_REF_CLK_ENABLE_VLV | DPLL_VGA_MODE_DIS;
> @@ -1580,17 +1580,17 @@ static void chv_disable_pll(struct drm_i915_private *dev_priv, enum pipe pipe)
> I915_WRITE(DPLL(pipe), val);
> POSTING_READ(DPLL(pipe));
>
> - vlv_dpio_get(dev_priv);
> + vlv_dpio_get(i915);
>
> /* Disable 10bit clock to display controller */
> - val = vlv_dpio_read(dev_priv, pipe, CHV_CMN_DW14(port));
> + val = vlv_dpio_read(i915, pipe, CHV_CMN_DW14(port));
> val &= ~DPIO_DCLKP_EN;
> - vlv_dpio_write(dev_priv, pipe, CHV_CMN_DW14(port), val);
> + vlv_dpio_write(i915, pipe, CHV_CMN_DW14(port), val);
>
> - vlv_dpio_put(dev_priv);
> + vlv_dpio_put(i915);
> }
>
> -void vlv_wait_port_ready(struct drm_i915_private *dev_priv,
> +void vlv_wait_port_ready(struct drm_i915_private *i915,
> struct intel_digital_port *dport,
> unsigned int expected_mask)
> {
> @@ -1615,7 +1615,7 @@ void vlv_wait_port_ready(struct drm_i915_private *dev_priv,
> BUG();
> }
>
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> dpll_reg, port_mask, expected_mask,
> 1000))
> WARN(1, "timed out waiting for port %c ready: got 0x%x, expected 0x%x\n",
> @@ -1626,19 +1626,19 @@ void vlv_wait_port_ready(struct drm_i915_private *dev_priv,
> static void ironlake_enable_pch_transcoder(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
> i915_reg_t reg;
> u32 val, pipeconf_val;
>
> /* Make sure PCH DPLL is enabled */
> - assert_shared_dpll_enabled(dev_priv, crtc_state->shared_dpll);
> + assert_shared_dpll_enabled(i915, crtc_state->shared_dpll);
>
> /* FDI must be feeding us bits for PCH ports */
> - assert_fdi_tx_enabled(dev_priv, pipe);
> - assert_fdi_rx_enabled(dev_priv, pipe);
> + assert_fdi_tx_enabled(i915, pipe);
> + assert_fdi_rx_enabled(i915, pipe);
>
> - if (HAS_PCH_CPT(dev_priv)) {
> + if (HAS_PCH_CPT(i915)) {
> /* Workaround: Set the timing override bit before enabling the
> * pch transcoder. */
> reg = TRANS_CHICKEN2(pipe);
> @@ -1651,7 +1651,7 @@ static void ironlake_enable_pch_transcoder(const struct intel_crtc_state *crtc_s
> val = I915_READ(reg);
> pipeconf_val = I915_READ(PIPECONF(pipe));
>
> - if (HAS_PCH_IBX(dev_priv)) {
> + if (HAS_PCH_IBX(i915)) {
> /*
> * Make the BPC in transcoder be consistent with
> * that in pipeconf reg. For HDMI we must use 8bpc
> @@ -1666,7 +1666,7 @@ static void ironlake_enable_pch_transcoder(const struct intel_crtc_state *crtc_s
>
> val &= ~TRANS_INTERLACE_MASK;
> if ((pipeconf_val & PIPECONF_INTERLACE_MASK) == PIPECONF_INTERLACED_ILK) {
> - if (HAS_PCH_IBX(dev_priv) &&
> + if (HAS_PCH_IBX(i915) &&
> intel_crtc_has_type(crtc_state, INTEL_OUTPUT_SDVO))
> val |= TRANS_LEGACY_INTERLACED_ILK;
> else
> @@ -1676,20 +1676,20 @@ static void ironlake_enable_pch_transcoder(const struct intel_crtc_state *crtc_s
> }
>
> I915_WRITE(reg, val | TRANS_ENABLE);
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> reg, TRANS_STATE_ENABLE, TRANS_STATE_ENABLE,
> 100))
> DRM_ERROR("failed to enable transcoder %c\n", pipe_name(pipe));
> }
>
> -static void lpt_enable_pch_transcoder(struct drm_i915_private *dev_priv,
> +static void lpt_enable_pch_transcoder(struct drm_i915_private *i915,
> enum transcoder cpu_transcoder)
> {
> u32 val, pipeconf_val;
>
> /* FDI must be feeding us bits for PCH ports */
> - assert_fdi_tx_enabled(dev_priv, (enum pipe) cpu_transcoder);
> - assert_fdi_rx_enabled(dev_priv, PIPE_A);
> + assert_fdi_tx_enabled(i915, (enum pipe) cpu_transcoder);
> + assert_fdi_rx_enabled(i915, PIPE_A);
>
> /* Workaround: set timing override bit. */
> val = I915_READ(TRANS_CHICKEN2(PIPE_A));
> @@ -1706,7 +1706,7 @@ static void lpt_enable_pch_transcoder(struct drm_i915_private *dev_priv,
> val |= TRANS_PROGRESSIVE;
>
> I915_WRITE(LPT_TRANSCONF, val);
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> LPT_TRANSCONF,
> TRANS_STATE_ENABLE,
> TRANS_STATE_ENABLE,
> @@ -1714,30 +1714,30 @@ static void lpt_enable_pch_transcoder(struct drm_i915_private *dev_priv,
> DRM_ERROR("Failed to enable PCH transcoder\n");
> }
>
> -static void ironlake_disable_pch_transcoder(struct drm_i915_private *dev_priv,
> +static void ironlake_disable_pch_transcoder(struct drm_i915_private *i915,
> enum pipe pipe)
> {
> i915_reg_t reg;
> u32 val;
>
> /* FDI relies on the transcoder */
> - assert_fdi_tx_disabled(dev_priv, pipe);
> - assert_fdi_rx_disabled(dev_priv, pipe);
> + assert_fdi_tx_disabled(i915, pipe);
> + assert_fdi_rx_disabled(i915, pipe);
>
> /* Ports must be off as well */
> - assert_pch_ports_disabled(dev_priv, pipe);
> + assert_pch_ports_disabled(i915, pipe);
>
> reg = PCH_TRANSCONF(pipe);
> val = I915_READ(reg);
> val &= ~TRANS_ENABLE;
> I915_WRITE(reg, val);
> /* wait for PCH transcoder off, transcoder state */
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> reg, TRANS_STATE_ENABLE, 0,
> 50))
> DRM_ERROR("failed to disable transcoder %c\n", pipe_name(pipe));
>
> - if (HAS_PCH_CPT(dev_priv)) {
> + if (HAS_PCH_CPT(i915)) {
> /* Workaround: Clear the timing override chicken bit again. */
> reg = TRANS_CHICKEN2(pipe);
> val = I915_READ(reg);
> @@ -1746,7 +1746,7 @@ static void ironlake_disable_pch_transcoder(struct drm_i915_private *dev_priv,
> }
> }
>
> -void lpt_disable_pch_transcoder(struct drm_i915_private *dev_priv)
> +void lpt_disable_pch_transcoder(struct drm_i915_private *i915)
> {
> u32 val;
>
> @@ -1754,7 +1754,7 @@ void lpt_disable_pch_transcoder(struct drm_i915_private *dev_priv)
> val &= ~TRANS_ENABLE;
> I915_WRITE(LPT_TRANSCONF, val);
> /* wait for PCH transcoder off, transcoder state */
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> LPT_TRANSCONF, TRANS_STATE_ENABLE, 0,
> 50))
> DRM_ERROR("Failed to disable PCH transcoder\n");
> @@ -1767,9 +1767,9 @@ void lpt_disable_pch_transcoder(struct drm_i915_private *dev_priv)
>
> enum pipe intel_crtc_pch_transcoder(struct intel_crtc *crtc)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
>
> - if (HAS_PCH_LPT(dev_priv))
> + if (HAS_PCH_LPT(i915))
> return PIPE_A;
> else
> return crtc->pipe;
> @@ -1777,19 +1777,19 @@ enum pipe intel_crtc_pch_transcoder(struct intel_crtc *crtc)
>
> static u32 intel_crtc_max_vblank_count(const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
>
> /*
> * On i965gm the hardware frame counter reads
> * zero when the TV encoder is enabled :(
> */
> - if (IS_I965GM(dev_priv) &&
> + if (IS_I965GM(i915) &&
> (crtc_state->output_types & BIT(INTEL_OUTPUT_TVOUT)))
> return 0;
>
> - if (INTEL_GEN(dev_priv) >= 5 || IS_G4X(dev_priv))
> + if (INTEL_GEN(i915) >= 5 || IS_G4X(i915))
> return 0xffffffff; /* full 32 bit counter */
> - else if (INTEL_GEN(dev_priv) >= 3)
> + else if (INTEL_GEN(i915) >= 3)
> return 0xffffff; /* only 24 bits of frame count */
> else
> return 0; /* Gen2 doesn't have a hardware frame counter */
> @@ -1807,7 +1807,7 @@ static void intel_crtc_vblank_on(const struct intel_crtc_state *crtc_state)
> static void intel_enable_pipe(const struct intel_crtc_state *new_crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(new_crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum transcoder cpu_transcoder = new_crtc_state->cpu_transcoder;
> enum pipe pipe = crtc->pipe;
> i915_reg_t reg;
> @@ -1822,29 +1822,29 @@ static void intel_enable_pipe(const struct intel_crtc_state *new_crtc_state)
> * a plane. On ILK+ the pipe PLLs are integrated, so we don't
> * need the check.
> */
> - if (HAS_GMCH(dev_priv)) {
> + if (HAS_GMCH(i915)) {
> if (intel_crtc_has_type(new_crtc_state, INTEL_OUTPUT_DSI))
> - assert_dsi_pll_enabled(dev_priv);
> + assert_dsi_pll_enabled(i915);
> else
> - assert_pll_enabled(dev_priv, pipe);
> + assert_pll_enabled(i915, pipe);
> } else {
> if (new_crtc_state->has_pch_encoder) {
> /* if driving the PCH, we need FDI enabled */
> - assert_fdi_rx_pll_enabled(dev_priv,
> + assert_fdi_rx_pll_enabled(i915,
> intel_crtc_pch_transcoder(crtc));
> - assert_fdi_tx_pll_enabled(dev_priv,
> + assert_fdi_tx_pll_enabled(i915,
> (enum pipe) cpu_transcoder);
> }
> /* FIXME: assert CPU port conditions for SNB+ */
> }
>
> - trace_intel_pipe_enable(dev_priv, pipe);
> + trace_intel_pipe_enable(i915, pipe);
>
> reg = PIPECONF(cpu_transcoder);
> val = I915_READ(reg);
> if (val & PIPECONF_ENABLE) {
> /* we keep both pipes enabled on 830 */
> - WARN_ON(!IS_I830(dev_priv));
> + WARN_ON(!IS_I830(i915));
> return;
> }
>
> @@ -1865,7 +1865,7 @@ static void intel_enable_pipe(const struct intel_crtc_state *new_crtc_state)
> static void intel_disable_pipe(const struct intel_crtc_state *old_crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum transcoder cpu_transcoder = old_crtc_state->cpu_transcoder;
> enum pipe pipe = crtc->pipe;
> i915_reg_t reg;
> @@ -1879,7 +1879,7 @@ static void intel_disable_pipe(const struct intel_crtc_state *old_crtc_state)
> */
> assert_planes_disabled(crtc);
>
> - trace_intel_pipe_disable(dev_priv, pipe);
> + trace_intel_pipe_disable(i915, pipe);
>
> reg = PIPECONF(cpu_transcoder);
> val = I915_READ(reg);
> @@ -1894,7 +1894,7 @@ static void intel_disable_pipe(const struct intel_crtc_state *old_crtc_state)
> val &= ~PIPECONF_DOUBLE_WIDE;
>
> /* Don't disable pipe or pipe PLLs if needed */
> - if (!IS_I830(dev_priv))
> + if (!IS_I830(i915))
> val &= ~PIPECONF_ENABLE;
>
> I915_WRITE(reg, val);
> @@ -1902,22 +1902,22 @@ static void intel_disable_pipe(const struct intel_crtc_state *old_crtc_state)
> intel_wait_for_pipe_off(old_crtc_state);
> }
>
> -static unsigned int intel_tile_size(const struct drm_i915_private *dev_priv)
> +static unsigned int intel_tile_size(const struct drm_i915_private *i915)
> {
> - return IS_GEN(dev_priv, 2) ? 2048 : 4096;
> + return IS_GEN(i915, 2) ? 2048 : 4096;
> }
>
> static unsigned int
> intel_tile_width_bytes(const struct drm_framebuffer *fb, int color_plane)
> {
> - struct drm_i915_private *dev_priv = to_i915(fb->dev);
> + struct drm_i915_private *i915 = to_i915(fb->dev);
> unsigned int cpp = fb->format->cpp[color_plane];
>
> switch (fb->modifier) {
> case DRM_FORMAT_MOD_LINEAR:
> - return intel_tile_size(dev_priv);
> + return intel_tile_size(i915);
> case I915_FORMAT_MOD_X_TILED:
> - if (IS_GEN(dev_priv, 2))
> + if (IS_GEN(i915, 2))
> return 128;
> else
> return 512;
> @@ -1926,7 +1926,7 @@ intel_tile_width_bytes(const struct drm_framebuffer *fb, int color_plane)
> return 128;
> /* fall through */
> case I915_FORMAT_MOD_Y_TILED:
> - if (IS_GEN(dev_priv, 2) || HAS_128_BYTE_Y_TILING(dev_priv))
> + if (IS_GEN(i915, 2) || HAS_128_BYTE_Y_TILING(i915))
> return 128;
> else
> return 512;
> @@ -2017,26 +2017,26 @@ intel_fill_fb_ggtt_view(struct i915_ggtt_view *view,
> }
> }
>
> -static unsigned int intel_cursor_alignment(const struct drm_i915_private *dev_priv)
> +static unsigned int intel_cursor_alignment(const struct drm_i915_private *i915)
> {
> - if (IS_I830(dev_priv))
> + if (IS_I830(i915))
> return 16 * 1024;
> - else if (IS_I85X(dev_priv))
> + else if (IS_I85X(i915))
> return 256;
> - else if (IS_I845G(dev_priv) || IS_I865G(dev_priv))
> + else if (IS_I845G(i915) || IS_I865G(i915))
> return 32;
> else
> return 4 * 1024;
> }
>
> -static unsigned int intel_linear_alignment(const struct drm_i915_private *dev_priv)
> +static unsigned int intel_linear_alignment(const struct drm_i915_private *i915)
> {
> - if (INTEL_GEN(dev_priv) >= 9)
> + if (INTEL_GEN(i915) >= 9)
> return 256 * 1024;
> - else if (IS_I965G(dev_priv) || IS_I965GM(dev_priv) ||
> - IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> + else if (IS_I965G(i915) || IS_I965GM(i915) ||
> + IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> return 128 * 1024;
> - else if (INTEL_GEN(dev_priv) >= 4)
> + else if (INTEL_GEN(i915) >= 4)
> return 4 * 1024;
> else
> return 0;
> @@ -2045,7 +2045,7 @@ static unsigned int intel_linear_alignment(const struct drm_i915_private *dev_pr
> static unsigned int intel_surf_alignment(const struct drm_framebuffer *fb,
> int color_plane)
> {
> - struct drm_i915_private *dev_priv = to_i915(fb->dev);
> + struct drm_i915_private *i915 = to_i915(fb->dev);
>
> /* AUX_DIST needs only 4K alignment */
> if (color_plane == 1)
> @@ -2053,9 +2053,9 @@ static unsigned int intel_surf_alignment(const struct drm_framebuffer *fb,
>
> switch (fb->modifier) {
> case DRM_FORMAT_MOD_LINEAR:
> - return intel_linear_alignment(dev_priv);
> + return intel_linear_alignment(i915);
> case I915_FORMAT_MOD_X_TILED:
> - if (INTEL_GEN(dev_priv) >= 9)
> + if (INTEL_GEN(i915) >= 9)
> return 256 * 1024;
> return 0;
> case I915_FORMAT_MOD_Y_TILED_CCS:
> @@ -2072,9 +2072,9 @@ static unsigned int intel_surf_alignment(const struct drm_framebuffer *fb,
> static bool intel_plane_uses_fence(const struct intel_plane_state *plane_state)
> {
> struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
>
> - return INTEL_GEN(dev_priv) < 4 ||
> + return INTEL_GEN(i915) < 4 ||
> (plane->has_fbc &&
> plane_state->view.type == I915_GGTT_VIEW_NORMAL);
> }
> @@ -2086,7 +2086,7 @@ intel_pin_and_fence_fb_obj(struct drm_framebuffer *fb,
> unsigned long *out_flags)
> {
> struct drm_device *dev = fb->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct drm_i915_gem_object *obj = intel_fb_obj(fb);
> intel_wakeref_t wakeref;
> struct i915_vma *vma;
> @@ -2102,7 +2102,7 @@ intel_pin_and_fence_fb_obj(struct drm_framebuffer *fb,
> * we should always have valid PTE following the scanout preventing
> * the VT-d warning.
> */
> - if (intel_scanout_needs_vtd_wa(dev_priv) && alignment < 256 * 1024)
> + if (intel_scanout_needs_vtd_wa(i915) && alignment < 256 * 1024)
> alignment = 256 * 1024;
>
> /*
> @@ -2112,10 +2112,10 @@ intel_pin_and_fence_fb_obj(struct drm_framebuffer *fb,
> * intel_runtime_pm_put(), so it is correct to wrap only the
> * pin/unpin/fence and not more.
> */
> - wakeref = intel_runtime_pm_get(dev_priv);
> + wakeref = intel_runtime_pm_get(i915);
> i915_gem_object_lock(obj);
>
> - atomic_inc(&dev_priv->gpu_error.pending_fb_pin);
> + atomic_inc(&i915->gpu_error.pending_fb_pin);
>
> pinctl = 0;
>
> @@ -2126,7 +2126,7 @@ intel_pin_and_fence_fb_obj(struct drm_framebuffer *fb,
> * complicated than this. For example, Cherryview appears quite
> * happy to scanout from anywhere within its global aperture.
> */
> - if (HAS_GMCH(dev_priv))
> + if (HAS_GMCH(i915))
> pinctl |= PIN_MAPPABLE;
>
> vma = i915_gem_object_pin_to_display_plane(obj,
> @@ -2154,7 +2154,7 @@ intel_pin_and_fence_fb_obj(struct drm_framebuffer *fb,
> * mode that matches the user configuration.
> */
> ret = i915_vma_pin_fence(vma);
> - if (ret != 0 && INTEL_GEN(dev_priv) < 4) {
> + if (ret != 0 && INTEL_GEN(i915) < 4) {
> i915_gem_object_unpin_from_display_plane(vma);
> vma = ERR_PTR(ret);
> goto err;
> @@ -2166,10 +2166,10 @@ intel_pin_and_fence_fb_obj(struct drm_framebuffer *fb,
>
> i915_vma_get(vma);
> err:
> - atomic_dec(&dev_priv->gpu_error.pending_fb_pin);
> + atomic_dec(&i915->gpu_error.pending_fb_pin);
>
> i915_gem_object_unlock(obj);
> - intel_runtime_pm_put(dev_priv, wakeref);
> + intel_runtime_pm_put(i915, wakeref);
> return vma;
> }
>
> @@ -2265,7 +2265,7 @@ static u32 intel_adjust_aligned_offset(int *x, int *y,
> unsigned int pitch,
> u32 old_offset, u32 new_offset)
> {
> - struct drm_i915_private *dev_priv = to_i915(fb->dev);
> + struct drm_i915_private *i915 = to_i915(fb->dev);
> unsigned int cpp = fb->format->cpp[color_plane];
>
> WARN_ON(new_offset > old_offset);
> @@ -2274,7 +2274,7 @@ static u32 intel_adjust_aligned_offset(int *x, int *y,
> unsigned int tile_size, tile_width, tile_height;
> unsigned int pitch_tiles;
>
> - tile_size = intel_tile_size(dev_priv);
> + tile_size = intel_tile_size(i915);
> intel_tile_dims(fb, color_plane, &tile_width, &tile_height);
>
> if (drm_rotation_90_or_270(rotation)) {
> @@ -2326,7 +2326,7 @@ static u32 intel_plane_adjust_aligned_offset(int *x, int *y,
> * used. This is why the user has to pass in the pitch since it
> * is specified in the rotated orientation.
> */
> -static u32 intel_compute_aligned_offset(struct drm_i915_private *dev_priv,
> +static u32 intel_compute_aligned_offset(struct drm_i915_private *i915,
> int *x, int *y,
> const struct drm_framebuffer *fb,
> int color_plane,
> @@ -2344,7 +2344,7 @@ static u32 intel_compute_aligned_offset(struct drm_i915_private *dev_priv,
> unsigned int tile_size, tile_width, tile_height;
> unsigned int tile_rows, tiles, pitch_tiles;
>
> - tile_size = intel_tile_size(dev_priv);
> + tile_size = intel_tile_size(i915);
> intel_tile_dims(fb, color_plane, &tile_width, &tile_height);
>
> if (drm_rotation_90_or_270(rotation)) {
> @@ -2382,18 +2382,18 @@ static u32 intel_plane_compute_aligned_offset(int *x, int *y,
> int color_plane)
> {
> struct intel_plane *intel_plane = to_intel_plane(state->base.plane);
> - struct drm_i915_private *dev_priv = to_i915(intel_plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(intel_plane->base.dev);
> const struct drm_framebuffer *fb = state->base.fb;
> unsigned int rotation = state->base.rotation;
> int pitch = state->color_plane[color_plane].stride;
> u32 alignment;
>
> if (intel_plane->id == PLANE_CURSOR)
> - alignment = intel_cursor_alignment(dev_priv);
> + alignment = intel_cursor_alignment(i915);
> else
> alignment = intel_surf_alignment(fb, color_plane);
>
> - return intel_compute_aligned_offset(dev_priv, x, y, fb, color_plane,
> + return intel_compute_aligned_offset(i915, x, y, fb, color_plane,
> pitch, rotation, alignment);
> }
>
> @@ -2402,11 +2402,11 @@ static int intel_fb_offset_to_xy(int *x, int *y,
> const struct drm_framebuffer *fb,
> int color_plane)
> {
> - struct drm_i915_private *dev_priv = to_i915(fb->dev);
> + struct drm_i915_private *i915 = to_i915(fb->dev);
> unsigned int height;
>
> if (fb->modifier != DRM_FORMAT_MOD_LINEAR &&
> - fb->offsets[color_plane] % intel_tile_size(dev_priv)) {
> + fb->offsets[color_plane] % intel_tile_size(i915)) {
> DRM_DEBUG_KMS("Misaligned offset 0x%08x for color plane %d\n",
> fb->offsets[color_plane], color_plane);
> return -EINVAL;
> @@ -2507,7 +2507,7 @@ bool is_ccs_modifier(u64 modifier)
> modifier == I915_FORMAT_MOD_Yf_TILED_CCS;
> }
>
> -u32 intel_plane_fb_max_stride(struct drm_i915_private *dev_priv,
> +u32 intel_plane_fb_max_stride(struct drm_i915_private *i915,
> u32 pixel_format, u64 modifier)
> {
> struct intel_crtc *crtc;
> @@ -2517,7 +2517,7 @@ u32 intel_plane_fb_max_stride(struct drm_i915_private *dev_priv,
> * We assume the primary plane for pipe A has
> * the highest stride limits of them all.
> */
> - crtc = intel_get_crtc_for_pipe(dev_priv, PIPE_A);
> + crtc = intel_get_crtc_for_pipe(i915, PIPE_A);
> plane = to_intel_plane(crtc->base.primary);
>
> return plane->max_stride(plane, pixel_format, modifier,
> @@ -2525,7 +2525,7 @@ u32 intel_plane_fb_max_stride(struct drm_i915_private *dev_priv,
> }
>
> static
> -u32 intel_fb_max_stride(struct drm_i915_private *dev_priv,
> +u32 intel_fb_max_stride(struct drm_i915_private *i915,
> u32 pixel_format, u64 modifier)
> {
> /*
> @@ -2535,22 +2535,22 @@ u32 intel_fb_max_stride(struct drm_i915_private *dev_priv,
> * The new CCS hash mode makes remapping impossible
> */
> if (!is_ccs_modifier(modifier)) {
> - if (INTEL_GEN(dev_priv) >= 7)
> + if (INTEL_GEN(i915) >= 7)
> return 256*1024;
> - else if (INTEL_GEN(dev_priv) >= 4)
> + else if (INTEL_GEN(i915) >= 4)
> return 128*1024;
> }
>
> - return intel_plane_fb_max_stride(dev_priv, pixel_format, modifier);
> + return intel_plane_fb_max_stride(i915, pixel_format, modifier);
> }
>
> static u32
> intel_fb_stride_alignment(const struct drm_framebuffer *fb, int color_plane)
> {
> - struct drm_i915_private *dev_priv = to_i915(fb->dev);
> + struct drm_i915_private *i915 = to_i915(fb->dev);
>
> if (fb->modifier == DRM_FORMAT_MOD_LINEAR) {
> - u32 max_stride = intel_plane_fb_max_stride(dev_priv,
> + u32 max_stride = intel_plane_fb_max_stride(i915,
> fb->format->format,
> fb->modifier);
>
> @@ -2559,7 +2559,7 @@ intel_fb_stride_alignment(const struct drm_framebuffer *fb, int color_plane)
> * we need the stride to be page aligned.
> */
> if (fb->pitches[color_plane] > max_stride)
> - return intel_tile_size(dev_priv);
> + return intel_tile_size(i915);
> else
> return 64;
> } else {
> @@ -2570,7 +2570,7 @@ intel_fb_stride_alignment(const struct drm_framebuffer *fb, int color_plane)
> bool intel_plane_can_remap(const struct intel_plane_state *plane_state)
> {
> struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> const struct drm_framebuffer *fb = plane_state->base.fb;
> int i;
>
> @@ -2584,7 +2584,7 @@ bool intel_plane_can_remap(const struct intel_plane_state *plane_state)
> * Would also need to deal with the fence POT alignment
> * and gen2 2KiB GTT tile size.
> */
> - if (INTEL_GEN(dev_priv) < 4)
> + if (INTEL_GEN(i915) < 4)
> return false;
>
> /*
> @@ -2596,7 +2596,7 @@ bool intel_plane_can_remap(const struct intel_plane_state *plane_state)
>
> /* Linear needs a page aligned stride for remapping */
> if (fb->modifier == DRM_FORMAT_MOD_LINEAR) {
> - unsigned int alignment = intel_tile_size(dev_priv) - 1;
> + unsigned int alignment = intel_tile_size(i915) - 1;
>
> for (i = 0; i < fb->format->num_planes; i++) {
> if (fb->pitches[i] & alignment)
> @@ -2636,7 +2636,7 @@ static bool intel_plane_needs_remap(const struct intel_plane_state *plane_state)
> }
>
> static int
> -intel_fill_fb_info(struct drm_i915_private *dev_priv,
> +intel_fill_fb_info(struct drm_i915_private *i915,
> struct drm_framebuffer *fb)
> {
> struct intel_framebuffer *intel_fb = to_intel_framebuffer(fb);
> @@ -2645,7 +2645,7 @@ intel_fill_fb_info(struct drm_i915_private *dev_priv,
> u32 gtt_offset_rotated = 0;
> unsigned int max_size = 0;
> int i, num_planes = fb->format->num_planes;
> - unsigned int tile_size = intel_tile_size(dev_priv);
> + unsigned int tile_size = intel_tile_size(i915);
>
> for (i = 0; i < num_planes; i++) {
> unsigned int width, height;
> @@ -2719,7 +2719,7 @@ intel_fill_fb_info(struct drm_i915_private *dev_priv,
> intel_fb->normal[i].x = x;
> intel_fb->normal[i].y = y;
>
> - offset = intel_compute_aligned_offset(dev_priv, &x, &y, fb, i,
> + offset = intel_compute_aligned_offset(i915, &x, &y, fb, i,
> fb->pitches[i],
> DRM_MODE_ROTATE_0,
> tile_size);
> @@ -2803,14 +2803,14 @@ intel_fill_fb_info(struct drm_i915_private *dev_priv,
> static void
> intel_plane_remap_gtt(struct intel_plane_state *plane_state)
> {
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> to_i915(plane_state->base.plane->dev);
> struct drm_framebuffer *fb = plane_state->base.fb;
> struct intel_framebuffer *intel_fb = to_intel_framebuffer(fb);
> struct intel_rotation_info *info = &plane_state->view.rotated;
> unsigned int rotation = plane_state->base.rotation;
> int i, num_planes = fb->format->num_planes;
> - unsigned int tile_size = intel_tile_size(dev_priv);
> + unsigned int tile_size = intel_tile_size(i915);
> unsigned int src_x, src_y;
> unsigned int src_w, src_h;
> u32 gtt_offset = 0;
> @@ -2860,7 +2860,7 @@ intel_plane_remap_gtt(struct intel_plane_state *plane_state)
> x += intel_fb->normal[i].x;
> y += intel_fb->normal[i].y;
>
> - offset = intel_compute_aligned_offset(dev_priv, &x, &y,
> + offset = intel_compute_aligned_offset(i915, &x, &y,
> fb, i, fb->pitches[i],
> DRM_MODE_ROTATE_0, tile_size);
> offset /= tile_size;
> @@ -3046,7 +3046,7 @@ intel_alloc_initial_plane_obj(struct intel_crtc *crtc,
> struct intel_initial_plane_config *plane_config)
> {
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct drm_i915_gem_object *obj = NULL;
> struct drm_mode_fb_cmd2 mode_cmd = { 0 };
> struct drm_framebuffer *fb = &plane_config->fb->base;
> @@ -3062,7 +3062,7 @@ intel_alloc_initial_plane_obj(struct intel_crtc *crtc,
> /* If the FB is too big, just don't use it since fbdev is not very
> * important and we should probably use that space with FBC or other
> * features. */
> - if (size_aligned * 2 > dev_priv->stolen_usable_size)
> + if (size_aligned * 2 > i915->stolen_usable_size)
> return false;
>
> switch (fb->modifier) {
> @@ -3077,7 +3077,7 @@ intel_alloc_initial_plane_obj(struct intel_crtc *crtc,
> }
>
> mutex_lock(&dev->struct_mutex);
> - obj = i915_gem_object_create_stolen_for_preallocated(dev_priv,
> + obj = i915_gem_object_create_stolen_for_preallocated(i915,
> base_aligned,
> base_aligned,
> size_aligned);
> @@ -3135,7 +3135,7 @@ intel_set_plane_visible(struct intel_crtc_state *crtc_state,
>
> static void fixup_active_planes(struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> struct drm_plane *plane;
>
> /*
> @@ -3145,7 +3145,7 @@ static void fixup_active_planes(struct intel_crtc_state *crtc_state)
> */
> crtc_state->active_planes = 0;
>
> - drm_for_each_plane_mask(plane, &dev_priv->drm,
> + drm_for_each_plane_mask(plane, &i915->drm,
> crtc_state->base.plane_mask)
> crtc_state->active_planes |= BIT(to_intel_plane(plane)->id);
> }
> @@ -3177,7 +3177,7 @@ intel_find_initial_plane_obj(struct intel_crtc *intel_crtc,
> struct intel_initial_plane_config *plane_config)
> {
> struct drm_device *dev = intel_crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct drm_crtc *c;
> struct drm_i915_gem_object *obj;
> struct drm_plane *primary = intel_crtc->base.primary;
> @@ -3272,7 +3272,7 @@ intel_find_initial_plane_obj(struct intel_crtc *intel_crtc,
> intel_state->base.dst = drm_plane_state_dest(plane_state);
>
> if (i915_gem_object_is_tiled(obj))
> - dev_priv->preserve_bios_swizzle = true;
> + i915->preserve_bios_swizzle = true;
>
> plane_state->fb = fb;
> plane_state->crtc = &intel_crtc->base;
> @@ -3381,7 +3381,7 @@ static bool skl_check_main_ccs_coordinates(struct intel_plane_state *plane_state
>
> static int skl_check_main_surface(struct intel_plane_state *plane_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(plane_state->base.plane->dev);
> + struct drm_i915_private *i915 = to_i915(plane_state->base.plane->dev);
> const struct drm_framebuffer *fb = plane_state->base.fb;
> unsigned int rotation = plane_state->base.rotation;
> int x = plane_state->base.src.x1 >> 16;
> @@ -3392,9 +3392,9 @@ static int skl_check_main_surface(struct intel_plane_state *plane_state)
> int max_height = 4096;
> u32 alignment, offset, aux_offset = plane_state->color_plane[1].offset;
>
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> max_width = icl_max_plane_width(fb, 0, rotation);
> - else if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> + else if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> max_width = glk_max_plane_width(fb, 0, rotation);
> else
> max_width = skl_max_plane_width(fb, 0, rotation);
> @@ -3564,16 +3564,16 @@ i9xx_plane_max_stride(struct intel_plane *plane,
> u32 pixel_format, u64 modifier,
> unsigned int rotation)
> {
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
>
> - if (!HAS_GMCH(dev_priv)) {
> + if (!HAS_GMCH(i915)) {
> return 32*1024;
> - } else if (INTEL_GEN(dev_priv) >= 4) {
> + } else if (INTEL_GEN(i915) >= 4) {
> if (modifier == I915_FORMAT_MOD_X_TILED)
> return 16*1024;
> else
> return 32*1024;
> - } else if (INTEL_GEN(dev_priv) >= 3) {
> + } else if (INTEL_GEN(i915) >= 3) {
> if (modifier == I915_FORMAT_MOD_X_TILED)
> return 8*1024;
> else
> @@ -3589,7 +3589,7 @@ i9xx_plane_max_stride(struct intel_plane *plane,
> static u32 i9xx_plane_ctl_crtc(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> u32 dspcntr = 0;
>
> if (crtc_state->gamma_enable)
> @@ -3598,7 +3598,7 @@ static u32 i9xx_plane_ctl_crtc(const struct intel_crtc_state *crtc_state)
> if (crtc_state->csc_enable)
> dspcntr |= DISPPLANE_PIPE_CSC_ENABLE;
>
> - if (INTEL_GEN(dev_priv) < 5)
> + if (INTEL_GEN(i915) < 5)
> dspcntr |= DISPPLANE_SEL_PIPE(crtc->pipe);
>
> return dspcntr;
> @@ -3607,7 +3607,7 @@ static u32 i9xx_plane_ctl_crtc(const struct intel_crtc_state *crtc_state)
> static u32 i9xx_plane_ctl(const struct intel_crtc_state *crtc_state,
> const struct intel_plane_state *plane_state)
> {
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> to_i915(plane_state->base.plane->dev);
> const struct drm_framebuffer *fb = plane_state->base.fb;
> unsigned int rotation = plane_state->base.rotation;
> @@ -3615,8 +3615,8 @@ static u32 i9xx_plane_ctl(const struct intel_crtc_state *crtc_state,
>
> dspcntr = DISPLAY_PLANE_ENABLE;
>
> - if (IS_G4X(dev_priv) || IS_GEN(dev_priv, 5) ||
> - IS_GEN(dev_priv, 6) || IS_IVYBRIDGE(dev_priv))
> + if (IS_G4X(i915) || IS_GEN(i915, 5) ||
> + IS_GEN(i915, 6) || IS_IVYBRIDGE(i915))
> dspcntr |= DISPPLANE_TRICKLE_FEED_DISABLE;
>
> switch (fb->format->format) {
> @@ -3646,7 +3646,7 @@ static u32 i9xx_plane_ctl(const struct intel_crtc_state *crtc_state,
> return 0;
> }
>
> - if (INTEL_GEN(dev_priv) >= 4 &&
> + if (INTEL_GEN(i915) >= 4 &&
> fb->modifier == I915_FORMAT_MOD_X_TILED)
> dspcntr |= DISPPLANE_TILED;
>
> @@ -3661,7 +3661,7 @@ static u32 i9xx_plane_ctl(const struct intel_crtc_state *crtc_state,
>
> int i9xx_check_plane_surface(struct intel_plane_state *plane_state)
> {
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> to_i915(plane_state->base.plane->dev);
> int src_x, src_y;
> u32 offset;
> @@ -3679,7 +3679,7 @@ int i9xx_check_plane_surface(struct intel_plane_state *plane_state)
>
> intel_add_fb_offsets(&src_x, &src_y, plane_state, 0);
>
> - if (INTEL_GEN(dev_priv) >= 4)
> + if (INTEL_GEN(i915) >= 4)
> offset = intel_plane_compute_aligned_offset(&src_x, &src_y,
> plane_state, 0);
> else
> @@ -3694,7 +3694,7 @@ int i9xx_check_plane_surface(struct intel_plane_state *plane_state)
> (src_y << 16) - plane_state->base.src.y1);
>
> /* HSW/BDW do this automagically in hardware */
> - if (!IS_HASWELL(dev_priv) && !IS_BROADWELL(dev_priv)) {
> + if (!IS_HASWELL(i915) && !IS_BROADWELL(i915)) {
> unsigned int rotation = plane_state->base.rotation;
> int src_w = drm_rect_width(&plane_state->base.src) >> 16;
> int src_h = drm_rect_height(&plane_state->base.src) >> 16;
> @@ -3752,7 +3752,7 @@ static void i9xx_update_plane(struct intel_plane *plane,
> const struct intel_crtc_state *crtc_state,
> const struct intel_plane_state *plane_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> enum i9xx_plane_id i9xx_plane = plane->i9xx_plane;
> u32 linear_offset;
> int x = plane_state->color_plane[0].x;
> @@ -3765,16 +3765,16 @@ static void i9xx_update_plane(struct intel_plane *plane,
>
> linear_offset = intel_fb_xy_to_linear(x, y, plane_state, 0);
>
> - if (INTEL_GEN(dev_priv) >= 4)
> + if (INTEL_GEN(i915) >= 4)
> dspaddr_offset = plane_state->color_plane[0].offset;
> else
> dspaddr_offset = linear_offset;
>
> - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> + spin_lock_irqsave(&i915->uncore.lock, irqflags);
>
> I915_WRITE_FW(DSPSTRIDE(i9xx_plane), plane_state->color_plane[0].stride);
>
> - if (INTEL_GEN(dev_priv) < 4) {
> + if (INTEL_GEN(i915) < 4) {
> /* pipesrc and dspsize control the size that is scaled from,
> * which should always be the user's requested size.
> */
> @@ -3782,7 +3782,7 @@ static void i9xx_update_plane(struct intel_plane *plane,
> I915_WRITE_FW(DSPSIZE(i9xx_plane),
> ((crtc_state->pipe_src_h - 1) << 16) |
> (crtc_state->pipe_src_w - 1));
> - } else if (IS_CHERRYVIEW(dev_priv) && i9xx_plane == PLANE_B) {
> + } else if (IS_CHERRYVIEW(i915) && i9xx_plane == PLANE_B) {
> I915_WRITE_FW(PRIMPOS(i9xx_plane), 0);
> I915_WRITE_FW(PRIMSIZE(i9xx_plane),
> ((crtc_state->pipe_src_h - 1) << 16) |
> @@ -3790,9 +3790,9 @@ static void i9xx_update_plane(struct intel_plane *plane,
> I915_WRITE_FW(PRIMCNSTALPHA(i9xx_plane), 0);
> }
>
> - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) {
> + if (IS_HASWELL(i915) || IS_BROADWELL(i915)) {
> I915_WRITE_FW(DSPOFFSET(i9xx_plane), (y << 16) | x);
> - } else if (INTEL_GEN(dev_priv) >= 4) {
> + } else if (INTEL_GEN(i915) >= 4) {
> I915_WRITE_FW(DSPLINOFF(i9xx_plane), linear_offset);
> I915_WRITE_FW(DSPTILEOFF(i9xx_plane), (y << 16) | x);
> }
> @@ -3803,7 +3803,7 @@ static void i9xx_update_plane(struct intel_plane *plane,
> * the control register just before the surface register.
> */
> I915_WRITE_FW(DSPCNTR(i9xx_plane), dspcntr);
> - if (INTEL_GEN(dev_priv) >= 4)
> + if (INTEL_GEN(i915) >= 4)
> I915_WRITE_FW(DSPSURF(i9xx_plane),
> intel_plane_ggtt_offset(plane_state) +
> dspaddr_offset);
> @@ -3812,13 +3812,13 @@ static void i9xx_update_plane(struct intel_plane *plane,
> intel_plane_ggtt_offset(plane_state) +
> dspaddr_offset);
>
> - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> + spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
> }
>
> static void i9xx_disable_plane(struct intel_plane *plane,
> const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> enum i9xx_plane_id i9xx_plane = plane->i9xx_plane;
> unsigned long irqflags;
> u32 dspcntr;
> @@ -3835,21 +3835,21 @@ static void i9xx_disable_plane(struct intel_plane *plane,
> */
> dspcntr = i9xx_plane_ctl_crtc(crtc_state);
>
> - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> + spin_lock_irqsave(&i915->uncore.lock, irqflags);
>
> I915_WRITE_FW(DSPCNTR(i9xx_plane), dspcntr);
> - if (INTEL_GEN(dev_priv) >= 4)
> + if (INTEL_GEN(i915) >= 4)
> I915_WRITE_FW(DSPSURF(i9xx_plane), 0);
> else
> I915_WRITE_FW(DSPADDR(i9xx_plane), 0);
>
> - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> + spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
> }
>
> static bool i9xx_plane_get_hw_state(struct intel_plane *plane,
> enum pipe *pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> enum intel_display_power_domain power_domain;
> enum i9xx_plane_id i9xx_plane = plane->i9xx_plane;
> intel_wakeref_t wakeref;
> @@ -3862,7 +3862,7 @@ static bool i9xx_plane_get_hw_state(struct intel_plane *plane,
> * display power wells.
> */
> power_domain = POWER_DOMAIN_PIPE(plane->pipe);
> - wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain);
> + wakeref = intel_display_power_get_if_enabled(i915, power_domain);
> if (!wakeref)
> return false;
>
> @@ -3870,13 +3870,13 @@ static bool i9xx_plane_get_hw_state(struct intel_plane *plane,
>
> ret = val & DISPLAY_PLANE_ENABLE;
>
> - if (INTEL_GEN(dev_priv) >= 5)
> + if (INTEL_GEN(i915) >= 5)
> *pipe = plane->pipe;
> else
> *pipe = (val & DISPPLANE_SEL_PIPE_MASK) >>
> DISPPLANE_SEL_PIPE_SHIFT;
>
> - intel_display_power_put(dev_priv, power_domain, wakeref);
> + intel_display_power_put(i915, power_domain, wakeref);
>
> return ret;
> }
> @@ -3884,7 +3884,7 @@ static bool i9xx_plane_get_hw_state(struct intel_plane *plane,
> static void skl_detach_scaler(struct intel_crtc *intel_crtc, int id)
> {
> struct drm_device *dev = intel_crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
>
> I915_WRITE(SKL_PS_CTRL(intel_crtc->pipe, id), 0);
> I915_WRITE(SKL_PS_WIN_POS(intel_crtc->pipe, id), 0);
> @@ -4091,10 +4091,10 @@ static u32 cnl_plane_ctl_flip(unsigned int reflect)
>
> u32 skl_plane_ctl_crtc(const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> u32 plane_ctl = 0;
>
> - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> return plane_ctl;
>
> if (crtc_state->gamma_enable)
> @@ -4109,7 +4109,7 @@ u32 skl_plane_ctl_crtc(const struct intel_crtc_state *crtc_state)
> u32 skl_plane_ctl(const struct intel_crtc_state *crtc_state,
> const struct intel_plane_state *plane_state)
> {
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> to_i915(plane_state->base.plane->dev);
> const struct drm_framebuffer *fb = plane_state->base.fb;
> unsigned int rotation = plane_state->base.rotation;
> @@ -4118,7 +4118,7 @@ u32 skl_plane_ctl(const struct intel_crtc_state *crtc_state,
>
> plane_ctl = PLANE_CTL_ENABLE;
>
> - if (INTEL_GEN(dev_priv) < 10 && !IS_GEMINILAKE(dev_priv)) {
> + if (INTEL_GEN(i915) < 10 && !IS_GEMINILAKE(i915)) {
> plane_ctl |= skl_plane_ctl_alpha(plane_state);
> plane_ctl |= PLANE_CTL_PLANE_GAMMA_DISABLE;
>
> @@ -4133,7 +4133,7 @@ u32 skl_plane_ctl(const struct intel_crtc_state *crtc_state,
> plane_ctl |= skl_plane_ctl_tiling(fb->modifier);
> plane_ctl |= skl_plane_ctl_rotate(rotation & DRM_MODE_ROTATE_MASK);
>
> - if (INTEL_GEN(dev_priv) >= 10)
> + if (INTEL_GEN(i915) >= 10)
> plane_ctl |= cnl_plane_ctl_flip(rotation &
> DRM_MODE_REFLECT_MASK);
>
> @@ -4147,10 +4147,10 @@ u32 skl_plane_ctl(const struct intel_crtc_state *crtc_state,
>
> u32 glk_plane_color_ctl_crtc(const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> u32 plane_color_ctl = 0;
>
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> return plane_color_ctl;
>
> if (crtc_state->gamma_enable)
> @@ -4165,7 +4165,7 @@ u32 glk_plane_color_ctl_crtc(const struct intel_crtc_state *crtc_state)
> u32 glk_plane_color_ctl(const struct intel_crtc_state *crtc_state,
> const struct intel_plane_state *plane_state)
> {
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> to_i915(plane_state->base.plane->dev);
> const struct drm_framebuffer *fb = plane_state->base.fb;
> struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> @@ -4174,7 +4174,7 @@ u32 glk_plane_color_ctl(const struct intel_crtc_state *crtc_state,
> plane_color_ctl |= PLANE_COLOR_PLANE_GAMMA_DISABLE;
> plane_color_ctl |= glk_plane_color_ctl_alpha(plane_state);
>
> - if (fb->format->is_yuv && !icl_is_hdr_plane(dev_priv, plane->id)) {
> + if (fb->format->is_yuv && !icl_is_hdr_plane(i915, plane->id)) {
> if (plane_state->base.color_encoding == DRM_COLOR_YCBCR_BT709)
> plane_color_ctl |= PLANE_COLOR_CSC_MODE_YUV709_TO_RGB709;
> else
> @@ -4228,31 +4228,31 @@ __intel_display_resume(struct drm_device *dev,
> return ret;
> }
>
> -static bool gpu_reset_clobbers_display(struct drm_i915_private *dev_priv)
> +static bool gpu_reset_clobbers_display(struct drm_i915_private *i915)
> {
> - return (INTEL_INFO(dev_priv)->gpu_reset_clobbers_display &&
> - intel_has_gpu_reset(dev_priv));
> + return (INTEL_INFO(i915)->gpu_reset_clobbers_display &&
> + intel_has_gpu_reset(i915));
> }
>
> -void intel_prepare_reset(struct drm_i915_private *dev_priv)
> +void intel_prepare_reset(struct drm_i915_private *i915)
> {
> - struct drm_device *dev = &dev_priv->drm;
> - struct drm_modeset_acquire_ctx *ctx = &dev_priv->reset_ctx;
> + struct drm_device *dev = &i915->drm;
> + struct drm_modeset_acquire_ctx *ctx = &i915->reset_ctx;
> struct drm_atomic_state *state;
> int ret;
>
> /* reset doesn't touch the display */
> if (!i915_modparams.force_reset_modeset_test &&
> - !gpu_reset_clobbers_display(dev_priv))
> + !gpu_reset_clobbers_display(i915))
> return;
>
> /* We have a modeset vs reset deadlock, defensively unbreak it. */
> - set_bit(I915_RESET_MODESET, &dev_priv->gpu_error.flags);
> - wake_up_all(&dev_priv->gpu_error.wait_queue);
> + set_bit(I915_RESET_MODESET, &i915->gpu_error.flags);
> + wake_up_all(&i915->gpu_error.wait_queue);
>
> - if (atomic_read(&dev_priv->gpu_error.pending_fb_pin)) {
> + if (atomic_read(&i915->gpu_error.pending_fb_pin)) {
> DRM_DEBUG_KMS("Modeset potentially stuck, unbreaking through wedging\n");
> - i915_gem_set_wedged(dev_priv);
> + i915_gem_set_wedged(i915);
> }
>
> /*
> @@ -4286,27 +4286,27 @@ void intel_prepare_reset(struct drm_i915_private *dev_priv)
> return;
> }
>
> - dev_priv->modeset_restore_state = state;
> + i915->modeset_restore_state = state;
> state->acquire_ctx = ctx;
> }
>
> -void intel_finish_reset(struct drm_i915_private *dev_priv)
> +void intel_finish_reset(struct drm_i915_private *i915)
> {
> - struct drm_device *dev = &dev_priv->drm;
> - struct drm_modeset_acquire_ctx *ctx = &dev_priv->reset_ctx;
> + struct drm_device *dev = &i915->drm;
> + struct drm_modeset_acquire_ctx *ctx = &i915->reset_ctx;
> struct drm_atomic_state *state;
> int ret;
>
> /* reset doesn't touch the display */
> - if (!test_bit(I915_RESET_MODESET, &dev_priv->gpu_error.flags))
> + if (!test_bit(I915_RESET_MODESET, &i915->gpu_error.flags))
> return;
>
> - state = fetch_and_zero(&dev_priv->modeset_restore_state);
> + state = fetch_and_zero(&i915->modeset_restore_state);
> if (!state)
> goto unlock;
>
> /* reset doesn't touch the display */
> - if (!gpu_reset_clobbers_display(dev_priv)) {
> + if (!gpu_reset_clobbers_display(i915)) {
> /* for testing only restore the display */
> ret = __intel_display_resume(dev, state, ctx);
> if (ret)
> @@ -4316,20 +4316,20 @@ void intel_finish_reset(struct drm_i915_private *dev_priv)
> * The display has been reset as well,
> * so need a full re-initialization.
> */
> - intel_pps_unlock_regs_wa(dev_priv);
> + intel_pps_unlock_regs_wa(i915);
> intel_modeset_init_hw(dev);
> - intel_init_clock_gating(dev_priv);
> + intel_init_clock_gating(i915);
>
> - spin_lock_irq(&dev_priv->irq_lock);
> - if (dev_priv->display.hpd_irq_setup)
> - dev_priv->display.hpd_irq_setup(dev_priv);
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
> + if (i915->display.hpd_irq_setup)
> + i915->display.hpd_irq_setup(i915);
> + spin_unlock_irq(&i915->irq_lock);
>
> ret = __intel_display_resume(dev, state, ctx);
> if (ret)
> DRM_ERROR("Restoring old state failed with %i\n", ret);
>
> - intel_hpd_init(dev_priv);
> + intel_hpd_init(i915);
> }
>
> drm_atomic_state_put(state);
> @@ -4338,12 +4338,12 @@ void intel_finish_reset(struct drm_i915_private *dev_priv)
> drm_modeset_acquire_fini(ctx);
> mutex_unlock(&dev->mode_config.mutex);
>
> - clear_bit(I915_RESET_MODESET, &dev_priv->gpu_error.flags);
> + clear_bit(I915_RESET_MODESET, &i915->gpu_error.flags);
> }
>
> static void icl_set_pipe_chicken(struct intel_crtc *crtc)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
> u32 tmp;
>
> @@ -4369,7 +4369,7 @@ static void intel_update_pipe_config(const struct intel_crtc_state *old_crtc_sta
> const struct intel_crtc_state *new_crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(new_crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
>
> /* drm_atomic_helper_update_legacy_modeset_state might not be called. */
> crtc->base.mode = new_crtc_state->base.mode;
> @@ -4388,26 +4388,26 @@ static void intel_update_pipe_config(const struct intel_crtc_state *old_crtc_sta
> (new_crtc_state->pipe_src_h - 1));
>
> /* on skylake this is done by detaching scalers */
> - if (INTEL_GEN(dev_priv) >= 9) {
> + if (INTEL_GEN(i915) >= 9) {
> skl_detach_scalers(new_crtc_state);
>
> if (new_crtc_state->pch_pfit.enabled)
> skylake_pfit_enable(new_crtc_state);
> - } else if (HAS_PCH_SPLIT(dev_priv)) {
> + } else if (HAS_PCH_SPLIT(i915)) {
> if (new_crtc_state->pch_pfit.enabled)
> ironlake_pfit_enable(new_crtc_state);
> else if (old_crtc_state->pch_pfit.enabled)
> ironlake_pfit_disable(old_crtc_state);
> }
>
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> icl_set_pipe_chicken(crtc);
> }
>
> static void intel_fdi_normal_train(struct intel_crtc *crtc)
> {
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> int pipe = crtc->pipe;
> i915_reg_t reg;
> u32 temp;
> @@ -4415,7 +4415,7 @@ static void intel_fdi_normal_train(struct intel_crtc *crtc)
> /* enable normal train */
> reg = FDI_TX_CTL(pipe);
> temp = I915_READ(reg);
> - if (IS_IVYBRIDGE(dev_priv)) {
> + if (IS_IVYBRIDGE(i915)) {
> temp &= ~FDI_LINK_TRAIN_NONE_IVB;
> temp |= FDI_LINK_TRAIN_NONE_IVB | FDI_TX_ENHANCE_FRAME_ENABLE;
> } else {
> @@ -4426,7 +4426,7 @@ static void intel_fdi_normal_train(struct intel_crtc *crtc)
>
> reg = FDI_RX_CTL(pipe);
> temp = I915_READ(reg);
> - if (HAS_PCH_CPT(dev_priv)) {
> + if (HAS_PCH_CPT(i915)) {
> temp &= ~FDI_LINK_TRAIN_PATTERN_MASK_CPT;
> temp |= FDI_LINK_TRAIN_NORMAL_CPT;
> } else {
> @@ -4440,7 +4440,7 @@ static void intel_fdi_normal_train(struct intel_crtc *crtc)
> udelay(1000);
>
> /* IVB wants error correction enabled */
> - if (IS_IVYBRIDGE(dev_priv))
> + if (IS_IVYBRIDGE(i915))
> I915_WRITE(reg, I915_READ(reg) | FDI_FS_ERRC_ENABLE |
> FDI_FE_ERRC_ENABLE);
> }
> @@ -4450,13 +4450,13 @@ static void ironlake_fdi_link_train(struct intel_crtc *crtc,
> const struct intel_crtc_state *crtc_state)
> {
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> int pipe = crtc->pipe;
> i915_reg_t reg;
> u32 temp, tries;
>
> /* FDI needs bits from pipe first */
> - assert_pipe_enabled(dev_priv, pipe);
> + assert_pipe_enabled(i915, pipe);
>
> /* Train 1: umask FDI RX Interrupt symbol_lock and bit_lock bit
> for train result */
> @@ -4551,7 +4551,7 @@ static void gen6_fdi_link_train(struct intel_crtc *crtc,
> const struct intel_crtc_state *crtc_state)
> {
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> int pipe = crtc->pipe;
> i915_reg_t reg;
> u32 temp, i, retry;
> @@ -4584,7 +4584,7 @@ static void gen6_fdi_link_train(struct intel_crtc *crtc,
>
> reg = FDI_RX_CTL(pipe);
> temp = I915_READ(reg);
> - if (HAS_PCH_CPT(dev_priv)) {
> + if (HAS_PCH_CPT(i915)) {
> temp &= ~FDI_LINK_TRAIN_PATTERN_MASK_CPT;
> temp |= FDI_LINK_TRAIN_PATTERN_1_CPT;
> } else {
> @@ -4628,7 +4628,7 @@ static void gen6_fdi_link_train(struct intel_crtc *crtc,
> temp = I915_READ(reg);
> temp &= ~FDI_LINK_TRAIN_NONE;
> temp |= FDI_LINK_TRAIN_PATTERN_2;
> - if (IS_GEN(dev_priv, 6)) {
> + if (IS_GEN(i915, 6)) {
> temp &= ~FDI_LINK_TRAIN_VOL_EMP_MASK;
> /* SNB-B */
> temp |= FDI_LINK_TRAIN_400MV_0DB_SNB_B;
> @@ -4637,7 +4637,7 @@ static void gen6_fdi_link_train(struct intel_crtc *crtc,
>
> reg = FDI_RX_CTL(pipe);
> temp = I915_READ(reg);
> - if (HAS_PCH_CPT(dev_priv)) {
> + if (HAS_PCH_CPT(i915)) {
> temp &= ~FDI_LINK_TRAIN_PATTERN_MASK_CPT;
> temp |= FDI_LINK_TRAIN_PATTERN_2_CPT;
> } else {
> @@ -4684,7 +4684,7 @@ static void ivb_manual_fdi_link_train(struct intel_crtc *crtc,
> const struct intel_crtc_state *crtc_state)
> {
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> int pipe = crtc->pipe;
> i915_reg_t reg;
> u32 temp, i, j;
> @@ -4802,7 +4802,7 @@ static void ivb_manual_fdi_link_train(struct intel_crtc *crtc,
> static void ironlake_fdi_pll_enable(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(intel_crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(intel_crtc->base.dev);
> int pipe = intel_crtc->pipe;
> i915_reg_t reg;
> u32 temp;
> @@ -4839,7 +4839,7 @@ static void ironlake_fdi_pll_enable(const struct intel_crtc_state *crtc_state)
> static void ironlake_fdi_pll_disable(struct intel_crtc *intel_crtc)
> {
> struct drm_device *dev = intel_crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> int pipe = intel_crtc->pipe;
> i915_reg_t reg;
> u32 temp;
> @@ -4869,7 +4869,7 @@ static void ironlake_fdi_pll_disable(struct intel_crtc *intel_crtc)
> static void ironlake_fdi_disable(struct drm_crtc *crtc)
> {
> struct drm_device *dev = crtc->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> int pipe = intel_crtc->pipe;
> i915_reg_t reg;
> @@ -4891,7 +4891,7 @@ static void ironlake_fdi_disable(struct drm_crtc *crtc)
> udelay(100);
>
> /* Ironlake workaround, disable clock pointer after downing FDI */
> - if (HAS_PCH_IBX(dev_priv))
> + if (HAS_PCH_IBX(i915))
> I915_WRITE(FDI_RX_CHICKEN(pipe), FDI_RX_PHASE_SYNC_POINTER_OVR);
>
> /* still set train pattern 1 */
> @@ -4903,7 +4903,7 @@ static void ironlake_fdi_disable(struct drm_crtc *crtc)
>
> reg = FDI_RX_CTL(pipe);
> temp = I915_READ(reg);
> - if (HAS_PCH_CPT(dev_priv)) {
> + if (HAS_PCH_CPT(i915)) {
> temp &= ~FDI_LINK_TRAIN_PATTERN_MASK_CPT;
> temp |= FDI_LINK_TRAIN_PATTERN_1_CPT;
> } else {
> @@ -4919,12 +4919,12 @@ static void ironlake_fdi_disable(struct drm_crtc *crtc)
> udelay(100);
> }
>
> -bool intel_has_pending_fb_unpin(struct drm_i915_private *dev_priv)
> +bool intel_has_pending_fb_unpin(struct drm_i915_private *i915)
> {
> struct drm_crtc *crtc;
> bool cleanup_done;
>
> - drm_for_each_crtc(crtc, &dev_priv->drm) {
> + drm_for_each_crtc(crtc, &i915->drm) {
> struct drm_crtc_commit *commit;
> spin_lock(&crtc->commit_lock);
> commit = list_first_entry_or_null(&crtc->commit_list,
> @@ -4944,31 +4944,31 @@ bool intel_has_pending_fb_unpin(struct drm_i915_private *dev_priv)
> return false;
> }
>
> -void lpt_disable_iclkip(struct drm_i915_private *dev_priv)
> +void lpt_disable_iclkip(struct drm_i915_private *i915)
> {
> u32 temp;
>
> I915_WRITE(PIXCLK_GATE, PIXCLK_GATE_GATE);
>
> - mutex_lock(&dev_priv->sb_lock);
> + mutex_lock(&i915->sb_lock);
>
> - temp = intel_sbi_read(dev_priv, SBI_SSCCTL6, SBI_ICLK);
> + temp = intel_sbi_read(i915, SBI_SSCCTL6, SBI_ICLK);
> temp |= SBI_SSCCTL_DISABLE;
> - intel_sbi_write(dev_priv, SBI_SSCCTL6, temp, SBI_ICLK);
> + intel_sbi_write(i915, SBI_SSCCTL6, temp, SBI_ICLK);
>
> - mutex_unlock(&dev_priv->sb_lock);
> + mutex_unlock(&i915->sb_lock);
> }
>
> /* Program iCLKIP clock to the desired frequency */
> static void lpt_program_iclkip(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> int clock = crtc_state->base.adjusted_mode.crtc_clock;
> u32 divsel, phaseinc, auxdiv, phasedir = 0;
> u32 temp;
>
> - lpt_disable_iclkip(dev_priv);
> + lpt_disable_iclkip(i915);
>
> /* The iCLK virtual clock root frequency is in MHz,
> * but the adjusted_mode->crtc_clock in in KHz. To get the
> @@ -5007,30 +5007,30 @@ static void lpt_program_iclkip(const struct intel_crtc_state *crtc_state)
> phasedir,
> phaseinc);
>
> - mutex_lock(&dev_priv->sb_lock);
> + mutex_lock(&i915->sb_lock);
>
> /* Program SSCDIVINTPHASE6 */
> - temp = intel_sbi_read(dev_priv, SBI_SSCDIVINTPHASE6, SBI_ICLK);
> + temp = intel_sbi_read(i915, SBI_SSCDIVINTPHASE6, SBI_ICLK);
> temp &= ~SBI_SSCDIVINTPHASE_DIVSEL_MASK;
> temp |= SBI_SSCDIVINTPHASE_DIVSEL(divsel);
> temp &= ~SBI_SSCDIVINTPHASE_INCVAL_MASK;
> temp |= SBI_SSCDIVINTPHASE_INCVAL(phaseinc);
> temp |= SBI_SSCDIVINTPHASE_DIR(phasedir);
> temp |= SBI_SSCDIVINTPHASE_PROPAGATE;
> - intel_sbi_write(dev_priv, SBI_SSCDIVINTPHASE6, temp, SBI_ICLK);
> + intel_sbi_write(i915, SBI_SSCDIVINTPHASE6, temp, SBI_ICLK);
>
> /* Program SSCAUXDIV */
> - temp = intel_sbi_read(dev_priv, SBI_SSCAUXDIV6, SBI_ICLK);
> + temp = intel_sbi_read(i915, SBI_SSCAUXDIV6, SBI_ICLK);
> temp &= ~SBI_SSCAUXDIV_FINALDIV2SEL(1);
> temp |= SBI_SSCAUXDIV_FINALDIV2SEL(auxdiv);
> - intel_sbi_write(dev_priv, SBI_SSCAUXDIV6, temp, SBI_ICLK);
> + intel_sbi_write(i915, SBI_SSCAUXDIV6, temp, SBI_ICLK);
>
> /* Enable modulator and associated divider */
> - temp = intel_sbi_read(dev_priv, SBI_SSCCTL6, SBI_ICLK);
> + temp = intel_sbi_read(i915, SBI_SSCCTL6, SBI_ICLK);
> temp &= ~SBI_SSCCTL_DISABLE;
> - intel_sbi_write(dev_priv, SBI_SSCCTL6, temp, SBI_ICLK);
> + intel_sbi_write(i915, SBI_SSCCTL6, temp, SBI_ICLK);
>
> - mutex_unlock(&dev_priv->sb_lock);
> + mutex_unlock(&i915->sb_lock);
>
> /* Wait for initialization time */
> udelay(24);
> @@ -5038,7 +5038,7 @@ static void lpt_program_iclkip(const struct intel_crtc_state *crtc_state)
> I915_WRITE(PIXCLK_GATE, PIXCLK_GATE_UNGATE);
> }
>
> -int lpt_get_iclkip(struct drm_i915_private *dev_priv)
> +int lpt_get_iclkip(struct drm_i915_private *i915)
> {
> u32 divsel, phaseinc, auxdiv;
> u32 iclk_virtual_root_freq = 172800 * 1000;
> @@ -5049,25 +5049,25 @@ int lpt_get_iclkip(struct drm_i915_private *dev_priv)
> if ((I915_READ(PIXCLK_GATE) & PIXCLK_GATE_UNGATE) == 0)
> return 0;
>
> - mutex_lock(&dev_priv->sb_lock);
> + mutex_lock(&i915->sb_lock);
>
> - temp = intel_sbi_read(dev_priv, SBI_SSCCTL6, SBI_ICLK);
> + temp = intel_sbi_read(i915, SBI_SSCCTL6, SBI_ICLK);
> if (temp & SBI_SSCCTL_DISABLE) {
> - mutex_unlock(&dev_priv->sb_lock);
> + mutex_unlock(&i915->sb_lock);
> return 0;
> }
>
> - temp = intel_sbi_read(dev_priv, SBI_SSCDIVINTPHASE6, SBI_ICLK);
> + temp = intel_sbi_read(i915, SBI_SSCDIVINTPHASE6, SBI_ICLK);
> divsel = (temp & SBI_SSCDIVINTPHASE_DIVSEL_MASK) >>
> SBI_SSCDIVINTPHASE_DIVSEL_SHIFT;
> phaseinc = (temp & SBI_SSCDIVINTPHASE_INCVAL_MASK) >>
> SBI_SSCDIVINTPHASE_INCVAL_SHIFT;
>
> - temp = intel_sbi_read(dev_priv, SBI_SSCAUXDIV6, SBI_ICLK);
> + temp = intel_sbi_read(i915, SBI_SSCAUXDIV6, SBI_ICLK);
> auxdiv = (temp & SBI_SSCAUXDIV_FINALDIV2SEL_MASK) >>
> SBI_SSCAUXDIV_FINALDIV2SEL_SHIFT;
>
> - mutex_unlock(&dev_priv->sb_lock);
> + mutex_unlock(&i915->sb_lock);
>
> desired_divisor = (divsel + 2) * iclk_pi_range + phaseinc;
>
> @@ -5079,7 +5079,7 @@ static void ironlake_pch_transcoder_set_timings(const struct intel_crtc_state *c
> enum pipe pch_transcoder)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
>
> I915_WRITE(PCH_TRANS_HTOTAL(pch_transcoder),
> @@ -5099,7 +5099,7 @@ static void ironlake_pch_transcoder_set_timings(const struct intel_crtc_state *c
> I915_READ(VSYNCSHIFT(cpu_transcoder)));
> }
>
> -static void cpt_set_fdi_bc_bifurcation(struct drm_i915_private *dev_priv, bool enable)
> +static void cpt_set_fdi_bc_bifurcation(struct drm_i915_private *i915, bool enable)
> {
> u32 temp;
>
> @@ -5122,20 +5122,20 @@ static void cpt_set_fdi_bc_bifurcation(struct drm_i915_private *dev_priv, bool e
> static void ivybridge_update_fdi_bc_bifurcation(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
>
> switch (crtc->pipe) {
> case PIPE_A:
> break;
> case PIPE_B:
> if (crtc_state->fdi_lanes > 2)
> - cpt_set_fdi_bc_bifurcation(dev_priv, false);
> + cpt_set_fdi_bc_bifurcation(i915, false);
> else
> - cpt_set_fdi_bc_bifurcation(dev_priv, true);
> + cpt_set_fdi_bc_bifurcation(i915, true);
>
> break;
> case PIPE_C:
> - cpt_set_fdi_bc_bifurcation(dev_priv, true);
> + cpt_set_fdi_bc_bifurcation(i915, true);
>
> break;
> default:
> @@ -5185,13 +5185,13 @@ static void ironlake_pch_enable(const struct intel_atomic_state *state,
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> int pipe = crtc->pipe;
> u32 temp;
>
> - assert_pch_transcoder_disabled(dev_priv, pipe);
> + assert_pch_transcoder_disabled(i915, pipe);
>
> - if (IS_IVYBRIDGE(dev_priv))
> + if (IS_IVYBRIDGE(i915))
> ivybridge_update_fdi_bc_bifurcation(crtc_state);
>
> /* Write the TU size bits before fdi link training, so that error
> @@ -5200,18 +5200,18 @@ static void ironlake_pch_enable(const struct intel_atomic_state *state,
> I915_READ(PIPE_DATA_M1(pipe)) & TU_SIZE_MASK);
>
> /* For PCH output, training FDI link */
> - dev_priv->display.fdi_link_train(crtc, crtc_state);
> + i915->display.fdi_link_train(crtc, crtc_state);
>
> /* We need to program the right clock selection before writing the pixel
> * mutliplier into the DPLL. */
> - if (HAS_PCH_CPT(dev_priv)) {
> + if (HAS_PCH_CPT(i915)) {
> u32 sel;
>
> temp = I915_READ(PCH_DPLL_SEL);
> temp |= TRANS_DPLL_ENABLE(pipe);
> sel = TRANS_DPLLB_SEL(pipe);
> if (crtc_state->shared_dpll ==
> - intel_get_shared_dpll_by_id(dev_priv, DPLL_ID_PCH_PLL_B))
> + intel_get_shared_dpll_by_id(i915, DPLL_ID_PCH_PLL_B))
> temp |= sel;
> else
> temp &= ~sel;
> @@ -5228,13 +5228,13 @@ static void ironlake_pch_enable(const struct intel_atomic_state *state,
> intel_enable_shared_dpll(crtc_state);
>
> /* set transcoder timing, panel must allow it */
> - assert_panel_unlocked(dev_priv, pipe);
> + assert_panel_unlocked(i915, pipe);
> ironlake_pch_transcoder_set_timings(crtc_state, pipe);
>
> intel_fdi_normal_train(crtc);
>
> /* For PCH DP, enable TRANS_DP_CTL */
> - if (HAS_PCH_CPT(dev_priv) &&
> + if (HAS_PCH_CPT(i915) &&
> intel_crtc_has_dp_encoder(crtc_state)) {
> const struct drm_display_mode *adjusted_mode =
> &crtc_state->base.adjusted_mode;
> @@ -5268,22 +5268,22 @@ static void lpt_pch_enable(const struct intel_atomic_state *state,
> const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
>
> - assert_pch_transcoder_disabled(dev_priv, PIPE_A);
> + assert_pch_transcoder_disabled(i915, PIPE_A);
>
> lpt_program_iclkip(crtc_state);
>
> /* Set transcoder timing. */
> ironlake_pch_transcoder_set_timings(crtc_state, PIPE_A);
>
> - lpt_enable_pch_transcoder(dev_priv, cpu_transcoder);
> + lpt_enable_pch_transcoder(i915, cpu_transcoder);
> }
>
> static void cpt_verify_modeset(struct drm_device *dev, int pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> i915_reg_t dslreg = PIPEDSL(pipe);
> u32 temp;
>
> @@ -5385,7 +5385,7 @@ skl_update_scaler(struct intel_crtc_state *crtc_state, bool force_detach,
> &crtc_state->scaler_state;
> struct intel_crtc *intel_crtc =
> to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(intel_crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(intel_crtc->base.dev);
> const struct drm_display_mode *adjusted_mode =
> &crtc_state->base.adjusted_mode;
>
> @@ -5403,7 +5403,7 @@ skl_update_scaler(struct intel_crtc_state *crtc_state, bool force_detach,
> * Once NV12 is enabled, handle it here while allocating scaler
> * for NV12.
> */
> - if (INTEL_GEN(dev_priv) >= 9 && crtc_state->base.enable &&
> + if (INTEL_GEN(i915) >= 9 && crtc_state->base.enable &&
> need_scaler && adjusted_mode->flags & DRM_MODE_FLAG_INTERLACE) {
> DRM_DEBUG_KMS("Pipe/Plane scaling not supported with IF-ID mode\n");
> return -EINVAL;
> @@ -5442,10 +5442,10 @@ skl_update_scaler(struct intel_crtc_state *crtc_state, bool force_detach,
> /* range checks */
> if (src_w < SKL_MIN_SRC_W || src_h < SKL_MIN_SRC_H ||
> dst_w < SKL_MIN_DST_W || dst_h < SKL_MIN_DST_H ||
> - (INTEL_GEN(dev_priv) >= 11 &&
> + (INTEL_GEN(i915) >= 11 &&
> (src_w > ICL_MAX_SRC_W || src_h > ICL_MAX_SRC_H ||
> dst_w > ICL_MAX_DST_W || dst_h > ICL_MAX_DST_H)) ||
> - (INTEL_GEN(dev_priv) < 11 &&
> + (INTEL_GEN(i915) < 11 &&
> (src_w > SKL_MAX_SRC_W || src_h > SKL_MAX_SRC_H ||
> dst_w > SKL_MAX_DST_W || dst_h > SKL_MAX_DST_H))) {
> DRM_DEBUG_KMS("scaler_user index %u.%u: src %ux%u dst %ux%u "
> @@ -5502,14 +5502,14 @@ static int skl_update_scaler_plane(struct intel_crtc_state *crtc_state,
> {
> struct intel_plane *intel_plane =
> to_intel_plane(plane_state->base.plane);
> - struct drm_i915_private *dev_priv = to_i915(intel_plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(intel_plane->base.dev);
> struct drm_framebuffer *fb = plane_state->base.fb;
> int ret;
> bool force_detach = !fb || !plane_state->base.visible;
> bool need_scaler = false;
>
> /* Pre-gen11 and SDR planes always need a scaler for planar formats. */
> - if (!icl_is_hdr_plane(dev_priv, intel_plane->id) &&
> + if (!icl_is_hdr_plane(i915, intel_plane->id) &&
> fb && is_planar_yuv_format(fb->format->format))
> need_scaler = true;
>
> @@ -5582,7 +5582,7 @@ static void skylake_scaler_disable(struct intel_crtc *crtc)
> static void skylake_pfit_enable(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
> const struct intel_crtc_scaler_state *scaler_state =
> &crtc_state->scaler_state;
> @@ -5619,7 +5619,7 @@ static void skylake_pfit_enable(const struct intel_crtc_state *crtc_state)
> static void ironlake_pfit_enable(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> int pipe = crtc->pipe;
>
> if (crtc_state->pch_pfit.enabled) {
> @@ -5627,7 +5627,7 @@ static void ironlake_pfit_enable(const struct intel_crtc_state *crtc_state)
> * as some pre-programmed values are broken,
> * e.g. x201.
> */
> - if (IS_IVYBRIDGE(dev_priv) || IS_HASWELL(dev_priv))
> + if (IS_IVYBRIDGE(i915) || IS_HASWELL(i915))
> I915_WRITE(PF_CTL(pipe), PF_ENABLE | PF_FILTER_MED_3x3 |
> PF_PIPE_SEL_IVB(pipe));
> else
> @@ -5641,7 +5641,7 @@ void hsw_enable_ips(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
>
> if (!crtc_state->ips_enabled)
> return;
> @@ -5653,8 +5653,8 @@ void hsw_enable_ips(const struct intel_crtc_state *crtc_state)
> */
> WARN_ON(!(crtc_state->active_planes & ~BIT(PLANE_CURSOR)));
>
> - if (IS_BROADWELL(dev_priv)) {
> - WARN_ON(sandybridge_pcode_write(dev_priv, DISPLAY_IPS_CONTROL,
> + if (IS_BROADWELL(i915)) {
> + WARN_ON(sandybridge_pcode_write(i915, DISPLAY_IPS_CONTROL,
> IPS_ENABLE | IPS_PCODE_CONTROL));
> /* Quoting Art Runyan: "its not safe to expect any particular
> * value in IPS_CTL bit 31 after enabling IPS through the
> @@ -5668,7 +5668,7 @@ void hsw_enable_ips(const struct intel_crtc_state *crtc_state)
> * and don't wait for vblanks until the end of crtc_enable, then
> * the HW state readout code will complain that the expected
> * IPS_CTL value is not the one we read. */
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> IPS_CTL, IPS_ENABLE, IPS_ENABLE,
> 50))
> DRM_ERROR("Timed out waiting for IPS enable\n");
> @@ -5679,19 +5679,19 @@ void hsw_disable_ips(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
>
> if (!crtc_state->ips_enabled)
> return;
>
> - if (IS_BROADWELL(dev_priv)) {
> - WARN_ON(sandybridge_pcode_write(dev_priv, DISPLAY_IPS_CONTROL, 0));
> + if (IS_BROADWELL(i915)) {
> + WARN_ON(sandybridge_pcode_write(i915, DISPLAY_IPS_CONTROL, 0));
> /*
> * Wait for PCODE to finish disabling IPS. The BSpec specified
> * 42ms timeout value leads to occasional timeouts so use 100ms
> * instead.
> */
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> IPS_CTL, IPS_ENABLE, 0,
> 100))
> DRM_ERROR("Timed out waiting for IPS disable\n");
> @@ -5701,7 +5701,7 @@ void hsw_disable_ips(const struct intel_crtc_state *crtc_state)
> }
>
> /* We need to wait for a vblank before we can disable the plane. */
> - intel_wait_for_vblank(dev_priv, crtc->pipe);
> + intel_wait_for_vblank(i915, crtc->pipe);
> }
>
> static void intel_crtc_dpms_overlay_disable(struct intel_crtc *intel_crtc)
> @@ -5735,7 +5735,7 @@ intel_post_enable_primary(struct drm_crtc *crtc,
> const struct intel_crtc_state *new_crtc_state)
> {
> struct drm_device *dev = crtc->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> int pipe = intel_crtc->pipe;
>
> @@ -5746,12 +5746,12 @@ intel_post_enable_primary(struct drm_crtc *crtc,
> * FIXME: Need to fix the logic to work when we turn off all planes
> * but leave the pipe running.
> */
> - if (IS_GEN(dev_priv, 2))
> - intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, true);
> + if (IS_GEN(i915, 2))
> + intel_set_cpu_fifo_underrun_reporting(i915, pipe, true);
>
> /* Underruns don't always raise interrupts, so check manually. */
> - intel_check_cpu_fifo_underruns(dev_priv);
> - intel_check_pch_fifo_underruns(dev_priv);
> + intel_check_cpu_fifo_underruns(i915);
> + intel_check_pch_fifo_underruns(i915);
> }
>
> /* FIXME get rid of this and use pre_plane_update */
> @@ -5759,7 +5759,7 @@ static void
> intel_pre_disable_primary_noatomic(struct drm_crtc *crtc)
> {
> struct drm_device *dev = crtc->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> int pipe = intel_crtc->pipe;
>
> @@ -5767,8 +5767,8 @@ intel_pre_disable_primary_noatomic(struct drm_crtc *crtc)
> * Gen2 reports pipe underruns whenever all planes are disabled.
> * So disable underrun reporting before all the planes get disabled.
> */
> - if (IS_GEN(dev_priv, 2))
> - intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, false);
> + if (IS_GEN(i915, 2))
> + intel_set_cpu_fifo_underrun_reporting(i915, pipe, false);
>
> hsw_disable_ips(to_intel_crtc_state(crtc->state));
>
> @@ -5781,16 +5781,16 @@ intel_pre_disable_primary_noatomic(struct drm_crtc *crtc)
> * event which is after the vblank start event, so we need to have a
> * wait-for-vblank between disabling the plane and the pipe.
> */
> - if (HAS_GMCH(dev_priv) &&
> - intel_set_memory_cxsr(dev_priv, false))
> - intel_wait_for_vblank(dev_priv, pipe);
> + if (HAS_GMCH(i915) &&
> + intel_set_memory_cxsr(i915, false))
> + intel_wait_for_vblank(i915, pipe);
> }
>
> static bool hsw_pre_update_disable_ips(const struct intel_crtc_state *old_crtc_state,
> const struct intel_crtc_state *new_crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(new_crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
>
> if (!old_crtc_state->ips_enabled)
> return false;
> @@ -5804,7 +5804,7 @@ static bool hsw_pre_update_disable_ips(const struct intel_crtc_state *old_crtc_s
> *
> * Disable IPS before we program the LUT.
> */
> - if (IS_HASWELL(dev_priv) &&
> + if (IS_HASWELL(i915) &&
> (new_crtc_state->base.color_mgmt_changed ||
> new_crtc_state->update_pipe) &&
> new_crtc_state->gamma_mode == GAMMA_MODE_MODE_SPLIT)
> @@ -5817,7 +5817,7 @@ static bool hsw_post_update_enable_ips(const struct intel_crtc_state *old_crtc_s
> const struct intel_crtc_state *new_crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(new_crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
>
> if (!new_crtc_state->ips_enabled)
> return false;
> @@ -5831,7 +5831,7 @@ static bool hsw_post_update_enable_ips(const struct intel_crtc_state *old_crtc_s
> *
> * Re-enable IPS after the LUT has been programmed.
> */
> - if (IS_HASWELL(dev_priv) &&
> + if (IS_HASWELL(i915) &&
> (new_crtc_state->base.color_mgmt_changed ||
> new_crtc_state->update_pipe) &&
> new_crtc_state->gamma_mode == GAMMA_MODE_MODE_SPLIT)
> @@ -5848,24 +5848,24 @@ static bool hsw_post_update_enable_ips(const struct intel_crtc_state *old_crtc_s
> return !old_crtc_state->ips_enabled;
> }
>
> -static bool needs_nv12_wa(struct drm_i915_private *dev_priv,
> +static bool needs_nv12_wa(struct drm_i915_private *i915,
> const struct intel_crtc_state *crtc_state)
> {
> if (!crtc_state->nv12_planes)
> return false;
>
> /* WA Display #0827: Gen9:all */
> - if (IS_GEN(dev_priv, 9) && !IS_GEMINILAKE(dev_priv))
> + if (IS_GEN(i915, 9) && !IS_GEMINILAKE(i915))
> return true;
>
> return false;
> }
>
> -static bool needs_scalerclk_wa(struct drm_i915_private *dev_priv,
> +static bool needs_scalerclk_wa(struct drm_i915_private *i915,
> const struct intel_crtc_state *crtc_state)
> {
> /* Wa_2006604312:icl */
> - if (crtc_state->scaler_state.scaler_users > 0 && IS_ICELAKE(dev_priv))
> + if (crtc_state->scaler_state.scaler_users > 0 && IS_ICELAKE(i915))
> return true;
>
> return false;
> @@ -5875,7 +5875,7 @@ static void intel_post_plane_update(struct intel_crtc_state *old_crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc);
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct drm_atomic_state *old_state = old_crtc_state->base.state;
> struct intel_crtc_state *pipe_config =
> intel_atomic_get_new_crtc_state(to_intel_atomic_state(old_state),
> @@ -5904,13 +5904,13 @@ static void intel_post_plane_update(struct intel_crtc_state *old_crtc_state)
> intel_post_enable_primary(&crtc->base, pipe_config);
> }
>
> - if (needs_nv12_wa(dev_priv, old_crtc_state) &&
> - !needs_nv12_wa(dev_priv, pipe_config))
> - skl_wa_827(dev_priv, crtc->pipe, false);
> + if (needs_nv12_wa(i915, old_crtc_state) &&
> + !needs_nv12_wa(i915, pipe_config))
> + skl_wa_827(i915, crtc->pipe, false);
>
> - if (needs_scalerclk_wa(dev_priv, old_crtc_state) &&
> - !needs_scalerclk_wa(dev_priv, pipe_config))
> - icl_wa_scalerclkgating(dev_priv, crtc->pipe, false);
> + if (needs_scalerclk_wa(i915, old_crtc_state) &&
> + !needs_scalerclk_wa(i915, pipe_config))
> + icl_wa_scalerclkgating(i915, crtc->pipe, false);
> }
>
> static void intel_pre_plane_update(struct intel_crtc_state *old_crtc_state,
> @@ -5918,7 +5918,7 @@ static void intel_pre_plane_update(struct intel_crtc_state *old_crtc_state,
> {
> struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc);
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct drm_atomic_state *old_state = old_crtc_state->base.state;
> struct drm_plane *primary = crtc->base.primary;
> struct drm_plane_state *old_primary_state =
> @@ -5940,20 +5940,20 @@ static void intel_pre_plane_update(struct intel_crtc_state *old_crtc_state,
> * Gen2 reports pipe underruns whenever all planes are disabled.
> * So disable underrun reporting before all the planes get disabled.
> */
> - if (IS_GEN(dev_priv, 2) && old_primary_state->visible &&
> + if (IS_GEN(i915, 2) && old_primary_state->visible &&
> (modeset || !new_primary_state->base.visible))
> - intel_set_cpu_fifo_underrun_reporting(dev_priv, crtc->pipe, false);
> + intel_set_cpu_fifo_underrun_reporting(i915, crtc->pipe, false);
> }
>
> /* Display WA 827 */
> - if (!needs_nv12_wa(dev_priv, old_crtc_state) &&
> - needs_nv12_wa(dev_priv, pipe_config))
> - skl_wa_827(dev_priv, crtc->pipe, true);
> + if (!needs_nv12_wa(i915, old_crtc_state) &&
> + needs_nv12_wa(i915, pipe_config))
> + skl_wa_827(i915, crtc->pipe, true);
>
> /* Wa_2006604312:icl */
> - if (!needs_scalerclk_wa(dev_priv, old_crtc_state) &&
> - needs_scalerclk_wa(dev_priv, pipe_config))
> - icl_wa_scalerclkgating(dev_priv, crtc->pipe, true);
> + if (!needs_scalerclk_wa(i915, old_crtc_state) &&
> + needs_scalerclk_wa(i915, pipe_config))
> + icl_wa_scalerclkgating(i915, crtc->pipe, true);
>
> /*
> * Vblank time updates from the shadow to live plane control register
> @@ -5964,9 +5964,9 @@ static void intel_pre_plane_update(struct intel_crtc_state *old_crtc_state,
> * event which is after the vblank start event, so we need to have a
> * wait-for-vblank between disabling the plane and the pipe.
> */
> - if (HAS_GMCH(dev_priv) && old_crtc_state->base.active &&
> - pipe_config->disable_cxsr && intel_set_memory_cxsr(dev_priv, false))
> - intel_wait_for_vblank(dev_priv, crtc->pipe);
> + if (HAS_GMCH(i915) && old_crtc_state->base.active &&
> + pipe_config->disable_cxsr && intel_set_memory_cxsr(i915, false))
> + intel_wait_for_vblank(i915, crtc->pipe);
>
> /*
> * IVB workaround: must disable low power watermarks for at least
> @@ -5977,7 +5977,7 @@ static void intel_pre_plane_update(struct intel_crtc_state *old_crtc_state,
> */
> if (pipe_config->disable_lp_wm && ilk_disable_lp_wm(dev) &&
> old_crtc_state->base.active)
> - intel_wait_for_vblank(dev_priv, crtc->pipe);
> + intel_wait_for_vblank(i915, crtc->pipe);
>
> /*
> * If we're doing a modeset, we're done. No need to do any pre-vblank
> @@ -6000,8 +6000,8 @@ static void intel_pre_plane_update(struct intel_crtc_state *old_crtc_state,
> * we'll continue to update watermarks the old way, if flags tell
> * us to.
> */
> - if (dev_priv->display.initial_watermarks != NULL)
> - dev_priv->display.initial_watermarks(old_intel_state,
> + if (i915->display.initial_watermarks != NULL)
> + i915->display.initial_watermarks(old_intel_state,
> pipe_config);
> else if (pipe_config->update_wm_pre)
> intel_update_watermarks(crtc);
> @@ -6010,7 +6010,7 @@ static void intel_pre_plane_update(struct intel_crtc_state *old_crtc_state,
> static void intel_crtc_disable_planes(struct intel_atomic_state *state,
> struct intel_crtc *crtc)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> const struct intel_crtc_state *new_crtc_state =
> intel_atomic_get_new_crtc_state(state, crtc);
> unsigned int update_mask = new_crtc_state->update_planes;
> @@ -6032,7 +6032,7 @@ static void intel_crtc_disable_planes(struct intel_atomic_state *state,
> fb_bits |= plane->frontbuffer_bit;
> }
>
> - intel_frontbuffer_flip(dev_priv, fb_bits);
> + intel_frontbuffer_flip(i915, fb_bits);
> }
>
> static void intel_encoders_pre_pll_enable(struct drm_crtc *crtc,
> @@ -6190,7 +6190,7 @@ static void ironlake_crtc_enable(struct intel_crtc_state *pipe_config,
> {
> struct drm_crtc *crtc = pipe_config->base.crtc;
> struct drm_device *dev = crtc->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> int pipe = intel_crtc->pipe;
> struct intel_atomic_state *old_intel_state =
> @@ -6209,8 +6209,8 @@ static void ironlake_crtc_enable(struct intel_crtc_state *pipe_config,
> *
> * Spurious PCH underruns also occur during PCH enabling.
> */
> - intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, false);
> - intel_set_pch_fifo_underrun_reporting(dev_priv, pipe, false);
> + intel_set_cpu_fifo_underrun_reporting(i915, pipe, false);
> + intel_set_pch_fifo_underrun_reporting(i915, pipe, false);
>
> if (pipe_config->has_pch_encoder)
> intel_prepare_shared_dpll(pipe_config);
> @@ -6238,8 +6238,8 @@ static void ironlake_crtc_enable(struct intel_crtc_state *pipe_config,
> * enabling. */
> ironlake_fdi_pll_enable(pipe_config);
> } else {
> - assert_fdi_tx_disabled(dev_priv, pipe);
> - assert_fdi_rx_disabled(dev_priv, pipe);
> + assert_fdi_tx_disabled(i915, pipe);
> + assert_fdi_rx_disabled(i915, pipe);
> }
>
> ironlake_pfit_enable(pipe_config);
> @@ -6253,8 +6253,8 @@ static void ironlake_crtc_enable(struct intel_crtc_state *pipe_config,
> /* update DSPCNTR to configure gamma for pipe bottom color */
> intel_disable_primary_plane(pipe_config);
>
> - if (dev_priv->display.initial_watermarks != NULL)
> - dev_priv->display.initial_watermarks(old_intel_state, pipe_config);
> + if (i915->display.initial_watermarks != NULL)
> + i915->display.initial_watermarks(old_intel_state, pipe_config);
> intel_enable_pipe(pipe_config);
>
> if (pipe_config->has_pch_encoder)
> @@ -6265,7 +6265,7 @@ static void ironlake_crtc_enable(struct intel_crtc_state *pipe_config,
>
> intel_encoders_enable(crtc, pipe_config, old_state);
>
> - if (HAS_PCH_CPT(dev_priv))
> + if (HAS_PCH_CPT(i915))
> cpt_verify_modeset(dev, intel_crtc->pipe);
>
> /*
> @@ -6275,11 +6275,11 @@ static void ironlake_crtc_enable(struct intel_crtc_state *pipe_config,
> * in case there are more corner cases we don't know about.
> */
> if (pipe_config->has_pch_encoder) {
> - intel_wait_for_vblank(dev_priv, pipe);
> - intel_wait_for_vblank(dev_priv, pipe);
> + intel_wait_for_vblank(i915, pipe);
> + intel_wait_for_vblank(i915, pipe);
> }
> - intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, true);
> - intel_set_pch_fifo_underrun_reporting(dev_priv, pipe, true);
> + intel_set_cpu_fifo_underrun_reporting(i915, pipe, true);
> + intel_set_pch_fifo_underrun_reporting(i915, pipe, true);
> }
>
> /* IPS only exists on ULT machines and is tied to pipe A. */
> @@ -6288,7 +6288,7 @@ static bool hsw_crtc_supports_ips(struct intel_crtc *crtc)
> return HAS_IPS(to_i915(crtc->base.dev)) && crtc->pipe == PIPE_A;
> }
>
> -static void glk_pipe_scaler_clock_gating_wa(struct drm_i915_private *dev_priv,
> +static void glk_pipe_scaler_clock_gating_wa(struct drm_i915_private *i915,
> enum pipe pipe, bool apply)
> {
> u32 val = I915_READ(CLKGATE_DIS_PSL(pipe));
> @@ -6304,7 +6304,7 @@ static void glk_pipe_scaler_clock_gating_wa(struct drm_i915_private *dev_priv,
>
> static void icl_pipe_mbus_enable(struct intel_crtc *crtc)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
> u32 val;
>
> @@ -6319,7 +6319,7 @@ static void haswell_crtc_enable(struct intel_crtc_state *pipe_config,
> struct drm_atomic_state *old_state)
> {
> struct drm_crtc *crtc = pipe_config->base.crtc;
> - struct drm_i915_private *dev_priv = to_i915(crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc->dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> int pipe = intel_crtc->pipe, hsw_workaround_pipe;
> enum transcoder cpu_transcoder = pipe_config->cpu_transcoder;
> @@ -6359,18 +6359,18 @@ static void haswell_crtc_enable(struct intel_crtc_state *pipe_config,
> if (!transcoder_is_dsi(cpu_transcoder))
> haswell_set_pipeconf(pipe_config);
>
> - if (INTEL_GEN(dev_priv) >= 9 || IS_BROADWELL(dev_priv))
> + if (INTEL_GEN(i915) >= 9 || IS_BROADWELL(i915))
> bdw_set_pipemisc(pipe_config);
>
> intel_crtc->active = true;
>
> /* Display WA #1180: WaDisableScalarClockGating: glk, cnl */
> - psl_clkgate_wa = (IS_GEMINILAKE(dev_priv) || IS_CANNONLAKE(dev_priv)) &&
> + psl_clkgate_wa = (IS_GEMINILAKE(i915) || IS_CANNONLAKE(i915)) &&
> pipe_config->pch_pfit.enabled;
> if (psl_clkgate_wa)
> - glk_pipe_scaler_clock_gating_wa(dev_priv, pipe, true);
> + glk_pipe_scaler_clock_gating_wa(i915, pipe, true);
>
> - if (INTEL_GEN(dev_priv) >= 9)
> + if (INTEL_GEN(i915) >= 9)
> skylake_pfit_enable(pipe_config);
> else
> ironlake_pfit_enable(pipe_config);
> @@ -6382,20 +6382,20 @@ static void haswell_crtc_enable(struct intel_crtc_state *pipe_config,
> intel_color_load_luts(pipe_config);
> intel_color_commit(pipe_config);
> /* update DSPCNTR to configure gamma/csc for pipe bottom color */
> - if (INTEL_GEN(dev_priv) < 9)
> + if (INTEL_GEN(i915) < 9)
> intel_disable_primary_plane(pipe_config);
>
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> icl_set_pipe_chicken(intel_crtc);
>
> intel_ddi_set_pipe_settings(pipe_config);
> if (!transcoder_is_dsi(cpu_transcoder))
> intel_ddi_enable_transcoder_func(pipe_config);
>
> - if (dev_priv->display.initial_watermarks != NULL)
> - dev_priv->display.initial_watermarks(old_intel_state, pipe_config);
> + if (i915->display.initial_watermarks != NULL)
> + i915->display.initial_watermarks(old_intel_state, pipe_config);
>
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> icl_pipe_mbus_enable(intel_crtc);
>
> /* XXX: Do the pipe assertions at the right place for BXT DSI. */
> @@ -6414,23 +6414,23 @@ static void haswell_crtc_enable(struct intel_crtc_state *pipe_config,
> intel_encoders_enable(crtc, pipe_config, old_state);
>
> if (psl_clkgate_wa) {
> - intel_wait_for_vblank(dev_priv, pipe);
> - glk_pipe_scaler_clock_gating_wa(dev_priv, pipe, false);
> + intel_wait_for_vblank(i915, pipe);
> + glk_pipe_scaler_clock_gating_wa(i915, pipe, false);
> }
>
> /* If we change the relative order between pipe/planes enabling, we need
> * to change the workaround. */
> hsw_workaround_pipe = pipe_config->hsw_workaround_pipe;
> - if (IS_HASWELL(dev_priv) && hsw_workaround_pipe != INVALID_PIPE) {
> - intel_wait_for_vblank(dev_priv, hsw_workaround_pipe);
> - intel_wait_for_vblank(dev_priv, hsw_workaround_pipe);
> + if (IS_HASWELL(i915) && hsw_workaround_pipe != INVALID_PIPE) {
> + intel_wait_for_vblank(i915, hsw_workaround_pipe);
> + intel_wait_for_vblank(i915, hsw_workaround_pipe);
> }
> }
>
> static void ironlake_pfit_disable(const struct intel_crtc_state *old_crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
>
> /* To avoid upsetting the power well on haswell only disable the pfit if
> @@ -6447,7 +6447,7 @@ static void ironlake_crtc_disable(struct intel_crtc_state *old_crtc_state,
> {
> struct drm_crtc *crtc = old_crtc_state->base.crtc;
> struct drm_device *dev = crtc->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> int pipe = intel_crtc->pipe;
>
> @@ -6456,8 +6456,8 @@ static void ironlake_crtc_disable(struct intel_crtc_state *old_crtc_state,
> * pipe is already disabled, but FDI RX/TX is still enabled.
> * Happens at least with VGA+HDMI cloning. Suppress them.
> */
> - intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, false);
> - intel_set_pch_fifo_underrun_reporting(dev_priv, pipe, false);
> + intel_set_cpu_fifo_underrun_reporting(i915, pipe, false);
> + intel_set_pch_fifo_underrun_reporting(i915, pipe, false);
>
> intel_encoders_disable(crtc, old_crtc_state, old_state);
>
> @@ -6474,9 +6474,9 @@ static void ironlake_crtc_disable(struct intel_crtc_state *old_crtc_state,
> intel_encoders_post_disable(crtc, old_crtc_state, old_state);
>
> if (old_crtc_state->has_pch_encoder) {
> - ironlake_disable_pch_transcoder(dev_priv, pipe);
> + ironlake_disable_pch_transcoder(i915, pipe);
>
> - if (HAS_PCH_CPT(dev_priv)) {
> + if (HAS_PCH_CPT(i915)) {
> i915_reg_t reg;
> u32 temp;
>
> @@ -6497,15 +6497,15 @@ static void ironlake_crtc_disable(struct intel_crtc_state *old_crtc_state,
> ironlake_fdi_pll_disable(intel_crtc);
> }
>
> - intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, true);
> - intel_set_pch_fifo_underrun_reporting(dev_priv, pipe, true);
> + intel_set_cpu_fifo_underrun_reporting(i915, pipe, true);
> + intel_set_pch_fifo_underrun_reporting(i915, pipe, true);
> }
>
> static void haswell_crtc_disable(struct intel_crtc_state *old_crtc_state,
> struct drm_atomic_state *old_state)
> {
> struct drm_crtc *crtc = old_crtc_state->base.crtc;
> - struct drm_i915_private *dev_priv = to_i915(crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc->dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> enum transcoder cpu_transcoder = old_crtc_state->cpu_transcoder;
>
> @@ -6526,7 +6526,7 @@ static void haswell_crtc_disable(struct intel_crtc_state *old_crtc_state,
>
> intel_dsc_disable(old_crtc_state);
>
> - if (INTEL_GEN(dev_priv) >= 9)
> + if (INTEL_GEN(i915) >= 9)
> skylake_scaler_disable(intel_crtc);
> else
> ironlake_pfit_disable(old_crtc_state);
> @@ -6539,7 +6539,7 @@ static void haswell_crtc_disable(struct intel_crtc_state *old_crtc_state,
> static void i9xx_pfit_enable(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
>
> if (!crtc_state->gmch_pfit.control)
> return;
> @@ -6549,7 +6549,7 @@ static void i9xx_pfit_enable(const struct intel_crtc_state *crtc_state)
> * according to register description and PRM.
> */
> WARN_ON(I915_READ(PFIT_CONTROL) & PFIT_ENABLE);
> - assert_pipe_disabled(dev_priv, crtc->pipe);
> + assert_pipe_disabled(i915, crtc->pipe);
>
> I915_WRITE(PFIT_PGM_RATIOS, crtc_state->gmch_pfit.pgm_ratios);
> I915_WRITE(PFIT_CONTROL, crtc_state->gmch_pfit.control);
> @@ -6559,31 +6559,31 @@ static void i9xx_pfit_enable(const struct intel_crtc_state *crtc_state)
> I915_WRITE(BCLRPAT(crtc->pipe), 0);
> }
>
> -bool intel_port_is_combophy(struct drm_i915_private *dev_priv, enum port port)
> +bool intel_port_is_combophy(struct drm_i915_private *i915, enum port port)
> {
> if (port == PORT_NONE)
> return false;
>
> - if (IS_ELKHARTLAKE(dev_priv))
> + if (IS_ELKHARTLAKE(i915))
> return port <= PORT_C;
>
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> return port <= PORT_B;
>
> return false;
> }
>
> -bool intel_port_is_tc(struct drm_i915_private *dev_priv, enum port port)
> +bool intel_port_is_tc(struct drm_i915_private *i915, enum port port)
> {
> - if (INTEL_GEN(dev_priv) >= 11 && !IS_ELKHARTLAKE(dev_priv))
> + if (INTEL_GEN(i915) >= 11 && !IS_ELKHARTLAKE(i915))
> return port >= PORT_C && port <= PORT_F;
>
> return false;
> }
>
> -enum tc_port intel_port_to_tc(struct drm_i915_private *dev_priv, enum port port)
> +enum tc_port intel_port_to_tc(struct drm_i915_private *i915, enum port port)
> {
> - if (!intel_port_is_tc(dev_priv, port))
> + if (!intel_port_is_tc(i915, port))
> return PORT_TC_NONE;
>
> return port - PORT_C;
> @@ -6636,7 +6636,7 @@ static u64 get_crtc_power_domains(struct drm_crtc *crtc,
> struct intel_crtc_state *crtc_state)
> {
> struct drm_device *dev = crtc->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct drm_encoder *encoder;
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> enum pipe pipe = intel_crtc->pipe;
> @@ -6658,7 +6658,7 @@ static u64 get_crtc_power_domains(struct drm_crtc *crtc,
> mask |= BIT_ULL(intel_encoder->power_domain);
> }
>
> - if (HAS_DDI(dev_priv) && crtc_state->has_audio)
> + if (HAS_DDI(i915) && crtc_state->has_audio)
> mask |= BIT_ULL(POWER_DOMAIN_AUDIO);
>
> if (crtc_state->shared_dpll)
> @@ -6671,7 +6671,7 @@ static u64
> modeset_get_crtc_power_domains(struct drm_crtc *crtc,
> struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc->dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> enum intel_display_power_domain domain;
> u64 domains, new_domains, old_domains;
> @@ -6683,18 +6683,18 @@ modeset_get_crtc_power_domains(struct drm_crtc *crtc,
> domains = new_domains & ~old_domains;
>
> for_each_power_domain(domain, domains)
> - intel_display_power_get(dev_priv, domain);
> + intel_display_power_get(i915, domain);
>
> return old_domains & ~new_domains;
> }
>
> -static void modeset_put_power_domains(struct drm_i915_private *dev_priv,
> +static void modeset_put_power_domains(struct drm_i915_private *i915,
> u64 domains)
> {
> enum intel_display_power_domain domain;
>
> for_each_power_domain(domain, domains)
> - intel_display_power_put_unchecked(dev_priv, domain);
> + intel_display_power_put_unchecked(i915, domain);
> }
>
> static void valleyview_crtc_enable(struct intel_crtc_state *pipe_config,
> @@ -6704,7 +6704,7 @@ static void valleyview_crtc_enable(struct intel_crtc_state *pipe_config,
> to_intel_atomic_state(old_state);
> struct drm_crtc *crtc = pipe_config->base.crtc;
> struct drm_device *dev = crtc->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> int pipe = intel_crtc->pipe;
>
> @@ -6717,7 +6717,7 @@ static void valleyview_crtc_enable(struct intel_crtc_state *pipe_config,
> intel_set_pipe_timings(pipe_config);
> intel_set_pipe_src_size(pipe_config);
>
> - if (IS_CHERRYVIEW(dev_priv) && pipe == PIPE_B) {
> + if (IS_CHERRYVIEW(i915) && pipe == PIPE_B) {
> I915_WRITE(CHV_BLEND(pipe), CHV_BLEND_LEGACY);
> I915_WRITE(CHV_CANVAS(pipe), 0);
> }
> @@ -6726,11 +6726,11 @@ static void valleyview_crtc_enable(struct intel_crtc_state *pipe_config,
>
> intel_crtc->active = true;
>
> - intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, true);
> + intel_set_cpu_fifo_underrun_reporting(i915, pipe, true);
>
> intel_encoders_pre_pll_enable(crtc, pipe_config, old_state);
>
> - if (IS_CHERRYVIEW(dev_priv)) {
> + if (IS_CHERRYVIEW(i915)) {
> chv_prepare_pll(intel_crtc, pipe_config);
> chv_enable_pll(intel_crtc, pipe_config);
> } else {
> @@ -6747,7 +6747,7 @@ static void valleyview_crtc_enable(struct intel_crtc_state *pipe_config,
> /* update DSPCNTR to configure gamma for pipe bottom color */
> intel_disable_primary_plane(pipe_config);
>
> - dev_priv->display.initial_watermarks(old_intel_state,
> + i915->display.initial_watermarks(old_intel_state,
> pipe_config);
> intel_enable_pipe(pipe_config);
>
> @@ -6760,7 +6760,7 @@ static void valleyview_crtc_enable(struct intel_crtc_state *pipe_config,
> static void i9xx_set_pll_dividers(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
>
> I915_WRITE(FP0(crtc->pipe), crtc_state->dpll_hw_state.fp0);
> I915_WRITE(FP1(crtc->pipe), crtc_state->dpll_hw_state.fp1);
> @@ -6773,7 +6773,7 @@ static void i9xx_crtc_enable(struct intel_crtc_state *pipe_config,
> to_intel_atomic_state(old_state);
> struct drm_crtc *crtc = pipe_config->base.crtc;
> struct drm_device *dev = crtc->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> enum pipe pipe = intel_crtc->pipe;
>
> @@ -6792,8 +6792,8 @@ static void i9xx_crtc_enable(struct intel_crtc_state *pipe_config,
>
> intel_crtc->active = true;
>
> - if (!IS_GEN(dev_priv, 2))
> - intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, true);
> + if (!IS_GEN(i915, 2))
> + intel_set_cpu_fifo_underrun_reporting(i915, pipe, true);
>
> intel_encoders_pre_enable(crtc, pipe_config, old_state);
>
> @@ -6806,8 +6806,8 @@ static void i9xx_crtc_enable(struct intel_crtc_state *pipe_config,
> /* update DSPCNTR to configure gamma for pipe bottom color */
> intel_disable_primary_plane(pipe_config);
>
> - if (dev_priv->display.initial_watermarks != NULL)
> - dev_priv->display.initial_watermarks(old_intel_state,
> + if (i915->display.initial_watermarks != NULL)
> + i915->display.initial_watermarks(old_intel_state,
> pipe_config);
> else
> intel_update_watermarks(intel_crtc);
> @@ -6822,12 +6822,12 @@ static void i9xx_crtc_enable(struct intel_crtc_state *pipe_config,
> static void i9xx_pfit_disable(const struct intel_crtc_state *old_crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
>
> if (!old_crtc_state->gmch_pfit.control)
> return;
>
> - assert_pipe_disabled(dev_priv, crtc->pipe);
> + assert_pipe_disabled(i915, crtc->pipe);
>
> DRM_DEBUG_KMS("disabling pfit, current: 0x%08x\n",
> I915_READ(PFIT_CONTROL));
> @@ -6839,7 +6839,7 @@ static void i9xx_crtc_disable(struct intel_crtc_state *old_crtc_state,
> {
> struct drm_crtc *crtc = old_crtc_state->base.crtc;
> struct drm_device *dev = crtc->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> int pipe = intel_crtc->pipe;
>
> @@ -6847,8 +6847,8 @@ static void i9xx_crtc_disable(struct intel_crtc_state *old_crtc_state,
> * On gen2 planes are double buffered but the pipe isn't, so we must
> * wait for planes to fully turn off before disabling the pipe.
> */
> - if (IS_GEN(dev_priv, 2))
> - intel_wait_for_vblank(dev_priv, pipe);
> + if (IS_GEN(i915, 2))
> + intel_wait_for_vblank(i915, pipe);
>
> intel_encoders_disable(crtc, old_crtc_state, old_state);
>
> @@ -6862,25 +6862,25 @@ static void i9xx_crtc_disable(struct intel_crtc_state *old_crtc_state,
> intel_encoders_post_disable(crtc, old_crtc_state, old_state);
>
> if (!intel_crtc_has_type(old_crtc_state, INTEL_OUTPUT_DSI)) {
> - if (IS_CHERRYVIEW(dev_priv))
> - chv_disable_pll(dev_priv, pipe);
> - else if (IS_VALLEYVIEW(dev_priv))
> - vlv_disable_pll(dev_priv, pipe);
> + if (IS_CHERRYVIEW(i915))
> + chv_disable_pll(i915, pipe);
> + else if (IS_VALLEYVIEW(i915))
> + vlv_disable_pll(i915, pipe);
> else
> i9xx_disable_pll(old_crtc_state);
> }
>
> intel_encoders_post_pll_disable(crtc, old_crtc_state, old_state);
>
> - if (!IS_GEN(dev_priv, 2))
> - intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, false);
> + if (!IS_GEN(i915, 2))
> + intel_set_cpu_fifo_underrun_reporting(i915, pipe, false);
>
> - if (!dev_priv->display.initial_watermarks)
> + if (!i915->display.initial_watermarks)
> intel_update_watermarks(intel_crtc);
>
> /* clock the pipe down to 640x480 at 60 to potentially save power */
> - if (IS_I830(dev_priv))
> - i830_enable_pipe(dev_priv, pipe);
> + if (IS_I830(i915))
> + i830_enable_pipe(i915, pipe);
> }
>
> static void intel_crtc_disable_noatomic(struct drm_crtc *crtc,
> @@ -6888,9 +6888,9 @@ static void intel_crtc_disable_noatomic(struct drm_crtc *crtc,
> {
> struct intel_encoder *encoder;
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc->dev);
> struct intel_bw_state *bw_state =
> - to_intel_bw_state(dev_priv->bw_obj.state);
> + to_intel_bw_state(i915->bw_obj.state);
> enum intel_display_power_domain domain;
> struct intel_plane *plane;
> u64 domains;
> @@ -6901,7 +6901,7 @@ static void intel_crtc_disable_noatomic(struct drm_crtc *crtc,
> if (!intel_crtc->active)
> return;
>
> - for_each_intel_plane_on_crtc(&dev_priv->drm, intel_crtc, plane) {
> + for_each_intel_plane_on_crtc(&i915->drm, intel_crtc, plane) {
> const struct intel_plane_state *plane_state =
> to_intel_plane_state(plane->base.state);
>
> @@ -6924,7 +6924,7 @@ static void intel_crtc_disable_noatomic(struct drm_crtc *crtc,
>
> WARN_ON(IS_ERR(crtc_state) || ret);
>
> - dev_priv->display.crtc_disable(crtc_state, state);
> + i915->display.crtc_disable(crtc_state, state);
>
> drm_atomic_state_put(state);
>
> @@ -6947,12 +6947,12 @@ static void intel_crtc_disable_noatomic(struct drm_crtc *crtc,
>
> domains = intel_crtc->enabled_power_domains;
> for_each_power_domain(domain, domains)
> - intel_display_power_put_unchecked(dev_priv, domain);
> + intel_display_power_put_unchecked(i915, domain);
> intel_crtc->enabled_power_domains = 0;
>
> - dev_priv->active_crtcs &= ~(1 << intel_crtc->pipe);
> - dev_priv->min_cdclk[intel_crtc->pipe] = 0;
> - dev_priv->min_voltage_level[intel_crtc->pipe] = 0;
> + i915->active_crtcs &= ~(1 << intel_crtc->pipe);
> + i915->min_cdclk[intel_crtc->pipe] = 0;
> + i915->min_voltage_level[intel_crtc->pipe] = 0;
>
> bw_state->data_rate[intel_crtc->pipe] = 0;
> bw_state->num_active_planes[intel_crtc->pipe] = 0;
> @@ -6964,7 +6964,7 @@ static void intel_crtc_disable_noatomic(struct drm_crtc *crtc,
> */
> int intel_display_suspend(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct drm_atomic_state *state;
> int ret;
>
> @@ -6973,7 +6973,7 @@ int intel_display_suspend(struct drm_device *dev)
> if (ret)
> DRM_ERROR("Suspending crtc's failed with %i\n", ret);
> else
> - dev_priv->modeset_restore_state = state;
> + i915->modeset_restore_state = state;
> return ret;
> }
>
> @@ -7035,7 +7035,7 @@ static int pipe_required_fdi_lanes(struct intel_crtc_state *crtc_state)
> static int ironlake_check_fdi_lanes(struct drm_device *dev, enum pipe pipe,
> struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct drm_atomic_state *state = pipe_config->base.state;
> struct intel_crtc *other_crtc;
> struct intel_crtc_state *other_crtc_state;
> @@ -7048,7 +7048,7 @@ static int ironlake_check_fdi_lanes(struct drm_device *dev, enum pipe pipe,
> return -EINVAL;
> }
>
> - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) {
> + if (IS_HASWELL(i915) || IS_BROADWELL(i915)) {
> if (pipe_config->fdi_lanes > 2) {
> DRM_DEBUG_KMS("only 2 lanes on haswell, required: %i lanes\n",
> pipe_config->fdi_lanes);
> @@ -7058,7 +7058,7 @@ static int ironlake_check_fdi_lanes(struct drm_device *dev, enum pipe pipe,
> }
> }
>
> - if (INTEL_INFO(dev_priv)->num_pipes == 2)
> + if (INTEL_INFO(i915)->num_pipes == 2)
> return 0;
>
> /* Ivybridge 3 pipe is really complicated */
> @@ -7069,7 +7069,7 @@ static int ironlake_check_fdi_lanes(struct drm_device *dev, enum pipe pipe,
> if (pipe_config->fdi_lanes <= 2)
> return 0;
>
> - other_crtc = intel_get_crtc_for_pipe(dev_priv, PIPE_C);
> + other_crtc = intel_get_crtc_for_pipe(i915, PIPE_C);
> other_crtc_state =
> intel_atomic_get_crtc_state(state, other_crtc);
> if (IS_ERR(other_crtc_state))
> @@ -7088,7 +7088,7 @@ static int ironlake_check_fdi_lanes(struct drm_device *dev, enum pipe pipe,
> return -EINVAL;
> }
>
> - other_crtc = intel_get_crtc_for_pipe(dev_priv, PIPE_B);
> + other_crtc = intel_get_crtc_for_pipe(i915, PIPE_B);
> other_crtc_state =
> intel_atomic_get_crtc_state(state, other_crtc);
> if (IS_ERR(other_crtc_state))
> @@ -7156,7 +7156,7 @@ static int ironlake_fdi_compute_config(struct intel_crtc *intel_crtc,
> bool hsw_crtc_state_ips_capable(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
>
> /* IPS only exists on ULT machines and is tied to pipe A. */
> if (!hsw_crtc_supports_ips(crtc))
> @@ -7175,8 +7175,8 @@ bool hsw_crtc_state_ips_capable(const struct intel_crtc_state *crtc_state)
> *
> * Should measure whether using a lower cdclk w/o IPS
> */
> - if (IS_BROADWELL(dev_priv) &&
> - crtc_state->pixel_rate > dev_priv->max_cdclk_freq * 95 / 100)
> + if (IS_BROADWELL(i915) &&
> + crtc_state->pixel_rate > i915->max_cdclk_freq * 95 / 100)
> return false;
>
> return true;
> @@ -7184,7 +7184,7 @@ bool hsw_crtc_state_ips_capable(const struct intel_crtc_state *crtc_state)
>
> static bool hsw_compute_ips_config(struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> to_i915(crtc_state->base.crtc->dev);
> struct intel_atomic_state *intel_state =
> to_intel_atomic_state(crtc_state->base.state);
> @@ -7206,7 +7206,7 @@ static bool hsw_compute_ips_config(struct intel_crtc_state *crtc_state)
> return false;
>
> /* pixel rate mustn't exceed 95% of cdclk with IPS on BDW */
> - if (IS_BROADWELL(dev_priv) &&
> + if (IS_BROADWELL(i915) &&
> crtc_state->pixel_rate > intel_state->cdclk.logical.cdclk * 95 / 100)
> return false;
>
> @@ -7215,11 +7215,11 @@ static bool hsw_compute_ips_config(struct intel_crtc_state *crtc_state)
>
> static bool intel_crtc_supports_double_wide(const struct intel_crtc *crtc)
> {
> - const struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + const struct drm_i915_private *i915 = to_i915(crtc->base.dev);
>
> /* GDG double wide on either pipe, otherwise pipe A only */
> - return INTEL_GEN(dev_priv) < 4 &&
> - (crtc->pipe == PIPE_A || IS_I915G(dev_priv));
> + return INTEL_GEN(i915) < 4 &&
> + (crtc->pipe == PIPE_A || IS_I915G(i915));
> }
>
> static u32 ilk_pipe_pixel_rate(const struct intel_crtc_state *pipe_config)
> @@ -7259,9 +7259,9 @@ static u32 ilk_pipe_pixel_rate(const struct intel_crtc_state *pipe_config)
>
> static void intel_crtc_compute_pixel_rate(struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
>
> - if (HAS_GMCH(dev_priv))
> + if (HAS_GMCH(i915))
> /* FIXME calculate proper pipe pixel rate for GMCH pfit */
> crtc_state->pixel_rate =
> crtc_state->base.adjusted_mode.crtc_clock;
> @@ -7273,12 +7273,12 @@ static void intel_crtc_compute_pixel_rate(struct intel_crtc_state *crtc_state)
> static int intel_crtc_compute_config(struct intel_crtc *crtc,
> struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> const struct drm_display_mode *adjusted_mode = &pipe_config->base.adjusted_mode;
> - int clock_limit = dev_priv->max_dotclk_freq;
> + int clock_limit = i915->max_dotclk_freq;
>
> - if (INTEL_GEN(dev_priv) < 4) {
> - clock_limit = dev_priv->max_cdclk_freq * 9 / 10;
> + if (INTEL_GEN(i915) < 4) {
> + clock_limit = i915->max_cdclk_freq * 9 / 10;
>
> /*
> * Enable double wide mode when the dot clock
> @@ -7286,7 +7286,7 @@ static int intel_crtc_compute_config(struct intel_crtc *crtc,
> */
> if (intel_crtc_supports_double_wide(crtc) &&
> adjusted_mode->crtc_clock > clock_limit) {
> - clock_limit = dev_priv->max_dotclk_freq;
> + clock_limit = i915->max_dotclk_freq;
> pipe_config->double_wide = true;
> }
> }
> @@ -7323,7 +7323,7 @@ static int intel_crtc_compute_config(struct intel_crtc *crtc,
> }
>
> if (intel_crtc_has_type(pipe_config, INTEL_OUTPUT_LVDS) &&
> - intel_is_dual_link_lvds(dev_priv)) {
> + intel_is_dual_link_lvds(i915)) {
> DRM_DEBUG_KMS("Odd pipe source width not supported with dual link LVDS\n");
> return -EINVAL;
> }
> @@ -7332,7 +7332,7 @@ static int intel_crtc_compute_config(struct intel_crtc *crtc,
> /* Cantiga+ cannot handle modes with a hsync front porch of 0.
> * WaPruneModeWithIncorrectHsyncOffset:ctg,elk,ilk,snb,ivb,vlv,hsw.
> */
> - if ((INTEL_GEN(dev_priv) > 4 || IS_G4X(dev_priv)) &&
> + if ((INTEL_GEN(i915) > 4 || IS_G4X(i915)) &&
> adjusted_mode->crtc_hsync_start == adjusted_mode->crtc_hdisplay)
> return -EINVAL;
>
> @@ -7392,12 +7392,12 @@ intel_link_compute_m_n(u16 bits_per_pixel, int nlanes,
> constant_n);
> }
>
> -static inline bool intel_panel_use_ssc(struct drm_i915_private *dev_priv)
> +static inline bool intel_panel_use_ssc(struct drm_i915_private *i915)
> {
> if (i915_modparams.panel_use_ssc >= 0)
> return i915_modparams.panel_use_ssc != 0;
> - return dev_priv->vbt.lvds_use_ssc
> - && !(dev_priv->quirks & QUIRK_LVDS_SSC_DISABLE);
> + return i915->vbt.lvds_use_ssc
> + && !(i915->quirks & QUIRK_LVDS_SSC_DISABLE);
> }
>
> static u32 pnv_dpll_compute_fp(struct dpll *dpll)
> @@ -7414,10 +7414,10 @@ static void i9xx_update_pll_dividers(struct intel_crtc *crtc,
> struct intel_crtc_state *crtc_state,
> struct dpll *reduced_clock)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> u32 fp, fp2 = 0;
>
> - if (IS_PINEVIEW(dev_priv)) {
> + if (IS_PINEVIEW(i915)) {
> fp = pnv_dpll_compute_fp(&crtc_state->dpll);
> if (reduced_clock)
> fp2 = pnv_dpll_compute_fp(reduced_clock);
> @@ -7437,7 +7437,7 @@ static void i9xx_update_pll_dividers(struct intel_crtc *crtc,
> }
> }
>
> -static void vlv_pllb_recal_opamp(struct drm_i915_private *dev_priv, enum pipe
> +static void vlv_pllb_recal_opamp(struct drm_i915_private *i915, enum pipe
> pipe)
> {
> u32 reg_val;
> @@ -7446,31 +7446,31 @@ static void vlv_pllb_recal_opamp(struct drm_i915_private *dev_priv, enum pipe
> * PLLB opamp always calibrates to max value of 0x3f, force enable it
> * and set it to a reasonable value instead.
> */
> - reg_val = vlv_dpio_read(dev_priv, pipe, VLV_PLL_DW9(1));
> + reg_val = vlv_dpio_read(i915, pipe, VLV_PLL_DW9(1));
> reg_val &= 0xffffff00;
> reg_val |= 0x00000030;
> - vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW9(1), reg_val);
> + vlv_dpio_write(i915, pipe, VLV_PLL_DW9(1), reg_val);
>
> - reg_val = vlv_dpio_read(dev_priv, pipe, VLV_REF_DW13);
> + reg_val = vlv_dpio_read(i915, pipe, VLV_REF_DW13);
> reg_val &= 0x00ffffff;
> reg_val |= 0x8c000000;
> - vlv_dpio_write(dev_priv, pipe, VLV_REF_DW13, reg_val);
> + vlv_dpio_write(i915, pipe, VLV_REF_DW13, reg_val);
>
> - reg_val = vlv_dpio_read(dev_priv, pipe, VLV_PLL_DW9(1));
> + reg_val = vlv_dpio_read(i915, pipe, VLV_PLL_DW9(1));
> reg_val &= 0xffffff00;
> - vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW9(1), reg_val);
> + vlv_dpio_write(i915, pipe, VLV_PLL_DW9(1), reg_val);
>
> - reg_val = vlv_dpio_read(dev_priv, pipe, VLV_REF_DW13);
> + reg_val = vlv_dpio_read(i915, pipe, VLV_REF_DW13);
> reg_val &= 0x00ffffff;
> reg_val |= 0xb0000000;
> - vlv_dpio_write(dev_priv, pipe, VLV_REF_DW13, reg_val);
> + vlv_dpio_write(i915, pipe, VLV_REF_DW13, reg_val);
> }
>
> static void intel_pch_transcoder_set_m_n(const struct intel_crtc_state *crtc_state,
> const struct intel_link_m_n *m_n)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
>
> I915_WRITE(PCH_TRANS_DATA_M1(pipe), TU_SIZE(m_n->tu) | m_n->gmch_m);
> @@ -7479,17 +7479,17 @@ static void intel_pch_transcoder_set_m_n(const struct intel_crtc_state *crtc_sta
> I915_WRITE(PCH_TRANS_LINK_N1(pipe), m_n->link_n);
> }
>
> -static bool transcoder_has_m2_n2(struct drm_i915_private *dev_priv,
> +static bool transcoder_has_m2_n2(struct drm_i915_private *i915,
> enum transcoder transcoder)
> {
> - if (IS_HASWELL(dev_priv))
> + if (IS_HASWELL(i915))
> return transcoder == TRANSCODER_EDP;
>
> /*
> * Strictly speaking some registers are available before
> * gen7, but we only support DRRS on gen7+
> */
> - return IS_GEN(dev_priv, 7) || IS_CHERRYVIEW(dev_priv);
> + return IS_GEN(i915, 7) || IS_CHERRYVIEW(i915);
> }
>
> static void intel_cpu_transcoder_set_m_n(const struct intel_crtc_state *crtc_state,
> @@ -7497,11 +7497,11 @@ static void intel_cpu_transcoder_set_m_n(const struct intel_crtc_state *crtc_sta
> const struct intel_link_m_n *m2_n2)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
> enum transcoder transcoder = crtc_state->cpu_transcoder;
>
> - if (INTEL_GEN(dev_priv) >= 5) {
> + if (INTEL_GEN(i915) >= 5) {
> I915_WRITE(PIPE_DATA_M1(transcoder), TU_SIZE(m_n->tu) | m_n->gmch_m);
> I915_WRITE(PIPE_DATA_N1(transcoder), m_n->gmch_n);
> I915_WRITE(PIPE_LINK_M1(transcoder), m_n->link_m);
> @@ -7511,7 +7511,7 @@ static void intel_cpu_transcoder_set_m_n(const struct intel_crtc_state *crtc_sta
> * (to make sure the registers are not unnecessarily accessed).
> */
> if (m2_n2 && crtc_state->has_drrs &&
> - transcoder_has_m2_n2(dev_priv, transcoder)) {
> + transcoder_has_m2_n2(i915, transcoder)) {
> I915_WRITE(PIPE_DATA_M2(transcoder),
> TU_SIZE(m2_n2->tu) | m2_n2->gmch_m);
> I915_WRITE(PIPE_DATA_N2(transcoder), m2_n2->gmch_n);
> @@ -7588,7 +7588,7 @@ static void vlv_prepare_pll(struct intel_crtc *crtc,
> const struct intel_crtc_state *pipe_config)
> {
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> enum pipe pipe = crtc->pipe;
> u32 mdiv;
> u32 bestn, bestm1, bestm2, bestp1, bestp2;
> @@ -7603,7 +7603,7 @@ static void vlv_prepare_pll(struct intel_crtc *crtc,
> if ((pipe_config->dpll_hw_state.dpll & DPLL_VCO_ENABLE) == 0)
> return;
>
> - vlv_dpio_get(dev_priv);
> + vlv_dpio_get(i915);
>
> bestn = pipe_config->dpll.n;
> bestm1 = pipe_config->dpll.m1;
> @@ -7615,18 +7615,18 @@ static void vlv_prepare_pll(struct intel_crtc *crtc,
>
> /* PLL B needs special handling */
> if (pipe == PIPE_B)
> - vlv_pllb_recal_opamp(dev_priv, pipe);
> + vlv_pllb_recal_opamp(i915, pipe);
>
> /* Set up Tx target for periodic Rcomp update */
> - vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW9_BCAST, 0x0100000f);
> + vlv_dpio_write(i915, pipe, VLV_PLL_DW9_BCAST, 0x0100000f);
>
> /* Disable target IRef on PLL */
> - reg_val = vlv_dpio_read(dev_priv, pipe, VLV_PLL_DW8(pipe));
> + reg_val = vlv_dpio_read(i915, pipe, VLV_PLL_DW8(pipe));
> reg_val &= 0x00ffffff;
> - vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW8(pipe), reg_val);
> + vlv_dpio_write(i915, pipe, VLV_PLL_DW8(pipe), reg_val);
>
> /* Disable fast lock */
> - vlv_dpio_write(dev_priv, pipe, VLV_CMN_DW0, 0x610);
> + vlv_dpio_write(i915, pipe, VLV_CMN_DW0, 0x610);
>
> /* Set idtafcrecal before PLL is enabled */
> mdiv = ((bestm1 << DPIO_M1DIV_SHIFT) | (bestm2 & DPIO_M2DIV_MASK));
> @@ -7640,55 +7640,55 @@ static void vlv_prepare_pll(struct intel_crtc *crtc,
> * Note: don't use the DAC post divider as it seems unstable.
> */
> mdiv |= (DPIO_POST_DIV_HDMIDP << DPIO_POST_DIV_SHIFT);
> - vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW3(pipe), mdiv);
> + vlv_dpio_write(i915, pipe, VLV_PLL_DW3(pipe), mdiv);
>
> mdiv |= DPIO_ENABLE_CALIBRATION;
> - vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW3(pipe), mdiv);
> + vlv_dpio_write(i915, pipe, VLV_PLL_DW3(pipe), mdiv);
>
> /* Set HBR and RBR LPF coefficients */
> if (pipe_config->port_clock == 162000 ||
> intel_crtc_has_type(pipe_config, INTEL_OUTPUT_ANALOG) ||
> intel_crtc_has_type(pipe_config, INTEL_OUTPUT_HDMI))
> - vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW10(pipe),
> + vlv_dpio_write(i915, pipe, VLV_PLL_DW10(pipe),
> 0x009f0003);
> else
> - vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW10(pipe),
> + vlv_dpio_write(i915, pipe, VLV_PLL_DW10(pipe),
> 0x00d0000f);
>
> if (intel_crtc_has_dp_encoder(pipe_config)) {
> /* Use SSC source */
> if (pipe == PIPE_A)
> - vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW5(pipe),
> + vlv_dpio_write(i915, pipe, VLV_PLL_DW5(pipe),
> 0x0df40000);
> else
> - vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW5(pipe),
> + vlv_dpio_write(i915, pipe, VLV_PLL_DW5(pipe),
> 0x0df70000);
> } else { /* HDMI or VGA */
> /* Use bend source */
> if (pipe == PIPE_A)
> - vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW5(pipe),
> + vlv_dpio_write(i915, pipe, VLV_PLL_DW5(pipe),
> 0x0df70000);
> else
> - vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW5(pipe),
> + vlv_dpio_write(i915, pipe, VLV_PLL_DW5(pipe),
> 0x0df40000);
> }
>
> - coreclk = vlv_dpio_read(dev_priv, pipe, VLV_PLL_DW7(pipe));
> + coreclk = vlv_dpio_read(i915, pipe, VLV_PLL_DW7(pipe));
> coreclk = (coreclk & 0x0000ff00) | 0x01c00000;
> if (intel_crtc_has_dp_encoder(pipe_config))
> coreclk |= 0x01000000;
> - vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW7(pipe), coreclk);
> + vlv_dpio_write(i915, pipe, VLV_PLL_DW7(pipe), coreclk);
>
> - vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW11(pipe), 0x87871000);
> + vlv_dpio_write(i915, pipe, VLV_PLL_DW11(pipe), 0x87871000);
>
> - vlv_dpio_put(dev_priv);
> + vlv_dpio_put(i915);
> }
>
> static void chv_prepare_pll(struct intel_crtc *crtc,
> const struct intel_crtc_state *pipe_config)
> {
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> enum pipe pipe = crtc->pipe;
> enum dpio_channel port = vlv_pipe_to_channel(pipe);
> u32 loopfilter, tribuf_calcntr;
> @@ -7714,42 +7714,42 @@ static void chv_prepare_pll(struct intel_crtc *crtc,
> dpio_val = 0;
> loopfilter = 0;
>
> - vlv_dpio_get(dev_priv);
> + vlv_dpio_get(i915);
>
> /* p1 and p2 divider */
> - vlv_dpio_write(dev_priv, pipe, CHV_CMN_DW13(port),
> + vlv_dpio_write(i915, pipe, CHV_CMN_DW13(port),
> 5 << DPIO_CHV_S1_DIV_SHIFT |
> bestp1 << DPIO_CHV_P1_DIV_SHIFT |
> bestp2 << DPIO_CHV_P2_DIV_SHIFT |
> 1 << DPIO_CHV_K_DIV_SHIFT);
>
> /* Feedback post-divider - m2 */
> - vlv_dpio_write(dev_priv, pipe, CHV_PLL_DW0(port), bestm2);
> + vlv_dpio_write(i915, pipe, CHV_PLL_DW0(port), bestm2);
>
> /* Feedback refclk divider - n and m1 */
> - vlv_dpio_write(dev_priv, pipe, CHV_PLL_DW1(port),
> + vlv_dpio_write(i915, pipe, CHV_PLL_DW1(port),
> DPIO_CHV_M1_DIV_BY_2 |
> 1 << DPIO_CHV_N_DIV_SHIFT);
>
> /* M2 fraction division */
> - vlv_dpio_write(dev_priv, pipe, CHV_PLL_DW2(port), bestm2_frac);
> + vlv_dpio_write(i915, pipe, CHV_PLL_DW2(port), bestm2_frac);
>
> /* M2 fraction division enable */
> - dpio_val = vlv_dpio_read(dev_priv, pipe, CHV_PLL_DW3(port));
> + dpio_val = vlv_dpio_read(i915, pipe, CHV_PLL_DW3(port));
> dpio_val &= ~(DPIO_CHV_FEEDFWD_GAIN_MASK | DPIO_CHV_FRAC_DIV_EN);
> dpio_val |= (2 << DPIO_CHV_FEEDFWD_GAIN_SHIFT);
> if (bestm2_frac)
> dpio_val |= DPIO_CHV_FRAC_DIV_EN;
> - vlv_dpio_write(dev_priv, pipe, CHV_PLL_DW3(port), dpio_val);
> + vlv_dpio_write(i915, pipe, CHV_PLL_DW3(port), dpio_val);
>
> /* Program digital lock detect threshold */
> - dpio_val = vlv_dpio_read(dev_priv, pipe, CHV_PLL_DW9(port));
> + dpio_val = vlv_dpio_read(i915, pipe, CHV_PLL_DW9(port));
> dpio_val &= ~(DPIO_CHV_INT_LOCK_THRESHOLD_MASK |
> DPIO_CHV_INT_LOCK_THRESHOLD_SEL_COARSE);
> dpio_val |= (0x5 << DPIO_CHV_INT_LOCK_THRESHOLD_SHIFT);
> if (!bestm2_frac)
> dpio_val |= DPIO_CHV_INT_LOCK_THRESHOLD_SEL_COARSE;
> - vlv_dpio_write(dev_priv, pipe, CHV_PLL_DW9(port), dpio_val);
> + vlv_dpio_write(i915, pipe, CHV_PLL_DW9(port), dpio_val);
>
> /* Loop filter */
> if (vco == 5400000) {
> @@ -7774,24 +7774,24 @@ static void chv_prepare_pll(struct intel_crtc *crtc,
> loopfilter |= (0x3 << DPIO_CHV_GAIN_CTRL_SHIFT);
> tribuf_calcntr = 0;
> }
> - vlv_dpio_write(dev_priv, pipe, CHV_PLL_DW6(port), loopfilter);
> + vlv_dpio_write(i915, pipe, CHV_PLL_DW6(port), loopfilter);
>
> - dpio_val = vlv_dpio_read(dev_priv, pipe, CHV_PLL_DW8(port));
> + dpio_val = vlv_dpio_read(i915, pipe, CHV_PLL_DW8(port));
> dpio_val &= ~DPIO_CHV_TDC_TARGET_CNT_MASK;
> dpio_val |= (tribuf_calcntr << DPIO_CHV_TDC_TARGET_CNT_SHIFT);
> - vlv_dpio_write(dev_priv, pipe, CHV_PLL_DW8(port), dpio_val);
> + vlv_dpio_write(i915, pipe, CHV_PLL_DW8(port), dpio_val);
>
> /* AFC Recal */
> - vlv_dpio_write(dev_priv, pipe, CHV_CMN_DW14(port),
> - vlv_dpio_read(dev_priv, pipe, CHV_CMN_DW14(port)) |
> + vlv_dpio_write(i915, pipe, CHV_CMN_DW14(port),
> + vlv_dpio_read(i915, pipe, CHV_CMN_DW14(port)) |
> DPIO_AFC_RECAL);
>
> - vlv_dpio_put(dev_priv);
> + vlv_dpio_put(i915);
> }
>
> /**
> * vlv_force_pll_on - forcibly enable just the PLL
> - * @dev_priv: i915 private structure
> + * @i915: i915 private structure
> * @pipe: pipe PLL to enable
> * @dpll: PLL configuration
> *
> @@ -7799,10 +7799,10 @@ static void chv_prepare_pll(struct intel_crtc *crtc,
> * in cases where we need the PLL enabled even when @pipe is not going to
> * be enabled.
> */
> -int vlv_force_pll_on(struct drm_i915_private *dev_priv, enum pipe pipe,
> +int vlv_force_pll_on(struct drm_i915_private *i915, enum pipe pipe,
> const struct dpll *dpll)
> {
> - struct intel_crtc *crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> + struct intel_crtc *crtc = intel_get_crtc_for_pipe(i915, pipe);
> struct intel_crtc_state *pipe_config;
>
> pipe_config = kzalloc(sizeof(*pipe_config), GFP_KERNEL);
> @@ -7813,7 +7813,7 @@ int vlv_force_pll_on(struct drm_i915_private *dev_priv, enum pipe pipe,
> pipe_config->pixel_multiplier = 1;
> pipe_config->dpll = *dpll;
>
> - if (IS_CHERRYVIEW(dev_priv)) {
> + if (IS_CHERRYVIEW(i915)) {
> chv_compute_dpll(crtc, pipe_config);
> chv_prepare_pll(crtc, pipe_config);
> chv_enable_pll(crtc, pipe_config);
> @@ -7830,25 +7830,25 @@ int vlv_force_pll_on(struct drm_i915_private *dev_priv, enum pipe pipe,
>
> /**
> * vlv_force_pll_off - forcibly disable just the PLL
> - * @dev_priv: i915 private structure
> + * @i915: i915 private structure
> * @pipe: pipe PLL to disable
> *
> * Disable the PLL for @pipe. To be used in cases where we need
> * the PLL enabled even when @pipe is not going to be enabled.
> */
> -void vlv_force_pll_off(struct drm_i915_private *dev_priv, enum pipe pipe)
> +void vlv_force_pll_off(struct drm_i915_private *i915, enum pipe pipe)
> {
> - if (IS_CHERRYVIEW(dev_priv))
> - chv_disable_pll(dev_priv, pipe);
> + if (IS_CHERRYVIEW(i915))
> + chv_disable_pll(i915, pipe);
> else
> - vlv_disable_pll(dev_priv, pipe);
> + vlv_disable_pll(i915, pipe);
> }
>
> static void i9xx_compute_dpll(struct intel_crtc *crtc,
> struct intel_crtc_state *crtc_state,
> struct dpll *reduced_clock)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> u32 dpll;
> struct dpll *clock = &crtc_state->dpll;
>
> @@ -7861,8 +7861,8 @@ static void i9xx_compute_dpll(struct intel_crtc *crtc,
> else
> dpll |= DPLLB_MODE_DAC_SERIAL;
>
> - if (IS_I945G(dev_priv) || IS_I945GM(dev_priv) ||
> - IS_G33(dev_priv) || IS_PINEVIEW(dev_priv)) {
> + if (IS_I945G(i915) || IS_I945GM(i915) ||
> + IS_G33(i915) || IS_PINEVIEW(i915)) {
> dpll |= (crtc_state->pixel_multiplier - 1)
> << SDVO_MULTIPLIER_SHIFT_HIRES;
> }
> @@ -7875,11 +7875,11 @@ static void i9xx_compute_dpll(struct intel_crtc *crtc,
> dpll |= DPLL_SDVO_HIGH_SPEED;
>
> /* compute bitmask from p1 value */
> - if (IS_PINEVIEW(dev_priv))
> + if (IS_PINEVIEW(i915))
> dpll |= (1 << (clock->p1 - 1)) << DPLL_FPA01_P1_POST_DIV_SHIFT_PINEVIEW;
> else {
> dpll |= (1 << (clock->p1 - 1)) << DPLL_FPA01_P1_POST_DIV_SHIFT;
> - if (IS_G4X(dev_priv) && reduced_clock)
> + if (IS_G4X(i915) && reduced_clock)
> dpll |= (1 << (reduced_clock->p1 - 1)) << DPLL_FPA1_P1_POST_DIV_SHIFT;
> }
> switch (clock->p2) {
> @@ -7896,13 +7896,13 @@ static void i9xx_compute_dpll(struct intel_crtc *crtc,
> dpll |= DPLLB_LVDS_P2_CLOCK_DIV_14;
> break;
> }
> - if (INTEL_GEN(dev_priv) >= 4)
> + if (INTEL_GEN(i915) >= 4)
> dpll |= (6 << PLL_LOAD_PULSE_PHASE_SHIFT);
>
> if (crtc_state->sdvo_tv_clock)
> dpll |= PLL_REF_INPUT_TVCLKINBC;
> else if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS) &&
> - intel_panel_use_ssc(dev_priv))
> + intel_panel_use_ssc(i915))
> dpll |= PLLB_REF_INPUT_SPREADSPECTRUMIN;
> else
> dpll |= PLL_REF_INPUT_DREFCLK;
> @@ -7910,7 +7910,7 @@ static void i9xx_compute_dpll(struct intel_crtc *crtc,
> dpll |= DPLL_VCO_ENABLE;
> crtc_state->dpll_hw_state.dpll = dpll;
>
> - if (INTEL_GEN(dev_priv) >= 4) {
> + if (INTEL_GEN(i915) >= 4) {
> u32 dpll_md = (crtc_state->pixel_multiplier - 1)
> << DPLL_MD_UDI_MULTIPLIER_SHIFT;
> crtc_state->dpll_hw_state.dpll_md = dpll_md;
> @@ -7922,7 +7922,7 @@ static void i8xx_compute_dpll(struct intel_crtc *crtc,
> struct dpll *reduced_clock)
> {
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> u32 dpll;
> struct dpll *clock = &crtc_state->dpll;
>
> @@ -7953,12 +7953,12 @@ static void i8xx_compute_dpll(struct intel_crtc *crtc,
> * both DPLLS. The spec says we should disable the DVO 2X clock
> * when not needed, but this seems to work fine in practice.
> */
> - if (IS_I830(dev_priv) ||
> + if (IS_I830(i915) ||
> intel_crtc_has_type(crtc_state, INTEL_OUTPUT_DVO))
> dpll |= DPLL_DVO_2X_MODE;
>
> if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS) &&
> - intel_panel_use_ssc(dev_priv))
> + intel_panel_use_ssc(i915))
> dpll |= PLLB_REF_INPUT_SPREADSPECTRUMIN;
> else
> dpll |= PLL_REF_INPUT_DREFCLK;
> @@ -7970,7 +7970,7 @@ static void i8xx_compute_dpll(struct intel_crtc *crtc,
> static void intel_set_pipe_timings(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
> enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> const struct drm_display_mode *adjusted_mode = &crtc_state->base.adjusted_mode;
> @@ -7996,7 +7996,7 @@ static void intel_set_pipe_timings(const struct intel_crtc_state *crtc_state)
> vsyncshift += adjusted_mode->crtc_htotal;
> }
>
> - if (INTEL_GEN(dev_priv) > 3)
> + if (INTEL_GEN(i915) > 3)
> I915_WRITE(VSYNCSHIFT(cpu_transcoder), vsyncshift);
>
> I915_WRITE(HTOTAL(cpu_transcoder),
> @@ -8023,7 +8023,7 @@ static void intel_set_pipe_timings(const struct intel_crtc_state *crtc_state)
> * programmed with the VTOTAL_EDP value. Same for VTOTAL_C. This is
> * documented on the DDI_FUNC_CTL register description, EDP Input Select
> * bits. */
> - if (IS_HASWELL(dev_priv) && cpu_transcoder == TRANSCODER_EDP &&
> + if (IS_HASWELL(i915) && cpu_transcoder == TRANSCODER_EDP &&
> (pipe == PIPE_B || pipe == PIPE_C))
> I915_WRITE(VTOTAL(pipe), I915_READ(VTOTAL(cpu_transcoder)));
>
> @@ -8032,7 +8032,7 @@ static void intel_set_pipe_timings(const struct intel_crtc_state *crtc_state)
> static void intel_set_pipe_src_size(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
>
> /* pipesrc controls the size that is scaled from, which should
> @@ -8047,7 +8047,7 @@ static void intel_get_pipe_timings(struct intel_crtc *crtc,
> struct intel_crtc_state *pipe_config)
> {
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> enum transcoder cpu_transcoder = pipe_config->cpu_transcoder;
> u32 tmp;
>
> @@ -8092,7 +8092,7 @@ static void intel_get_pipe_src_size(struct intel_crtc *crtc,
> struct intel_crtc_state *pipe_config)
> {
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> u32 tmp;
>
> tmp = I915_READ(PIPESRC(crtc->pipe));
> @@ -8129,21 +8129,21 @@ void intel_mode_from_pipe_config(struct drm_display_mode *mode,
> static void i9xx_set_pipeconf(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> u32 pipeconf;
>
> pipeconf = 0;
>
> /* we keep both pipes enabled on 830 */
> - if (IS_I830(dev_priv))
> + if (IS_I830(i915))
> pipeconf |= I915_READ(PIPECONF(crtc->pipe)) & PIPECONF_ENABLE;
>
> if (crtc_state->double_wide)
> pipeconf |= PIPECONF_DOUBLE_WIDE;
>
> /* only g4x and later have fancy bpc/dither controls */
> - if (IS_G4X(dev_priv) || IS_VALLEYVIEW(dev_priv) ||
> - IS_CHERRYVIEW(dev_priv)) {
> + if (IS_G4X(i915) || IS_VALLEYVIEW(i915) ||
> + IS_CHERRYVIEW(i915)) {
> /* Bspec claims that we can't use dithering for 30bpp pipes. */
> if (crtc_state->dither && crtc_state->pipe_bpp != 30)
> pipeconf |= PIPECONF_DITHER_EN |
> @@ -8166,7 +8166,7 @@ static void i9xx_set_pipeconf(const struct intel_crtc_state *crtc_state)
> }
>
> if (crtc_state->base.adjusted_mode.flags & DRM_MODE_FLAG_INTERLACE) {
> - if (INTEL_GEN(dev_priv) < 4 ||
> + if (INTEL_GEN(i915) < 4 ||
> intel_crtc_has_type(crtc_state, INTEL_OUTPUT_SDVO))
> pipeconf |= PIPECONF_INTERLACE_W_FIELD_INDICATION;
> else
> @@ -8175,7 +8175,7 @@ static void i9xx_set_pipeconf(const struct intel_crtc_state *crtc_state)
> pipeconf |= PIPECONF_PROGRESSIVE;
> }
>
> - if ((IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) &&
> + if ((IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) &&
> crtc_state->limited_color_range)
> pipeconf |= PIPECONF_COLOR_RANGE_SELECT;
>
> @@ -8189,7 +8189,7 @@ static int i8xx_crtc_compute_clock(struct intel_crtc *crtc,
> struct intel_crtc_state *crtc_state)
> {
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> const struct intel_limit *limit;
> int refclk = 48000;
>
> @@ -8197,8 +8197,8 @@ static int i8xx_crtc_compute_clock(struct intel_crtc *crtc,
> sizeof(crtc_state->dpll_hw_state));
>
> if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS)) {
> - if (intel_panel_use_ssc(dev_priv)) {
> - refclk = dev_priv->vbt.lvds_ssc_freq;
> + if (intel_panel_use_ssc(i915)) {
> + refclk = i915->vbt.lvds_ssc_freq;
> DRM_DEBUG_KMS("using SSC reference clock of %d kHz\n", refclk);
> }
>
> @@ -8224,7 +8224,7 @@ static int i8xx_crtc_compute_clock(struct intel_crtc *crtc,
> static int g4x_crtc_compute_clock(struct intel_crtc *crtc,
> struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> const struct intel_limit *limit;
> int refclk = 96000;
>
> @@ -8232,12 +8232,12 @@ static int g4x_crtc_compute_clock(struct intel_crtc *crtc,
> sizeof(crtc_state->dpll_hw_state));
>
> if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS)) {
> - if (intel_panel_use_ssc(dev_priv)) {
> - refclk = dev_priv->vbt.lvds_ssc_freq;
> + if (intel_panel_use_ssc(i915)) {
> + refclk = i915->vbt.lvds_ssc_freq;
> DRM_DEBUG_KMS("using SSC reference clock of %d kHz\n", refclk);
> }
>
> - if (intel_is_dual_link_lvds(dev_priv))
> + if (intel_is_dual_link_lvds(i915))
> limit = &intel_limits_g4x_dual_channel_lvds;
> else
> limit = &intel_limits_g4x_single_channel_lvds;
> @@ -8267,7 +8267,7 @@ static int pnv_crtc_compute_clock(struct intel_crtc *crtc,
> struct intel_crtc_state *crtc_state)
> {
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> const struct intel_limit *limit;
> int refclk = 96000;
>
> @@ -8275,8 +8275,8 @@ static int pnv_crtc_compute_clock(struct intel_crtc *crtc,
> sizeof(crtc_state->dpll_hw_state));
>
> if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS)) {
> - if (intel_panel_use_ssc(dev_priv)) {
> - refclk = dev_priv->vbt.lvds_ssc_freq;
> + if (intel_panel_use_ssc(i915)) {
> + refclk = i915->vbt.lvds_ssc_freq;
> DRM_DEBUG_KMS("using SSC reference clock of %d kHz\n", refclk);
> }
>
> @@ -8301,7 +8301,7 @@ static int i9xx_crtc_compute_clock(struct intel_crtc *crtc,
> struct intel_crtc_state *crtc_state)
> {
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> const struct intel_limit *limit;
> int refclk = 96000;
>
> @@ -8309,8 +8309,8 @@ static int i9xx_crtc_compute_clock(struct intel_crtc *crtc,
> sizeof(crtc_state->dpll_hw_state));
>
> if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS)) {
> - if (intel_panel_use_ssc(dev_priv)) {
> - refclk = dev_priv->vbt.lvds_ssc_freq;
> + if (intel_panel_use_ssc(i915)) {
> + refclk = i915->vbt.lvds_ssc_freq;
> DRM_DEBUG_KMS("using SSC reference clock of %d kHz\n", refclk);
> }
>
> @@ -8373,22 +8373,22 @@ static int vlv_crtc_compute_clock(struct intel_crtc *crtc,
> return 0;
> }
>
> -static bool i9xx_has_pfit(struct drm_i915_private *dev_priv)
> +static bool i9xx_has_pfit(struct drm_i915_private *i915)
> {
> - if (IS_I830(dev_priv))
> + if (IS_I830(i915))
> return false;
>
> - return INTEL_GEN(dev_priv) >= 4 ||
> - IS_PINEVIEW(dev_priv) || IS_MOBILE(dev_priv);
> + return INTEL_GEN(i915) >= 4 ||
> + IS_PINEVIEW(i915) || IS_MOBILE(i915);
> }
>
> static void i9xx_get_pfit_config(struct intel_crtc *crtc,
> struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> u32 tmp;
>
> - if (!i9xx_has_pfit(dev_priv))
> + if (!i9xx_has_pfit(i915))
> return;
>
> tmp = I915_READ(PFIT_CONTROL);
> @@ -8396,7 +8396,7 @@ static void i9xx_get_pfit_config(struct intel_crtc *crtc,
> return;
>
> /* Check whether the pfit is attached to our pipe. */
> - if (INTEL_GEN(dev_priv) < 4) {
> + if (INTEL_GEN(i915) < 4) {
> if (crtc->pipe != PIPE_B)
> return;
> } else {
> @@ -8412,7 +8412,7 @@ static void vlv_crtc_clock_get(struct intel_crtc *crtc,
> struct intel_crtc_state *pipe_config)
> {
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> int pipe = pipe_config->cpu_transcoder;
> struct dpll clock;
> u32 mdiv;
> @@ -8422,9 +8422,9 @@ static void vlv_crtc_clock_get(struct intel_crtc *crtc,
> if ((pipe_config->dpll_hw_state.dpll & DPLL_VCO_ENABLE) == 0)
> return;
>
> - vlv_dpio_get(dev_priv);
> - mdiv = vlv_dpio_read(dev_priv, pipe, VLV_PLL_DW3(pipe));
> - vlv_dpio_put(dev_priv);
> + vlv_dpio_get(i915);
> + mdiv = vlv_dpio_read(i915, pipe, VLV_PLL_DW3(pipe));
> + vlv_dpio_put(i915);
>
> clock.m1 = (mdiv >> DPIO_M1DIV_SHIFT) & 7;
> clock.m2 = mdiv & DPIO_M2DIV_MASK;
> @@ -8440,7 +8440,7 @@ i9xx_get_initial_plane_config(struct intel_crtc *crtc,
> struct intel_initial_plane_config *plane_config)
> {
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_plane *plane = to_intel_plane(crtc->base.primary);
> enum i9xx_plane_id i9xx_plane = plane->i9xx_plane;
> enum pipe pipe;
> @@ -8467,7 +8467,7 @@ i9xx_get_initial_plane_config(struct intel_crtc *crtc,
>
> val = I915_READ(DSPCNTR(i9xx_plane));
>
> - if (INTEL_GEN(dev_priv) >= 4) {
> + if (INTEL_GEN(i915) >= 4) {
> if (val & DISPPLANE_TILED) {
> plane_config->tiling = I915_TILING_X;
> fb->modifier = I915_FORMAT_MOD_X_TILED;
> @@ -8477,7 +8477,7 @@ i9xx_get_initial_plane_config(struct intel_crtc *crtc,
> plane_config->rotation = DRM_MODE_ROTATE_180;
> }
>
> - if (IS_CHERRYVIEW(dev_priv) && pipe == PIPE_B &&
> + if (IS_CHERRYVIEW(i915) && pipe == PIPE_B &&
> val & DISPPLANE_MIRROR)
> plane_config->rotation |= DRM_MODE_REFLECT_X;
>
> @@ -8485,10 +8485,10 @@ i9xx_get_initial_plane_config(struct intel_crtc *crtc,
> fourcc = i9xx_format_to_fourcc(pixel_format);
> fb->format = drm_format_info(fourcc);
>
> - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) {
> + if (IS_HASWELL(i915) || IS_BROADWELL(i915)) {
> offset = I915_READ(DSPOFFSET(i9xx_plane));
> base = I915_READ(DSPSURF(i9xx_plane)) & 0xfffff000;
> - } else if (INTEL_GEN(dev_priv) >= 4) {
> + } else if (INTEL_GEN(i915) >= 4) {
> if (plane_config->tiling)
> offset = I915_READ(DSPTILEOFF(i9xx_plane));
> else
> @@ -8522,7 +8522,7 @@ static void chv_crtc_clock_get(struct intel_crtc *crtc,
> struct intel_crtc_state *pipe_config)
> {
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> int pipe = pipe_config->cpu_transcoder;
> enum dpio_channel port = vlv_pipe_to_channel(pipe);
> struct dpll clock;
> @@ -8533,13 +8533,13 @@ static void chv_crtc_clock_get(struct intel_crtc *crtc,
> if ((pipe_config->dpll_hw_state.dpll & DPLL_VCO_ENABLE) == 0)
> return;
>
> - vlv_dpio_get(dev_priv);
> - cmn_dw13 = vlv_dpio_read(dev_priv, pipe, CHV_CMN_DW13(port));
> - pll_dw0 = vlv_dpio_read(dev_priv, pipe, CHV_PLL_DW0(port));
> - pll_dw1 = vlv_dpio_read(dev_priv, pipe, CHV_PLL_DW1(port));
> - pll_dw2 = vlv_dpio_read(dev_priv, pipe, CHV_PLL_DW2(port));
> - pll_dw3 = vlv_dpio_read(dev_priv, pipe, CHV_PLL_DW3(port));
> - vlv_dpio_put(dev_priv);
> + vlv_dpio_get(i915);
> + cmn_dw13 = vlv_dpio_read(i915, pipe, CHV_CMN_DW13(port));
> + pll_dw0 = vlv_dpio_read(i915, pipe, CHV_PLL_DW0(port));
> + pll_dw1 = vlv_dpio_read(i915, pipe, CHV_PLL_DW1(port));
> + pll_dw2 = vlv_dpio_read(i915, pipe, CHV_PLL_DW2(port));
> + pll_dw3 = vlv_dpio_read(i915, pipe, CHV_PLL_DW3(port));
> + vlv_dpio_put(i915);
>
> clock.m1 = (pll_dw1 & 0x7) == DPIO_CHV_M1_DIV_BY_2 ? 2 : 0;
> clock.m2 = (pll_dw0 & 0xff) << 22;
> @@ -8555,12 +8555,12 @@ static void chv_crtc_clock_get(struct intel_crtc *crtc,
> static void intel_get_crtc_ycbcr_config(struct intel_crtc *crtc,
> struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum intel_output_format output = INTEL_OUTPUT_FORMAT_RGB;
>
> pipe_config->lspcon_downsampling = false;
>
> - if (IS_BROADWELL(dev_priv) || INTEL_GEN(dev_priv) >= 9) {
> + if (IS_BROADWELL(i915) || INTEL_GEN(i915) >= 9) {
> u32 tmp = I915_READ(PIPEMISC(crtc->pipe));
>
> if (tmp & PIPEMISC_OUTPUT_COLORSPACE_YUV) {
> @@ -8571,8 +8571,8 @@ static void intel_get_crtc_ycbcr_config(struct intel_crtc *crtc,
> /* We support 4:2:0 in full blend mode only */
> if (!blend)
> output = INTEL_OUTPUT_FORMAT_INVALID;
> - else if (!(IS_GEMINILAKE(dev_priv) ||
> - INTEL_GEN(dev_priv) >= 10))
> + else if (!(IS_GEMINILAKE(i915) ||
> + INTEL_GEN(i915) >= 10))
> output = INTEL_OUTPUT_FORMAT_INVALID;
> else
> output = INTEL_OUTPUT_FORMAT_YCBCR420;
> @@ -8599,7 +8599,7 @@ static void i9xx_get_pipe_color_config(struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> struct intel_plane *plane = to_intel_plane(crtc->base.primary);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum i9xx_plane_id i9xx_plane = plane->i9xx_plane;
> u32 tmp;
>
> @@ -8608,7 +8608,7 @@ static void i9xx_get_pipe_color_config(struct intel_crtc_state *crtc_state)
> if (tmp & DISPPLANE_GAMMA_ENABLE)
> crtc_state->gamma_enable = true;
>
> - if (!HAS_GMCH(dev_priv) &&
> + if (!HAS_GMCH(i915) &&
> tmp & DISPPLANE_PIPE_CSC_ENABLE)
> crtc_state->csc_enable = true;
> }
> @@ -8616,14 +8616,14 @@ static void i9xx_get_pipe_color_config(struct intel_crtc_state *crtc_state)
> static bool i9xx_get_pipe_config(struct intel_crtc *crtc,
> struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum intel_display_power_domain power_domain;
> intel_wakeref_t wakeref;
> u32 tmp;
> bool ret;
>
> power_domain = POWER_DOMAIN_PIPE(crtc->pipe);
> - wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain);
> + wakeref = intel_display_power_get_if_enabled(i915, power_domain);
> if (!wakeref)
> return false;
>
> @@ -8637,8 +8637,8 @@ static bool i9xx_get_pipe_config(struct intel_crtc *crtc,
> if (!(tmp & PIPECONF_ENABLE))
> goto out;
>
> - if (IS_G4X(dev_priv) || IS_VALLEYVIEW(dev_priv) ||
> - IS_CHERRYVIEW(dev_priv)) {
> + if (IS_G4X(i915) || IS_VALLEYVIEW(i915) ||
> + IS_CHERRYVIEW(i915)) {
> switch (tmp & PIPECONF_BPC_MASK) {
> case PIPECONF_6BPC:
> pipe_config->pipe_bpp = 18;
> @@ -8654,20 +8654,20 @@ static bool i9xx_get_pipe_config(struct intel_crtc *crtc,
> }
> }
>
> - if ((IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) &&
> + if ((IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) &&
> (tmp & PIPECONF_COLOR_RANGE_SELECT))
> pipe_config->limited_color_range = true;
>
> pipe_config->gamma_mode = (tmp & PIPECONF_GAMMA_MODE_MASK_I9XX) >>
> PIPECONF_GAMMA_MODE_SHIFT;
>
> - if (IS_CHERRYVIEW(dev_priv))
> + if (IS_CHERRYVIEW(i915))
> pipe_config->cgm_mode = I915_READ(CGM_PIPE_MODE(crtc->pipe));
>
> i9xx_get_pipe_color_config(pipe_config);
> intel_color_get_config(pipe_config);
>
> - if (INTEL_GEN(dev_priv) < 4)
> + if (INTEL_GEN(i915) < 4)
> pipe_config->double_wide = tmp & PIPECONF_DOUBLE_WIDE;
>
> intel_get_pipe_timings(crtc, pipe_config);
> @@ -8675,18 +8675,18 @@ static bool i9xx_get_pipe_config(struct intel_crtc *crtc,
>
> i9xx_get_pfit_config(crtc, pipe_config);
>
> - if (INTEL_GEN(dev_priv) >= 4) {
> + if (INTEL_GEN(i915) >= 4) {
> /* No way to read it out on pipes B and C */
> - if (IS_CHERRYVIEW(dev_priv) && crtc->pipe != PIPE_A)
> - tmp = dev_priv->chv_dpll_md[crtc->pipe];
> + if (IS_CHERRYVIEW(i915) && crtc->pipe != PIPE_A)
> + tmp = i915->chv_dpll_md[crtc->pipe];
> else
> tmp = I915_READ(DPLL_MD(crtc->pipe));
> pipe_config->pixel_multiplier =
> ((tmp & DPLL_MD_UDI_MULTIPLIER_MASK)
> >> DPLL_MD_UDI_MULTIPLIER_SHIFT) + 1;
> pipe_config->dpll_hw_state.dpll_md = tmp;
> - } else if (IS_I945G(dev_priv) || IS_I945GM(dev_priv) ||
> - IS_G33(dev_priv) || IS_PINEVIEW(dev_priv)) {
> + } else if (IS_I945G(i915) || IS_I945GM(i915) ||
> + IS_G33(i915) || IS_PINEVIEW(i915)) {
> tmp = I915_READ(DPLL(crtc->pipe));
> pipe_config->pixel_multiplier =
> ((tmp & SDVO_MULTIPLIER_MASK)
> @@ -8698,7 +8698,7 @@ static bool i9xx_get_pipe_config(struct intel_crtc *crtc,
> pipe_config->pixel_multiplier = 1;
> }
> pipe_config->dpll_hw_state.dpll = I915_READ(DPLL(crtc->pipe));
> - if (!IS_VALLEYVIEW(dev_priv) && !IS_CHERRYVIEW(dev_priv)) {
> + if (!IS_VALLEYVIEW(i915) && !IS_CHERRYVIEW(i915)) {
> pipe_config->dpll_hw_state.fp0 = I915_READ(FP0(crtc->pipe));
> pipe_config->dpll_hw_state.fp1 = I915_READ(FP1(crtc->pipe));
> } else {
> @@ -8708,9 +8708,9 @@ static bool i9xx_get_pipe_config(struct intel_crtc *crtc,
> DPLL_PORTB_READY_MASK);
> }
>
> - if (IS_CHERRYVIEW(dev_priv))
> + if (IS_CHERRYVIEW(i915))
> chv_crtc_clock_get(crtc, pipe_config);
> - else if (IS_VALLEYVIEW(dev_priv))
> + else if (IS_VALLEYVIEW(i915))
> vlv_crtc_clock_get(crtc, pipe_config);
> else
> i9xx_crtc_clock_get(crtc, pipe_config);
> @@ -8726,12 +8726,12 @@ static bool i9xx_get_pipe_config(struct intel_crtc *crtc,
> ret = true;
>
> out:
> - intel_display_power_put(dev_priv, power_domain, wakeref);
> + intel_display_power_put(i915, power_domain, wakeref);
>
> return ret;
> }
>
> -static void ironlake_init_pch_refclk(struct drm_i915_private *dev_priv)
> +static void ironlake_init_pch_refclk(struct drm_i915_private *i915)
> {
> struct intel_encoder *encoder;
> int i;
> @@ -8744,7 +8744,7 @@ static void ironlake_init_pch_refclk(struct drm_i915_private *dev_priv)
> bool using_ssc_source = false;
>
> /* We need to take the global config into account */
> - for_each_intel_encoder(&dev_priv->drm, encoder) {
> + for_each_intel_encoder(&i915->drm, encoder) {
> switch (encoder->type) {
> case INTEL_OUTPUT_LVDS:
> has_panel = true;
> @@ -8760,8 +8760,8 @@ static void ironlake_init_pch_refclk(struct drm_i915_private *dev_priv)
> }
> }
>
> - if (HAS_PCH_IBX(dev_priv)) {
> - has_ck505 = dev_priv->vbt.display_clock_mode;
> + if (HAS_PCH_IBX(i915)) {
> + has_ck505 = i915->vbt.display_clock_mode;
> can_ssc = has_ck505;
> } else {
> has_ck505 = false;
> @@ -8769,7 +8769,7 @@ static void ironlake_init_pch_refclk(struct drm_i915_private *dev_priv)
> }
>
> /* Check if any DPLLs are using the SSC source */
> - for (i = 0; i < dev_priv->num_shared_dpll; i++) {
> + for (i = 0; i < i915->num_shared_dpll; i++) {
> u32 temp = I915_READ(PCH_DPLL(i));
>
> if (!(temp & DPLL_VCO_ENABLE))
> @@ -8810,11 +8810,11 @@ static void ironlake_init_pch_refclk(struct drm_i915_private *dev_priv)
> if (has_panel) {
> final |= DREF_SSC_SOURCE_ENABLE;
>
> - if (intel_panel_use_ssc(dev_priv) && can_ssc)
> + if (intel_panel_use_ssc(i915) && can_ssc)
> final |= DREF_SSC1_ENABLE;
>
> if (has_cpu_edp) {
> - if (intel_panel_use_ssc(dev_priv) && can_ssc)
> + if (intel_panel_use_ssc(i915) && can_ssc)
> final |= DREF_CPU_SOURCE_OUTPUT_DOWNSPREAD;
> else
> final |= DREF_CPU_SOURCE_OUTPUT_NONSPREAD;
> @@ -8841,7 +8841,7 @@ static void ironlake_init_pch_refclk(struct drm_i915_private *dev_priv)
> val |= DREF_SSC_SOURCE_ENABLE;
>
> /* SSC must be turned on before enabling the CPU output */
> - if (intel_panel_use_ssc(dev_priv) && can_ssc) {
> + if (intel_panel_use_ssc(i915) && can_ssc) {
> DRM_DEBUG_KMS("Using SSC on panel\n");
> val |= DREF_SSC1_ENABLE;
> } else
> @@ -8856,7 +8856,7 @@ static void ironlake_init_pch_refclk(struct drm_i915_private *dev_priv)
>
> /* Enable CPU source on CPU attached eDP */
> if (has_cpu_edp) {
> - if (intel_panel_use_ssc(dev_priv) && can_ssc) {
> + if (intel_panel_use_ssc(i915) && can_ssc) {
> DRM_DEBUG_KMS("Using SSC on eDP\n");
> val |= DREF_CPU_SOURCE_OUTPUT_DOWNSPREAD;
> } else
> @@ -8898,7 +8898,7 @@ static void ironlake_init_pch_refclk(struct drm_i915_private *dev_priv)
> BUG_ON(val != final);
> }
>
> -static void lpt_reset_fdi_mphy(struct drm_i915_private *dev_priv)
> +static void lpt_reset_fdi_mphy(struct drm_i915_private *i915)
> {
> u32 tmp;
>
> @@ -8920,78 +8920,78 @@ static void lpt_reset_fdi_mphy(struct drm_i915_private *dev_priv)
> }
>
> /* WaMPhyProgramming:hsw */
> -static void lpt_program_fdi_mphy(struct drm_i915_private *dev_priv)
> +static void lpt_program_fdi_mphy(struct drm_i915_private *i915)
> {
> u32 tmp;
>
> - tmp = intel_sbi_read(dev_priv, 0x8008, SBI_MPHY);
> + tmp = intel_sbi_read(i915, 0x8008, SBI_MPHY);
> tmp &= ~(0xFF << 24);
> tmp |= (0x12 << 24);
> - intel_sbi_write(dev_priv, 0x8008, tmp, SBI_MPHY);
> + intel_sbi_write(i915, 0x8008, tmp, SBI_MPHY);
>
> - tmp = intel_sbi_read(dev_priv, 0x2008, SBI_MPHY);
> + tmp = intel_sbi_read(i915, 0x2008, SBI_MPHY);
> tmp |= (1 << 11);
> - intel_sbi_write(dev_priv, 0x2008, tmp, SBI_MPHY);
> + intel_sbi_write(i915, 0x2008, tmp, SBI_MPHY);
>
> - tmp = intel_sbi_read(dev_priv, 0x2108, SBI_MPHY);
> + tmp = intel_sbi_read(i915, 0x2108, SBI_MPHY);
> tmp |= (1 << 11);
> - intel_sbi_write(dev_priv, 0x2108, tmp, SBI_MPHY);
> + intel_sbi_write(i915, 0x2108, tmp, SBI_MPHY);
>
> - tmp = intel_sbi_read(dev_priv, 0x206C, SBI_MPHY);
> + tmp = intel_sbi_read(i915, 0x206C, SBI_MPHY);
> tmp |= (1 << 24) | (1 << 21) | (1 << 18);
> - intel_sbi_write(dev_priv, 0x206C, tmp, SBI_MPHY);
> + intel_sbi_write(i915, 0x206C, tmp, SBI_MPHY);
>
> - tmp = intel_sbi_read(dev_priv, 0x216C, SBI_MPHY);
> + tmp = intel_sbi_read(i915, 0x216C, SBI_MPHY);
> tmp |= (1 << 24) | (1 << 21) | (1 << 18);
> - intel_sbi_write(dev_priv, 0x216C, tmp, SBI_MPHY);
> + intel_sbi_write(i915, 0x216C, tmp, SBI_MPHY);
>
> - tmp = intel_sbi_read(dev_priv, 0x2080, SBI_MPHY);
> + tmp = intel_sbi_read(i915, 0x2080, SBI_MPHY);
> tmp &= ~(7 << 13);
> tmp |= (5 << 13);
> - intel_sbi_write(dev_priv, 0x2080, tmp, SBI_MPHY);
> + intel_sbi_write(i915, 0x2080, tmp, SBI_MPHY);
>
> - tmp = intel_sbi_read(dev_priv, 0x2180, SBI_MPHY);
> + tmp = intel_sbi_read(i915, 0x2180, SBI_MPHY);
> tmp &= ~(7 << 13);
> tmp |= (5 << 13);
> - intel_sbi_write(dev_priv, 0x2180, tmp, SBI_MPHY);
> + intel_sbi_write(i915, 0x2180, tmp, SBI_MPHY);
>
> - tmp = intel_sbi_read(dev_priv, 0x208C, SBI_MPHY);
> + tmp = intel_sbi_read(i915, 0x208C, SBI_MPHY);
> tmp &= ~0xFF;
> tmp |= 0x1C;
> - intel_sbi_write(dev_priv, 0x208C, tmp, SBI_MPHY);
> + intel_sbi_write(i915, 0x208C, tmp, SBI_MPHY);
>
> - tmp = intel_sbi_read(dev_priv, 0x218C, SBI_MPHY);
> + tmp = intel_sbi_read(i915, 0x218C, SBI_MPHY);
> tmp &= ~0xFF;
> tmp |= 0x1C;
> - intel_sbi_write(dev_priv, 0x218C, tmp, SBI_MPHY);
> + intel_sbi_write(i915, 0x218C, tmp, SBI_MPHY);
>
> - tmp = intel_sbi_read(dev_priv, 0x2098, SBI_MPHY);
> + tmp = intel_sbi_read(i915, 0x2098, SBI_MPHY);
> tmp &= ~(0xFF << 16);
> tmp |= (0x1C << 16);
> - intel_sbi_write(dev_priv, 0x2098, tmp, SBI_MPHY);
> + intel_sbi_write(i915, 0x2098, tmp, SBI_MPHY);
>
> - tmp = intel_sbi_read(dev_priv, 0x2198, SBI_MPHY);
> + tmp = intel_sbi_read(i915, 0x2198, SBI_MPHY);
> tmp &= ~(0xFF << 16);
> tmp |= (0x1C << 16);
> - intel_sbi_write(dev_priv, 0x2198, tmp, SBI_MPHY);
> + intel_sbi_write(i915, 0x2198, tmp, SBI_MPHY);
>
> - tmp = intel_sbi_read(dev_priv, 0x20C4, SBI_MPHY);
> + tmp = intel_sbi_read(i915, 0x20C4, SBI_MPHY);
> tmp |= (1 << 27);
> - intel_sbi_write(dev_priv, 0x20C4, tmp, SBI_MPHY);
> + intel_sbi_write(i915, 0x20C4, tmp, SBI_MPHY);
>
> - tmp = intel_sbi_read(dev_priv, 0x21C4, SBI_MPHY);
> + tmp = intel_sbi_read(i915, 0x21C4, SBI_MPHY);
> tmp |= (1 << 27);
> - intel_sbi_write(dev_priv, 0x21C4, tmp, SBI_MPHY);
> + intel_sbi_write(i915, 0x21C4, tmp, SBI_MPHY);
>
> - tmp = intel_sbi_read(dev_priv, 0x20EC, SBI_MPHY);
> + tmp = intel_sbi_read(i915, 0x20EC, SBI_MPHY);
> tmp &= ~(0xF << 28);
> tmp |= (4 << 28);
> - intel_sbi_write(dev_priv, 0x20EC, tmp, SBI_MPHY);
> + intel_sbi_write(i915, 0x20EC, tmp, SBI_MPHY);
>
> - tmp = intel_sbi_read(dev_priv, 0x21EC, SBI_MPHY);
> + tmp = intel_sbi_read(i915, 0x21EC, SBI_MPHY);
> tmp &= ~(0xF << 28);
> tmp |= (4 << 28);
> - intel_sbi_write(dev_priv, 0x21EC, tmp, SBI_MPHY);
> + intel_sbi_write(i915, 0x21EC, tmp, SBI_MPHY);
> }
>
> /* Implements 3 different sequences from BSpec chapter "Display iCLK
> @@ -9000,69 +9000,69 @@ static void lpt_program_fdi_mphy(struct drm_i915_private *dev_priv)
> * - Sequence to enable CLKOUT_DP without spread
> * - Sequence to enable CLKOUT_DP for FDI usage and configure PCH FDI I/O
> */
> -static void lpt_enable_clkout_dp(struct drm_i915_private *dev_priv,
> +static void lpt_enable_clkout_dp(struct drm_i915_private *i915,
> bool with_spread, bool with_fdi)
> {
> u32 reg, tmp;
>
> if (WARN(with_fdi && !with_spread, "FDI requires downspread\n"))
> with_spread = true;
> - if (WARN(HAS_PCH_LPT_LP(dev_priv) &&
> + if (WARN(HAS_PCH_LPT_LP(i915) &&
> with_fdi, "LP PCH doesn't have FDI\n"))
> with_fdi = false;
>
> - mutex_lock(&dev_priv->sb_lock);
> + mutex_lock(&i915->sb_lock);
>
> - tmp = intel_sbi_read(dev_priv, SBI_SSCCTL, SBI_ICLK);
> + tmp = intel_sbi_read(i915, SBI_SSCCTL, SBI_ICLK);
> tmp &= ~SBI_SSCCTL_DISABLE;
> tmp |= SBI_SSCCTL_PATHALT;
> - intel_sbi_write(dev_priv, SBI_SSCCTL, tmp, SBI_ICLK);
> + intel_sbi_write(i915, SBI_SSCCTL, tmp, SBI_ICLK);
>
> udelay(24);
>
> if (with_spread) {
> - tmp = intel_sbi_read(dev_priv, SBI_SSCCTL, SBI_ICLK);
> + tmp = intel_sbi_read(i915, SBI_SSCCTL, SBI_ICLK);
> tmp &= ~SBI_SSCCTL_PATHALT;
> - intel_sbi_write(dev_priv, SBI_SSCCTL, tmp, SBI_ICLK);
> + intel_sbi_write(i915, SBI_SSCCTL, tmp, SBI_ICLK);
>
> if (with_fdi) {
> - lpt_reset_fdi_mphy(dev_priv);
> - lpt_program_fdi_mphy(dev_priv);
> + lpt_reset_fdi_mphy(i915);
> + lpt_program_fdi_mphy(i915);
> }
> }
>
> - reg = HAS_PCH_LPT_LP(dev_priv) ? SBI_GEN0 : SBI_DBUFF0;
> - tmp = intel_sbi_read(dev_priv, reg, SBI_ICLK);
> + reg = HAS_PCH_LPT_LP(i915) ? SBI_GEN0 : SBI_DBUFF0;
> + tmp = intel_sbi_read(i915, reg, SBI_ICLK);
> tmp |= SBI_GEN0_CFG_BUFFENABLE_DISABLE;
> - intel_sbi_write(dev_priv, reg, tmp, SBI_ICLK);
> + intel_sbi_write(i915, reg, tmp, SBI_ICLK);
>
> - mutex_unlock(&dev_priv->sb_lock);
> + mutex_unlock(&i915->sb_lock);
> }
>
> /* Sequence to disable CLKOUT_DP */
> -void lpt_disable_clkout_dp(struct drm_i915_private *dev_priv)
> +void lpt_disable_clkout_dp(struct drm_i915_private *i915)
> {
> u32 reg, tmp;
>
> - mutex_lock(&dev_priv->sb_lock);
> + mutex_lock(&i915->sb_lock);
>
> - reg = HAS_PCH_LPT_LP(dev_priv) ? SBI_GEN0 : SBI_DBUFF0;
> - tmp = intel_sbi_read(dev_priv, reg, SBI_ICLK);
> + reg = HAS_PCH_LPT_LP(i915) ? SBI_GEN0 : SBI_DBUFF0;
> + tmp = intel_sbi_read(i915, reg, SBI_ICLK);
> tmp &= ~SBI_GEN0_CFG_BUFFENABLE_DISABLE;
> - intel_sbi_write(dev_priv, reg, tmp, SBI_ICLK);
> + intel_sbi_write(i915, reg, tmp, SBI_ICLK);
>
> - tmp = intel_sbi_read(dev_priv, SBI_SSCCTL, SBI_ICLK);
> + tmp = intel_sbi_read(i915, SBI_SSCCTL, SBI_ICLK);
> if (!(tmp & SBI_SSCCTL_DISABLE)) {
> if (!(tmp & SBI_SSCCTL_PATHALT)) {
> tmp |= SBI_SSCCTL_PATHALT;
> - intel_sbi_write(dev_priv, SBI_SSCCTL, tmp, SBI_ICLK);
> + intel_sbi_write(i915, SBI_SSCCTL, tmp, SBI_ICLK);
> udelay(32);
> }
> tmp |= SBI_SSCCTL_DISABLE;
> - intel_sbi_write(dev_priv, SBI_SSCCTL, tmp, SBI_ICLK);
> + intel_sbi_write(i915, SBI_SSCCTL, tmp, SBI_ICLK);
> }
>
> - mutex_unlock(&dev_priv->sb_lock);
> + mutex_unlock(&i915->sb_lock);
> }
>
> #define BEND_IDX(steps) ((50 + (steps)) / 5)
> @@ -9097,7 +9097,7 @@ static const u16 sscdivintphase[] = {
> * < 0 slow down the clock, > 0 speed up the clock, 0 == no bend (135MHz)
> * change in clock period = -(steps / 10) * 5.787 ps
> */
> -static void lpt_bend_clkout_dp(struct drm_i915_private *dev_priv, int steps)
> +static void lpt_bend_clkout_dp(struct drm_i915_private *i915, int steps)
> {
> u32 tmp;
> int idx = BEND_IDX(steps);
> @@ -9108,25 +9108,25 @@ static void lpt_bend_clkout_dp(struct drm_i915_private *dev_priv, int steps)
> if (WARN_ON(idx >= ARRAY_SIZE(sscdivintphase)))
> return;
>
> - mutex_lock(&dev_priv->sb_lock);
> + mutex_lock(&i915->sb_lock);
>
> if (steps % 10 != 0)
> tmp = 0xAAAAAAAB;
> else
> tmp = 0x00000000;
> - intel_sbi_write(dev_priv, SBI_SSCDITHPHASE, tmp, SBI_ICLK);
> + intel_sbi_write(i915, SBI_SSCDITHPHASE, tmp, SBI_ICLK);
>
> - tmp = intel_sbi_read(dev_priv, SBI_SSCDIVINTPHASE, SBI_ICLK);
> + tmp = intel_sbi_read(i915, SBI_SSCDIVINTPHASE, SBI_ICLK);
> tmp &= 0xffff0000;
> tmp |= sscdivintphase[idx];
> - intel_sbi_write(dev_priv, SBI_SSCDIVINTPHASE, tmp, SBI_ICLK);
> + intel_sbi_write(i915, SBI_SSCDIVINTPHASE, tmp, SBI_ICLK);
>
> - mutex_unlock(&dev_priv->sb_lock);
> + mutex_unlock(&i915->sb_lock);
> }
>
> #undef BEND_IDX
>
> -static bool spll_uses_pch_ssc(struct drm_i915_private *dev_priv)
> +static bool spll_uses_pch_ssc(struct drm_i915_private *i915)
> {
> u32 fuse_strap = I915_READ(FUSE_STRAP);
> u32 ctl = I915_READ(SPLL_CTL);
> @@ -9138,14 +9138,14 @@ static bool spll_uses_pch_ssc(struct drm_i915_private *dev_priv)
> (fuse_strap & HSW_CPU_SSC_ENABLE) == 0)
> return true;
>
> - if (IS_BROADWELL(dev_priv) &&
> + if (IS_BROADWELL(i915) &&
> (ctl & SPLL_REF_MASK) == SPLL_REF_PCH_SSC_BDW)
> return true;
>
> return false;
> }
>
> -static bool wrpll_uses_pch_ssc(struct drm_i915_private *dev_priv,
> +static bool wrpll_uses_pch_ssc(struct drm_i915_private *i915,
> enum intel_dpll_id id)
> {
> u32 fuse_strap = I915_READ(FUSE_STRAP);
> @@ -9157,7 +9157,7 @@ static bool wrpll_uses_pch_ssc(struct drm_i915_private *dev_priv,
> if ((ctl & WRPLL_REF_MASK) == WRPLL_REF_PCH_SSC)
> return true;
>
> - if ((IS_BROADWELL(dev_priv) || IS_HSW_ULT(dev_priv)) &&
> + if ((IS_BROADWELL(i915) || IS_HSW_ULT(i915)) &&
> (ctl & WRPLL_REF_MASK) == WRPLL_REF_MUXED_SSC_BDW &&
> (fuse_strap & HSW_CPU_SSC_ENABLE) == 0)
> return true;
> @@ -9165,13 +9165,13 @@ static bool wrpll_uses_pch_ssc(struct drm_i915_private *dev_priv,
> return false;
> }
>
> -static void lpt_init_pch_refclk(struct drm_i915_private *dev_priv)
> +static void lpt_init_pch_refclk(struct drm_i915_private *i915)
> {
> struct intel_encoder *encoder;
> bool pch_ssc_in_use = false;
> bool has_fdi = false;
>
> - for_each_intel_encoder(&dev_priv->drm, encoder) {
> + for_each_intel_encoder(&i915->drm, encoder) {
> switch (encoder->type) {
> case INTEL_OUTPUT_ANALOG:
> has_fdi = true;
> @@ -9196,17 +9196,17 @@ static void lpt_init_pch_refclk(struct drm_i915_private *dev_priv)
> * clock hierarchy. That would also allow us to do
> * clock bending finally.
> */
> - if (spll_uses_pch_ssc(dev_priv)) {
> + if (spll_uses_pch_ssc(i915)) {
> DRM_DEBUG_KMS("SPLL using PCH SSC\n");
> pch_ssc_in_use = true;
> }
>
> - if (wrpll_uses_pch_ssc(dev_priv, DPLL_ID_WRPLL1)) {
> + if (wrpll_uses_pch_ssc(i915, DPLL_ID_WRPLL1)) {
> DRM_DEBUG_KMS("WRPLL1 using PCH SSC\n");
> pch_ssc_in_use = true;
> }
>
> - if (wrpll_uses_pch_ssc(dev_priv, DPLL_ID_WRPLL2)) {
> + if (wrpll_uses_pch_ssc(i915, DPLL_ID_WRPLL2)) {
> DRM_DEBUG_KMS("WRPLL2 using PCH SSC\n");
> pch_ssc_in_use = true;
> }
> @@ -9215,28 +9215,28 @@ static void lpt_init_pch_refclk(struct drm_i915_private *dev_priv)
> return;
>
> if (has_fdi) {
> - lpt_bend_clkout_dp(dev_priv, 0);
> - lpt_enable_clkout_dp(dev_priv, true, true);
> + lpt_bend_clkout_dp(i915, 0);
> + lpt_enable_clkout_dp(i915, true, true);
> } else {
> - lpt_disable_clkout_dp(dev_priv);
> + lpt_disable_clkout_dp(i915);
> }
> }
>
> /*
> * Initialize reference clocks when the driver loads
> */
> -void intel_init_pch_refclk(struct drm_i915_private *dev_priv)
> +void intel_init_pch_refclk(struct drm_i915_private *i915)
> {
> - if (HAS_PCH_IBX(dev_priv) || HAS_PCH_CPT(dev_priv))
> - ironlake_init_pch_refclk(dev_priv);
> - else if (HAS_PCH_LPT(dev_priv))
> - lpt_init_pch_refclk(dev_priv);
> + if (HAS_PCH_IBX(i915) || HAS_PCH_CPT(i915))
> + ironlake_init_pch_refclk(i915);
> + else if (HAS_PCH_LPT(i915))
> + lpt_init_pch_refclk(i915);
> }
>
> static void ironlake_set_pipeconf(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
> u32 val;
>
> @@ -9280,11 +9280,11 @@ static void ironlake_set_pipeconf(const struct intel_crtc_state *crtc_state)
> static void haswell_set_pipeconf(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> u32 val = 0;
>
> - if (IS_HASWELL(dev_priv) && crtc_state->dither)
> + if (IS_HASWELL(i915) && crtc_state->dither)
> val |= (PIPECONF_DITHER_EN | PIPECONF_DITHER_TYPE_SP);
>
> if (crtc_state->base.adjusted_mode.flags & DRM_MODE_FLAG_INTERLACE)
> @@ -9299,7 +9299,7 @@ static void haswell_set_pipeconf(const struct intel_crtc_state *crtc_state)
> static void bdw_set_pipemisc(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> u32 val = 0;
>
> switch (crtc_state->pipe_bpp) {
> @@ -9331,7 +9331,7 @@ static void bdw_set_pipemisc(const struct intel_crtc_state *crtc_state)
> val |= PIPEMISC_YUV420_ENABLE |
> PIPEMISC_YUV420_MODE_FULL_BLEND;
>
> - if (INTEL_GEN(dev_priv) >= 11 &&
> + if (INTEL_GEN(i915) >= 11 &&
> (crtc_state->active_planes & ~(icl_hdr_plane_mask() |
> BIT(PLANE_CURSOR))) == 0)
> val |= PIPEMISC_HDR_MODE_PRECISION;
> @@ -9341,7 +9341,7 @@ static void bdw_set_pipemisc(const struct intel_crtc_state *crtc_state)
>
> int bdw_get_pipemisc_bpp(struct intel_crtc *crtc)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> u32 tmp;
>
> tmp = I915_READ(PIPEMISC(crtc->pipe));
> @@ -9381,17 +9381,17 @@ static void ironlake_compute_dpll(struct intel_crtc *crtc,
> struct intel_crtc_state *crtc_state,
> struct dpll *reduced_clock)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> u32 dpll, fp, fp2;
> int factor;
>
> /* Enable autotuning of the PLL clock (if permissible) */
> factor = 21;
> if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS)) {
> - if ((intel_panel_use_ssc(dev_priv) &&
> - dev_priv->vbt.lvds_ssc_freq == 100000) ||
> - (HAS_PCH_IBX(dev_priv) &&
> - intel_is_dual_link_lvds(dev_priv)))
> + if ((intel_panel_use_ssc(i915) &&
> + i915->vbt.lvds_ssc_freq == 100000) ||
> + (HAS_PCH_IBX(i915) &&
> + intel_is_dual_link_lvds(i915)))
> factor = 25;
> } else if (crtc_state->sdvo_tv_clock) {
> factor = 20;
> @@ -9442,7 +9442,7 @@ static void ironlake_compute_dpll(struct intel_crtc *crtc,
> * clear if it''s a win or loss power wise. No point in doing
> * this on ILK at all since it has a fixed DPLL<->pipe mapping.
> */
> - if (INTEL_INFO(dev_priv)->num_pipes == 3 &&
> + if (INTEL_INFO(i915)->num_pipes == 3 &&
> intel_crtc_has_type(crtc_state, INTEL_OUTPUT_ANALOG))
> dpll |= DPLL_SDVO_HIGH_SPEED;
>
> @@ -9467,7 +9467,7 @@ static void ironlake_compute_dpll(struct intel_crtc *crtc,
> }
>
> if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS) &&
> - intel_panel_use_ssc(dev_priv))
> + intel_panel_use_ssc(i915))
> dpll |= PLLB_REF_INPUT_SPREADSPECTRUMIN;
> else
> dpll |= PLL_REF_INPUT_DREFCLK;
> @@ -9482,7 +9482,7 @@ static void ironlake_compute_dpll(struct intel_crtc *crtc,
> static int ironlake_crtc_compute_clock(struct intel_crtc *crtc,
> struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> const struct intel_limit *limit;
> int refclk = 120000;
>
> @@ -9494,13 +9494,13 @@ static int ironlake_crtc_compute_clock(struct intel_crtc *crtc,
> return 0;
>
> if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS)) {
> - if (intel_panel_use_ssc(dev_priv)) {
> + if (intel_panel_use_ssc(i915)) {
> DRM_DEBUG_KMS("using SSC reference clock of %d kHz\n",
> - dev_priv->vbt.lvds_ssc_freq);
> - refclk = dev_priv->vbt.lvds_ssc_freq;
> + i915->vbt.lvds_ssc_freq);
> + refclk = i915->vbt.lvds_ssc_freq;
> }
>
> - if (intel_is_dual_link_lvds(dev_priv)) {
> + if (intel_is_dual_link_lvds(i915)) {
> if (refclk == 100000)
> limit = &intel_limits_ironlake_dual_lvds_100m;
> else
> @@ -9537,7 +9537,7 @@ static void intel_pch_transcoder_get_m_n(struct intel_crtc *crtc,
> struct intel_link_m_n *m_n)
> {
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> enum pipe pipe = crtc->pipe;
>
> m_n->link_m = I915_READ(PCH_TRANS_LINK_M1(pipe));
> @@ -9554,10 +9554,10 @@ static void intel_cpu_transcoder_get_m_n(struct intel_crtc *crtc,
> struct intel_link_m_n *m_n,
> struct intel_link_m_n *m2_n2)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
>
> - if (INTEL_GEN(dev_priv) >= 5) {
> + if (INTEL_GEN(i915) >= 5) {
> m_n->link_m = I915_READ(PIPE_LINK_M1(transcoder));
> m_n->link_n = I915_READ(PIPE_LINK_N1(transcoder));
> m_n->gmch_m = I915_READ(PIPE_DATA_M1(transcoder))
> @@ -9566,7 +9566,7 @@ static void intel_cpu_transcoder_get_m_n(struct intel_crtc *crtc,
> m_n->tu = ((I915_READ(PIPE_DATA_M1(transcoder))
> & TU_SIZE_MASK) >> TU_SIZE_SHIFT) + 1;
>
> - if (m2_n2 && transcoder_has_m2_n2(dev_priv, transcoder)) {
> + if (m2_n2 && transcoder_has_m2_n2(i915, transcoder)) {
> m2_n2->link_m = I915_READ(PIPE_LINK_M2(transcoder));
> m2_n2->link_n = I915_READ(PIPE_LINK_N2(transcoder));
> m2_n2->gmch_m = I915_READ(PIPE_DATA_M2(transcoder))
> @@ -9608,7 +9608,7 @@ static void skylake_get_pfit_config(struct intel_crtc *crtc,
> struct intel_crtc_state *pipe_config)
> {
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_crtc_scaler_state *scaler_state = &pipe_config->scaler_state;
> u32 ps_ctrl = 0;
> int id = -1;
> @@ -9640,7 +9640,7 @@ skylake_get_initial_plane_config(struct intel_crtc *crtc,
> struct intel_initial_plane_config *plane_config)
> {
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_plane *plane = to_intel_plane(crtc->base.primary);
> enum plane_id plane_id = plane->id;
> enum pipe pipe;
> @@ -9667,12 +9667,12 @@ skylake_get_initial_plane_config(struct intel_crtc *crtc,
>
> val = I915_READ(PLANE_CTL(pipe, plane_id));
>
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> pixel_format = val & ICL_PLANE_CTL_FORMAT_MASK;
> else
> pixel_format = val & PLANE_CTL_FORMAT_MASK;
>
> - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) {
> + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) {
> alpha = I915_READ(PLANE_COLOR_CTL(pipe, plane_id));
> alpha &= PLANE_COLOR_ALPHA_MASK;
> } else {
> @@ -9729,7 +9729,7 @@ skylake_get_initial_plane_config(struct intel_crtc *crtc,
> break;
> }
>
> - if (INTEL_GEN(dev_priv) >= 10 &&
> + if (INTEL_GEN(i915) >= 10 &&
> val & PLANE_CTL_FLIP_HORIZONTAL)
> plane_config->rotation |= DRM_MODE_REFLECT_X;
>
> @@ -9766,7 +9766,7 @@ static void ironlake_get_pfit_config(struct intel_crtc *crtc,
> struct intel_crtc_state *pipe_config)
> {
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> u32 tmp;
>
> tmp = I915_READ(PF_CTL(crtc->pipe));
> @@ -9779,7 +9779,7 @@ static void ironlake_get_pfit_config(struct intel_crtc *crtc,
> /* We currently do not free assignements of panel fitters on
> * ivb/hsw (since we don't use the higher upscaling modes which
> * differentiates them) so just WARN about this case for now. */
> - if (IS_GEN(dev_priv, 7)) {
> + if (IS_GEN(i915, 7)) {
> WARN_ON((tmp & PF_PIPE_SEL_MASK_IVB) !=
> PF_PIPE_SEL_IVB(crtc->pipe));
> }
> @@ -9790,14 +9790,14 @@ static bool ironlake_get_pipe_config(struct intel_crtc *crtc,
> struct intel_crtc_state *pipe_config)
> {
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> enum intel_display_power_domain power_domain;
> intel_wakeref_t wakeref;
> u32 tmp;
> bool ret;
>
> power_domain = POWER_DOMAIN_PIPE(crtc->pipe);
> - wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain);
> + wakeref = intel_display_power_get_if_enabled(i915, power_domain);
> if (!wakeref)
> return false;
>
> @@ -9850,7 +9850,7 @@ static bool ironlake_get_pipe_config(struct intel_crtc *crtc,
>
> ironlake_get_fdi_m_n_config(crtc, pipe_config);
>
> - if (HAS_PCH_IBX(dev_priv)) {
> + if (HAS_PCH_IBX(i915)) {
> /*
> * The pipe->pch transcoder and pch transcoder->pll
> * mapping is fixed.
> @@ -9865,10 +9865,10 @@ static bool ironlake_get_pipe_config(struct intel_crtc *crtc,
> }
>
> pipe_config->shared_dpll =
> - intel_get_shared_dpll_by_id(dev_priv, pll_id);
> + intel_get_shared_dpll_by_id(i915, pll_id);
> pll = pipe_config->shared_dpll;
>
> - WARN_ON(!pll->info->funcs->get_hw_state(dev_priv, pll,
> + WARN_ON(!pll->info->funcs->get_hw_state(i915, pll,
> &pipe_config->dpll_hw_state));
>
> tmp = pipe_config->dpll_hw_state.dpll;
> @@ -9889,19 +9889,19 @@ static bool ironlake_get_pipe_config(struct intel_crtc *crtc,
> ret = true;
>
> out:
> - intel_display_power_put(dev_priv, power_domain, wakeref);
> + intel_display_power_put(i915, power_domain, wakeref);
>
> return ret;
> }
> static int haswell_crtc_compute_clock(struct intel_crtc *crtc,
> struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> struct intel_atomic_state *state =
> to_intel_atomic_state(crtc_state->base.state);
>
> if (!intel_crtc_has_type(crtc_state, INTEL_OUTPUT_DSI) ||
> - INTEL_GEN(dev_priv) >= 11) {
> + INTEL_GEN(i915) >= 11) {
> struct intel_encoder *encoder =
> intel_get_crtc_new_encoder(state, crtc_state);
>
> @@ -9915,7 +9915,7 @@ static int haswell_crtc_compute_clock(struct intel_crtc *crtc,
> return 0;
> }
>
> -static void cannonlake_get_ddi_pll(struct drm_i915_private *dev_priv,
> +static void cannonlake_get_ddi_pll(struct drm_i915_private *i915,
> enum port port,
> struct intel_crtc_state *pipe_config)
> {
> @@ -9928,10 +9928,10 @@ static void cannonlake_get_ddi_pll(struct drm_i915_private *dev_priv,
> if (WARN_ON(id < SKL_DPLL0 || id > SKL_DPLL2))
> return;
>
> - pipe_config->shared_dpll = intel_get_shared_dpll_by_id(dev_priv, id);
> + pipe_config->shared_dpll = intel_get_shared_dpll_by_id(i915, id);
> }
>
> -static void icelake_get_ddi_pll(struct drm_i915_private *dev_priv,
> +static void icelake_get_ddi_pll(struct drm_i915_private *i915,
> enum port port,
> struct intel_crtc_state *pipe_config)
> {
> @@ -9939,21 +9939,21 @@ static void icelake_get_ddi_pll(struct drm_i915_private *dev_priv,
> u32 temp;
>
> /* TODO: TBT pll not implemented. */
> - if (intel_port_is_combophy(dev_priv, port)) {
> + if (intel_port_is_combophy(i915, port)) {
> temp = I915_READ(DPCLKA_CFGCR0_ICL) &
> DPCLKA_CFGCR0_DDI_CLK_SEL_MASK(port);
> id = temp >> DPCLKA_CFGCR0_DDI_CLK_SEL_SHIFT(port);
> - } else if (intel_port_is_tc(dev_priv, port)) {
> - id = icl_tc_port_to_pll_id(intel_port_to_tc(dev_priv, port));
> + } else if (intel_port_is_tc(i915, port)) {
> + id = icl_tc_port_to_pll_id(intel_port_to_tc(i915, port));
> } else {
> WARN(1, "Invalid port %x\n", port);
> return;
> }
>
> - pipe_config->shared_dpll = intel_get_shared_dpll_by_id(dev_priv, id);
> + pipe_config->shared_dpll = intel_get_shared_dpll_by_id(i915, id);
> }
>
> -static void bxt_get_ddi_pll(struct drm_i915_private *dev_priv,
> +static void bxt_get_ddi_pll(struct drm_i915_private *i915,
> enum port port,
> struct intel_crtc_state *pipe_config)
> {
> @@ -9974,10 +9974,10 @@ static void bxt_get_ddi_pll(struct drm_i915_private *dev_priv,
> return;
> }
>
> - pipe_config->shared_dpll = intel_get_shared_dpll_by_id(dev_priv, id);
> + pipe_config->shared_dpll = intel_get_shared_dpll_by_id(i915, id);
> }
>
> -static void skylake_get_ddi_pll(struct drm_i915_private *dev_priv,
> +static void skylake_get_ddi_pll(struct drm_i915_private *i915,
> enum port port,
> struct intel_crtc_state *pipe_config)
> {
> @@ -9990,10 +9990,10 @@ static void skylake_get_ddi_pll(struct drm_i915_private *dev_priv,
> if (WARN_ON(id < SKL_DPLL0 || id > SKL_DPLL3))
> return;
>
> - pipe_config->shared_dpll = intel_get_shared_dpll_by_id(dev_priv, id);
> + pipe_config->shared_dpll = intel_get_shared_dpll_by_id(i915, id);
> }
>
> -static void haswell_get_ddi_pll(struct drm_i915_private *dev_priv,
> +static void haswell_get_ddi_pll(struct drm_i915_private *i915,
> enum port port,
> struct intel_crtc_state *pipe_config)
> {
> @@ -10026,7 +10026,7 @@ static void haswell_get_ddi_pll(struct drm_i915_private *dev_priv,
> return;
> }
>
> - pipe_config->shared_dpll = intel_get_shared_dpll_by_id(dev_priv, id);
> + pipe_config->shared_dpll = intel_get_shared_dpll_by_id(i915, id);
> }
>
> static bool hsw_get_transcoder_state(struct intel_crtc *crtc,
> @@ -10035,7 +10035,7 @@ static bool hsw_get_transcoder_state(struct intel_crtc *crtc,
> intel_wakeref_t *wakerefs)
> {
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> enum intel_display_power_domain power_domain;
> unsigned long panel_transcoder_mask = 0;
> unsigned long enabled_panel_transcoders = 0;
> @@ -10043,11 +10043,11 @@ static bool hsw_get_transcoder_state(struct intel_crtc *crtc,
> intel_wakeref_t wf;
> u32 tmp;
>
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> panel_transcoder_mask |=
> BIT(TRANSCODER_DSI_0) | BIT(TRANSCODER_DSI_1);
>
> - if (HAS_TRANSCODER_EDP(dev_priv))
> + if (HAS_TRANSCODER_EDP(i915))
> panel_transcoder_mask |= BIT(TRANSCODER_EDP);
>
> /*
> @@ -10062,7 +10062,7 @@ static bool hsw_get_transcoder_state(struct intel_crtc *crtc,
> */
> for_each_set_bit(panel_transcoder,
> &panel_transcoder_mask,
> - ARRAY_SIZE(INTEL_INFO(dev_priv)->trans_offsets)) {
> + ARRAY_SIZE(INTEL_INFO(i915)->trans_offsets)) {
> bool force_thru = false;
> enum pipe trans_pipe;
>
> @@ -10113,7 +10113,7 @@ static bool hsw_get_transcoder_state(struct intel_crtc *crtc,
> power_domain = POWER_DOMAIN_TRANSCODER(pipe_config->cpu_transcoder);
> WARN_ON(*power_domain_mask & BIT_ULL(power_domain));
>
> - wf = intel_display_power_get_if_enabled(dev_priv, power_domain);
> + wf = intel_display_power_get_if_enabled(i915, power_domain);
> if (!wf)
> return false;
>
> @@ -10131,7 +10131,7 @@ static bool bxt_get_dsi_transcoder_state(struct intel_crtc *crtc,
> intel_wakeref_t *wakerefs)
> {
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> enum intel_display_power_domain power_domain;
> enum transcoder cpu_transcoder;
> intel_wakeref_t wf;
> @@ -10147,7 +10147,7 @@ static bool bxt_get_dsi_transcoder_state(struct intel_crtc *crtc,
> power_domain = POWER_DOMAIN_TRANSCODER(cpu_transcoder);
> WARN_ON(*power_domain_mask & BIT_ULL(power_domain));
>
> - wf = intel_display_power_get_if_enabled(dev_priv, power_domain);
> + wf = intel_display_power_get_if_enabled(i915, power_domain);
> if (!wf)
> continue;
>
> @@ -10161,7 +10161,7 @@ static bool bxt_get_dsi_transcoder_state(struct intel_crtc *crtc,
> * registers/MIPI[BXT]. We can break out here early, since we
> * need the same DSI PLL to be enabled for both DSI ports.
> */
> - if (!bxt_dsi_pll_is_enabled(dev_priv))
> + if (!bxt_dsi_pll_is_enabled(i915))
> break;
>
> /* XXX: this works for video mode only */
> @@ -10183,7 +10183,7 @@ static bool bxt_get_dsi_transcoder_state(struct intel_crtc *crtc,
> static void haswell_get_ddi_port_state(struct intel_crtc *crtc,
> struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> struct intel_shared_dpll *pll;
> enum port port;
> u32 tmp;
> @@ -10192,20 +10192,20 @@ static void haswell_get_ddi_port_state(struct intel_crtc *crtc,
>
> port = (tmp & TRANS_DDI_PORT_MASK) >> TRANS_DDI_PORT_SHIFT;
>
> - if (INTEL_GEN(dev_priv) >= 11)
> - icelake_get_ddi_pll(dev_priv, port, pipe_config);
> - else if (IS_CANNONLAKE(dev_priv))
> - cannonlake_get_ddi_pll(dev_priv, port, pipe_config);
> - else if (IS_GEN9_BC(dev_priv))
> - skylake_get_ddi_pll(dev_priv, port, pipe_config);
> - else if (IS_GEN9_LP(dev_priv))
> - bxt_get_ddi_pll(dev_priv, port, pipe_config);
> + if (INTEL_GEN(i915) >= 11)
> + icelake_get_ddi_pll(i915, port, pipe_config);
> + else if (IS_CANNONLAKE(i915))
> + cannonlake_get_ddi_pll(i915, port, pipe_config);
> + else if (IS_GEN9_BC(i915))
> + skylake_get_ddi_pll(i915, port, pipe_config);
> + else if (IS_GEN9_LP(i915))
> + bxt_get_ddi_pll(i915, port, pipe_config);
> else
> - haswell_get_ddi_pll(dev_priv, port, pipe_config);
> + haswell_get_ddi_pll(i915, port, pipe_config);
>
> pll = pipe_config->shared_dpll;
> if (pll) {
> - WARN_ON(!pll->info->funcs->get_hw_state(dev_priv, pll,
> + WARN_ON(!pll->info->funcs->get_hw_state(i915, pll,
> &pipe_config->dpll_hw_state));
> }
>
> @@ -10214,7 +10214,7 @@ static void haswell_get_ddi_port_state(struct intel_crtc *crtc,
> * DDI E. So just check whether this pipe is wired to DDI E and whether
> * the PCH transcoder is on.
> */
> - if (INTEL_GEN(dev_priv) < 9 &&
> + if (INTEL_GEN(i915) < 9 &&
> (port == PORT_E) && I915_READ(LPT_TRANSCONF) & TRANS_ENABLE) {
> pipe_config->has_pch_encoder = true;
>
> @@ -10229,7 +10229,7 @@ static void haswell_get_ddi_port_state(struct intel_crtc *crtc,
> static bool haswell_get_pipe_config(struct intel_crtc *crtc,
> struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> intel_wakeref_t wakerefs[POWER_DOMAIN_NUM], wf;
> enum intel_display_power_domain power_domain;
> u64 power_domain_mask;
> @@ -10238,7 +10238,7 @@ static bool haswell_get_pipe_config(struct intel_crtc *crtc,
> intel_crtc_init_scalers(crtc, pipe_config);
>
> power_domain = POWER_DOMAIN_PIPE(crtc->pipe);
> - wf = intel_display_power_get_if_enabled(dev_priv, power_domain);
> + wf = intel_display_power_get_if_enabled(i915, power_domain);
> if (!wf)
> return false;
>
> @@ -10250,7 +10250,7 @@ static bool haswell_get_pipe_config(struct intel_crtc *crtc,
> active = hsw_get_transcoder_state(crtc, pipe_config,
> &power_domain_mask, wakerefs);
>
> - if (IS_GEN9_LP(dev_priv) &&
> + if (IS_GEN9_LP(i915) &&
> bxt_get_dsi_transcoder_state(crtc, pipe_config,
> &power_domain_mask, wakerefs)) {
> WARN_ON(active);
> @@ -10261,7 +10261,7 @@ static bool haswell_get_pipe_config(struct intel_crtc *crtc,
> goto out;
>
> if (!transcoder_is_dsi(pipe_config->cpu_transcoder) ||
> - INTEL_GEN(dev_priv) >= 11) {
> + INTEL_GEN(i915) >= 11) {
> haswell_get_ddi_port_state(crtc, pipe_config);
> intel_get_pipe_timings(crtc, pipe_config);
> }
> @@ -10273,7 +10273,7 @@ static bool haswell_get_pipe_config(struct intel_crtc *crtc,
>
> pipe_config->csc_mode = I915_READ(PIPE_CSC_MODE(crtc->pipe));
>
> - if (INTEL_GEN(dev_priv) >= 9) {
> + if (INTEL_GEN(i915) >= 9) {
> u32 tmp = I915_READ(SKL_BOTTOM_COLOR(crtc->pipe));
>
> if (tmp & SKL_BOTTOM_COLOR_GAMMA_ENABLE)
> @@ -10290,19 +10290,19 @@ static bool haswell_get_pipe_config(struct intel_crtc *crtc,
> power_domain = POWER_DOMAIN_PIPE_PANEL_FITTER(crtc->pipe);
> WARN_ON(power_domain_mask & BIT_ULL(power_domain));
>
> - wf = intel_display_power_get_if_enabled(dev_priv, power_domain);
> + wf = intel_display_power_get_if_enabled(i915, power_domain);
> if (wf) {
> wakerefs[power_domain] = wf;
> power_domain_mask |= BIT_ULL(power_domain);
>
> - if (INTEL_GEN(dev_priv) >= 9)
> + if (INTEL_GEN(i915) >= 9)
> skylake_get_pfit_config(crtc, pipe_config);
> else
> ironlake_get_pfit_config(crtc, pipe_config);
> }
>
> if (hsw_crtc_supports_ips(crtc)) {
> - if (IS_HASWELL(dev_priv))
> + if (IS_HASWELL(i915))
> pipe_config->ips_enabled = I915_READ(IPS_CTL) & IPS_ENABLE;
> else {
> /*
> @@ -10324,7 +10324,7 @@ static bool haswell_get_pipe_config(struct intel_crtc *crtc,
>
> out:
> for_each_power_domain(power_domain, power_domain_mask)
> - intel_display_power_put(dev_priv,
> + intel_display_power_put(i915,
> power_domain, wakerefs[power_domain]);
>
> return active;
> @@ -10332,13 +10332,13 @@ static bool haswell_get_pipe_config(struct intel_crtc *crtc,
>
> static u32 intel_cursor_base(const struct intel_plane_state *plane_state)
> {
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> to_i915(plane_state->base.plane->dev);
> const struct drm_framebuffer *fb = plane_state->base.fb;
> const struct drm_i915_gem_object *obj = intel_fb_obj(fb);
> u32 base;
>
> - if (INTEL_INFO(dev_priv)->display.cursor_needs_physical)
> + if (INTEL_INFO(i915)->display.cursor_needs_physical)
> base = obj->phys_handle->busaddr;
> else
> base = intel_plane_ggtt_offset(plane_state);
> @@ -10346,7 +10346,7 @@ static u32 intel_cursor_base(const struct intel_plane_state *plane_state)
> base += plane_state->color_plane[0].offset;
>
> /* ILK+ do this automagically */
> - if (HAS_GMCH(dev_priv) &&
> + if (HAS_GMCH(i915) &&
> plane_state->base.rotation & DRM_MODE_ROTATE_180)
> base += (plane_state->base.crtc_h *
> plane_state->base.crtc_w - 1) * fb->format->cpp[0];
> @@ -10532,7 +10532,7 @@ static void i845_update_cursor(struct intel_plane *plane,
> const struct intel_crtc_state *crtc_state,
> const struct intel_plane_state *plane_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> u32 cntl = 0, base = 0, pos = 0, size = 0;
> unsigned long irqflags;
>
> @@ -10549,7 +10549,7 @@ static void i845_update_cursor(struct intel_plane *plane,
> pos = intel_cursor_position(plane_state);
> }
>
> - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> + spin_lock_irqsave(&i915->uncore.lock, irqflags);
>
> /* On these chipsets we can only modify the base/size/stride
> * whilst the cursor is disabled.
> @@ -10570,7 +10570,7 @@ static void i845_update_cursor(struct intel_plane *plane,
> I915_WRITE_FW(CURPOS(PIPE_A), pos);
> }
>
> - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> + spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
> }
>
> static void i845_disable_cursor(struct intel_plane *plane,
> @@ -10582,13 +10582,13 @@ static void i845_disable_cursor(struct intel_plane *plane,
> static bool i845_cursor_get_hw_state(struct intel_plane *plane,
> enum pipe *pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> enum intel_display_power_domain power_domain;
> intel_wakeref_t wakeref;
> bool ret;
>
> power_domain = POWER_DOMAIN_PIPE(PIPE_A);
> - wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain);
> + wakeref = intel_display_power_get_if_enabled(i915, power_domain);
> if (!wakeref)
> return false;
>
> @@ -10596,7 +10596,7 @@ static bool i845_cursor_get_hw_state(struct intel_plane *plane,
>
> *pipe = PIPE_A;
>
> - intel_display_power_put(dev_priv, power_domain, wakeref);
> + intel_display_power_put(i915, power_domain, wakeref);
>
> return ret;
> }
> @@ -10612,10 +10612,10 @@ i9xx_cursor_max_stride(struct intel_plane *plane,
> static u32 i9xx_cursor_ctl_crtc(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> u32 cntl = 0;
>
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> return cntl;
>
> if (crtc_state->gamma_enable)
> @@ -10624,7 +10624,7 @@ static u32 i9xx_cursor_ctl_crtc(const struct intel_crtc_state *crtc_state)
> if (crtc_state->csc_enable)
> cntl |= MCURSOR_PIPE_CSC_ENABLE;
>
> - if (INTEL_GEN(dev_priv) < 5 && !IS_G4X(dev_priv))
> + if (INTEL_GEN(i915) < 5 && !IS_G4X(i915))
> cntl |= MCURSOR_PIPE_SELECT(crtc->pipe);
>
> return cntl;
> @@ -10633,11 +10633,11 @@ static u32 i9xx_cursor_ctl_crtc(const struct intel_crtc_state *crtc_state)
> static u32 i9xx_cursor_ctl(const struct intel_crtc_state *crtc_state,
> const struct intel_plane_state *plane_state)
> {
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> to_i915(plane_state->base.plane->dev);
> u32 cntl = 0;
>
> - if (IS_GEN(dev_priv, 6) || IS_IVYBRIDGE(dev_priv))
> + if (IS_GEN(i915, 6) || IS_IVYBRIDGE(i915))
> cntl |= MCURSOR_TRICKLE_FEED_DISABLE;
>
> switch (plane_state->base.crtc_w) {
> @@ -10663,7 +10663,7 @@ static u32 i9xx_cursor_ctl(const struct intel_crtc_state *crtc_state,
>
> static bool i9xx_cursor_size_ok(const struct intel_plane_state *plane_state)
> {
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> to_i915(plane_state->base.plane->dev);
> int width = plane_state->base.crtc_w;
> int height = plane_state->base.crtc_h;
> @@ -10687,7 +10687,7 @@ static bool i9xx_cursor_size_ok(const struct intel_plane_state *plane_state)
> * cursor is not rotated. Everything else requires square
> * cursors.
> */
> - if (HAS_CUR_FBC(dev_priv) &&
> + if (HAS_CUR_FBC(i915) &&
> plane_state->base.rotation & DRM_MODE_ROTATE_0) {
> if (height < 8 || height > width)
> return false;
> @@ -10703,7 +10703,7 @@ static int i9xx_check_cursor(struct intel_crtc_state *crtc_state,
> struct intel_plane_state *plane_state)
> {
> struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> const struct drm_framebuffer *fb = plane_state->base.fb;
> enum pipe pipe = plane->pipe;
> int ret;
> @@ -10743,7 +10743,7 @@ static int i9xx_check_cursor(struct intel_crtc_state *crtc_state,
> * display power well must be turned off and on again.
> * Refuse the put the cursor into that compromised position.
> */
> - if (IS_CHERRYVIEW(dev_priv) && pipe == PIPE_C &&
> + if (IS_CHERRYVIEW(i915) && pipe == PIPE_C &&
> plane_state->base.visible && plane_state->base.crtc_x < 0) {
> DRM_DEBUG_KMS("CHV cursor C not allowed to straddle the left screen edge\n");
> return -EINVAL;
> @@ -10758,7 +10758,7 @@ static void i9xx_update_cursor(struct intel_plane *plane,
> const struct intel_crtc_state *crtc_state,
> const struct intel_plane_state *plane_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> enum pipe pipe = plane->pipe;
> u32 cntl = 0, base = 0, pos = 0, fbc_ctl = 0;
> unsigned long irqflags;
> @@ -10774,7 +10774,7 @@ static void i9xx_update_cursor(struct intel_plane *plane,
> pos = intel_cursor_position(plane_state);
> }
>
> - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> + spin_lock_irqsave(&i915->uncore.lock, irqflags);
>
> /*
> * On some platforms writing CURCNTR first will also
> @@ -10796,13 +10796,13 @@ static void i9xx_update_cursor(struct intel_plane *plane,
> * the CURCNTR write arms the update.
> */
>
> - if (INTEL_GEN(dev_priv) >= 9)
> + if (INTEL_GEN(i915) >= 9)
> skl_write_cursor_wm(plane, crtc_state);
>
> if (plane->cursor.base != base ||
> plane->cursor.size != fbc_ctl ||
> plane->cursor.cntl != cntl) {
> - if (HAS_CUR_FBC(dev_priv))
> + if (HAS_CUR_FBC(i915))
> I915_WRITE_FW(CUR_FBC_CTL(pipe), fbc_ctl);
> I915_WRITE_FW(CURCNTR(pipe), cntl);
> I915_WRITE_FW(CURPOS(pipe), pos);
> @@ -10816,7 +10816,7 @@ static void i9xx_update_cursor(struct intel_plane *plane,
> I915_WRITE_FW(CURBASE(pipe), base);
> }
>
> - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> + spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
> }
>
> static void i9xx_disable_cursor(struct intel_plane *plane,
> @@ -10828,7 +10828,7 @@ static void i9xx_disable_cursor(struct intel_plane *plane,
> static bool i9xx_cursor_get_hw_state(struct intel_plane *plane,
> enum pipe *pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> enum intel_display_power_domain power_domain;
> intel_wakeref_t wakeref;
> bool ret;
> @@ -10840,7 +10840,7 @@ static bool i9xx_cursor_get_hw_state(struct intel_plane *plane,
> * display power wells.
> */
> power_domain = POWER_DOMAIN_PIPE(plane->pipe);
> - wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain);
> + wakeref = intel_display_power_get_if_enabled(i915, power_domain);
> if (!wakeref)
> return false;
>
> @@ -10848,13 +10848,13 @@ static bool i9xx_cursor_get_hw_state(struct intel_plane *plane,
>
> ret = val & MCURSOR_MODE;
>
> - if (INTEL_GEN(dev_priv) >= 5 || IS_G4X(dev_priv))
> + if (INTEL_GEN(i915) >= 5 || IS_G4X(i915))
> *pipe = plane->pipe;
> else
> *pipe = (val & MCURSOR_PIPE_SELECT_MASK) >>
> MCURSOR_PIPE_SELECT_SHIFT;
>
> - intel_display_power_put(dev_priv, power_domain, wakeref);
> + intel_display_power_put(i915, power_domain, wakeref);
>
> return ret;
> }
> @@ -10924,7 +10924,7 @@ int intel_get_load_detect_pipe(struct drm_connector *connector,
> struct drm_encoder *encoder = &intel_encoder->base;
> struct drm_crtc *crtc = NULL;
> struct drm_device *dev = encoder->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct drm_mode_config *config = &dev->mode_config;
> struct drm_atomic_state *state = NULL, *restore_state = NULL;
> struct drm_connector_state *connector_state;
> @@ -11051,7 +11051,7 @@ int intel_get_load_detect_pipe(struct drm_connector *connector,
> drm_atomic_state_put(state);
>
> /* let the connector get through one full cycle before testing */
> - intel_wait_for_vblank(dev_priv, intel_crtc->pipe);
> + intel_wait_for_vblank(i915, intel_crtc->pipe);
> return true;
>
> fail:
> @@ -11096,14 +11096,14 @@ void intel_release_load_detect_pipe(struct drm_connector *connector,
> static int i9xx_pll_refclk(struct drm_device *dev,
> const struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> u32 dpll = pipe_config->dpll_hw_state.dpll;
>
> if ((dpll & PLL_REF_INPUT_MASK) == PLLB_REF_INPUT_SPREADSPECTRUMIN)
> - return dev_priv->vbt.lvds_ssc_freq;
> - else if (HAS_PCH_SPLIT(dev_priv))
> + return i915->vbt.lvds_ssc_freq;
> + else if (HAS_PCH_SPLIT(i915))
> return 120000;
> - else if (!IS_GEN(dev_priv, 2))
> + else if (!IS_GEN(i915, 2))
> return 96000;
> else
> return 48000;
> @@ -11114,7 +11114,7 @@ static void i9xx_crtc_clock_get(struct intel_crtc *crtc,
> struct intel_crtc_state *pipe_config)
> {
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> int pipe = pipe_config->cpu_transcoder;
> u32 dpll = pipe_config->dpll_hw_state.dpll;
> u32 fp;
> @@ -11128,7 +11128,7 @@ static void i9xx_crtc_clock_get(struct intel_crtc *crtc,
> fp = pipe_config->dpll_hw_state.fp1;
>
> clock.m1 = (fp & FP_M1_DIV_MASK) >> FP_M1_DIV_SHIFT;
> - if (IS_PINEVIEW(dev_priv)) {
> + if (IS_PINEVIEW(i915)) {
> clock.n = ffs((fp & FP_N_PINEVIEW_DIV_MASK) >> FP_N_DIV_SHIFT) - 1;
> clock.m2 = (fp & FP_M2_PINEVIEW_DIV_MASK) >> FP_M2_DIV_SHIFT;
> } else {
> @@ -11136,8 +11136,8 @@ static void i9xx_crtc_clock_get(struct intel_crtc *crtc,
> clock.m2 = (fp & FP_M2_DIV_MASK) >> FP_M2_DIV_SHIFT;
> }
>
> - if (!IS_GEN(dev_priv, 2)) {
> - if (IS_PINEVIEW(dev_priv))
> + if (!IS_GEN(i915, 2)) {
> + if (IS_PINEVIEW(i915))
> clock.p1 = ffs((dpll & DPLL_FPA01_P1_POST_DIV_MASK_PINEVIEW) >>
> DPLL_FPA01_P1_POST_DIV_SHIFT_PINEVIEW);
> else
> @@ -11159,12 +11159,12 @@ static void i9xx_crtc_clock_get(struct intel_crtc *crtc,
> return;
> }
>
> - if (IS_PINEVIEW(dev_priv))
> + if (IS_PINEVIEW(i915))
> port_clock = pnv_calc_dpll_params(refclk, &clock);
> else
> port_clock = i9xx_calc_dpll_params(refclk, &clock);
> } else {
> - u32 lvds = IS_I830(dev_priv) ? 0 : I915_READ(LVDS);
> + u32 lvds = IS_I830(i915) ? 0 : I915_READ(LVDS);
> bool is_lvds = (pipe == 1) && (lvds & LVDS_PORT_EN);
>
> if (is_lvds) {
> @@ -11221,7 +11221,7 @@ int intel_dotclock_calculate(int link_freq,
> static void ironlake_pch_clock_get(struct intel_crtc *crtc,
> struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
>
> /* read out port_clock from the DPLL */
> i9xx_crtc_clock_get(crtc, pipe_config);
> @@ -11232,7 +11232,7 @@ static void ironlake_pch_clock_get(struct intel_crtc *crtc,
> * Calculate one based on the FDI configuration.
> */
> pipe_config->base.adjusted_mode.crtc_clock =
> - intel_dotclock_calculate(intel_fdi_link_freq(dev_priv, pipe_config),
> + intel_dotclock_calculate(intel_fdi_link_freq(i915, pipe_config),
> &pipe_config->fdi_m_n);
> }
>
> @@ -11240,7 +11240,7 @@ static void ironlake_pch_clock_get(struct intel_crtc *crtc,
> struct drm_display_mode *
> intel_encoder_current_mode(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crtc_state *crtc_state;
> struct drm_display_mode *mode;
> struct intel_crtc *crtc;
> @@ -11249,7 +11249,7 @@ intel_encoder_current_mode(struct intel_encoder *encoder)
> if (!encoder->get_hw_state(encoder, &pipe))
> return NULL;
>
> - crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> + crtc = intel_get_crtc_for_pipe(i915, pipe);
>
> mode = kzalloc(sizeof(*mode), GFP_KERNEL);
> if (!mode)
> @@ -11263,7 +11263,7 @@ intel_encoder_current_mode(struct intel_encoder *encoder)
>
> crtc_state->base.crtc = &crtc->base;
>
> - if (!dev_priv->display.get_pipe_config(crtc, crtc_state)) {
> + if (!i915->display.get_pipe_config(crtc, crtc_state)) {
> kfree(crtc_state);
> kfree(mode);
> return NULL;
> @@ -11337,7 +11337,7 @@ int intel_plane_atomic_calc_changes(const struct intel_crtc_state *old_crtc_stat
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> struct intel_plane *plane = to_intel_plane(plane_state->plane);
> struct drm_device *dev = crtc->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> bool mode_changed = needs_modeset(crtc_state);
> bool was_crtc_enabled = old_crtc_state->base.active;
> bool is_crtc_enabled = crtc_state->active;
> @@ -11345,7 +11345,7 @@ int intel_plane_atomic_calc_changes(const struct intel_crtc_state *old_crtc_stat
> struct drm_framebuffer *fb = plane_state->fb;
> int ret;
>
> - if (INTEL_GEN(dev_priv) >= 9 && plane->id != PLANE_CURSOR) {
> + if (INTEL_GEN(i915) >= 9 && plane->id != PLANE_CURSOR) {
> ret = skl_update_scaler_plane(
> to_intel_crtc_state(crtc_state),
> to_intel_plane_state(plane_state));
> @@ -11395,14 +11395,14 @@ int intel_plane_atomic_calc_changes(const struct intel_crtc_state *old_crtc_stat
> turn_off, turn_on, mode_changed);
>
> if (turn_on) {
> - if (INTEL_GEN(dev_priv) < 5 && !IS_G4X(dev_priv))
> + if (INTEL_GEN(i915) < 5 && !IS_G4X(i915))
> pipe_config->update_wm_pre = true;
>
> /* must disable cxsr around plane enable/disable */
> if (plane->id != PLANE_CURSOR)
> pipe_config->disable_cxsr = true;
> } else if (turn_off) {
> - if (INTEL_GEN(dev_priv) < 5 && !IS_G4X(dev_priv))
> + if (INTEL_GEN(i915) < 5 && !IS_G4X(i915))
> pipe_config->update_wm_post = true;
>
> /* must disable cxsr around plane enable/disable */
> @@ -11410,7 +11410,7 @@ int intel_plane_atomic_calc_changes(const struct intel_crtc_state *old_crtc_stat
> pipe_config->disable_cxsr = true;
> } else if (intel_wm_need_update(to_intel_plane_state(plane->base.state),
> to_intel_plane_state(plane_state))) {
> - if (INTEL_GEN(dev_priv) < 5 && !IS_G4X(dev_priv)) {
> + if (INTEL_GEN(i915) < 5 && !IS_G4X(i915)) {
> /* FIXME bollocks */
> pipe_config->update_wm_pre = true;
> pipe_config->update_wm_post = true;
> @@ -11454,8 +11454,8 @@ int intel_plane_atomic_calc_changes(const struct intel_crtc_state *old_crtc_stat
> * plane, not only sprite plane.
> */
> if (plane->id != PLANE_CURSOR &&
> - (IS_GEN_RANGE(dev_priv, 5, 6) ||
> - IS_IVYBRIDGE(dev_priv)) &&
> + (IS_GEN_RANGE(i915, 5, 6) ||
> + IS_IVYBRIDGE(i915)) &&
> (turn_on || (!needs_scaling(old_plane_state) &&
> needs_scaling(to_intel_plane_state(plane_state)))))
> pipe_config->disable_lp_wm = true;
> @@ -11519,13 +11519,13 @@ static int icl_add_linked_planes(struct intel_atomic_state *state)
> static int icl_check_nv12_planes(struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> struct intel_atomic_state *state = to_intel_atomic_state(crtc_state->base.state);
> struct intel_plane *plane, *linked;
> struct intel_plane_state *plane_state;
> int i;
>
> - if (INTEL_GEN(dev_priv) < 11)
> + if (INTEL_GEN(i915) < 11)
> return 0;
>
> /*
> @@ -11555,7 +11555,7 @@ static int icl_check_nv12_planes(struct intel_crtc_state *crtc_state)
> !(crtc_state->nv12_planes & BIT(plane->id)))
> continue;
>
> - for_each_intel_plane_on_crtc(&dev_priv->drm, crtc, linked) {
> + for_each_intel_plane_on_crtc(&i915->drm, crtc, linked) {
> if (!icl_is_nv12_y_plane(linked->id))
> continue;
>
> @@ -11602,21 +11602,21 @@ static bool c8_planes_changed(const struct intel_crtc_state *new_crtc_state)
> static int intel_crtc_atomic_check(struct drm_crtc *crtc,
> struct drm_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc->dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> struct intel_crtc_state *pipe_config =
> to_intel_crtc_state(crtc_state);
> int ret;
> bool mode_changed = needs_modeset(crtc_state);
>
> - if (INTEL_GEN(dev_priv) < 5 && !IS_G4X(dev_priv) &&
> + if (INTEL_GEN(i915) < 5 && !IS_G4X(i915) &&
> mode_changed && !crtc_state->active)
> pipe_config->update_wm_post = true;
>
> if (mode_changed && crtc_state->enable &&
> - dev_priv->display.crtc_compute_clock &&
> + i915->display.crtc_compute_clock &&
> !WARN_ON(pipe_config->shared_dpll)) {
> - ret = dev_priv->display.crtc_compute_clock(intel_crtc,
> + ret = i915->display.crtc_compute_clock(intel_crtc,
> pipe_config);
> if (ret)
> return ret;
> @@ -11637,16 +11637,16 @@ static int intel_crtc_atomic_check(struct drm_crtc *crtc,
> }
>
> ret = 0;
> - if (dev_priv->display.compute_pipe_wm) {
> - ret = dev_priv->display.compute_pipe_wm(pipe_config);
> + if (i915->display.compute_pipe_wm) {
> + ret = i915->display.compute_pipe_wm(pipe_config);
> if (ret) {
> DRM_DEBUG_KMS("Target pipe watermarks are invalid\n");
> return ret;
> }
> }
>
> - if (dev_priv->display.compute_intermediate_wm) {
> - if (WARN_ON(!dev_priv->display.compute_pipe_wm))
> + if (i915->display.compute_intermediate_wm) {
> + if (WARN_ON(!i915->display.compute_pipe_wm))
> return 0;
>
> /*
> @@ -11654,14 +11654,14 @@ static int intel_crtc_atomic_check(struct drm_crtc *crtc,
> * old state and the new state. We can program these
> * immediately.
> */
> - ret = dev_priv->display.compute_intermediate_wm(pipe_config);
> + ret = i915->display.compute_intermediate_wm(pipe_config);
> if (ret) {
> DRM_DEBUG_KMS("No valid intermediate pipe watermarks are possible\n");
> return ret;
> }
> }
>
> - if (INTEL_GEN(dev_priv) >= 9) {
> + if (INTEL_GEN(i915) >= 9) {
> if (mode_changed || pipe_config->update_pipe)
> ret = skl_update_scaler_crtc(pipe_config);
>
> @@ -11671,11 +11671,11 @@ static int intel_crtc_atomic_check(struct drm_crtc *crtc,
> ret = skl_check_pipe_max_pixel_rate(intel_crtc,
> pipe_config);
> if (!ret)
> - ret = intel_atomic_setup_scalers(dev_priv, intel_crtc,
> + ret = intel_atomic_setup_scalers(i915, intel_crtc,
> pipe_config);
> }
>
> - if (HAS_IPS(dev_priv))
> + if (HAS_IPS(i915))
> pipe_config->ips_enabled = hsw_compute_ips_config(pipe_config);
>
> return ret;
> @@ -11752,16 +11752,16 @@ static int
> compute_baseline_pipe_bpp(struct intel_crtc *crtc,
> struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> struct drm_atomic_state *state = pipe_config->base.state;
> struct drm_connector *connector;
> struct drm_connector_state *connector_state;
> int bpp, i;
>
> - if ((IS_G4X(dev_priv) || IS_VALLEYVIEW(dev_priv) ||
> - IS_CHERRYVIEW(dev_priv)))
> + if ((IS_G4X(i915) || IS_VALLEYVIEW(i915) ||
> + IS_CHERRYVIEW(i915)))
> bpp = 10*3;
> - else if (INTEL_GEN(dev_priv) >= 5)
> + else if (INTEL_GEN(i915) >= 5)
> bpp = 12*3;
> else
> bpp = 8*3;
> @@ -11807,13 +11807,13 @@ intel_dump_m_n_config(const struct intel_crtc_state *pipe_config,
> }
>
> static void
> -intel_dump_infoframe(struct drm_i915_private *dev_priv,
> +intel_dump_infoframe(struct drm_i915_private *i915,
> const union hdmi_infoframe *frame)
> {
> if ((drm_debug & DRM_UT_KMS) == 0)
> return;
>
> - hdmi_infoframe_log(KERN_DEBUG, dev_priv->drm.dev, frame);
> + hdmi_infoframe_log(KERN_DEBUG, i915->drm.dev, frame);
> }
>
> #define OUTPUT_TYPE(x) [INTEL_OUTPUT_ ## x] = #x
> @@ -11907,7 +11907,7 @@ static void intel_dump_pipe_config(const struct intel_crtc_state *pipe_config,
> const char *context)
> {
> struct intel_crtc *crtc = to_intel_crtc(pipe_config->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> const struct intel_plane_state *plane_state;
> struct intel_plane *plane;
> char buf[64];
> @@ -11953,13 +11953,13 @@ static void intel_dump_pipe_config(const struct intel_crtc_state *pipe_config,
> DRM_DEBUG_KMS("GCP: 0x%x\n", pipe_config->infoframes.gcp);
> if (pipe_config->infoframes.enable &
> intel_hdmi_infoframe_enable(HDMI_INFOFRAME_TYPE_AVI))
> - intel_dump_infoframe(dev_priv, &pipe_config->infoframes.avi);
> + intel_dump_infoframe(i915, &pipe_config->infoframes.avi);
> if (pipe_config->infoframes.enable &
> intel_hdmi_infoframe_enable(HDMI_INFOFRAME_TYPE_SPD))
> - intel_dump_infoframe(dev_priv, &pipe_config->infoframes.spd);
> + intel_dump_infoframe(i915, &pipe_config->infoframes.spd);
> if (pipe_config->infoframes.enable &
> intel_hdmi_infoframe_enable(HDMI_INFOFRAME_TYPE_VENDOR))
> - intel_dump_infoframe(dev_priv, &pipe_config->infoframes.hdmi);
> + intel_dump_infoframe(i915, &pipe_config->infoframes.hdmi);
>
> DRM_DEBUG_KMS("requested mode:\n");
> drm_mode_debug_printmodeline(&pipe_config->base.mode);
> @@ -11971,13 +11971,13 @@ static void intel_dump_pipe_config(const struct intel_crtc_state *pipe_config,
> pipe_config->pipe_src_w, pipe_config->pipe_src_h,
> pipe_config->pixel_rate);
>
> - if (INTEL_GEN(dev_priv) >= 9)
> + if (INTEL_GEN(i915) >= 9)
> DRM_DEBUG_KMS("num_scalers: %d, scaler_users: 0x%x, scaler_id: %d\n",
> crtc->num_scalers,
> pipe_config->scaler_state.scaler_users,
> pipe_config->scaler_state.scaler_id);
>
> - if (HAS_GMCH(dev_priv))
> + if (HAS_GMCH(i915))
> DRM_DEBUG_KMS("gmch pfit: control: 0x%08x, ratios: 0x%08x, lvds border: 0x%08x\n",
> pipe_config->gmch_pfit.control,
> pipe_config->gmch_pfit.pgm_ratios,
> @@ -11992,7 +11992,7 @@ static void intel_dump_pipe_config(const struct intel_crtc_state *pipe_config,
> DRM_DEBUG_KMS("ips: %i, double wide: %i\n",
> pipe_config->ips_enabled, pipe_config->double_wide);
>
> - intel_dpll_dump_hw_state(dev_priv, &pipe_config->dpll_hw_state);
> + intel_dpll_dump_hw_state(i915, &pipe_config->dpll_hw_state);
>
> dump_planes:
> if (!state)
> @@ -12073,7 +12073,7 @@ static bool check_digital_port_conflicts(struct intel_atomic_state *state)
> static int
> clear_intel_crtc_state(struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> to_i915(crtc_state->base.crtc->dev);
> struct intel_crtc_state *saved_state;
>
> @@ -12090,8 +12090,8 @@ clear_intel_crtc_state(struct intel_crtc_state *crtc_state)
> saved_state->shared_dpll = crtc_state->shared_dpll;
> saved_state->dpll_hw_state = crtc_state->dpll_hw_state;
> saved_state->crc_enabled = crtc_state->crc_enabled;
> - if (IS_G4X(dev_priv) ||
> - IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> + if (IS_G4X(i915) ||
> + IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> saved_state->wm = crtc_state->wm;
>
> /* Keep base drm_crtc_state intact, only clear our extended struct */
> @@ -12316,7 +12316,7 @@ intel_compare_infoframe(const union hdmi_infoframe *a,
> }
>
> static void
> -pipe_config_infoframe_err(struct drm_i915_private *dev_priv,
> +pipe_config_infoframe_err(struct drm_i915_private *i915,
> bool adjust, const char *name,
> const union hdmi_infoframe *a,
> const union hdmi_infoframe *b)
> @@ -12327,15 +12327,15 @@ pipe_config_infoframe_err(struct drm_i915_private *dev_priv,
>
> drm_dbg(DRM_UT_KMS, "mismatch in %s infoframe", name);
> drm_dbg(DRM_UT_KMS, "expected:");
> - hdmi_infoframe_log(KERN_DEBUG, dev_priv->drm.dev, a);
> + hdmi_infoframe_log(KERN_DEBUG, i915->drm.dev, a);
> drm_dbg(DRM_UT_KMS, "found");
> - hdmi_infoframe_log(KERN_DEBUG, dev_priv->drm.dev, b);
> + hdmi_infoframe_log(KERN_DEBUG, i915->drm.dev, b);
> } else {
> drm_err("mismatch in %s infoframe", name);
> drm_err("expected:");
> - hdmi_infoframe_log(KERN_ERR, dev_priv->drm.dev, a);
> + hdmi_infoframe_log(KERN_ERR, i915->drm.dev, a);
> drm_err("found");
> - hdmi_infoframe_log(KERN_ERR, dev_priv->drm.dev, b);
> + hdmi_infoframe_log(KERN_ERR, i915->drm.dev, b);
> }
> }
>
> @@ -12357,17 +12357,17 @@ pipe_config_err(bool adjust, const char *name, const char *format, ...)
> va_end(args);
> }
>
> -static bool fastboot_enabled(struct drm_i915_private *dev_priv)
> +static bool fastboot_enabled(struct drm_i915_private *i915)
> {
> if (i915_modparams.fastboot != -1)
> return i915_modparams.fastboot;
>
> /* Enable fastboot by default on Skylake and newer */
> - if (INTEL_GEN(dev_priv) >= 9)
> + if (INTEL_GEN(i915) >= 9)
> return true;
>
> /* Enable fastboot by default on VLV and CHV */
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> return true;
>
> /* Disabled by default on all others */
> @@ -12375,7 +12375,7 @@ static bool fastboot_enabled(struct drm_i915_private *dev_priv)
> }
>
> static bool
> -intel_pipe_config_compare(struct drm_i915_private *dev_priv,
> +intel_pipe_config_compare(struct drm_i915_private *i915,
> struct intel_crtc_state *current_config,
> struct intel_crtc_state *pipe_config,
> bool adjust)
> @@ -12385,7 +12385,7 @@ intel_pipe_config_compare(struct drm_i915_private *dev_priv,
> (current_config->base.mode.private_flags & I915_MODE_FLAG_INHERITED) &&
> !(pipe_config->base.mode.private_flags & I915_MODE_FLAG_INHERITED);
>
> - if (fixup_inherited && !fastboot_enabled(dev_priv)) {
> + if (fixup_inherited && !fastboot_enabled(i915)) {
> DRM_DEBUG_KMS("initial modeset and fastboot not set\n");
> ret = false;
> }
> @@ -12525,7 +12525,7 @@ intel_pipe_config_compare(struct drm_i915_private *dev_priv,
> #define PIPE_CONF_CHECK_INFOFRAME(name) do { \
> if (!intel_compare_infoframe(¤t_config->infoframes.name, \
> &pipe_config->infoframes.name)) { \
> - pipe_config_infoframe_err(dev_priv, adjust, __stringify(name), \
> + pipe_config_infoframe_err(i915, adjust, __stringify(name), \
> ¤t_config->infoframes.name, \
> &pipe_config->infoframes.name); \
> ret = false; \
> @@ -12544,7 +12544,7 @@ intel_pipe_config_compare(struct drm_i915_private *dev_priv,
> PIPE_CONF_CHECK_I(lane_count);
> PIPE_CONF_CHECK_X(lane_lat_optim_mask);
>
> - if (INTEL_GEN(dev_priv) < 8) {
> + if (INTEL_GEN(i915) < 8) {
> PIPE_CONF_CHECK_M_N(dp_m_n);
>
> if (current_config->has_drrs)
> @@ -12571,8 +12571,8 @@ intel_pipe_config_compare(struct drm_i915_private *dev_priv,
> PIPE_CONF_CHECK_I(pixel_multiplier);
> PIPE_CONF_CHECK_I(output_format);
> PIPE_CONF_CHECK_BOOL(has_hdmi_sink);
> - if ((INTEL_GEN(dev_priv) < 8 && !IS_HASWELL(dev_priv)) ||
> - IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> + if ((INTEL_GEN(i915) < 8 && !IS_HASWELL(i915)) ||
> + IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> PIPE_CONF_CHECK_BOOL(limited_color_range);
>
> PIPE_CONF_CHECK_BOOL(hdmi_scrambling);
> @@ -12597,7 +12597,7 @@ intel_pipe_config_compare(struct drm_i915_private *dev_priv,
>
> PIPE_CONF_CHECK_X(gmch_pfit.control);
> /* pfit ratios are autocomputed by the hw on gen4+ */
> - if (INTEL_GEN(dev_priv) < 4)
> + if (INTEL_GEN(i915) < 4)
> PIPE_CONF_CHECK_X(gmch_pfit.pgm_ratios);
> PIPE_CONF_CHECK_X(gmch_pfit.lvds_border_bits);
>
> @@ -12621,7 +12621,7 @@ intel_pipe_config_compare(struct drm_i915_private *dev_priv,
> PIPE_CONF_CHECK_CLOCK_FUZZY(pixel_rate);
>
> PIPE_CONF_CHECK_X(gamma_mode);
> - if (IS_CHERRYVIEW(dev_priv))
> + if (IS_CHERRYVIEW(i915))
> PIPE_CONF_CHECK_X(cgm_mode);
> else
> PIPE_CONF_CHECK_X(csc_mode);
> @@ -12667,7 +12667,7 @@ intel_pipe_config_compare(struct drm_i915_private *dev_priv,
> PIPE_CONF_CHECK_X(dsi_pll.ctrl);
> PIPE_CONF_CHECK_X(dsi_pll.div);
>
> - if (IS_G4X(dev_priv) || INTEL_GEN(dev_priv) >= 5)
> + if (IS_G4X(i915) || INTEL_GEN(i915) >= 5)
> PIPE_CONF_CHECK_I(pipe_bpp);
>
> PIPE_CONF_CHECK_CLOCK_FUZZY(base.adjusted_mode.crtc_clock);
> @@ -12694,11 +12694,11 @@ intel_pipe_config_compare(struct drm_i915_private *dev_priv,
> return ret;
> }
>
> -static void intel_pipe_config_sanity_check(struct drm_i915_private *dev_priv,
> +static void intel_pipe_config_sanity_check(struct drm_i915_private *i915,
> const struct intel_crtc_state *pipe_config)
> {
> if (pipe_config->has_pch_encoder) {
> - int fdi_dotclock = intel_dotclock_calculate(intel_fdi_link_freq(dev_priv, pipe_config),
> + int fdi_dotclock = intel_dotclock_calculate(intel_fdi_link_freq(i915, pipe_config),
> &pipe_config->fdi_m_n);
> int dotclock = pipe_config->base.adjusted_mode.crtc_clock;
>
> @@ -12715,7 +12715,7 @@ static void intel_pipe_config_sanity_check(struct drm_i915_private *dev_priv,
> static void verify_wm_state(struct drm_crtc *crtc,
> struct drm_crtc_state *new_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc->dev);
> struct skl_hw_state {
> struct skl_ddb_entry ddb_y[I915_MAX_PLANES];
> struct skl_ddb_entry ddb_uv[I915_MAX_PLANES];
> @@ -12727,9 +12727,9 @@ static void verify_wm_state(struct drm_crtc *crtc,
> struct skl_ddb_entry *hw_ddb_entry, *sw_ddb_entry;
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> const enum pipe pipe = intel_crtc->pipe;
> - int plane, level, max_level = ilk_wm_max_level(dev_priv);
> + int plane, level, max_level = ilk_wm_max_level(i915);
>
> - if (INTEL_GEN(dev_priv) < 9 || !new_state->active)
> + if (INTEL_GEN(i915) < 9 || !new_state->active)
> return;
>
> hw = kzalloc(sizeof(*hw), GFP_KERNEL);
> @@ -12741,17 +12741,17 @@ static void verify_wm_state(struct drm_crtc *crtc,
>
> skl_pipe_ddb_get_hw_state(intel_crtc, hw->ddb_y, hw->ddb_uv);
>
> - skl_ddb_get_hw_state(dev_priv, &hw->ddb);
> - sw_ddb = &dev_priv->wm.skl_hw.ddb;
> + skl_ddb_get_hw_state(i915, &hw->ddb);
> + sw_ddb = &i915->wm.skl_hw.ddb;
>
> - if (INTEL_GEN(dev_priv) >= 11 &&
> + if (INTEL_GEN(i915) >= 11 &&
> hw->ddb.enabled_slices != sw_ddb->enabled_slices)
> DRM_ERROR("mismatch in DBUF Slices (expected %u, got %u)\n",
> sw_ddb->enabled_slices,
> hw->ddb.enabled_slices);
>
> /* planes */
> - for_each_universal_plane(dev_priv, pipe, plane) {
> + for_each_universal_plane(i915, pipe, plane) {
> struct skl_plane_wm *hw_plane_wm, *sw_plane_wm;
>
> hw_plane_wm = &hw->wm.planes[plane];
> @@ -12933,7 +12933,7 @@ verify_crtc_state(struct drm_crtc *crtc,
> struct drm_crtc_state *new_crtc_state)
> {
> struct drm_device *dev = crtc->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_encoder *encoder;
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> struct intel_crtc_state *pipe_config, *sw_config;
> @@ -12949,10 +12949,10 @@ verify_crtc_state(struct drm_crtc *crtc,
>
> DRM_DEBUG_KMS("[CRTC:%d:%s]\n", crtc->base.id, crtc->name);
>
> - active = dev_priv->display.get_pipe_config(intel_crtc, pipe_config);
> + active = i915->display.get_pipe_config(intel_crtc, pipe_config);
>
> /* we keep both pipes enabled on 830 */
> - if (IS_I830(dev_priv))
> + if (IS_I830(i915))
> active = new_crtc_state->active;
>
> I915_STATE_WARN(new_crtc_state->active != active,
> @@ -12984,10 +12984,10 @@ verify_crtc_state(struct drm_crtc *crtc,
> if (!new_crtc_state->active)
> return;
>
> - intel_pipe_config_sanity_check(dev_priv, pipe_config);
> + intel_pipe_config_sanity_check(i915, pipe_config);
>
> sw_config = to_intel_crtc_state(new_crtc_state);
> - if (!intel_pipe_config_compare(dev_priv, sw_config,
> + if (!intel_pipe_config_compare(i915, sw_config,
> pipe_config, false)) {
> I915_STATE_WARN(1, "pipe state doesn't match!\n");
> intel_dump_pipe_config(pipe_config, NULL, "[hw state]");
> @@ -13009,7 +13009,7 @@ intel_verify_planes(struct intel_atomic_state *state)
> }
>
> static void
> -verify_single_dpll_state(struct drm_i915_private *dev_priv,
> +verify_single_dpll_state(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll,
> struct drm_crtc *crtc,
> struct drm_crtc_state *new_state)
> @@ -13022,7 +13022,7 @@ verify_single_dpll_state(struct drm_i915_private *dev_priv,
>
> DRM_DEBUG_KMS("%s\n", pll->info->name);
>
> - active = pll->info->funcs->get_hw_state(dev_priv, pll, &dpll_hw_state);
> + active = pll->info->funcs->get_hw_state(i915, pll, &dpll_hw_state);
>
> if (!(pll->info->flags & INTEL_DPLL_ALWAYS_ON)) {
> I915_STATE_WARN(!pll->on && pll->active_mask,
> @@ -13068,12 +13068,12 @@ verify_shared_dpll_state(struct drm_device *dev, struct drm_crtc *crtc,
> struct drm_crtc_state *old_crtc_state,
> struct drm_crtc_state *new_crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_crtc_state *old_state = to_intel_crtc_state(old_crtc_state);
> struct intel_crtc_state *new_state = to_intel_crtc_state(new_crtc_state);
>
> if (new_state->shared_dpll)
> - verify_single_dpll_state(dev_priv, new_state->shared_dpll, crtc, new_crtc_state);
> + verify_single_dpll_state(i915, new_state->shared_dpll, crtc, new_crtc_state);
>
> if (old_state->shared_dpll &&
> old_state->shared_dpll != new_state->shared_dpll) {
> @@ -13108,11 +13108,11 @@ intel_modeset_verify_crtc(struct drm_crtc *crtc,
> static void
> verify_disabled_dpll_state(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> int i;
>
> - for (i = 0; i < dev_priv->num_shared_dpll; i++)
> - verify_single_dpll_state(dev_priv, &dev_priv->shared_dplls[i], NULL, NULL);
> + for (i = 0; i < i915->num_shared_dpll; i++)
> + verify_single_dpll_state(i915, &i915->shared_dplls[i], NULL, NULL);
> }
>
> static void
> @@ -13127,7 +13127,7 @@ intel_modeset_verify_disabled(struct drm_device *dev,
> static void update_scanline_offset(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
>
> /*
> * The scanline counter increments at the leading edge of hsync.
> @@ -13156,7 +13156,7 @@ static void update_scanline_offset(const struct intel_crtc_state *crtc_state)
> * However if queried just before the start of vblank we'll get an
> * answer that's slightly in the future.
> */
> - if (IS_GEN(dev_priv, 2)) {
> + if (IS_GEN(i915, 2)) {
> const struct drm_display_mode *adjusted_mode = &crtc_state->base.adjusted_mode;
> int vtotal;
>
> @@ -13165,7 +13165,7 @@ static void update_scanline_offset(const struct intel_crtc_state *crtc_state)
> vtotal /= 2;
>
> crtc->scanline_offset = vtotal - 1;
> - } else if (HAS_DDI(dev_priv) &&
> + } else if (HAS_DDI(i915) &&
> intel_crtc_has_type(crtc_state, INTEL_OUTPUT_HDMI)) {
> crtc->scanline_offset = 2;
> } else
> @@ -13174,12 +13174,12 @@ static void update_scanline_offset(const struct intel_crtc_state *crtc_state)
>
> static void intel_modeset_clear_plls(struct intel_atomic_state *state)
> {
> - struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> + struct drm_i915_private *i915 = to_i915(state->base.dev);
> struct intel_crtc_state *old_crtc_state, *new_crtc_state;
> struct intel_crtc *crtc;
> int i;
>
> - if (!dev_priv->display.crtc_compute_clock)
> + if (!i915->display.crtc_compute_clock)
> return;
>
> for_each_oldnew_intel_crtc_in_state(state, crtc, old_crtc_state,
> @@ -13311,7 +13311,7 @@ static int intel_modeset_all_pipes(struct drm_atomic_state *state)
>
> static int intel_modeset_checks(struct intel_atomic_state *state)
> {
> - struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> + struct drm_i915_private *i915 = to_i915(state->base.dev);
> struct intel_crtc_state *old_crtc_state, *new_crtc_state;
> struct intel_crtc *crtc;
> int ret = 0, i;
> @@ -13323,12 +13323,12 @@ static int intel_modeset_checks(struct intel_atomic_state *state)
>
> /* keep the current setting */
> if (!state->cdclk.force_min_cdclk_changed)
> - state->cdclk.force_min_cdclk = dev_priv->cdclk.force_min_cdclk;
> + state->cdclk.force_min_cdclk = i915->cdclk.force_min_cdclk;
>
> state->modeset = true;
> - state->active_crtcs = dev_priv->active_crtcs;
> - state->cdclk.logical = dev_priv->cdclk.logical;
> - state->cdclk.actual = dev_priv->cdclk.actual;
> + state->active_crtcs = i915->active_crtcs;
> + state->cdclk.logical = i915->cdclk.logical;
> + state->cdclk.actual = i915->cdclk.actual;
> state->cdclk.pipe = INVALID_PIPE;
>
> for_each_oldnew_intel_crtc_in_state(state, crtc, old_crtc_state,
> @@ -13349,19 +13349,19 @@ static int intel_modeset_checks(struct intel_atomic_state *state)
> * mode set on this crtc. For other crtcs we need to use the
> * adjusted_mode bits in the crtc directly.
> */
> - if (dev_priv->display.modeset_calc_cdclk) {
> + if (i915->display.modeset_calc_cdclk) {
> enum pipe pipe;
>
> - ret = dev_priv->display.modeset_calc_cdclk(state);
> + ret = i915->display.modeset_calc_cdclk(state);
> if (ret < 0)
> return ret;
>
> /*
> - * Writes to dev_priv->cdclk.logical must protected by
> + * Writes to i915->cdclk.logical must protected by
> * holding all the crtc locks, even if we don't end up
> * touching the hardware
> */
> - if (intel_cdclk_changed(&dev_priv->cdclk.logical,
> + if (intel_cdclk_changed(&i915->cdclk.logical,
> &state->cdclk.logical)) {
> ret = intel_lock_all_pipes(&state->base);
> if (ret < 0)
> @@ -13373,7 +13373,7 @@ static int intel_modeset_checks(struct intel_atomic_state *state)
> struct drm_crtc_state *crtc_state;
>
> pipe = ilog2(state->active_crtcs);
> - crtc = &intel_get_crtc_for_pipe(dev_priv, pipe)->base;
> + crtc = &intel_get_crtc_for_pipe(i915, pipe)->base;
> crtc_state = drm_atomic_get_new_crtc_state(&state->base, crtc);
> if (crtc_state && needs_modeset(crtc_state))
> pipe = INVALID_PIPE;
> @@ -13383,15 +13383,15 @@ static int intel_modeset_checks(struct intel_atomic_state *state)
>
> /* All pipes must be switched off while we change the cdclk. */
> if (pipe != INVALID_PIPE &&
> - intel_cdclk_needs_cd2x_update(dev_priv,
> - &dev_priv->cdclk.actual,
> + intel_cdclk_needs_cd2x_update(i915,
> + &i915->cdclk.actual,
> &state->cdclk.actual)) {
> ret = intel_lock_all_pipes(&state->base);
> if (ret < 0)
> return ret;
>
> state->cdclk.pipe = pipe;
> - } else if (intel_cdclk_needs_modeset(&dev_priv->cdclk.actual,
> + } else if (intel_cdclk_needs_modeset(&i915->cdclk.actual,
> &state->cdclk.actual)) {
> ret = intel_modeset_all_pipes(&state->base);
> if (ret < 0)
> @@ -13410,7 +13410,7 @@ static int intel_modeset_checks(struct intel_atomic_state *state)
>
> intel_modeset_clear_plls(state);
>
> - if (IS_HASWELL(dev_priv))
> + if (IS_HASWELL(i915))
> return haswell_mode_set_planes_workaround(state);
>
> return 0;
> @@ -13424,11 +13424,11 @@ static int intel_modeset_checks(struct intel_atomic_state *state)
> static int calc_watermark_data(struct intel_atomic_state *state)
> {
> struct drm_device *dev = state->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
>
> /* Is there platform-specific watermark information to calculate? */
> - if (dev_priv->display.compute_global_watermarks)
> - return dev_priv->display.compute_global_watermarks(state);
> + if (i915->display.compute_global_watermarks)
> + return i915->display.compute_global_watermarks(state);
>
> return 0;
> }
> @@ -13441,7 +13441,7 @@ static int calc_watermark_data(struct intel_atomic_state *state)
> static int intel_atomic_check(struct drm_device *dev,
> struct drm_atomic_state *_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_atomic_state *state = to_intel_atomic_state(_state);
> struct intel_crtc_state *old_crtc_state, *new_crtc_state;
> struct intel_crtc *crtc;
> @@ -13474,7 +13474,7 @@ static int intel_atomic_check(struct drm_device *dev,
> if (ret)
> goto fail;
>
> - if (intel_pipe_config_compare(dev_priv, old_crtc_state,
> + if (intel_pipe_config_compare(i915, old_crtc_state,
> new_crtc_state, true)) {
> new_crtc_state->base.mode_changed = false;
> new_crtc_state->update_pipe = true;
> @@ -13493,7 +13493,7 @@ static int intel_atomic_check(struct drm_device *dev,
> if (ret)
> goto fail;
> } else {
> - state->cdclk.logical = dev_priv->cdclk.logical;
> + state->cdclk.logical = i915->cdclk.logical;
> }
>
> ret = icl_add_linked_planes(state);
> @@ -13504,7 +13504,7 @@ static int intel_atomic_check(struct drm_device *dev,
> if (ret)
> goto fail;
>
> - intel_fbc_choose_crtc(dev_priv, state);
> + intel_fbc_choose_crtc(i915, state);
> ret = calc_watermark_data(state);
> if (ret)
> goto fail;
> @@ -13564,7 +13564,7 @@ static void intel_update_crtc(struct drm_crtc *crtc,
> struct drm_crtc_state *new_crtc_state)
> {
> struct drm_device *dev = crtc->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> struct intel_crtc_state *pipe_config = to_intel_crtc_state(new_crtc_state);
> bool modeset = needs_modeset(new_crtc_state);
> @@ -13574,7 +13574,7 @@ static void intel_update_crtc(struct drm_crtc *crtc,
>
> if (modeset) {
> update_scanline_offset(pipe_config);
> - dev_priv->display.crtc_enable(pipe_config, state);
> + i915->display.crtc_enable(pipe_config, state);
>
> /* vblanks work again, re-enable pipe CRC. */
> intel_crtc_enable_pipe_crc(intel_crtc);
> @@ -13593,7 +13593,7 @@ static void intel_update_crtc(struct drm_crtc *crtc,
>
> intel_begin_crtc_commit(to_intel_atomic_state(state), intel_crtc);
>
> - if (INTEL_GEN(dev_priv) >= 9)
> + if (INTEL_GEN(i915) >= 9)
> skl_update_planes_on_crtc(to_intel_atomic_state(state), intel_crtc);
> else
> i9xx_update_planes_on_crtc(to_intel_atomic_state(state), intel_crtc);
> @@ -13618,7 +13618,7 @@ static void intel_update_crtcs(struct drm_atomic_state *state)
>
> static void skl_update_crtcs(struct drm_atomic_state *state)
> {
> - struct drm_i915_private *dev_priv = to_i915(state->dev);
> + struct drm_i915_private *i915 = to_i915(state->dev);
> struct intel_atomic_state *intel_state = to_intel_atomic_state(state);
> struct drm_crtc *crtc;
> struct intel_crtc *intel_crtc;
> @@ -13628,7 +13628,7 @@ static void skl_update_crtcs(struct drm_atomic_state *state)
> bool progress;
> enum pipe pipe;
> int i;
> - u8 hw_enabled_slices = dev_priv->wm.skl_hw.ddb.enabled_slices;
> + u8 hw_enabled_slices = i915->wm.skl_hw.ddb.enabled_slices;
> u8 required_slices = intel_state->wm_results.ddb.enabled_slices;
> struct skl_ddb_entry entries[I915_MAX_PIPES] = {};
>
> @@ -13638,8 +13638,8 @@ static void skl_update_crtcs(struct drm_atomic_state *state)
> entries[i] = to_intel_crtc_state(old_crtc_state)->wm.skl.ddb;
>
> /* If 2nd DBuf slice required, enable it here */
> - if (INTEL_GEN(dev_priv) >= 11 && required_slices > hw_enabled_slices)
> - icl_dbuf_slices_update(dev_priv, required_slices);
> + if (INTEL_GEN(i915) >= 11 && required_slices > hw_enabled_slices)
> + icl_dbuf_slices_update(i915, required_slices);
>
> /*
> * Whenever the number of active pipes changes, we need to make sure we
> @@ -13663,7 +13663,7 @@ static void skl_update_crtcs(struct drm_atomic_state *state)
>
> if (skl_ddb_allocation_overlaps(&cstate->wm.skl.ddb,
> entries,
> - INTEL_INFO(dev_priv)->num_pipes, i))
> + INTEL_INFO(i915)->num_pipes, i))
> continue;
>
> updated |= cmask;
> @@ -13685,57 +13685,57 @@ static void skl_update_crtcs(struct drm_atomic_state *state)
> new_crtc_state);
>
> if (vbl_wait)
> - intel_wait_for_vblank(dev_priv, pipe);
> + intel_wait_for_vblank(i915, pipe);
>
> progress = true;
> }
> } while (progress);
>
> /* If 2nd DBuf slice is no more required disable it */
> - if (INTEL_GEN(dev_priv) >= 11 && required_slices < hw_enabled_slices)
> - icl_dbuf_slices_update(dev_priv, required_slices);
> + if (INTEL_GEN(i915) >= 11 && required_slices < hw_enabled_slices)
> + icl_dbuf_slices_update(i915, required_slices);
> }
>
> -static void intel_atomic_helper_free_state(struct drm_i915_private *dev_priv)
> +static void intel_atomic_helper_free_state(struct drm_i915_private *i915)
> {
> struct intel_atomic_state *state, *next;
> struct llist_node *freed;
>
> - freed = llist_del_all(&dev_priv->atomic_helper.free_list);
> + freed = llist_del_all(&i915->atomic_helper.free_list);
> llist_for_each_entry_safe(state, next, freed, freed)
> drm_atomic_state_put(&state->base);
> }
>
> static void intel_atomic_helper_free_state_worker(struct work_struct *work)
> {
> - struct drm_i915_private *dev_priv =
> - container_of(work, typeof(*dev_priv), atomic_helper.free_work);
> + struct drm_i915_private *i915 =
> + container_of(work, typeof(*i915), atomic_helper.free_work);
>
> - intel_atomic_helper_free_state(dev_priv);
> + intel_atomic_helper_free_state(i915);
> }
>
> static void intel_atomic_commit_fence_wait(struct intel_atomic_state *intel_state)
> {
> struct wait_queue_entry wait_fence, wait_reset;
> - struct drm_i915_private *dev_priv = to_i915(intel_state->base.dev);
> + struct drm_i915_private *i915 = to_i915(intel_state->base.dev);
>
> init_wait_entry(&wait_fence, 0);
> init_wait_entry(&wait_reset, 0);
> for (;;) {
> prepare_to_wait(&intel_state->commit_ready.wait,
> &wait_fence, TASK_UNINTERRUPTIBLE);
> - prepare_to_wait(&dev_priv->gpu_error.wait_queue,
> + prepare_to_wait(&i915->gpu_error.wait_queue,
> &wait_reset, TASK_UNINTERRUPTIBLE);
>
>
> if (i915_sw_fence_done(&intel_state->commit_ready)
> - || test_bit(I915_RESET_MODESET, &dev_priv->gpu_error.flags))
> + || test_bit(I915_RESET_MODESET, &i915->gpu_error.flags))
> break;
>
> schedule();
> }
> finish_wait(&intel_state->commit_ready.wait, &wait_fence);
> - finish_wait(&dev_priv->gpu_error.wait_queue, &wait_reset);
> + finish_wait(&i915->gpu_error.wait_queue, &wait_reset);
> }
>
> static void intel_atomic_cleanup_work(struct work_struct *work)
> @@ -13755,7 +13755,7 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state)
> {
> struct drm_device *dev = state->dev;
> struct intel_atomic_state *intel_state = to_intel_atomic_state(state);
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct drm_crtc_state *old_crtc_state, *new_crtc_state;
> struct intel_crtc_state *new_intel_crtc_state, *old_intel_crtc_state;
> struct drm_crtc *crtc;
> @@ -13769,7 +13769,7 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state)
> drm_atomic_helper_wait_for_dependencies(state);
>
> if (intel_state->modeset)
> - wakeref = intel_display_power_get(dev_priv, POWER_DOMAIN_MODESET);
> + wakeref = intel_display_power_get(i915, POWER_DOMAIN_MODESET);
>
> for_each_oldnew_crtc_in_state(state, crtc, old_crtc_state, new_crtc_state, i) {
> old_intel_crtc_state = to_intel_crtc_state(old_crtc_state);
> @@ -13798,7 +13798,7 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state)
> */
> intel_crtc_disable_pipe_crc(intel_crtc);
>
> - dev_priv->display.crtc_disable(old_intel_crtc_state, state);
> + i915->display.crtc_disable(old_intel_crtc_state, state);
> intel_crtc->active = false;
> intel_fbc_disable(intel_crtc);
> intel_disable_shared_dpll(old_intel_crtc_state);
> @@ -13807,14 +13807,14 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state)
> * Underruns don't always raise
> * interrupts, so check manually.
> */
> - intel_check_cpu_fifo_underruns(dev_priv);
> - intel_check_pch_fifo_underruns(dev_priv);
> + intel_check_cpu_fifo_underruns(i915);
> + intel_check_pch_fifo_underruns(i915);
>
> /* FIXME unify this for all platforms */
> if (!new_crtc_state->active &&
> - !HAS_GMCH(dev_priv) &&
> - dev_priv->display.initial_watermarks)
> - dev_priv->display.initial_watermarks(intel_state,
> + !HAS_GMCH(i915) &&
> + i915->display.initial_watermarks)
> + i915->display.initial_watermarks(intel_state,
> new_intel_crtc_state);
> }
> }
> @@ -13826,9 +13826,9 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state)
> if (intel_state->modeset) {
> drm_atomic_helper_update_legacy_modeset_state(state->dev, state);
>
> - intel_set_cdclk_pre_plane_update(dev_priv,
> + intel_set_cdclk_pre_plane_update(i915,
> &intel_state->cdclk.actual,
> - &dev_priv->cdclk.actual,
> + &i915->cdclk.actual,
> intel_state->cdclk.pipe);
>
> /*
> @@ -13836,7 +13836,7 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state)
> * have more then one pipe enabled
> */
> if (!intel_can_enable_sagv(state))
> - intel_disable_sagv(dev_priv);
> + intel_disable_sagv(i915);
>
> intel_modeset_verify_disabled(dev, state);
> }
> @@ -13856,12 +13856,12 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state)
> }
>
> /* Now enable the clocks, plane, pipe, and connectors that we set up. */
> - dev_priv->display.update_crtcs(state);
> + i915->display.update_crtcs(state);
>
> if (intel_state->modeset)
> - intel_set_cdclk_post_plane_update(dev_priv,
> + intel_set_cdclk_post_plane_update(i915,
> &intel_state->cdclk.actual,
> - &dev_priv->cdclk.actual,
> + &i915->cdclk.actual,
> intel_state->cdclk.pipe);
>
> /* FIXME: We should call drm_atomic_helper_commit_hw_done() here
> @@ -13895,8 +13895,8 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state)
> for_each_new_crtc_in_state(state, crtc, new_crtc_state, i) {
> new_intel_crtc_state = to_intel_crtc_state(new_crtc_state);
>
> - if (dev_priv->display.optimize_watermarks)
> - dev_priv->display.optimize_watermarks(intel_state,
> + if (i915->display.optimize_watermarks)
> + i915->display.optimize_watermarks(intel_state,
> new_intel_crtc_state);
> }
>
> @@ -13904,7 +13904,7 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state)
> intel_post_plane_update(to_intel_crtc_state(old_crtc_state));
>
> if (put_domains[i])
> - modeset_put_power_domains(dev_priv, put_domains[i]);
> + modeset_put_power_domains(i915, put_domains[i]);
>
> intel_modeset_verify_crtc(crtc, state, old_crtc_state, new_crtc_state);
> }
> @@ -13913,7 +13913,7 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state)
> intel_verify_planes(intel_state);
>
> if (intel_state->modeset && intel_can_enable_sagv(state))
> - intel_enable_sagv(dev_priv);
> + intel_enable_sagv(i915);
>
> drm_atomic_helper_commit_hw_done(state);
>
> @@ -13924,10 +13924,10 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state)
> * so enable debugging for the next modeset - and hope we catch
> * the culprit.
> */
> - intel_uncore_arm_unclaimed_mmio_detection(&dev_priv->uncore);
> - intel_display_power_put(dev_priv, POWER_DOMAIN_MODESET, wakeref);
> + intel_uncore_arm_unclaimed_mmio_detection(&i915->uncore);
> + intel_display_power_put(i915, POWER_DOMAIN_MODESET, wakeref);
> }
> - intel_runtime_pm_put(dev_priv, intel_state->wakeref);
> + intel_runtime_pm_put(i915, intel_state->wakeref);
>
> /*
> * Defer the cleanup of the old state to a separate worker to not
> @@ -14003,10 +14003,10 @@ static int intel_atomic_commit(struct drm_device *dev,
> bool nonblock)
> {
> struct intel_atomic_state *intel_state = to_intel_atomic_state(state);
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> int ret = 0;
>
> - intel_state->wakeref = intel_runtime_pm_get(dev_priv);
> + intel_state->wakeref = intel_runtime_pm_get(i915);
>
> drm_atomic_state_get(state);
> i915_sw_fence_init(&intel_state->commit_ready,
> @@ -14029,7 +14029,7 @@ static int intel_atomic_commit(struct drm_device *dev,
> * FIXME doing watermarks and fb cleanup from a vblank worker
> * (assuming we had any) would solve these problems.
> */
> - if (INTEL_GEN(dev_priv) < 9 && state->legacy_cursor_update) {
> + if (INTEL_GEN(i915) < 9 && state->legacy_cursor_update) {
> struct intel_crtc_state *new_crtc_state;
> struct intel_crtc *crtc;
> int i;
> @@ -14044,7 +14044,7 @@ static int intel_atomic_commit(struct drm_device *dev,
> if (ret) {
> DRM_DEBUG_ATOMIC("Preparing state failed with %i\n", ret);
> i915_sw_fence_commit(&intel_state->commit_ready);
> - intel_runtime_pm_put(dev_priv, intel_state->wakeref);
> + intel_runtime_pm_put(i915, intel_state->wakeref);
> return ret;
> }
>
> @@ -14056,21 +14056,21 @@ static int intel_atomic_commit(struct drm_device *dev,
> i915_sw_fence_commit(&intel_state->commit_ready);
>
> drm_atomic_helper_cleanup_planes(dev, state);
> - intel_runtime_pm_put(dev_priv, intel_state->wakeref);
> + intel_runtime_pm_put(i915, intel_state->wakeref);
> return ret;
> }
> - dev_priv->wm.distrust_bios_wm = false;
> + i915->wm.distrust_bios_wm = false;
> intel_shared_dpll_swap_state(state);
> intel_atomic_track_fbs(state);
>
> if (intel_state->modeset) {
> - memcpy(dev_priv->min_cdclk, intel_state->min_cdclk,
> + memcpy(i915->min_cdclk, intel_state->min_cdclk,
> sizeof(intel_state->min_cdclk));
> - memcpy(dev_priv->min_voltage_level,
> + memcpy(i915->min_voltage_level,
> intel_state->min_voltage_level,
> sizeof(intel_state->min_voltage_level));
> - dev_priv->active_crtcs = intel_state->active_crtcs;
> - dev_priv->cdclk.force_min_cdclk =
> + i915->active_crtcs = intel_state->active_crtcs;
> + i915->cdclk.force_min_cdclk =
> intel_state->cdclk.force_min_cdclk;
>
> intel_cdclk_swap_state(intel_state);
> @@ -14081,12 +14081,12 @@ static int intel_atomic_commit(struct drm_device *dev,
>
> i915_sw_fence_commit(&intel_state->commit_ready);
> if (nonblock && intel_state->modeset) {
> - queue_work(dev_priv->modeset_wq, &state->commit_work);
> + queue_work(i915->modeset_wq, &state->commit_work);
> } else if (nonblock) {
> queue_work(system_unbound_wq, &state->commit_work);
> } else {
> if (intel_state->modeset)
> - flush_workqueue(dev_priv->modeset_wq);
> + flush_workqueue(i915->modeset_wq);
> intel_atomic_commit_tail(state);
> }
>
> @@ -14166,14 +14166,14 @@ static void add_rps_boost_after_vblank(struct drm_crtc *crtc,
> static int intel_plane_pin_fb(struct intel_plane_state *plane_state)
> {
> struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> struct drm_framebuffer *fb = plane_state->base.fb;
> struct i915_vma *vma;
>
> if (plane->id == PLANE_CURSOR &&
> - INTEL_INFO(dev_priv)->display.cursor_needs_physical) {
> + INTEL_INFO(i915)->display.cursor_needs_physical) {
> struct drm_i915_gem_object *obj = intel_fb_obj(fb);
> - const int align = intel_cursor_alignment(dev_priv);
> + const int align = intel_cursor_alignment(i915);
> int err;
>
> err = i915_gem_object_attach_phys(obj, align);
> @@ -14231,7 +14231,7 @@ intel_prepare_plane_fb(struct drm_plane *plane,
> {
> struct intel_atomic_state *intel_state =
> to_intel_atomic_state(new_state->state);
> - struct drm_i915_private *dev_priv = to_i915(plane->dev);
> + struct drm_i915_private *i915 = to_i915(plane->dev);
> struct drm_framebuffer *fb = new_state->fb;
> struct drm_i915_gem_object *obj = intel_fb_obj(fb);
> struct drm_i915_gem_object *old_obj = intel_fb_obj(plane->state->fb);
> @@ -14279,7 +14279,7 @@ intel_prepare_plane_fb(struct drm_plane *plane,
> if (ret)
> return ret;
>
> - ret = mutex_lock_interruptible(&dev_priv->drm.struct_mutex);
> + ret = mutex_lock_interruptible(&i915->drm.struct_mutex);
> if (ret) {
> i915_gem_object_unpin_pages(obj);
> return ret;
> @@ -14287,7 +14287,7 @@ intel_prepare_plane_fb(struct drm_plane *plane,
>
> ret = intel_plane_pin_fb(to_intel_plane_state(new_state));
>
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + mutex_unlock(&i915->drm.struct_mutex);
> i915_gem_object_unpin_pages(obj);
> if (ret)
> return ret;
> @@ -14323,7 +14323,7 @@ intel_prepare_plane_fb(struct drm_plane *plane,
> * maximum clocks following a vblank miss (see do_rps_boost()).
> */
> if (!intel_state->rps_interactive) {
> - intel_rps_mark_interactive(dev_priv, true);
> + intel_rps_mark_interactive(i915, true);
> intel_state->rps_interactive = true;
> }
>
> @@ -14345,17 +14345,17 @@ intel_cleanup_plane_fb(struct drm_plane *plane,
> {
> struct intel_atomic_state *intel_state =
> to_intel_atomic_state(old_state->state);
> - struct drm_i915_private *dev_priv = to_i915(plane->dev);
> + struct drm_i915_private *i915 = to_i915(plane->dev);
>
> if (intel_state->rps_interactive) {
> - intel_rps_mark_interactive(dev_priv, false);
> + intel_rps_mark_interactive(i915, false);
> intel_state->rps_interactive = false;
> }
>
> /* Should only be called after a successful intel_prepare_plane_fb()! */
> - mutex_lock(&dev_priv->drm.struct_mutex);
> + mutex_lock(&i915->drm.struct_mutex);
> intel_plane_unpin_fb(to_intel_plane_state(old_state));
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + mutex_unlock(&i915->drm.struct_mutex);
> }
>
> int
> @@ -14363,7 +14363,7 @@ skl_max_scale(const struct intel_crtc_state *crtc_state,
> u32 pixel_format)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> int max_scale, mult;
> int crtc_clock, max_dotclk, tmpclk1, tmpclk2;
>
> @@ -14373,7 +14373,7 @@ skl_max_scale(const struct intel_crtc_state *crtc_state,
> crtc_clock = crtc_state->base.adjusted_mode.crtc_clock;
> max_dotclk = to_intel_atomic_state(crtc_state->base.state)->cdclk.logical.cdclk;
>
> - if (IS_GEMINILAKE(dev_priv) || INTEL_GEN(dev_priv) >= 10)
> + if (IS_GEMINILAKE(i915) || INTEL_GEN(i915) >= 10)
> max_dotclk *= 2;
>
> if (WARN_ON_ONCE(!crtc_clock || max_dotclk < crtc_clock))
> @@ -14396,7 +14396,7 @@ skl_max_scale(const struct intel_crtc_state *crtc_state,
> static void intel_begin_crtc_commit(struct intel_atomic_state *state,
> struct intel_crtc *crtc)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> struct intel_crtc_state *old_crtc_state =
> intel_atomic_get_old_crtc_state(state, crtc);
> struct intel_crtc_state *new_crtc_state =
> @@ -14415,31 +14415,31 @@ static void intel_begin_crtc_commit(struct intel_atomic_state *state,
>
> if (new_crtc_state->update_pipe)
> intel_update_pipe_config(old_crtc_state, new_crtc_state);
> - else if (INTEL_GEN(dev_priv) >= 9)
> + else if (INTEL_GEN(i915) >= 9)
> skl_detach_scalers(new_crtc_state);
>
> - if (INTEL_GEN(dev_priv) >= 9 || IS_BROADWELL(dev_priv))
> + if (INTEL_GEN(i915) >= 9 || IS_BROADWELL(i915))
> bdw_set_pipemisc(new_crtc_state);
>
> out:
> - if (dev_priv->display.atomic_update_watermarks)
> - dev_priv->display.atomic_update_watermarks(state,
> + if (i915->display.atomic_update_watermarks)
> + i915->display.atomic_update_watermarks(state,
> new_crtc_state);
> }
>
> void intel_crtc_arm_fifo_underrun(struct intel_crtc *crtc,
> struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
>
> - if (!IS_GEN(dev_priv, 2))
> - intel_set_cpu_fifo_underrun_reporting(dev_priv, crtc->pipe, true);
> + if (!IS_GEN(i915, 2))
> + intel_set_cpu_fifo_underrun_reporting(i915, crtc->pipe, true);
>
> if (crtc_state->has_pch_encoder) {
> enum pipe pch_transcoder =
> intel_crtc_pch_transcoder(crtc);
>
> - intel_set_pch_fifo_underrun_reporting(dev_priv, pch_transcoder, true);
> + intel_set_pch_fifo_underrun_reporting(i915, pch_transcoder, true);
> }
> }
>
> @@ -14559,7 +14559,7 @@ intel_legacy_cursor_update(struct drm_plane *plane,
> u32 src_w, u32 src_h,
> struct drm_modeset_acquire_ctx *ctx)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc->dev);
> int ret;
> struct drm_plane_state *old_plane_state, *new_plane_state;
> struct intel_plane *intel_plane = to_intel_plane(plane);
> @@ -14626,7 +14626,7 @@ intel_legacy_cursor_update(struct drm_plane *plane,
> if (ret)
> goto out_free;
>
> - ret = mutex_lock_interruptible(&dev_priv->drm.struct_mutex);
> + ret = mutex_lock_interruptible(&i915->drm.struct_mutex);
> if (ret)
> goto out_free;
>
> @@ -14664,7 +14664,7 @@ intel_legacy_cursor_update(struct drm_plane *plane,
> intel_plane_unpin_fb(to_intel_plane_state(old_plane_state));
>
> out_unlock:
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + mutex_unlock(&i915->drm.struct_mutex);
> out_free:
> if (new_crtc_state)
> intel_crtc_destroy_state(crtc, &new_crtc_state->base);
> @@ -14691,25 +14691,25 @@ static const struct drm_plane_funcs intel_cursor_plane_funcs = {
> .format_mod_supported = intel_cursor_format_mod_supported,
> };
>
> -static bool i9xx_plane_has_fbc(struct drm_i915_private *dev_priv,
> +static bool i9xx_plane_has_fbc(struct drm_i915_private *i915,
> enum i9xx_plane_id i9xx_plane)
> {
> - if (!HAS_FBC(dev_priv))
> + if (!HAS_FBC(i915))
> return false;
>
> - if (IS_BROADWELL(dev_priv) || IS_HASWELL(dev_priv))
> + if (IS_BROADWELL(i915) || IS_HASWELL(i915))
> return i9xx_plane == PLANE_A; /* tied to pipe A */
> - else if (IS_IVYBRIDGE(dev_priv))
> + else if (IS_IVYBRIDGE(i915))
> return i9xx_plane == PLANE_A || i9xx_plane == PLANE_B ||
> i9xx_plane == PLANE_C;
> - else if (INTEL_GEN(dev_priv) >= 4)
> + else if (INTEL_GEN(i915) >= 4)
> return i9xx_plane == PLANE_A || i9xx_plane == PLANE_B;
> else
> return i9xx_plane == PLANE_A;
> }
>
> static struct intel_plane *
> -intel_primary_plane_create(struct drm_i915_private *dev_priv, enum pipe pipe)
> +intel_primary_plane_create(struct drm_i915_private *i915, enum pipe pipe)
> {
> struct intel_plane *plane;
> const struct drm_plane_funcs *plane_funcs;
> @@ -14720,8 +14720,8 @@ intel_primary_plane_create(struct drm_i915_private *dev_priv, enum pipe pipe)
> int num_formats;
> int ret;
>
> - if (INTEL_GEN(dev_priv) >= 9)
> - return skl_universal_plane_create(dev_priv, pipe,
> + if (INTEL_GEN(i915) >= 9)
> + return skl_universal_plane_create(i915, pipe,
> PLANE_PRIMARY);
>
> plane = intel_plane_alloc();
> @@ -14733,21 +14733,21 @@ intel_primary_plane_create(struct drm_i915_private *dev_priv, enum pipe pipe)
> * On gen2/3 only plane A can do FBC, but the panel fitter and LVDS
> * port is hooked to pipe B. Hence we want plane A feeding pipe B.
> */
> - if (HAS_FBC(dev_priv) && INTEL_GEN(dev_priv) < 4)
> + if (HAS_FBC(i915) && INTEL_GEN(i915) < 4)
> plane->i9xx_plane = (enum i9xx_plane_id) !pipe;
> else
> plane->i9xx_plane = (enum i9xx_plane_id) pipe;
> plane->id = PLANE_PRIMARY;
> plane->frontbuffer_bit = INTEL_FRONTBUFFER(pipe, plane->id);
>
> - plane->has_fbc = i9xx_plane_has_fbc(dev_priv, plane->i9xx_plane);
> + plane->has_fbc = i9xx_plane_has_fbc(i915, plane->i9xx_plane);
> if (plane->has_fbc) {
> - struct intel_fbc *fbc = &dev_priv->fbc;
> + struct intel_fbc *fbc = &i915->fbc;
>
> fbc->possible_framebuffer_bits |= plane->frontbuffer_bit;
> }
>
> - if (INTEL_GEN(dev_priv) >= 4) {
> + if (INTEL_GEN(i915) >= 4) {
> formats = i965_primary_formats;
> num_formats = ARRAY_SIZE(i965_primary_formats);
> modifiers = i9xx_format_modifiers;
> @@ -14775,14 +14775,14 @@ intel_primary_plane_create(struct drm_i915_private *dev_priv, enum pipe pipe)
>
> possible_crtcs = BIT(pipe);
>
> - if (INTEL_GEN(dev_priv) >= 5 || IS_G4X(dev_priv))
> - ret = drm_universal_plane_init(&dev_priv->drm, &plane->base,
> + if (INTEL_GEN(i915) >= 5 || IS_G4X(i915))
> + ret = drm_universal_plane_init(&i915->drm, &plane->base,
> possible_crtcs, plane_funcs,
> formats, num_formats, modifiers,
> DRM_PLANE_TYPE_PRIMARY,
> "primary %c", pipe_name(pipe));
> else
> - ret = drm_universal_plane_init(&dev_priv->drm, &plane->base,
> + ret = drm_universal_plane_init(&i915->drm, &plane->base,
> possible_crtcs, plane_funcs,
> formats, num_formats, modifiers,
> DRM_PLANE_TYPE_PRIMARY,
> @@ -14791,18 +14791,18 @@ intel_primary_plane_create(struct drm_i915_private *dev_priv, enum pipe pipe)
> if (ret)
> goto fail;
>
> - if (IS_CHERRYVIEW(dev_priv) && pipe == PIPE_B) {
> + if (IS_CHERRYVIEW(i915) && pipe == PIPE_B) {
> supported_rotations =
> DRM_MODE_ROTATE_0 | DRM_MODE_ROTATE_180 |
> DRM_MODE_REFLECT_X;
> - } else if (INTEL_GEN(dev_priv) >= 4) {
> + } else if (INTEL_GEN(i915) >= 4) {
> supported_rotations =
> DRM_MODE_ROTATE_0 | DRM_MODE_ROTATE_180;
> } else {
> supported_rotations = DRM_MODE_ROTATE_0;
> }
>
> - if (INTEL_GEN(dev_priv) >= 4)
> + if (INTEL_GEN(i915) >= 4)
> drm_plane_create_rotation_property(&plane->base,
> DRM_MODE_ROTATE_0,
> supported_rotations);
> @@ -14818,7 +14818,7 @@ intel_primary_plane_create(struct drm_i915_private *dev_priv, enum pipe pipe)
> }
>
> static struct intel_plane *
> -intel_cursor_plane_create(struct drm_i915_private *dev_priv,
> +intel_cursor_plane_create(struct drm_i915_private *i915,
> enum pipe pipe)
> {
> unsigned int possible_crtcs;
> @@ -14834,7 +14834,7 @@ intel_cursor_plane_create(struct drm_i915_private *dev_priv,
> cursor->id = PLANE_CURSOR;
> cursor->frontbuffer_bit = INTEL_FRONTBUFFER(pipe, cursor->id);
>
> - if (IS_I845G(dev_priv) || IS_I865G(dev_priv)) {
> + if (IS_I845G(i915) || IS_I865G(i915)) {
> cursor->max_stride = i845_cursor_max_stride;
> cursor->update_plane = i845_update_cursor;
> cursor->disable_plane = i845_disable_cursor;
> @@ -14851,12 +14851,12 @@ intel_cursor_plane_create(struct drm_i915_private *dev_priv,
> cursor->cursor.base = ~0;
> cursor->cursor.cntl = ~0;
>
> - if (IS_I845G(dev_priv) || IS_I865G(dev_priv) || HAS_CUR_FBC(dev_priv))
> + if (IS_I845G(i915) || IS_I865G(i915) || HAS_CUR_FBC(i915))
> cursor->cursor.size = ~0;
>
> possible_crtcs = BIT(pipe);
>
> - ret = drm_universal_plane_init(&dev_priv->drm, &cursor->base,
> + ret = drm_universal_plane_init(&i915->drm, &cursor->base,
> possible_crtcs, &intel_cursor_plane_funcs,
> intel_cursor_formats,
> ARRAY_SIZE(intel_cursor_formats),
> @@ -14866,7 +14866,7 @@ intel_cursor_plane_create(struct drm_i915_private *dev_priv,
> if (ret)
> goto fail;
>
> - if (INTEL_GEN(dev_priv) >= 4)
> + if (INTEL_GEN(i915) >= 4)
> drm_plane_create_rotation_property(&cursor->base,
> DRM_MODE_ROTATE_0,
> DRM_MODE_ROTATE_0 |
> @@ -14887,10 +14887,10 @@ static void intel_crtc_init_scalers(struct intel_crtc *crtc,
> {
> struct intel_crtc_scaler_state *scaler_state =
> &crtc_state->scaler_state;
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> int i;
>
> - crtc->num_scalers = RUNTIME_INFO(dev_priv)->num_scalers[crtc->pipe];
> + crtc->num_scalers = RUNTIME_INFO(i915)->num_scalers[crtc->pipe];
> if (!crtc->num_scalers)
> return;
>
> @@ -14904,7 +14904,7 @@ static void intel_crtc_init_scalers(struct intel_crtc *crtc,
> scaler_state->scaler_id = -1;
> }
>
> -static int intel_crtc_init(struct drm_i915_private *dev_priv, enum pipe pipe)
> +static int intel_crtc_init(struct drm_i915_private *i915, enum pipe pipe)
> {
> struct intel_crtc *intel_crtc;
> struct intel_crtc_state *crtc_state = NULL;
> @@ -14924,17 +14924,17 @@ static int intel_crtc_init(struct drm_i915_private *dev_priv, enum pipe pipe)
> __drm_atomic_helper_crtc_reset(&intel_crtc->base, &crtc_state->base);
> intel_crtc->config = crtc_state;
>
> - primary = intel_primary_plane_create(dev_priv, pipe);
> + primary = intel_primary_plane_create(i915, pipe);
> if (IS_ERR(primary)) {
> ret = PTR_ERR(primary);
> goto fail;
> }
> intel_crtc->plane_ids_mask |= BIT(primary->id);
>
> - for_each_sprite(dev_priv, pipe, sprite) {
> + for_each_sprite(i915, pipe, sprite) {
> struct intel_plane *plane;
>
> - plane = intel_sprite_plane_create(dev_priv, pipe, sprite);
> + plane = intel_sprite_plane_create(i915, pipe, sprite);
> if (IS_ERR(plane)) {
> ret = PTR_ERR(plane);
> goto fail;
> @@ -14942,14 +14942,14 @@ static int intel_crtc_init(struct drm_i915_private *dev_priv, enum pipe pipe)
> intel_crtc->plane_ids_mask |= BIT(plane->id);
> }
>
> - cursor = intel_cursor_plane_create(dev_priv, pipe);
> + cursor = intel_cursor_plane_create(i915, pipe);
> if (IS_ERR(cursor)) {
> ret = PTR_ERR(cursor);
> goto fail;
> }
> intel_crtc->plane_ids_mask |= BIT(cursor->id);
>
> - ret = drm_crtc_init_with_planes(&dev_priv->drm, &intel_crtc->base,
> + ret = drm_crtc_init_with_planes(&i915->drm, &intel_crtc->base,
> &primary->base, &cursor->base,
> &intel_crtc_funcs,
> "pipe %c", pipe_name(pipe));
> @@ -14961,16 +14961,16 @@ static int intel_crtc_init(struct drm_i915_private *dev_priv, enum pipe pipe)
> /* initialize shared scalers */
> intel_crtc_init_scalers(intel_crtc, crtc_state);
>
> - BUG_ON(pipe >= ARRAY_SIZE(dev_priv->pipe_to_crtc_mapping) ||
> - dev_priv->pipe_to_crtc_mapping[pipe] != NULL);
> - dev_priv->pipe_to_crtc_mapping[pipe] = intel_crtc;
> + BUG_ON(pipe >= ARRAY_SIZE(i915->pipe_to_crtc_mapping) ||
> + i915->pipe_to_crtc_mapping[pipe] != NULL);
> + i915->pipe_to_crtc_mapping[pipe] = intel_crtc;
>
> - if (INTEL_GEN(dev_priv) < 9) {
> + if (INTEL_GEN(i915) < 9) {
> enum i9xx_plane_id i9xx_plane = primary->i9xx_plane;
>
> - BUG_ON(i9xx_plane >= ARRAY_SIZE(dev_priv->plane_to_crtc_mapping) ||
> - dev_priv->plane_to_crtc_mapping[i9xx_plane] != NULL);
> - dev_priv->plane_to_crtc_mapping[i9xx_plane] = intel_crtc;
> + BUG_ON(i9xx_plane >= ARRAY_SIZE(i915->plane_to_crtc_mapping) ||
> + i915->plane_to_crtc_mapping[i9xx_plane] != NULL);
> + i915->plane_to_crtc_mapping[i9xx_plane] = intel_crtc;
> }
>
> drm_crtc_helper_add(&intel_crtc->base, &intel_helper_funcs);
> @@ -15026,29 +15026,29 @@ static int intel_encoder_clones(struct intel_encoder *encoder)
> return index_mask;
> }
>
> -static bool ilk_has_edp_a(struct drm_i915_private *dev_priv)
> +static bool ilk_has_edp_a(struct drm_i915_private *i915)
> {
> - if (!IS_MOBILE(dev_priv))
> + if (!IS_MOBILE(i915))
> return false;
>
> if ((I915_READ(DP_A) & DP_DETECTED) == 0)
> return false;
>
> - if (IS_GEN(dev_priv, 5) && (I915_READ(FUSE_STRAP) & ILK_eDP_A_DISABLE))
> + if (IS_GEN(i915, 5) && (I915_READ(FUSE_STRAP) & ILK_eDP_A_DISABLE))
> return false;
>
> return true;
> }
>
> -static bool intel_ddi_crt_present(struct drm_i915_private *dev_priv)
> +static bool intel_ddi_crt_present(struct drm_i915_private *i915)
> {
> - if (INTEL_GEN(dev_priv) >= 9)
> + if (INTEL_GEN(i915) >= 9)
> return false;
>
> - if (IS_HSW_ULT(dev_priv) || IS_BDW_ULT(dev_priv))
> + if (IS_HSW_ULT(i915) || IS_BDW_ULT(i915))
> return false;
>
> - if (HAS_PCH_LPT_H(dev_priv) &&
> + if (HAS_PCH_LPT_H(i915) &&
> I915_READ(SFUSE_STRAP) & SFUSE_STRAP_CRT_DISABLED)
> return false;
>
> @@ -15056,24 +15056,24 @@ static bool intel_ddi_crt_present(struct drm_i915_private *dev_priv)
> if (I915_READ(DDI_BUF_CTL(PORT_A)) & DDI_A_4_LANES)
> return false;
>
> - if (!dev_priv->vbt.int_crt_support)
> + if (!i915->vbt.int_crt_support)
> return false;
>
> return true;
> }
>
> -void intel_pps_unlock_regs_wa(struct drm_i915_private *dev_priv)
> +void intel_pps_unlock_regs_wa(struct drm_i915_private *i915)
> {
> int pps_num;
> int pps_idx;
>
> - if (HAS_DDI(dev_priv))
> + if (HAS_DDI(i915))
> return;
> /*
> * This w/a is needed at least on CPT/PPT, but to be sure apply it
> * everywhere where registers can be write protected.
> */
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> pps_num = 2;
> else
> pps_num = 1;
> @@ -15086,65 +15086,65 @@ void intel_pps_unlock_regs_wa(struct drm_i915_private *dev_priv)
> }
> }
>
> -static void intel_pps_init(struct drm_i915_private *dev_priv)
> +static void intel_pps_init(struct drm_i915_private *i915)
> {
> - if (HAS_PCH_SPLIT(dev_priv) || IS_GEN9_LP(dev_priv))
> - dev_priv->pps_mmio_base = PCH_PPS_BASE;
> - else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> - dev_priv->pps_mmio_base = VLV_PPS_BASE;
> + if (HAS_PCH_SPLIT(i915) || IS_GEN9_LP(i915))
> + i915->pps_mmio_base = PCH_PPS_BASE;
> + else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> + i915->pps_mmio_base = VLV_PPS_BASE;
> else
> - dev_priv->pps_mmio_base = PPS_BASE;
> + i915->pps_mmio_base = PPS_BASE;
>
> - intel_pps_unlock_regs_wa(dev_priv);
> + intel_pps_unlock_regs_wa(i915);
> }
>
> -static void intel_setup_outputs(struct drm_i915_private *dev_priv)
> +static void intel_setup_outputs(struct drm_i915_private *i915)
> {
> struct intel_encoder *encoder;
> bool dpd_is_edp = false;
>
> - intel_pps_init(dev_priv);
> + intel_pps_init(i915);
>
> - if (!HAS_DISPLAY(dev_priv))
> + if (!HAS_DISPLAY(i915))
> return;
>
> - if (IS_ELKHARTLAKE(dev_priv)) {
> - intel_ddi_init(dev_priv, PORT_A);
> - intel_ddi_init(dev_priv, PORT_B);
> - intel_ddi_init(dev_priv, PORT_C);
> - icl_dsi_init(dev_priv);
> - } else if (INTEL_GEN(dev_priv) >= 11) {
> - intel_ddi_init(dev_priv, PORT_A);
> - intel_ddi_init(dev_priv, PORT_B);
> - intel_ddi_init(dev_priv, PORT_C);
> - intel_ddi_init(dev_priv, PORT_D);
> - intel_ddi_init(dev_priv, PORT_E);
> + if (IS_ELKHARTLAKE(i915)) {
> + intel_ddi_init(i915, PORT_A);
> + intel_ddi_init(i915, PORT_B);
> + intel_ddi_init(i915, PORT_C);
> + icl_dsi_init(i915);
> + } else if (INTEL_GEN(i915) >= 11) {
> + intel_ddi_init(i915, PORT_A);
> + intel_ddi_init(i915, PORT_B);
> + intel_ddi_init(i915, PORT_C);
> + intel_ddi_init(i915, PORT_D);
> + intel_ddi_init(i915, PORT_E);
> /*
> * On some ICL SKUs port F is not present. No strap bits for
> * this, so rely on VBT.
> * Work around broken VBTs on SKUs known to have no port F.
> */
> - if (IS_ICL_WITH_PORT_F(dev_priv) &&
> - intel_bios_is_port_present(dev_priv, PORT_F))
> - intel_ddi_init(dev_priv, PORT_F);
> + if (IS_ICL_WITH_PORT_F(i915) &&
> + intel_bios_is_port_present(i915, PORT_F))
> + intel_ddi_init(i915, PORT_F);
>
> - icl_dsi_init(dev_priv);
> - } else if (IS_GEN9_LP(dev_priv)) {
> + icl_dsi_init(i915);
> + } else if (IS_GEN9_LP(i915)) {
> /*
> * FIXME: Broxton doesn't support port detection via the
> * DDI_BUF_CTL_A or SFUSE_STRAP registers, find another way to
> * detect the ports.
> */
> - intel_ddi_init(dev_priv, PORT_A);
> - intel_ddi_init(dev_priv, PORT_B);
> - intel_ddi_init(dev_priv, PORT_C);
> + intel_ddi_init(i915, PORT_A);
> + intel_ddi_init(i915, PORT_B);
> + intel_ddi_init(i915, PORT_C);
>
> - vlv_dsi_init(dev_priv);
> - } else if (HAS_DDI(dev_priv)) {
> + vlv_dsi_init(i915);
> + } else if (HAS_DDI(i915)) {
> int found;
>
> - if (intel_ddi_crt_present(dev_priv))
> - intel_crt_init(dev_priv);
> + if (intel_ddi_crt_present(i915))
> + intel_crt_init(i915);
>
> /*
> * Haswell uses DDI functions to detect digital outputs.
> @@ -15153,29 +15153,29 @@ static void intel_setup_outputs(struct drm_i915_private *dev_priv)
> */
> found = I915_READ(DDI_BUF_CTL(PORT_A)) & DDI_INIT_DISPLAY_DETECTED;
> /* WaIgnoreDDIAStrap: skl */
> - if (found || IS_GEN9_BC(dev_priv))
> - intel_ddi_init(dev_priv, PORT_A);
> + if (found || IS_GEN9_BC(i915))
> + intel_ddi_init(i915, PORT_A);
>
> /* DDI B, C, D, and F detection is indicated by the SFUSE_STRAP
> * register */
> found = I915_READ(SFUSE_STRAP);
>
> if (found & SFUSE_STRAP_DDIB_DETECTED)
> - intel_ddi_init(dev_priv, PORT_B);
> + intel_ddi_init(i915, PORT_B);
> if (found & SFUSE_STRAP_DDIC_DETECTED)
> - intel_ddi_init(dev_priv, PORT_C);
> + intel_ddi_init(i915, PORT_C);
> if (found & SFUSE_STRAP_DDID_DETECTED)
> - intel_ddi_init(dev_priv, PORT_D);
> + intel_ddi_init(i915, PORT_D);
> if (found & SFUSE_STRAP_DDIF_DETECTED)
> - intel_ddi_init(dev_priv, PORT_F);
> + intel_ddi_init(i915, PORT_F);
> /*
> * On SKL we don't have a way to detect DDI-E so we rely on VBT.
> */
> - if (IS_GEN9_BC(dev_priv) &&
> - intel_bios_is_port_present(dev_priv, PORT_E))
> - intel_ddi_init(dev_priv, PORT_E);
> + if (IS_GEN9_BC(i915) &&
> + intel_bios_is_port_present(i915, PORT_E))
> + intel_ddi_init(i915, PORT_E);
>
> - } else if (HAS_PCH_SPLIT(dev_priv)) {
> + } else if (HAS_PCH_SPLIT(i915)) {
> int found;
>
> /*
> @@ -15183,39 +15183,39 @@ static void intel_setup_outputs(struct drm_i915_private *dev_priv)
> * to prevent the registration of both eDP and LVDS and the
> * incorrect sharing of the PPS.
> */
> - intel_lvds_init(dev_priv);
> - intel_crt_init(dev_priv);
> + intel_lvds_init(i915);
> + intel_crt_init(i915);
>
> - dpd_is_edp = intel_dp_is_port_edp(dev_priv, PORT_D);
> + dpd_is_edp = intel_dp_is_port_edp(i915, PORT_D);
>
> - if (ilk_has_edp_a(dev_priv))
> - intel_dp_init(dev_priv, DP_A, PORT_A);
> + if (ilk_has_edp_a(i915))
> + intel_dp_init(i915, DP_A, PORT_A);
>
> if (I915_READ(PCH_HDMIB) & SDVO_DETECTED) {
> /* PCH SDVOB multiplex with HDMIB */
> - found = intel_sdvo_init(dev_priv, PCH_SDVOB, PORT_B);
> + found = intel_sdvo_init(i915, PCH_SDVOB, PORT_B);
> if (!found)
> - intel_hdmi_init(dev_priv, PCH_HDMIB, PORT_B);
> + intel_hdmi_init(i915, PCH_HDMIB, PORT_B);
> if (!found && (I915_READ(PCH_DP_B) & DP_DETECTED))
> - intel_dp_init(dev_priv, PCH_DP_B, PORT_B);
> + intel_dp_init(i915, PCH_DP_B, PORT_B);
> }
>
> if (I915_READ(PCH_HDMIC) & SDVO_DETECTED)
> - intel_hdmi_init(dev_priv, PCH_HDMIC, PORT_C);
> + intel_hdmi_init(i915, PCH_HDMIC, PORT_C);
>
> if (!dpd_is_edp && I915_READ(PCH_HDMID) & SDVO_DETECTED)
> - intel_hdmi_init(dev_priv, PCH_HDMID, PORT_D);
> + intel_hdmi_init(i915, PCH_HDMID, PORT_D);
>
> if (I915_READ(PCH_DP_C) & DP_DETECTED)
> - intel_dp_init(dev_priv, PCH_DP_C, PORT_C);
> + intel_dp_init(i915, PCH_DP_C, PORT_C);
>
> if (I915_READ(PCH_DP_D) & DP_DETECTED)
> - intel_dp_init(dev_priv, PCH_DP_D, PORT_D);
> - } else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> + intel_dp_init(i915, PCH_DP_D, PORT_D);
> + } else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> bool has_edp, has_port;
>
> - if (IS_VALLEYVIEW(dev_priv) && dev_priv->vbt.int_crt_support)
> - intel_crt_init(dev_priv);
> + if (IS_VALLEYVIEW(i915) && i915->vbt.int_crt_support)
> + intel_crt_init(i915);
>
> /*
> * The DP_DETECTED bit is the latched state of the DDC
> @@ -15232,97 +15232,97 @@ static void intel_setup_outputs(struct drm_i915_private *dev_priv)
> * trust the port type the VBT declares as we've seen at least
> * HDMI ports that the VBT claim are DP or eDP.
> */
> - has_edp = intel_dp_is_port_edp(dev_priv, PORT_B);
> - has_port = intel_bios_is_port_present(dev_priv, PORT_B);
> + has_edp = intel_dp_is_port_edp(i915, PORT_B);
> + has_port = intel_bios_is_port_present(i915, PORT_B);
> if (I915_READ(VLV_DP_B) & DP_DETECTED || has_port)
> - has_edp &= intel_dp_init(dev_priv, VLV_DP_B, PORT_B);
> + has_edp &= intel_dp_init(i915, VLV_DP_B, PORT_B);
> if ((I915_READ(VLV_HDMIB) & SDVO_DETECTED || has_port) && !has_edp)
> - intel_hdmi_init(dev_priv, VLV_HDMIB, PORT_B);
> + intel_hdmi_init(i915, VLV_HDMIB, PORT_B);
>
> - has_edp = intel_dp_is_port_edp(dev_priv, PORT_C);
> - has_port = intel_bios_is_port_present(dev_priv, PORT_C);
> + has_edp = intel_dp_is_port_edp(i915, PORT_C);
> + has_port = intel_bios_is_port_present(i915, PORT_C);
> if (I915_READ(VLV_DP_C) & DP_DETECTED || has_port)
> - has_edp &= intel_dp_init(dev_priv, VLV_DP_C, PORT_C);
> + has_edp &= intel_dp_init(i915, VLV_DP_C, PORT_C);
> if ((I915_READ(VLV_HDMIC) & SDVO_DETECTED || has_port) && !has_edp)
> - intel_hdmi_init(dev_priv, VLV_HDMIC, PORT_C);
> + intel_hdmi_init(i915, VLV_HDMIC, PORT_C);
>
> - if (IS_CHERRYVIEW(dev_priv)) {
> + if (IS_CHERRYVIEW(i915)) {
> /*
> * eDP not supported on port D,
> * so no need to worry about it
> */
> - has_port = intel_bios_is_port_present(dev_priv, PORT_D);
> + has_port = intel_bios_is_port_present(i915, PORT_D);
> if (I915_READ(CHV_DP_D) & DP_DETECTED || has_port)
> - intel_dp_init(dev_priv, CHV_DP_D, PORT_D);
> + intel_dp_init(i915, CHV_DP_D, PORT_D);
> if (I915_READ(CHV_HDMID) & SDVO_DETECTED || has_port)
> - intel_hdmi_init(dev_priv, CHV_HDMID, PORT_D);
> + intel_hdmi_init(i915, CHV_HDMID, PORT_D);
> }
>
> - vlv_dsi_init(dev_priv);
> - } else if (IS_PINEVIEW(dev_priv)) {
> - intel_lvds_init(dev_priv);
> - intel_crt_init(dev_priv);
> - } else if (IS_GEN_RANGE(dev_priv, 3, 4)) {
> + vlv_dsi_init(i915);
> + } else if (IS_PINEVIEW(i915)) {
> + intel_lvds_init(i915);
> + intel_crt_init(i915);
> + } else if (IS_GEN_RANGE(i915, 3, 4)) {
> bool found = false;
>
> - if (IS_MOBILE(dev_priv))
> - intel_lvds_init(dev_priv);
> + if (IS_MOBILE(i915))
> + intel_lvds_init(i915);
>
> - intel_crt_init(dev_priv);
> + intel_crt_init(i915);
>
> if (I915_READ(GEN3_SDVOB) & SDVO_DETECTED) {
> DRM_DEBUG_KMS("probing SDVOB\n");
> - found = intel_sdvo_init(dev_priv, GEN3_SDVOB, PORT_B);
> - if (!found && IS_G4X(dev_priv)) {
> + found = intel_sdvo_init(i915, GEN3_SDVOB, PORT_B);
> + if (!found && IS_G4X(i915)) {
> DRM_DEBUG_KMS("probing HDMI on SDVOB\n");
> - intel_hdmi_init(dev_priv, GEN4_HDMIB, PORT_B);
> + intel_hdmi_init(i915, GEN4_HDMIB, PORT_B);
> }
>
> - if (!found && IS_G4X(dev_priv))
> - intel_dp_init(dev_priv, DP_B, PORT_B);
> + if (!found && IS_G4X(i915))
> + intel_dp_init(i915, DP_B, PORT_B);
> }
>
> /* Before G4X SDVOC doesn't have its own detect register */
>
> if (I915_READ(GEN3_SDVOB) & SDVO_DETECTED) {
> DRM_DEBUG_KMS("probing SDVOC\n");
> - found = intel_sdvo_init(dev_priv, GEN3_SDVOC, PORT_C);
> + found = intel_sdvo_init(i915, GEN3_SDVOC, PORT_C);
> }
>
> if (!found && (I915_READ(GEN3_SDVOC) & SDVO_DETECTED)) {
>
> - if (IS_G4X(dev_priv)) {
> + if (IS_G4X(i915)) {
> DRM_DEBUG_KMS("probing HDMI on SDVOC\n");
> - intel_hdmi_init(dev_priv, GEN4_HDMIC, PORT_C);
> + intel_hdmi_init(i915, GEN4_HDMIC, PORT_C);
> }
> - if (IS_G4X(dev_priv))
> - intel_dp_init(dev_priv, DP_C, PORT_C);
> + if (IS_G4X(i915))
> + intel_dp_init(i915, DP_C, PORT_C);
> }
>
> - if (IS_G4X(dev_priv) && (I915_READ(DP_D) & DP_DETECTED))
> - intel_dp_init(dev_priv, DP_D, PORT_D);
> + if (IS_G4X(i915) && (I915_READ(DP_D) & DP_DETECTED))
> + intel_dp_init(i915, DP_D, PORT_D);
>
> - if (SUPPORTS_TV(dev_priv))
> - intel_tv_init(dev_priv);
> - } else if (IS_GEN(dev_priv, 2)) {
> - if (IS_I85X(dev_priv))
> - intel_lvds_init(dev_priv);
> + if (SUPPORTS_TV(i915))
> + intel_tv_init(i915);
> + } else if (IS_GEN(i915, 2)) {
> + if (IS_I85X(i915))
> + intel_lvds_init(i915);
>
> - intel_crt_init(dev_priv);
> - intel_dvo_init(dev_priv);
> + intel_crt_init(i915);
> + intel_dvo_init(i915);
> }
>
> - intel_psr_init(dev_priv);
> + intel_psr_init(i915);
>
> - for_each_intel_encoder(&dev_priv->drm, encoder) {
> + for_each_intel_encoder(&i915->drm, encoder) {
> encoder->base.possible_crtcs = encoder->crtc_mask;
> encoder->base.possible_clones =
> intel_encoder_clones(encoder);
> }
>
> - intel_init_pch_refclk(dev_priv);
> + intel_init_pch_refclk(i915);
>
> - drm_helper_move_panel_connectors_to_head(&dev_priv->drm);
> + drm_helper_move_panel_connectors_to_head(&i915->drm);
> }
>
> static void intel_user_framebuffer_destroy(struct drm_framebuffer *fb)
> @@ -15379,7 +15379,7 @@ static int intel_framebuffer_init(struct intel_framebuffer *intel_fb,
> struct drm_i915_gem_object *obj,
> struct drm_mode_fb_cmd2 *mode_cmd)
> {
> - struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
> + struct drm_i915_private *i915 = to_i915(obj->base.dev);
> struct drm_framebuffer *fb = &intel_fb->base;
> u32 max_stride;
> unsigned int tiling, stride;
> @@ -15411,7 +15411,7 @@ static int intel_framebuffer_init(struct intel_framebuffer *intel_fb,
> }
> }
>
> - if (!drm_any_plane_has_format(&dev_priv->drm,
> + if (!drm_any_plane_has_format(&i915->drm,
> mode_cmd->pixel_format,
> mode_cmd->modifier[0])) {
> struct drm_format_name_buf format_name;
> @@ -15427,13 +15427,13 @@ static int intel_framebuffer_init(struct intel_framebuffer *intel_fb,
> * gen2/3 display engine uses the fence if present,
> * so the tiling mode must match the fb modifier exactly.
> */
> - if (INTEL_GEN(dev_priv) < 4 &&
> + if (INTEL_GEN(i915) < 4 &&
> tiling != intel_fb_modifier_to_tiling(mode_cmd->modifier[0])) {
> DRM_DEBUG_KMS("tiling_mode must match fb modifier exactly on gen2/3\n");
> goto err;
> }
>
> - max_stride = intel_fb_max_stride(dev_priv, mode_cmd->pixel_format,
> + max_stride = intel_fb_max_stride(i915, mode_cmd->pixel_format,
> mode_cmd->modifier[0]);
> if (mode_cmd->pitches[0] > max_stride) {
> DRM_DEBUG_KMS("%s pitch (%u) must be at most %d\n",
> @@ -15457,7 +15457,7 @@ static int intel_framebuffer_init(struct intel_framebuffer *intel_fb,
> if (mode_cmd->offsets[0] != 0)
> goto err;
>
> - drm_helper_mode_fill_fb_struct(&dev_priv->drm, fb, mode_cmd);
> + drm_helper_mode_fill_fb_struct(&i915->drm, fb, mode_cmd);
>
> for (i = 0; i < fb->format->num_planes; i++) {
> u32 stride_alignment;
> @@ -15478,7 +15478,7 @@ static int intel_framebuffer_init(struct intel_framebuffer *intel_fb,
> * require the entire fb to accommodate that to avoid
> * potential runtime errors at plane configuration time.
> */
> - if (IS_GEN(dev_priv, 9) && i == 0 && fb->width > 3840 &&
> + if (IS_GEN(i915, 9) && i == 0 && fb->width > 3840 &&
> is_ccs_modifier(fb->modifier))
> stride_alignment *= 4;
>
> @@ -15491,11 +15491,11 @@ static int intel_framebuffer_init(struct intel_framebuffer *intel_fb,
> fb->obj[i] = &obj->base;
> }
>
> - ret = intel_fill_fb_info(dev_priv, fb);
> + ret = intel_fill_fb_info(i915, fb);
> if (ret)
> goto err;
>
> - ret = drm_framebuffer_init(&dev_priv->drm, fb, &intel_fb_funcs);
> + ret = drm_framebuffer_init(&i915->drm, fb, &intel_fb_funcs);
> if (ret) {
> DRM_ERROR("framebuffer init failed %d\n", ret);
> goto err;
> @@ -15545,7 +15545,7 @@ static enum drm_mode_status
> intel_mode_valid(struct drm_device *dev,
> const struct drm_display_mode *mode)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> int hdisplay_max, htotal_max;
> int vdisplay_max, vtotal_max;
>
> @@ -15578,13 +15578,13 @@ intel_mode_valid(struct drm_device *dev,
> DRM_MODE_FLAG_CLKDIV2))
> return MODE_BAD;
>
> - if (INTEL_GEN(dev_priv) >= 9 ||
> - IS_BROADWELL(dev_priv) || IS_HASWELL(dev_priv)) {
> + if (INTEL_GEN(i915) >= 9 ||
> + IS_BROADWELL(i915) || IS_HASWELL(i915)) {
> hdisplay_max = 8192; /* FDI max 4096 handled elsewhere */
> vdisplay_max = 4096;
> htotal_max = 8192;
> vtotal_max = 8192;
> - } else if (INTEL_GEN(dev_priv) >= 3) {
> + } else if (INTEL_GEN(i915) >= 3) {
> hdisplay_max = 4096;
> vdisplay_max = 4096;
> htotal_max = 8192;
> @@ -15625,113 +15625,113 @@ static const struct drm_mode_config_funcs intel_mode_funcs = {
>
> /**
> * intel_init_display_hooks - initialize the display modesetting hooks
> - * @dev_priv: device private
> + * @i915: device private
> */
> -void intel_init_display_hooks(struct drm_i915_private *dev_priv)
> +void intel_init_display_hooks(struct drm_i915_private *i915)
> {
> - intel_init_cdclk_hooks(dev_priv);
> + intel_init_cdclk_hooks(i915);
>
> - if (INTEL_GEN(dev_priv) >= 9) {
> - dev_priv->display.get_pipe_config = haswell_get_pipe_config;
> - dev_priv->display.get_initial_plane_config =
> + if (INTEL_GEN(i915) >= 9) {
> + i915->display.get_pipe_config = haswell_get_pipe_config;
> + i915->display.get_initial_plane_config =
> skylake_get_initial_plane_config;
> - dev_priv->display.crtc_compute_clock =
> + i915->display.crtc_compute_clock =
> haswell_crtc_compute_clock;
> - dev_priv->display.crtc_enable = haswell_crtc_enable;
> - dev_priv->display.crtc_disable = haswell_crtc_disable;
> - } else if (HAS_DDI(dev_priv)) {
> - dev_priv->display.get_pipe_config = haswell_get_pipe_config;
> - dev_priv->display.get_initial_plane_config =
> + i915->display.crtc_enable = haswell_crtc_enable;
> + i915->display.crtc_disable = haswell_crtc_disable;
> + } else if (HAS_DDI(i915)) {
> + i915->display.get_pipe_config = haswell_get_pipe_config;
> + i915->display.get_initial_plane_config =
> i9xx_get_initial_plane_config;
> - dev_priv->display.crtc_compute_clock =
> + i915->display.crtc_compute_clock =
> haswell_crtc_compute_clock;
> - dev_priv->display.crtc_enable = haswell_crtc_enable;
> - dev_priv->display.crtc_disable = haswell_crtc_disable;
> - } else if (HAS_PCH_SPLIT(dev_priv)) {
> - dev_priv->display.get_pipe_config = ironlake_get_pipe_config;
> - dev_priv->display.get_initial_plane_config =
> + i915->display.crtc_enable = haswell_crtc_enable;
> + i915->display.crtc_disable = haswell_crtc_disable;
> + } else if (HAS_PCH_SPLIT(i915)) {
> + i915->display.get_pipe_config = ironlake_get_pipe_config;
> + i915->display.get_initial_plane_config =
> i9xx_get_initial_plane_config;
> - dev_priv->display.crtc_compute_clock =
> + i915->display.crtc_compute_clock =
> ironlake_crtc_compute_clock;
> - dev_priv->display.crtc_enable = ironlake_crtc_enable;
> - dev_priv->display.crtc_disable = ironlake_crtc_disable;
> - } else if (IS_CHERRYVIEW(dev_priv)) {
> - dev_priv->display.get_pipe_config = i9xx_get_pipe_config;
> - dev_priv->display.get_initial_plane_config =
> + i915->display.crtc_enable = ironlake_crtc_enable;
> + i915->display.crtc_disable = ironlake_crtc_disable;
> + } else if (IS_CHERRYVIEW(i915)) {
> + i915->display.get_pipe_config = i9xx_get_pipe_config;
> + i915->display.get_initial_plane_config =
> i9xx_get_initial_plane_config;
> - dev_priv->display.crtc_compute_clock = chv_crtc_compute_clock;
> - dev_priv->display.crtc_enable = valleyview_crtc_enable;
> - dev_priv->display.crtc_disable = i9xx_crtc_disable;
> - } else if (IS_VALLEYVIEW(dev_priv)) {
> - dev_priv->display.get_pipe_config = i9xx_get_pipe_config;
> - dev_priv->display.get_initial_plane_config =
> + i915->display.crtc_compute_clock = chv_crtc_compute_clock;
> + i915->display.crtc_enable = valleyview_crtc_enable;
> + i915->display.crtc_disable = i9xx_crtc_disable;
> + } else if (IS_VALLEYVIEW(i915)) {
> + i915->display.get_pipe_config = i9xx_get_pipe_config;
> + i915->display.get_initial_plane_config =
> i9xx_get_initial_plane_config;
> - dev_priv->display.crtc_compute_clock = vlv_crtc_compute_clock;
> - dev_priv->display.crtc_enable = valleyview_crtc_enable;
> - dev_priv->display.crtc_disable = i9xx_crtc_disable;
> - } else if (IS_G4X(dev_priv)) {
> - dev_priv->display.get_pipe_config = i9xx_get_pipe_config;
> - dev_priv->display.get_initial_plane_config =
> + i915->display.crtc_compute_clock = vlv_crtc_compute_clock;
> + i915->display.crtc_enable = valleyview_crtc_enable;
> + i915->display.crtc_disable = i9xx_crtc_disable;
> + } else if (IS_G4X(i915)) {
> + i915->display.get_pipe_config = i9xx_get_pipe_config;
> + i915->display.get_initial_plane_config =
> i9xx_get_initial_plane_config;
> - dev_priv->display.crtc_compute_clock = g4x_crtc_compute_clock;
> - dev_priv->display.crtc_enable = i9xx_crtc_enable;
> - dev_priv->display.crtc_disable = i9xx_crtc_disable;
> - } else if (IS_PINEVIEW(dev_priv)) {
> - dev_priv->display.get_pipe_config = i9xx_get_pipe_config;
> - dev_priv->display.get_initial_plane_config =
> + i915->display.crtc_compute_clock = g4x_crtc_compute_clock;
> + i915->display.crtc_enable = i9xx_crtc_enable;
> + i915->display.crtc_disable = i9xx_crtc_disable;
> + } else if (IS_PINEVIEW(i915)) {
> + i915->display.get_pipe_config = i9xx_get_pipe_config;
> + i915->display.get_initial_plane_config =
> i9xx_get_initial_plane_config;
> - dev_priv->display.crtc_compute_clock = pnv_crtc_compute_clock;
> - dev_priv->display.crtc_enable = i9xx_crtc_enable;
> - dev_priv->display.crtc_disable = i9xx_crtc_disable;
> - } else if (!IS_GEN(dev_priv, 2)) {
> - dev_priv->display.get_pipe_config = i9xx_get_pipe_config;
> - dev_priv->display.get_initial_plane_config =
> + i915->display.crtc_compute_clock = pnv_crtc_compute_clock;
> + i915->display.crtc_enable = i9xx_crtc_enable;
> + i915->display.crtc_disable = i9xx_crtc_disable;
> + } else if (!IS_GEN(i915, 2)) {
> + i915->display.get_pipe_config = i9xx_get_pipe_config;
> + i915->display.get_initial_plane_config =
> i9xx_get_initial_plane_config;
> - dev_priv->display.crtc_compute_clock = i9xx_crtc_compute_clock;
> - dev_priv->display.crtc_enable = i9xx_crtc_enable;
> - dev_priv->display.crtc_disable = i9xx_crtc_disable;
> + i915->display.crtc_compute_clock = i9xx_crtc_compute_clock;
> + i915->display.crtc_enable = i9xx_crtc_enable;
> + i915->display.crtc_disable = i9xx_crtc_disable;
> } else {
> - dev_priv->display.get_pipe_config = i9xx_get_pipe_config;
> - dev_priv->display.get_initial_plane_config =
> + i915->display.get_pipe_config = i9xx_get_pipe_config;
> + i915->display.get_initial_plane_config =
> i9xx_get_initial_plane_config;
> - dev_priv->display.crtc_compute_clock = i8xx_crtc_compute_clock;
> - dev_priv->display.crtc_enable = i9xx_crtc_enable;
> - dev_priv->display.crtc_disable = i9xx_crtc_disable;
> + i915->display.crtc_compute_clock = i8xx_crtc_compute_clock;
> + i915->display.crtc_enable = i9xx_crtc_enable;
> + i915->display.crtc_disable = i9xx_crtc_disable;
> }
>
> - if (IS_GEN(dev_priv, 5)) {
> - dev_priv->display.fdi_link_train = ironlake_fdi_link_train;
> - } else if (IS_GEN(dev_priv, 6)) {
> - dev_priv->display.fdi_link_train = gen6_fdi_link_train;
> - } else if (IS_IVYBRIDGE(dev_priv)) {
> + if (IS_GEN(i915, 5)) {
> + i915->display.fdi_link_train = ironlake_fdi_link_train;
> + } else if (IS_GEN(i915, 6)) {
> + i915->display.fdi_link_train = gen6_fdi_link_train;
> + } else if (IS_IVYBRIDGE(i915)) {
> /* FIXME: detect B0+ stepping and use auto training */
> - dev_priv->display.fdi_link_train = ivb_manual_fdi_link_train;
> - } else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) {
> - dev_priv->display.fdi_link_train = hsw_fdi_link_train;
> + i915->display.fdi_link_train = ivb_manual_fdi_link_train;
> + } else if (IS_HASWELL(i915) || IS_BROADWELL(i915)) {
> + i915->display.fdi_link_train = hsw_fdi_link_train;
> }
>
> - if (INTEL_GEN(dev_priv) >= 9)
> - dev_priv->display.update_crtcs = skl_update_crtcs;
> + if (INTEL_GEN(i915) >= 9)
> + i915->display.update_crtcs = skl_update_crtcs;
> else
> - dev_priv->display.update_crtcs = intel_update_crtcs;
> + i915->display.update_crtcs = intel_update_crtcs;
> }
>
> -static i915_reg_t i915_vgacntrl_reg(struct drm_i915_private *dev_priv)
> +static i915_reg_t i915_vgacntrl_reg(struct drm_i915_private *i915)
> {
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> return VLV_VGACNTRL;
> - else if (INTEL_GEN(dev_priv) >= 5)
> + else if (INTEL_GEN(i915) >= 5)
> return CPU_VGACNTRL;
> else
> return VGACNTRL;
> }
>
> /* Disable the VGA plane that we never use */
> -static void i915_disable_vga(struct drm_i915_private *dev_priv)
> +static void i915_disable_vga(struct drm_i915_private *i915)
> {
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct pci_dev *pdev = i915->drm.pdev;
> u8 sr1;
> - i915_reg_t vga_reg = i915_vgacntrl_reg(dev_priv);
> + i915_reg_t vga_reg = i915_vgacntrl_reg(i915);
>
> /* WaEnableVGAAccessThroughIOPort:ctg,elk,ilk,snb,ivb,vlv,hsw */
> vga_get_uninterruptible(pdev, VGA_RSRC_LEGACY_IO);
> @@ -15747,11 +15747,11 @@ static void i915_disable_vga(struct drm_i915_private *dev_priv)
>
> void intel_modeset_init_hw(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
>
> - intel_update_cdclk(dev_priv);
> - intel_dump_cdclk_state(&dev_priv->cdclk.hw, "Current CDCLK");
> - dev_priv->cdclk.logical = dev_priv->cdclk.actual = dev_priv->cdclk.hw;
> + intel_update_cdclk(i915);
> + intel_dump_cdclk_state(&i915->cdclk.hw, "Current CDCLK");
> + i915->cdclk.logical = i915->cdclk.actual = i915->cdclk.hw;
> }
>
> /*
> @@ -15766,7 +15766,7 @@ void intel_modeset_init_hw(struct drm_device *dev)
> */
> static void sanitize_watermarks(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct drm_atomic_state *state;
> struct intel_atomic_state *intel_state;
> struct drm_crtc *crtc;
> @@ -15776,7 +15776,7 @@ static void sanitize_watermarks(struct drm_device *dev)
> int i;
>
> /* Only supported on platforms that use atomic watermark design */
> - if (!dev_priv->display.optimize_watermarks)
> + if (!i915->display.optimize_watermarks)
> return;
>
> /*
> @@ -15804,7 +15804,7 @@ static void sanitize_watermarks(struct drm_device *dev)
> * intermediate watermarks (since we don't trust the current
> * watermarks).
> */
> - if (!HAS_GMCH(dev_priv))
> + if (!HAS_GMCH(i915))
> intel_state->skip_intermediate_wm = true;
>
> ret = intel_atomic_check(dev, state);
> @@ -15829,7 +15829,7 @@ static void sanitize_watermarks(struct drm_device *dev)
> struct intel_crtc_state *cs = to_intel_crtc_state(cstate);
>
> cs->wm.need_postvbl_update = true;
> - dev_priv->display.optimize_watermarks(intel_state, cs);
> + i915->display.optimize_watermarks(intel_state, cs);
>
> to_intel_crtc_state(crtc->state)->wm = cs->wm;
> }
> @@ -15841,20 +15841,20 @@ static void sanitize_watermarks(struct drm_device *dev)
> drm_modeset_acquire_fini(&ctx);
> }
>
> -static void intel_update_fdi_pll_freq(struct drm_i915_private *dev_priv)
> +static void intel_update_fdi_pll_freq(struct drm_i915_private *i915)
> {
> - if (IS_GEN(dev_priv, 5)) {
> + if (IS_GEN(i915, 5)) {
> u32 fdi_pll_clk =
> I915_READ(FDI_PLL_BIOS_0) & FDI_PLL_FB_CLOCK_MASK;
>
> - dev_priv->fdi_pll_freq = (fdi_pll_clk + 2) * 10000;
> - } else if (IS_GEN(dev_priv, 6) || IS_IVYBRIDGE(dev_priv)) {
> - dev_priv->fdi_pll_freq = 270000;
> + i915->fdi_pll_freq = (fdi_pll_clk + 2) * 10000;
> + } else if (IS_GEN(i915, 6) || IS_IVYBRIDGE(i915)) {
> + i915->fdi_pll_freq = 270000;
> } else {
> return;
> }
>
> - DRM_DEBUG_DRIVER("FDI PLL freq=%d\n", dev_priv->fdi_pll_freq);
> + DRM_DEBUG_DRIVER("FDI PLL freq=%d\n", i915->fdi_pll_freq);
> }
>
> static int intel_initial_commit(struct drm_device *dev)
> @@ -15915,17 +15915,17 @@ static int intel_initial_commit(struct drm_device *dev)
>
> int intel_modeset_init(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct i915_ggtt *ggtt = &dev_priv->ggtt;
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct i915_ggtt *ggtt = &i915->ggtt;
> enum pipe pipe;
> struct intel_crtc *crtc;
> int ret;
>
> - dev_priv->modeset_wq = alloc_ordered_workqueue("i915_modeset", 0);
> + i915->modeset_wq = alloc_ordered_workqueue("i915_modeset", 0);
>
> drm_mode_config_init(dev);
>
> - ret = intel_bw_init(dev_priv);
> + ret = intel_bw_init(i915);
> if (ret)
> return ret;
>
> @@ -15939,15 +15939,15 @@ int intel_modeset_init(struct drm_device *dev)
>
> dev->mode_config.funcs = &intel_mode_funcs;
>
> - init_llist_head(&dev_priv->atomic_helper.free_list);
> - INIT_WORK(&dev_priv->atomic_helper.free_work,
> + init_llist_head(&i915->atomic_helper.free_list);
> + INIT_WORK(&i915->atomic_helper.free_work,
> intel_atomic_helper_free_state_worker);
>
> - intel_init_quirks(dev_priv);
> + intel_init_quirks(i915);
>
> - intel_fbc_init(dev_priv);
> + intel_fbc_init(i915);
>
> - intel_init_pm(dev_priv);
> + intel_init_pm(i915);
>
> /*
> * There may be no VBT; and if the BIOS enabled SSC we can
> @@ -15955,15 +15955,15 @@ int intel_modeset_init(struct drm_device *dev)
> * BIOS isn't using it, don't assume it will work even if the VBT
> * indicates as much.
> */
> - if (HAS_PCH_IBX(dev_priv) || HAS_PCH_CPT(dev_priv)) {
> + if (HAS_PCH_IBX(i915) || HAS_PCH_CPT(i915)) {
> bool bios_lvds_use_ssc = !!(I915_READ(PCH_DREF_CONTROL) &
> DREF_SSC1_ENABLE);
>
> - if (dev_priv->vbt.lvds_use_ssc != bios_lvds_use_ssc) {
> + if (i915->vbt.lvds_use_ssc != bios_lvds_use_ssc) {
> DRM_DEBUG_KMS("SSC %sabled by BIOS, overriding VBT which says %sabled\n",
> bios_lvds_use_ssc ? "en" : "dis",
> - dev_priv->vbt.lvds_use_ssc ? "en" : "dis");
> - dev_priv->vbt.lvds_use_ssc = bios_lvds_use_ssc;
> + i915->vbt.lvds_use_ssc ? "en" : "dis");
> + i915->vbt.lvds_use_ssc = bios_lvds_use_ssc;
> }
> }
>
> @@ -15971,13 +15971,13 @@ int intel_modeset_init(struct drm_device *dev)
> * Maximum framebuffer dimensions, chosen to match
> * the maximum render engine surface size on gen4+.
> */
> - if (INTEL_GEN(dev_priv) >= 7) {
> + if (INTEL_GEN(i915) >= 7) {
> dev->mode_config.max_width = 16384;
> dev->mode_config.max_height = 16384;
> - } else if (INTEL_GEN(dev_priv) >= 4) {
> + } else if (INTEL_GEN(i915) >= 4) {
> dev->mode_config.max_width = 8192;
> dev->mode_config.max_height = 8192;
> - } else if (IS_GEN(dev_priv, 3)) {
> + } else if (IS_GEN(i915, 3)) {
> dev->mode_config.max_width = 4096;
> dev->mode_config.max_height = 4096;
> } else {
> @@ -15985,10 +15985,10 @@ int intel_modeset_init(struct drm_device *dev)
> dev->mode_config.max_height = 2048;
> }
>
> - if (IS_I845G(dev_priv) || IS_I865G(dev_priv)) {
> - dev->mode_config.cursor_width = IS_I845G(dev_priv) ? 64 : 512;
> + if (IS_I845G(i915) || IS_I865G(i915)) {
> + dev->mode_config.cursor_width = IS_I845G(i915) ? 64 : 512;
> dev->mode_config.cursor_height = 1023;
> - } else if (IS_GEN(dev_priv, 2)) {
> + } else if (IS_GEN(i915, 2)) {
> dev->mode_config.cursor_width = 64;
> dev->mode_config.cursor_height = 64;
> } else {
> @@ -15999,11 +15999,11 @@ int intel_modeset_init(struct drm_device *dev)
> dev->mode_config.fb_base = ggtt->gmadr.start;
>
> DRM_DEBUG_KMS("%d display pipe%s available.\n",
> - INTEL_INFO(dev_priv)->num_pipes,
> - INTEL_INFO(dev_priv)->num_pipes > 1 ? "s" : "");
> + INTEL_INFO(i915)->num_pipes,
> + INTEL_INFO(i915)->num_pipes > 1 ? "s" : "");
>
> - for_each_pipe(dev_priv, pipe) {
> - ret = intel_crtc_init(dev_priv, pipe);
> + for_each_pipe(i915, pipe) {
> + ret = intel_crtc_init(i915, pipe);
> if (ret) {
> drm_mode_config_cleanup(dev);
> return ret;
> @@ -16011,19 +16011,19 @@ int intel_modeset_init(struct drm_device *dev)
> }
>
> intel_shared_dpll_init(dev);
> - intel_update_fdi_pll_freq(dev_priv);
> + intel_update_fdi_pll_freq(i915);
>
> - intel_update_czclk(dev_priv);
> + intel_update_czclk(i915);
> intel_modeset_init_hw(dev);
>
> - intel_hdcp_component_init(dev_priv);
> + intel_hdcp_component_init(i915);
>
> - if (dev_priv->max_cdclk_freq == 0)
> - intel_update_max_cdclk(dev_priv);
> + if (i915->max_cdclk_freq == 0)
> + intel_update_max_cdclk(i915);
>
> /* Just disable it once at startup */
> - i915_disable_vga(dev_priv);
> - intel_setup_outputs(dev_priv);
> + i915_disable_vga(i915);
> + intel_setup_outputs(i915);
>
> drm_modeset_lock_all(dev);
> intel_modeset_setup_hw_state(dev, dev->mode_config.acquire_ctx);
> @@ -16042,7 +16042,7 @@ int intel_modeset_init(struct drm_device *dev)
> * can even allow for smooth boot transitions if the BIOS
> * fb is large enough for the active pipe configuration.
> */
> - dev_priv->display.get_initial_plane_config(crtc,
> + i915->display.get_initial_plane_config(crtc,
> &plane_config);
>
> /*
> @@ -16057,7 +16057,7 @@ int intel_modeset_init(struct drm_device *dev)
> * Note that we need to do this after reconstructing the BIOS fb's
> * since the watermark calculation done here will use pstate->fb.
> */
> - if (!HAS_GMCH(dev_priv))
> + if (!HAS_GMCH(i915))
> sanitize_watermarks(dev);
>
> /*
> @@ -16073,9 +16073,9 @@ int intel_modeset_init(struct drm_device *dev)
> return 0;
> }
>
> -void i830_enable_pipe(struct drm_i915_private *dev_priv, enum pipe pipe)
> +void i830_enable_pipe(struct drm_i915_private *i915, enum pipe pipe)
> {
> - struct intel_crtc *crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> + struct intel_crtc *crtc = intel_get_crtc_for_pipe(i915, pipe);
> /* 640x480 at 60Hz, ~25175 kHz */
> struct dpll clock = {
> .m1 = 18,
> @@ -16143,9 +16143,9 @@ void i830_enable_pipe(struct drm_i915_private *dev_priv, enum pipe pipe)
> intel_wait_for_pipe_scanline_moving(crtc);
> }
>
> -void i830_disable_pipe(struct drm_i915_private *dev_priv, enum pipe pipe)
> +void i830_disable_pipe(struct drm_i915_private *i915, enum pipe pipe)
> {
> - struct intel_crtc *crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> + struct intel_crtc *crtc = intel_get_crtc_for_pipe(i915, pipe);
>
> DRM_DEBUG_KMS("disabling pipe %c due to force quirk\n",
> pipe_name(pipe));
> @@ -16166,14 +16166,14 @@ void i830_disable_pipe(struct drm_i915_private *dev_priv, enum pipe pipe)
> }
>
> static void
> -intel_sanitize_plane_mapping(struct drm_i915_private *dev_priv)
> +intel_sanitize_plane_mapping(struct drm_i915_private *i915)
> {
> struct intel_crtc *crtc;
>
> - if (INTEL_GEN(dev_priv) >= 4)
> + if (INTEL_GEN(i915) >= 4)
> return;
>
> - for_each_intel_crtc(&dev_priv->drm, crtc) {
> + for_each_intel_crtc(&i915->drm, crtc) {
> struct intel_plane *plane =
> to_intel_plane(crtc->base.primary);
> struct intel_crtc *plane_crtc;
> @@ -16188,7 +16188,7 @@ intel_sanitize_plane_mapping(struct drm_i915_private *dev_priv)
> DRM_DEBUG_KMS("[PLANE:%d:%s] attached to the wrong pipe, disabling plane\n",
> plane->base.base.id, plane->base.name);
>
> - plane_crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> + plane_crtc = intel_get_crtc_for_pipe(i915, pipe);
> intel_plane_disable_noatomic(plane_crtc, plane);
> }
> }
> @@ -16215,18 +16215,18 @@ static struct intel_connector *intel_encoder_find_connector(struct intel_encoder
> return NULL;
> }
>
> -static bool has_pch_trancoder(struct drm_i915_private *dev_priv,
> +static bool has_pch_trancoder(struct drm_i915_private *i915,
> enum pipe pch_transcoder)
> {
> - return HAS_PCH_IBX(dev_priv) || HAS_PCH_CPT(dev_priv) ||
> - (HAS_PCH_LPT_H(dev_priv) && pch_transcoder == PIPE_A);
> + return HAS_PCH_IBX(i915) || HAS_PCH_CPT(i915) ||
> + (HAS_PCH_LPT_H(i915) && pch_transcoder == PIPE_A);
> }
>
> static void intel_sanitize_crtc(struct intel_crtc *crtc,
> struct drm_modeset_acquire_ctx *ctx)
> {
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_crtc_state *crtc_state = to_intel_crtc_state(crtc->base.state);
> enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
>
> @@ -16255,7 +16255,7 @@ static void intel_sanitize_crtc(struct intel_crtc *crtc,
> * Disable any background color set by the BIOS, but enable the
> * gamma and CSC to match how we program our planes.
> */
> - if (INTEL_GEN(dev_priv) >= 9)
> + if (INTEL_GEN(i915) >= 9)
> I915_WRITE(SKL_BOTTOM_COLOR(crtc->pipe),
> SKL_BOTTOM_COLOR_GAMMA_ENABLE |
> SKL_BOTTOM_COLOR_CSC_ENABLE);
> @@ -16266,7 +16266,7 @@ static void intel_sanitize_crtc(struct intel_crtc *crtc,
> if (crtc_state->base.active && !intel_crtc_has_encoders(crtc))
> intel_crtc_disable_noatomic(&crtc->base, ctx);
>
> - if (crtc_state->base.active || HAS_GMCH(dev_priv)) {
> + if (crtc_state->base.active || HAS_GMCH(i915)) {
> /*
> * We start out with underrun reporting disabled to avoid races.
> * For correct bookkeeping mark this on active crtcs.
> @@ -16290,14 +16290,14 @@ static void intel_sanitize_crtc(struct intel_crtc *crtc,
> * PCH transcoders B and C would prevent enabling the south
> * error interrupt (see cpt_can_enable_serr_int()).
> */
> - if (has_pch_trancoder(dev_priv, crtc->pipe))
> + if (has_pch_trancoder(i915, crtc->pipe))
> crtc->pch_fifo_underrun_disabled = true;
> }
> }
>
> static bool has_bogus_dpll_config(const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
>
> /*
> * Some SNB BIOSen (eg. ASUS K53SV) are known to misprogram
> @@ -16309,7 +16309,7 @@ static bool has_bogus_dpll_config(const struct intel_crtc_state *crtc_state)
> * without several WARNs, but for now let's take the easy
> * road.
> */
> - return IS_GEN(dev_priv, 6) &&
> + return IS_GEN(i915, 6) &&
> crtc_state->base.active &&
> crtc_state->shared_dpll &&
> crtc_state->port_clock == 0;
> @@ -16317,7 +16317,7 @@ static bool has_bogus_dpll_config(const struct intel_crtc_state *crtc_state)
>
> static void intel_sanitize_encoder(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_connector *connector;
> struct intel_crtc *crtc = to_intel_crtc(encoder->base.crtc);
> struct intel_crtc_state *crtc_state = crtc ?
> @@ -16378,21 +16378,21 @@ static void intel_sanitize_encoder(struct intel_encoder *encoder)
> /* notify opregion of the sanitized encoder state */
> intel_opregion_notify_encoder(encoder, connector && has_active_crtc);
>
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> icl_sanitize_encoder_pll_mapping(encoder);
> }
>
> -void i915_redisable_vga_power_on(struct drm_i915_private *dev_priv)
> +void i915_redisable_vga_power_on(struct drm_i915_private *i915)
> {
> - i915_reg_t vga_reg = i915_vgacntrl_reg(dev_priv);
> + i915_reg_t vga_reg = i915_vgacntrl_reg(i915);
>
> if (!(I915_READ(vga_reg) & VGA_DISP_DISABLE)) {
> DRM_DEBUG_KMS("Something enabled VGA plane, disabling it\n");
> - i915_disable_vga(dev_priv);
> + i915_disable_vga(i915);
> }
> }
>
> -void i915_redisable_vga(struct drm_i915_private *dev_priv)
> +void i915_redisable_vga(struct drm_i915_private *i915)
> {
> intel_wakeref_t wakeref;
>
> @@ -16405,23 +16405,23 @@ void i915_redisable_vga(struct drm_i915_private *dev_priv)
> * follow the "don't touch the power well if we don't need it" policy
> * the rest of the driver uses.
> */
> - wakeref = intel_display_power_get_if_enabled(dev_priv,
> + wakeref = intel_display_power_get_if_enabled(i915,
> POWER_DOMAIN_VGA);
> if (!wakeref)
> return;
>
> - i915_redisable_vga_power_on(dev_priv);
> + i915_redisable_vga_power_on(i915);
>
> - intel_display_power_put(dev_priv, POWER_DOMAIN_VGA, wakeref);
> + intel_display_power_put(i915, POWER_DOMAIN_VGA, wakeref);
> }
>
> /* FIXME read out full plane state for all planes */
> -static void readout_plane_state(struct drm_i915_private *dev_priv)
> +static void readout_plane_state(struct drm_i915_private *i915)
> {
> struct intel_plane *plane;
> struct intel_crtc *crtc;
>
> - for_each_intel_plane(&dev_priv->drm, plane) {
> + for_each_intel_plane(&i915->drm, plane) {
> struct intel_plane_state *plane_state =
> to_intel_plane_state(plane->base.state);
> struct intel_crtc_state *crtc_state;
> @@ -16430,7 +16430,7 @@ static void readout_plane_state(struct drm_i915_private *dev_priv)
>
> visible = plane->get_hw_state(plane, &pipe);
>
> - crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> + crtc = intel_get_crtc_for_pipe(i915, pipe);
> crtc_state = to_intel_crtc_state(crtc->base.state);
>
> intel_set_plane_visible(crtc_state, plane_state, visible);
> @@ -16440,7 +16440,7 @@ static void readout_plane_state(struct drm_i915_private *dev_priv)
> enableddisabled(visible), pipe_name(pipe));
> }
>
> - for_each_intel_crtc(&dev_priv->drm, crtc) {
> + for_each_intel_crtc(&i915->drm, crtc) {
> struct intel_crtc_state *crtc_state =
> to_intel_crtc_state(crtc->base.state);
>
> @@ -16450,7 +16450,7 @@ static void readout_plane_state(struct drm_i915_private *dev_priv)
>
> static void intel_modeset_readout_hw_state(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> enum pipe pipe;
> struct intel_crtc *crtc;
> struct intel_encoder *encoder;
> @@ -16458,7 +16458,7 @@ static void intel_modeset_readout_hw_state(struct drm_device *dev)
> struct drm_connector_list_iter conn_iter;
> int i;
>
> - dev_priv->active_crtcs = 0;
> + i915->active_crtcs = 0;
>
> for_each_intel_crtc(dev, crtc) {
> struct intel_crtc_state *crtc_state =
> @@ -16469,25 +16469,25 @@ static void intel_modeset_readout_hw_state(struct drm_device *dev)
> __drm_atomic_helper_crtc_reset(&crtc->base, &crtc_state->base);
>
> crtc_state->base.active = crtc_state->base.enable =
> - dev_priv->display.get_pipe_config(crtc, crtc_state);
> + i915->display.get_pipe_config(crtc, crtc_state);
>
> crtc->base.enabled = crtc_state->base.enable;
> crtc->active = crtc_state->base.active;
>
> if (crtc_state->base.active)
> - dev_priv->active_crtcs |= 1 << crtc->pipe;
> + i915->active_crtcs |= 1 << crtc->pipe;
>
> DRM_DEBUG_KMS("[CRTC:%d:%s] hw state readout: %s\n",
> crtc->base.base.id, crtc->base.name,
> enableddisabled(crtc_state->base.active));
> }
>
> - readout_plane_state(dev_priv);
> + readout_plane_state(i915);
>
> - for (i = 0; i < dev_priv->num_shared_dpll; i++) {
> - struct intel_shared_dpll *pll = &dev_priv->shared_dplls[i];
> + for (i = 0; i < i915->num_shared_dpll; i++) {
> + struct intel_shared_dpll *pll = &i915->shared_dplls[i];
>
> - pll->on = pll->info->funcs->get_hw_state(dev_priv, pll,
> + pll->on = pll->info->funcs->get_hw_state(i915, pll,
> &pll->state.hw_state);
> pll->state.crtc_mask = 0;
> for_each_intel_crtc(dev, crtc) {
> @@ -16510,7 +16510,7 @@ static void intel_modeset_readout_hw_state(struct drm_device *dev)
> if (encoder->get_hw_state(encoder, &pipe)) {
> struct intel_crtc_state *crtc_state;
>
> - crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> + crtc = intel_get_crtc_for_pipe(i915, pipe);
> crtc_state = to_intel_crtc_state(crtc->base.state);
>
> encoder->base.crtc = &crtc->base;
> @@ -16558,7 +16558,7 @@ static void intel_modeset_readout_hw_state(struct drm_device *dev)
>
> for_each_intel_crtc(dev, crtc) {
> struct intel_bw_state *bw_state =
> - to_intel_bw_state(dev_priv->bw_obj.state);
> + to_intel_bw_state(i915->bw_obj.state);
> struct intel_crtc_state *crtc_state =
> to_intel_crtc_state(crtc->base.state);
> struct intel_plane *plane;
> @@ -16585,7 +16585,7 @@ static void intel_modeset_readout_hw_state(struct drm_device *dev)
>
> intel_crtc_compute_pixel_rate(crtc_state);
>
> - if (dev_priv->display.modeset_calc_cdclk) {
> + if (i915->display.modeset_calc_cdclk) {
> min_cdclk = intel_crtc_compute_min_cdclk(crtc_state);
> if (WARN_ON(min_cdclk < 0))
> min_cdclk = 0;
> @@ -16596,11 +16596,11 @@ static void intel_modeset_readout_hw_state(struct drm_device *dev)
> update_scanline_offset(crtc_state);
> }
>
> - dev_priv->min_cdclk[crtc->pipe] = min_cdclk;
> - dev_priv->min_voltage_level[crtc->pipe] =
> + i915->min_cdclk[crtc->pipe] = min_cdclk;
> + i915->min_voltage_level[crtc->pipe] =
> crtc_state->min_voltage_level;
>
> - for_each_intel_plane_on_crtc(&dev_priv->drm, crtc, plane) {
> + for_each_intel_plane_on_crtc(&i915->drm, crtc, plane) {
> const struct intel_plane_state *plane_state =
> to_intel_plane_state(plane->base.state);
>
> @@ -16615,16 +16615,16 @@ static void intel_modeset_readout_hw_state(struct drm_device *dev)
>
> intel_bw_crtc_update(bw_state, crtc_state);
>
> - intel_pipe_config_sanity_check(dev_priv, crtc_state);
> + intel_pipe_config_sanity_check(i915, crtc_state);
> }
> }
>
> static void
> -get_encoder_power_domains(struct drm_i915_private *dev_priv)
> +get_encoder_power_domains(struct drm_i915_private *i915)
> {
> struct intel_encoder *encoder;
>
> - for_each_intel_encoder(&dev_priv->drm, encoder) {
> + for_each_intel_encoder(&i915->drm, encoder) {
> struct intel_crtc_state *crtc_state;
>
> if (!encoder->get_power_domains)
> @@ -16642,14 +16642,14 @@ get_encoder_power_domains(struct drm_i915_private *dev_priv)
> }
> }
>
> -static void intel_early_display_was(struct drm_i915_private *dev_priv)
> +static void intel_early_display_was(struct drm_i915_private *i915)
> {
> /* Display WA #1185 WaDisableDARBFClkGating:cnl,glk */
> - if (IS_CANNONLAKE(dev_priv) || IS_GEMINILAKE(dev_priv))
> + if (IS_CANNONLAKE(i915) || IS_GEMINILAKE(i915))
> I915_WRITE(GEN9_CLKGATE_DIS_0, I915_READ(GEN9_CLKGATE_DIS_0) |
> DARBF_GATING_DIS);
>
> - if (IS_HASWELL(dev_priv)) {
> + if (IS_HASWELL(i915)) {
> /*
> * WaRsPkgCStateDisplayPMReq:hsw
> * System hang if this isn't done before disabling all planes!
> @@ -16659,7 +16659,7 @@ static void intel_early_display_was(struct drm_i915_private *dev_priv)
> }
> }
>
> -static void ibx_sanitize_pch_hdmi_port(struct drm_i915_private *dev_priv,
> +static void ibx_sanitize_pch_hdmi_port(struct drm_i915_private *i915,
> enum port port, i915_reg_t hdmi_reg)
> {
> u32 val = I915_READ(hdmi_reg);
> @@ -16677,7 +16677,7 @@ static void ibx_sanitize_pch_hdmi_port(struct drm_i915_private *dev_priv,
> I915_WRITE(hdmi_reg, val);
> }
>
> -static void ibx_sanitize_pch_dp_port(struct drm_i915_private *dev_priv,
> +static void ibx_sanitize_pch_dp_port(struct drm_i915_private *i915,
> enum port port, i915_reg_t dp_reg)
> {
> u32 val = I915_READ(dp_reg);
> @@ -16695,7 +16695,7 @@ static void ibx_sanitize_pch_dp_port(struct drm_i915_private *dev_priv,
> I915_WRITE(dp_reg, val);
> }
>
> -static void ibx_sanitize_pch_ports(struct drm_i915_private *dev_priv)
> +static void ibx_sanitize_pch_ports(struct drm_i915_private *i915)
> {
> /*
> * The BIOS may select transcoder B on some of the PCH
> @@ -16708,14 +16708,14 @@ static void ibx_sanitize_pch_ports(struct drm_i915_private *dev_priv)
> * (see. intel_dp_link_down(), intel_disable_hdmi(),
> * intel_disable_sdvo()).
> */
> - ibx_sanitize_pch_dp_port(dev_priv, PORT_B, PCH_DP_B);
> - ibx_sanitize_pch_dp_port(dev_priv, PORT_C, PCH_DP_C);
> - ibx_sanitize_pch_dp_port(dev_priv, PORT_D, PCH_DP_D);
> + ibx_sanitize_pch_dp_port(i915, PORT_B, PCH_DP_B);
> + ibx_sanitize_pch_dp_port(i915, PORT_C, PCH_DP_C);
> + ibx_sanitize_pch_dp_port(i915, PORT_D, PCH_DP_D);
>
> /* PCH SDVOB multiplex with HDMIB */
> - ibx_sanitize_pch_hdmi_port(dev_priv, PORT_B, PCH_HDMIB);
> - ibx_sanitize_pch_hdmi_port(dev_priv, PORT_C, PCH_HDMIC);
> - ibx_sanitize_pch_hdmi_port(dev_priv, PORT_D, PCH_HDMID);
> + ibx_sanitize_pch_hdmi_port(i915, PORT_B, PCH_HDMIB);
> + ibx_sanitize_pch_hdmi_port(i915, PORT_C, PCH_HDMIC);
> + ibx_sanitize_pch_hdmi_port(i915, PORT_D, PCH_HDMID);
> }
>
> /* Scan out the current hw modeset state,
> @@ -16725,29 +16725,29 @@ static void
> intel_modeset_setup_hw_state(struct drm_device *dev,
> struct drm_modeset_acquire_ctx *ctx)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_crtc_state *crtc_state;
> struct intel_encoder *encoder;
> struct intel_crtc *crtc;
> intel_wakeref_t wakeref;
> int i;
>
> - wakeref = intel_display_power_get(dev_priv, POWER_DOMAIN_INIT);
> + wakeref = intel_display_power_get(i915, POWER_DOMAIN_INIT);
>
> - intel_early_display_was(dev_priv);
> + intel_early_display_was(i915);
> intel_modeset_readout_hw_state(dev);
>
> /* HW state is read out, now we need to sanitize this mess. */
> - get_encoder_power_domains(dev_priv);
> + get_encoder_power_domains(i915);
>
> - if (HAS_PCH_IBX(dev_priv))
> - ibx_sanitize_pch_ports(dev_priv);
> + if (HAS_PCH_IBX(i915))
> + ibx_sanitize_pch_ports(i915);
>
> /*
> * intel_sanitize_plane_mapping() may need to do vblank
> * waits, so we need vblank interrupts restored beforehand.
> */
> - for_each_intel_crtc(&dev_priv->drm, crtc) {
> + for_each_intel_crtc(&i915->drm, crtc) {
> crtc_state = to_intel_crtc_state(crtc->base.state);
>
> drm_crtc_vblank_reset(&crtc->base);
> @@ -16756,12 +16756,12 @@ intel_modeset_setup_hw_state(struct drm_device *dev,
> intel_crtc_vblank_on(crtc_state);
> }
>
> - intel_sanitize_plane_mapping(dev_priv);
> + intel_sanitize_plane_mapping(i915);
>
> for_each_intel_encoder(dev, encoder)
> intel_sanitize_encoder(encoder);
>
> - for_each_intel_crtc(&dev_priv->drm, crtc) {
> + for_each_intel_crtc(&i915->drm, crtc) {
> crtc_state = to_intel_crtc_state(crtc->base.state);
> intel_sanitize_crtc(crtc, ctx);
> intel_dump_pipe_config(crtc_state, NULL, "[setup_hw_state]");
> @@ -16769,8 +16769,8 @@ intel_modeset_setup_hw_state(struct drm_device *dev,
>
> intel_modeset_update_connector_atomic_state(dev);
>
> - for (i = 0; i < dev_priv->num_shared_dpll; i++) {
> - struct intel_shared_dpll *pll = &dev_priv->shared_dplls[i];
> + for (i = 0; i < i915->num_shared_dpll; i++) {
> + struct intel_shared_dpll *pll = &i915->shared_dplls[i];
>
> if (!pll->on || pll->active_mask)
> continue;
> @@ -16778,20 +16778,20 @@ intel_modeset_setup_hw_state(struct drm_device *dev,
> DRM_DEBUG_KMS("%s enabled but not in use, disabling\n",
> pll->info->name);
>
> - pll->info->funcs->disable(dev_priv, pll);
> + pll->info->funcs->disable(i915, pll);
> pll->on = false;
> }
>
> - if (IS_G4X(dev_priv)) {
> - g4x_wm_get_hw_state(dev_priv);
> - g4x_wm_sanitize(dev_priv);
> - } else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> - vlv_wm_get_hw_state(dev_priv);
> - vlv_wm_sanitize(dev_priv);
> - } else if (INTEL_GEN(dev_priv) >= 9) {
> - skl_wm_get_hw_state(dev_priv);
> - } else if (HAS_PCH_SPLIT(dev_priv)) {
> - ilk_wm_get_hw_state(dev_priv);
> + if (IS_G4X(i915)) {
> + g4x_wm_get_hw_state(i915);
> + g4x_wm_sanitize(i915);
> + } else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> + vlv_wm_get_hw_state(i915);
> + vlv_wm_sanitize(i915);
> + } else if (INTEL_GEN(i915) >= 9) {
> + skl_wm_get_hw_state(i915);
> + } else if (HAS_PCH_SPLIT(i915)) {
> + ilk_wm_get_hw_state(i915);
> }
>
> for_each_intel_crtc(dev, crtc) {
> @@ -16800,22 +16800,22 @@ intel_modeset_setup_hw_state(struct drm_device *dev,
> crtc_state = to_intel_crtc_state(crtc->base.state);
> put_domains = modeset_get_crtc_power_domains(&crtc->base, crtc_state);
> if (WARN_ON(put_domains))
> - modeset_put_power_domains(dev_priv, put_domains);
> + modeset_put_power_domains(i915, put_domains);
> }
>
> - intel_display_power_put(dev_priv, POWER_DOMAIN_INIT, wakeref);
> + intel_display_power_put(i915, POWER_DOMAIN_INIT, wakeref);
>
> - intel_fbc_init_pipe_state(dev_priv);
> + intel_fbc_init_pipe_state(i915);
> }
>
> void intel_display_resume(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct drm_atomic_state *state = dev_priv->modeset_restore_state;
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct drm_atomic_state *state = i915->modeset_restore_state;
> struct drm_modeset_acquire_ctx ctx;
> int ret;
>
> - dev_priv->modeset_restore_state = NULL;
> + i915->modeset_restore_state = NULL;
> if (state)
> state->acquire_ctx = &ctx;
>
> @@ -16832,7 +16832,7 @@ void intel_display_resume(struct drm_device *dev)
> if (!ret)
> ret = __intel_display_resume(dev, state, &ctx);
>
> - intel_enable_ipc(dev_priv);
> + intel_enable_ipc(i915);
> drm_modeset_drop_locks(&ctx);
> drm_modeset_acquire_fini(&ctx);
>
> @@ -16862,19 +16862,19 @@ static void intel_hpd_poll_fini(struct drm_device *dev)
>
> void intel_modeset_cleanup(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
>
> - flush_workqueue(dev_priv->modeset_wq);
> + flush_workqueue(i915->modeset_wq);
>
> - flush_work(&dev_priv->atomic_helper.free_work);
> - WARN_ON(!llist_empty(&dev_priv->atomic_helper.free_list));
> + flush_work(&i915->atomic_helper.free_work);
> + WARN_ON(!llist_empty(&i915->atomic_helper.free_list));
>
> /*
> * Interrupts and polling as the first thing to avoid creating havoc.
> * Too much stuff here (turning of connectors, ...) would
> * experience fancy races otherwise.
> */
> - intel_irq_uninstall(dev_priv);
> + intel_irq_uninstall(i915);
>
> /*
> * Due to the hpd irq storm handling the hotplug work can re-arm the
> @@ -16883,37 +16883,37 @@ void intel_modeset_cleanup(struct drm_device *dev)
> intel_hpd_poll_fini(dev);
>
> /* poll work can call into fbdev, hence clean that up afterwards */
> - intel_fbdev_fini(dev_priv);
> + intel_fbdev_fini(i915);
>
> intel_unregister_dsm_handler();
>
> - intel_fbc_global_disable(dev_priv);
> + intel_fbc_global_disable(i915);
>
> /* flush any delayed tasks or pending work */
> flush_scheduled_work();
>
> - intel_hdcp_component_fini(dev_priv);
> + intel_hdcp_component_fini(i915);
>
> drm_mode_config_cleanup(dev);
>
> - intel_overlay_cleanup(dev_priv);
> + intel_overlay_cleanup(i915);
>
> - intel_gmbus_teardown(dev_priv);
> + intel_gmbus_teardown(i915);
>
> - destroy_workqueue(dev_priv->modeset_wq);
> + destroy_workqueue(i915->modeset_wq);
>
> - intel_fbc_cleanup_cfb(dev_priv);
> + intel_fbc_cleanup_cfb(i915);
> }
>
> /*
> * set vga decode state - true == enable VGA decode
> */
> -int intel_modeset_vga_set_state(struct drm_i915_private *dev_priv, bool state)
> +int intel_modeset_vga_set_state(struct drm_i915_private *i915, bool state)
> {
> - unsigned reg = INTEL_GEN(dev_priv) >= 6 ? SNB_GMCH_CTRL : INTEL_GMCH_CTRL;
> + unsigned reg = INTEL_GEN(i915) >= 6 ? SNB_GMCH_CTRL : INTEL_GMCH_CTRL;
> u16 gmch_ctrl;
>
> - if (pci_read_config_word(dev_priv->bridge_dev, reg, &gmch_ctrl)) {
> + if (pci_read_config_word(i915->bridge_dev, reg, &gmch_ctrl)) {
> DRM_ERROR("failed to read control word\n");
> return -EIO;
> }
> @@ -16926,7 +16926,7 @@ int intel_modeset_vga_set_state(struct drm_i915_private *dev_priv, bool state)
> else
> gmch_ctrl |= INTEL_GMCH_VGA_DISABLE;
>
> - if (pci_write_config_word(dev_priv->bridge_dev, reg, gmch_ctrl)) {
> + if (pci_write_config_word(i915->bridge_dev, reg, gmch_ctrl)) {
> DRM_ERROR("failed to write control word\n");
> return -EIO;
> }
> @@ -16980,7 +16980,7 @@ struct intel_display_error_state {
> };
>
> struct intel_display_error_state *
> -intel_display_capture_error_state(struct drm_i915_private *dev_priv)
> +intel_display_capture_error_state(struct drm_i915_private *i915)
> {
> struct intel_display_error_state *error;
> int transcoders[] = {
> @@ -16993,19 +16993,19 @@ intel_display_capture_error_state(struct drm_i915_private *dev_priv)
>
> BUILD_BUG_ON(ARRAY_SIZE(transcoders) != ARRAY_SIZE(error->transcoder));
>
> - if (!HAS_DISPLAY(dev_priv))
> + if (!HAS_DISPLAY(i915))
> return NULL;
>
> error = kzalloc(sizeof(*error), GFP_ATOMIC);
> if (error == NULL)
> return NULL;
>
> - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> + if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> error->power_well_driver = I915_READ(HSW_PWR_WELL_CTL2);
>
> - for_each_pipe(dev_priv, i) {
> + for_each_pipe(i915, i) {
> error->pipe[i].power_domain_on =
> - __intel_display_power_is_enabled(dev_priv,
> + __intel_display_power_is_enabled(i915,
> POWER_DOMAIN_PIPE(i));
> if (!error->pipe[i].power_domain_on)
> continue;
> @@ -17016,32 +17016,32 @@ intel_display_capture_error_state(struct drm_i915_private *dev_priv)
>
> error->plane[i].control = I915_READ(DSPCNTR(i));
> error->plane[i].stride = I915_READ(DSPSTRIDE(i));
> - if (INTEL_GEN(dev_priv) <= 3) {
> + if (INTEL_GEN(i915) <= 3) {
> error->plane[i].size = I915_READ(DSPSIZE(i));
> error->plane[i].pos = I915_READ(DSPPOS(i));
> }
> - if (INTEL_GEN(dev_priv) <= 7 && !IS_HASWELL(dev_priv))
> + if (INTEL_GEN(i915) <= 7 && !IS_HASWELL(i915))
> error->plane[i].addr = I915_READ(DSPADDR(i));
> - if (INTEL_GEN(dev_priv) >= 4) {
> + if (INTEL_GEN(i915) >= 4) {
> error->plane[i].surface = I915_READ(DSPSURF(i));
> error->plane[i].tile_offset = I915_READ(DSPTILEOFF(i));
> }
>
> error->pipe[i].source = I915_READ(PIPESRC(i));
>
> - if (HAS_GMCH(dev_priv))
> + if (HAS_GMCH(i915))
> error->pipe[i].stat = I915_READ(PIPESTAT(i));
> }
>
> for (i = 0; i < ARRAY_SIZE(error->transcoder); i++) {
> enum transcoder cpu_transcoder = transcoders[i];
>
> - if (!INTEL_INFO(dev_priv)->trans_offsets[cpu_transcoder])
> + if (!INTEL_INFO(i915)->trans_offsets[cpu_transcoder])
> continue;
>
> error->transcoder[i].available = true;
> error->transcoder[i].power_domain_on =
> - __intel_display_power_is_enabled(dev_priv,
> + __intel_display_power_is_enabled(i915,
> POWER_DOMAIN_TRANSCODER(cpu_transcoder));
> if (!error->transcoder[i].power_domain_on)
> continue;
> @@ -17066,17 +17066,17 @@ void
> intel_display_print_error_state(struct drm_i915_error_state_buf *m,
> struct intel_display_error_state *error)
> {
> - struct drm_i915_private *dev_priv = m->i915;
> + struct drm_i915_private *i915 = m->i915;
> int i;
>
> if (!error)
> return;
>
> - err_printf(m, "Num Pipes: %d\n", INTEL_INFO(dev_priv)->num_pipes);
> - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> + err_printf(m, "Num Pipes: %d\n", INTEL_INFO(i915)->num_pipes);
> + if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> err_printf(m, "PWR_WELL_CTL2: %08x\n",
> error->power_well_driver);
> - for_each_pipe(dev_priv, i) {
> + for_each_pipe(i915, i) {
> err_printf(m, "Pipe [%d]:\n", i);
> err_printf(m, " Power: %s\n",
> onoff(error->pipe[i].power_domain_on));
> @@ -17086,13 +17086,13 @@ intel_display_print_error_state(struct drm_i915_error_state_buf *m,
> err_printf(m, "Plane [%d]:\n", i);
> err_printf(m, " CNTR: %08x\n", error->plane[i].control);
> err_printf(m, " STRIDE: %08x\n", error->plane[i].stride);
> - if (INTEL_GEN(dev_priv) <= 3) {
> + if (INTEL_GEN(i915) <= 3) {
> err_printf(m, " SIZE: %08x\n", error->plane[i].size);
> err_printf(m, " POS: %08x\n", error->plane[i].pos);
> }
> - if (INTEL_GEN(dev_priv) <= 7 && !IS_HASWELL(dev_priv))
> + if (INTEL_GEN(i915) <= 7 && !IS_HASWELL(i915))
> err_printf(m, " ADDR: %08x\n", error->plane[i].addr);
> - if (INTEL_GEN(dev_priv) >= 4) {
> + if (INTEL_GEN(i915) >= 4) {
> err_printf(m, " SURF: %08x\n", error->plane[i].surface);
> err_printf(m, " TILEOFF: %08x\n", error->plane[i].tile_offset);
> }
> diff --git a/drivers/gpu/drm/i915/intel_display.h b/drivers/gpu/drm/i915/intel_display.h
> index ee6b8194a459..7e4b4b1c0ac6 100644
> --- a/drivers/gpu/drm/i915/intel_display.h
> +++ b/drivers/gpu/drm/i915/intel_display.h
> @@ -125,7 +125,7 @@ enum i9xx_plane_id {
> };
>
> #define plane_name(p) ((p) + 'A')
> -#define sprite_name(p, s) ((p) * RUNTIME_INFO(dev_priv)->num_sprites[(p)] + (s) + 'A')
> +#define sprite_name(p, s) ((p) * RUNTIME_INFO(i915)->num_sprites[(p)] + (s) + 'A')
>
> /*
> * Per-pipe plane identifier.
> @@ -353,8 +353,8 @@ void intel_link_compute_m_n(u16 bpp, int nlanes,
> struct intel_link_m_n *m_n,
> bool constant_n);
> bool is_ccs_modifier(u64 modifier);
> -void lpt_disable_clkout_dp(struct drm_i915_private *dev_priv);
> -u32 intel_plane_fb_max_stride(struct drm_i915_private *dev_priv,
> +void lpt_disable_clkout_dp(struct drm_i915_private *i915);
> +u32 intel_plane_fb_max_stride(struct drm_i915_private *i915,
> u32 pixel_format, u64 modifier);
> bool intel_plane_can_remap(const struct intel_plane_state *plane_state);
>
> diff --git a/drivers/gpu/drm/i915/intel_display_power.c b/drivers/gpu/drm/i915/intel_display_power.c
> index bb9ef1cea5db..9acfad5c83ef 100644
> --- a/drivers/gpu/drm/i915/intel_display_power.c
> +++ b/drivers/gpu/drm/i915/intel_display_power.c
> @@ -17,7 +17,7 @@
> #include "intel_hotplug.h"
> #include "intel_sideband.h"
>
> -bool intel_display_power_well_is_enabled(struct drm_i915_private *dev_priv,
> +bool intel_display_power_well_is_enabled(struct drm_i915_private *i915,
> enum i915_power_well_id power_well_id);
>
> const char *
> @@ -122,42 +122,42 @@ intel_display_power_domain_str(enum intel_display_power_domain domain)
> }
> }
>
> -static void intel_power_well_enable(struct drm_i915_private *dev_priv,
> +static void intel_power_well_enable(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> DRM_DEBUG_KMS("enabling %s\n", power_well->desc->name);
> - power_well->desc->ops->enable(dev_priv, power_well);
> + power_well->desc->ops->enable(i915, power_well);
> power_well->hw_enabled = true;
> }
>
> -static void intel_power_well_disable(struct drm_i915_private *dev_priv,
> +static void intel_power_well_disable(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> DRM_DEBUG_KMS("disabling %s\n", power_well->desc->name);
> power_well->hw_enabled = false;
> - power_well->desc->ops->disable(dev_priv, power_well);
> + power_well->desc->ops->disable(i915, power_well);
> }
>
> -static void intel_power_well_get(struct drm_i915_private *dev_priv,
> +static void intel_power_well_get(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> if (!power_well->count++)
> - intel_power_well_enable(dev_priv, power_well);
> + intel_power_well_enable(i915, power_well);
> }
>
> -static void intel_power_well_put(struct drm_i915_private *dev_priv,
> +static void intel_power_well_put(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> WARN(!power_well->count, "Use count on power well %s is already zero",
> power_well->desc->name);
>
> if (!--power_well->count)
> - intel_power_well_disable(dev_priv, power_well);
> + intel_power_well_disable(i915, power_well);
> }
>
> /**
> * __intel_display_power_is_enabled - unlocked check for a power domain
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> * @domain: power domain to check
> *
> * This is the unlocked version of intel_display_power_is_enabled() and should
> @@ -167,18 +167,18 @@ static void intel_power_well_put(struct drm_i915_private *dev_priv,
> * Returns:
> * True when the power domain is enabled, false otherwise.
> */
> -bool __intel_display_power_is_enabled(struct drm_i915_private *dev_priv,
> +bool __intel_display_power_is_enabled(struct drm_i915_private *i915,
> enum intel_display_power_domain domain)
> {
> struct i915_power_well *power_well;
> bool is_enabled;
>
> - if (dev_priv->runtime_pm.suspended)
> + if (i915->runtime_pm.suspended)
> return false;
>
> is_enabled = true;
>
> - for_each_power_domain_well_reverse(dev_priv, power_well, BIT_ULL(domain)) {
> + for_each_power_domain_well_reverse(i915, power_well, BIT_ULL(domain)) {
> if (power_well->desc->always_on)
> continue;
>
> @@ -193,7 +193,7 @@ bool __intel_display_power_is_enabled(struct drm_i915_private *dev_priv,
>
> /**
> * intel_display_power_is_enabled - check for a power domain
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> * @domain: power domain to check
> *
> * This function can be used to check the hw power domain state. It is mostly
> @@ -208,16 +208,16 @@ bool __intel_display_power_is_enabled(struct drm_i915_private *dev_priv,
> * Returns:
> * True when the power domain is enabled, false otherwise.
> */
> -bool intel_display_power_is_enabled(struct drm_i915_private *dev_priv,
> +bool intel_display_power_is_enabled(struct drm_i915_private *i915,
> enum intel_display_power_domain domain)
> {
> struct i915_power_domains *power_domains;
> bool ret;
>
> - power_domains = &dev_priv->power_domains;
> + power_domains = &i915->power_domains;
>
> mutex_lock(&power_domains->lock);
> - ret = __intel_display_power_is_enabled(dev_priv, domain);
> + ret = __intel_display_power_is_enabled(i915, domain);
> mutex_unlock(&power_domains->lock);
>
> return ret;
> @@ -229,10 +229,10 @@ bool intel_display_power_is_enabled(struct drm_i915_private *dev_priv,
> * to be enabled, and it will only be disabled if none of the registers is
> * requesting it to be enabled.
> */
> -static void hsw_power_well_post_enable(struct drm_i915_private *dev_priv,
> +static void hsw_power_well_post_enable(struct drm_i915_private *i915,
> u8 irq_pipe_mask, bool has_vga)
> {
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct pci_dev *pdev = i915->drm.pdev;
>
> /*
> * After we re-enable the power well, if we touch VGA register 0x3d5
> @@ -251,31 +251,31 @@ static void hsw_power_well_post_enable(struct drm_i915_private *dev_priv,
> }
>
> if (irq_pipe_mask)
> - gen8_irq_power_well_post_enable(dev_priv, irq_pipe_mask);
> + gen8_irq_power_well_post_enable(i915, irq_pipe_mask);
> }
>
> -static void hsw_power_well_pre_disable(struct drm_i915_private *dev_priv,
> +static void hsw_power_well_pre_disable(struct drm_i915_private *i915,
> u8 irq_pipe_mask)
> {
> if (irq_pipe_mask)
> - gen8_irq_power_well_pre_disable(dev_priv, irq_pipe_mask);
> + gen8_irq_power_well_pre_disable(i915, irq_pipe_mask);
> }
>
> -static void hsw_wait_for_power_well_enable(struct drm_i915_private *dev_priv,
> +static void hsw_wait_for_power_well_enable(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> const struct i915_power_well_regs *regs = power_well->desc->hsw.regs;
> int pw_idx = power_well->desc->hsw.idx;
>
> /* Timeout for PW1:10 us, AUX:not specified, other PWs:20 us. */
> - WARN_ON(intel_wait_for_register(&dev_priv->uncore,
> + WARN_ON(intel_wait_for_register(&i915->uncore,
> regs->driver,
> HSW_PWR_WELL_CTL_STATE(pw_idx),
> HSW_PWR_WELL_CTL_STATE(pw_idx),
> 1));
> }
>
> -static u32 hsw_power_well_requesters(struct drm_i915_private *dev_priv,
> +static u32 hsw_power_well_requesters(struct drm_i915_private *i915,
> const struct i915_power_well_regs *regs,
> int pw_idx)
> {
> @@ -291,7 +291,7 @@ static u32 hsw_power_well_requesters(struct drm_i915_private *dev_priv,
> return ret;
> }
>
> -static void hsw_wait_for_power_well_disable(struct drm_i915_private *dev_priv,
> +static void hsw_wait_for_power_well_disable(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> const struct i915_power_well_regs *regs = power_well->desc->hsw.regs;
> @@ -310,7 +310,7 @@ static void hsw_wait_for_power_well_disable(struct drm_i915_private *dev_priv,
> */
> wait_for((disabled = !(I915_READ(regs->driver) &
> HSW_PWR_WELL_CTL_STATE(pw_idx))) ||
> - (reqs = hsw_power_well_requesters(dev_priv, regs, pw_idx)), 1);
> + (reqs = hsw_power_well_requesters(i915, regs, pw_idx)), 1);
> if (disabled)
> return;
>
> @@ -319,16 +319,16 @@ static void hsw_wait_for_power_well_disable(struct drm_i915_private *dev_priv,
> !!(reqs & 1), !!(reqs & 2), !!(reqs & 4), !!(reqs & 8));
> }
>
> -static void gen9_wait_for_power_well_fuses(struct drm_i915_private *dev_priv,
> +static void gen9_wait_for_power_well_fuses(struct drm_i915_private *i915,
> enum skl_power_gate pg)
> {
> /* Timeout 5us for PG#0, for other PGs 1us */
> - WARN_ON(intel_wait_for_register(&dev_priv->uncore, SKL_FUSE_STATUS,
> + WARN_ON(intel_wait_for_register(&i915->uncore, SKL_FUSE_STATUS,
> SKL_FUSE_PG_DIST_STATUS(pg),
> SKL_FUSE_PG_DIST_STATUS(pg), 1));
> }
>
> -static void hsw_power_well_enable(struct drm_i915_private *dev_priv,
> +static void hsw_power_well_enable(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> const struct i915_power_well_regs *regs = power_well->desc->hsw.regs;
> @@ -338,7 +338,7 @@ static void hsw_power_well_enable(struct drm_i915_private *dev_priv,
> u32 val;
>
> if (wait_fuses) {
> - pg = INTEL_GEN(dev_priv) >= 11 ? ICL_PW_CTL_IDX_TO_PG(pw_idx) :
> + pg = INTEL_GEN(i915) >= 11 ? ICL_PW_CTL_IDX_TO_PG(pw_idx) :
> SKL_PW_CTL_IDX_TO_PG(pw_idx);
> /*
> * For PW1 we have to wait both for the PW0/PG0 fuse state
> @@ -348,15 +348,15 @@ static void hsw_power_well_enable(struct drm_i915_private *dev_priv,
> * after the enabling.
> */
> if (pg == SKL_PG1)
> - gen9_wait_for_power_well_fuses(dev_priv, SKL_PG0);
> + gen9_wait_for_power_well_fuses(i915, SKL_PG0);
> }
>
> val = I915_READ(regs->driver);
> I915_WRITE(regs->driver, val | HSW_PWR_WELL_CTL_REQ(pw_idx));
> - hsw_wait_for_power_well_enable(dev_priv, power_well);
> + hsw_wait_for_power_well_enable(i915, power_well);
>
> /* Display WA #1178: cnl */
> - if (IS_CANNONLAKE(dev_priv) &&
> + if (IS_CANNONLAKE(i915) &&
> pw_idx >= GLK_PW_CTL_IDX_AUX_B &&
> pw_idx <= CNL_PW_CTL_IDX_AUX_F) {
> val = I915_READ(CNL_AUX_ANAOVRD1(pw_idx));
> @@ -365,32 +365,32 @@ static void hsw_power_well_enable(struct drm_i915_private *dev_priv,
> }
>
> if (wait_fuses)
> - gen9_wait_for_power_well_fuses(dev_priv, pg);
> + gen9_wait_for_power_well_fuses(i915, pg);
>
> - hsw_power_well_post_enable(dev_priv,
> + hsw_power_well_post_enable(i915,
> power_well->desc->hsw.irq_pipe_mask,
> power_well->desc->hsw.has_vga);
> }
>
> -static void hsw_power_well_disable(struct drm_i915_private *dev_priv,
> +static void hsw_power_well_disable(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> const struct i915_power_well_regs *regs = power_well->desc->hsw.regs;
> int pw_idx = power_well->desc->hsw.idx;
> u32 val;
>
> - hsw_power_well_pre_disable(dev_priv,
> + hsw_power_well_pre_disable(i915,
> power_well->desc->hsw.irq_pipe_mask);
>
> val = I915_READ(regs->driver);
> I915_WRITE(regs->driver, val & ~HSW_PWR_WELL_CTL_REQ(pw_idx));
> - hsw_wait_for_power_well_disable(dev_priv, power_well);
> + hsw_wait_for_power_well_disable(i915, power_well);
> }
>
> #define ICL_AUX_PW_TO_PORT(pw_idx) ((pw_idx) - ICL_PW_CTL_IDX_AUX_A)
>
> static void
> -icl_combo_phy_aux_power_well_enable(struct drm_i915_private *dev_priv,
> +icl_combo_phy_aux_power_well_enable(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> const struct i915_power_well_regs *regs = power_well->desc->hsw.regs;
> @@ -404,12 +404,12 @@ icl_combo_phy_aux_power_well_enable(struct drm_i915_private *dev_priv,
> val = I915_READ(ICL_PORT_CL_DW12(port));
> I915_WRITE(ICL_PORT_CL_DW12(port), val | ICL_LANE_ENABLE_AUX);
>
> - hsw_wait_for_power_well_enable(dev_priv, power_well);
> + hsw_wait_for_power_well_enable(i915, power_well);
>
> /* Display WA #1178: icl */
> - if (IS_ICELAKE(dev_priv) &&
> + if (IS_ICELAKE(i915) &&
> pw_idx >= ICL_PW_CTL_IDX_AUX_A && pw_idx <= ICL_PW_CTL_IDX_AUX_B &&
> - !intel_bios_is_port_edp(dev_priv, port)) {
> + !intel_bios_is_port_edp(i915, port)) {
> val = I915_READ(ICL_AUX_ANAOVRD1(pw_idx));
> val |= ICL_AUX_ANAOVRD1_ENABLE | ICL_AUX_ANAOVRD1_LDO_BYPASS;
> I915_WRITE(ICL_AUX_ANAOVRD1(pw_idx), val);
> @@ -417,7 +417,7 @@ icl_combo_phy_aux_power_well_enable(struct drm_i915_private *dev_priv,
> }
>
> static void
> -icl_combo_phy_aux_power_well_disable(struct drm_i915_private *dev_priv,
> +icl_combo_phy_aux_power_well_disable(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> const struct i915_power_well_regs *regs = power_well->desc->hsw.regs;
> @@ -431,14 +431,14 @@ icl_combo_phy_aux_power_well_disable(struct drm_i915_private *dev_priv,
> val = I915_READ(regs->driver);
> I915_WRITE(regs->driver, val & ~HSW_PWR_WELL_CTL_REQ(pw_idx));
>
> - hsw_wait_for_power_well_disable(dev_priv, power_well);
> + hsw_wait_for_power_well_disable(i915, power_well);
> }
>
> #define ICL_AUX_PW_TO_CH(pw_idx) \
> ((pw_idx) - ICL_PW_CTL_IDX_AUX_A + AUX_CH_A)
>
> static void
> -icl_tc_phy_aux_power_well_enable(struct drm_i915_private *dev_priv,
> +icl_tc_phy_aux_power_well_enable(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> enum aux_ch aux_ch = ICL_AUX_PW_TO_CH(power_well->desc->hsw.idx);
> @@ -450,7 +450,7 @@ icl_tc_phy_aux_power_well_enable(struct drm_i915_private *dev_priv,
> val |= DP_AUX_CH_CTL_TBT_IO;
> I915_WRITE(DP_AUX_CH_CTL(aux_ch), val);
>
> - hsw_power_well_enable(dev_priv, power_well);
> + hsw_power_well_enable(i915, power_well);
> }
>
> /*
> @@ -458,7 +458,7 @@ icl_tc_phy_aux_power_well_enable(struct drm_i915_private *dev_priv,
> * enable it, so check if it's enabled and also check if we've requested it to
> * be enabled.
> */
> -static bool hsw_power_well_enabled(struct drm_i915_private *dev_priv,
> +static bool hsw_power_well_enabled(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> const struct i915_power_well_regs *regs = power_well->desc->hsw.regs;
> @@ -476,14 +476,14 @@ static bool hsw_power_well_enabled(struct drm_i915_private *dev_priv,
> * BIOS's own request bits, which are forced-on for these power wells
> * when exiting DC5/6.
> */
> - if (IS_GEN(dev_priv, 9) && !IS_GEN9_LP(dev_priv) &&
> + if (IS_GEN(i915, 9) && !IS_GEN9_LP(i915) &&
> (id == SKL_DISP_PW_1 || id == SKL_DISP_PW_MISC_IO))
> val |= I915_READ(regs->bios);
>
> return (val & mask) == mask;
> }
>
> -static void assert_can_enable_dc9(struct drm_i915_private *dev_priv)
> +static void assert_can_enable_dc9(struct drm_i915_private *i915)
> {
> WARN_ONCE((I915_READ(DC_STATE_EN) & DC_STATE_EN_DC9),
> "DC9 already programmed to be enabled.\n");
> @@ -492,7 +492,7 @@ static void assert_can_enable_dc9(struct drm_i915_private *dev_priv)
> WARN_ONCE(I915_READ(HSW_PWR_WELL_CTL2) &
> HSW_PWR_WELL_CTL_REQ(SKL_PW_CTL_IDX_PW_2),
> "Power well 2 on.\n");
> - WARN_ONCE(intel_irqs_enabled(dev_priv),
> + WARN_ONCE(intel_irqs_enabled(i915),
> "Interrupts not disabled yet.\n");
>
> /*
> @@ -504,9 +504,9 @@ static void assert_can_enable_dc9(struct drm_i915_private *dev_priv)
> */
> }
>
> -static void assert_can_disable_dc9(struct drm_i915_private *dev_priv)
> +static void assert_can_disable_dc9(struct drm_i915_private *i915)
> {
> - WARN_ONCE(intel_irqs_enabled(dev_priv),
> + WARN_ONCE(intel_irqs_enabled(i915),
> "Interrupts not disabled yet.\n");
> WARN_ONCE(I915_READ(DC_STATE_EN) & DC_STATE_EN_UPTO_DC5,
> "DC5 still not disabled.\n");
> @@ -520,7 +520,7 @@ static void assert_can_disable_dc9(struct drm_i915_private *dev_priv)
> */
> }
>
> -static void gen9_write_dc_state(struct drm_i915_private *dev_priv,
> +static void gen9_write_dc_state(struct drm_i915_private *i915,
> u32 state)
> {
> int rewrites = 0;
> @@ -557,14 +557,14 @@ static void gen9_write_dc_state(struct drm_i915_private *dev_priv,
> state, rewrites);
> }
>
> -static u32 gen9_dc_mask(struct drm_i915_private *dev_priv)
> +static u32 gen9_dc_mask(struct drm_i915_private *i915)
> {
> u32 mask;
>
> mask = DC_STATE_EN_UPTO_DC5;
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> mask |= DC_STATE_EN_UPTO_DC6 | DC_STATE_EN_DC9;
> - else if (IS_GEN9_LP(dev_priv))
> + else if (IS_GEN9_LP(i915))
> mask |= DC_STATE_EN_DC9;
> else
> mask |= DC_STATE_EN_UPTO_DC6;
> @@ -572,20 +572,20 @@ static u32 gen9_dc_mask(struct drm_i915_private *dev_priv)
> return mask;
> }
>
> -void gen9_sanitize_dc_state(struct drm_i915_private *dev_priv)
> +void gen9_sanitize_dc_state(struct drm_i915_private *i915)
> {
> u32 val;
>
> - val = I915_READ(DC_STATE_EN) & gen9_dc_mask(dev_priv);
> + val = I915_READ(DC_STATE_EN) & gen9_dc_mask(i915);
>
> DRM_DEBUG_KMS("Resetting DC state tracking from %02x to %02x\n",
> - dev_priv->csr.dc_state, val);
> - dev_priv->csr.dc_state = val;
> + i915->csr.dc_state, val);
> + i915->csr.dc_state = val;
> }
>
> /**
> * gen9_set_dc_state - set target display C power state
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> * @state: target DC power state
> * - DC_STATE_DISABLE
> * - DC_STATE_EN_UPTO_DC5
> @@ -606,35 +606,35 @@ void gen9_sanitize_dc_state(struct drm_i915_private *dev_priv)
> * back on and register state is restored. This is guaranteed by the MMIO write
> * to DC_STATE_EN blocking until the state is restored.
> */
> -static void gen9_set_dc_state(struct drm_i915_private *dev_priv, u32 state)
> +static void gen9_set_dc_state(struct drm_i915_private *i915, u32 state)
> {
> u32 val;
> u32 mask;
>
> - if (WARN_ON_ONCE(state & ~dev_priv->csr.allowed_dc_mask))
> - state &= dev_priv->csr.allowed_dc_mask;
> + if (WARN_ON_ONCE(state & ~i915->csr.allowed_dc_mask))
> + state &= i915->csr.allowed_dc_mask;
>
> val = I915_READ(DC_STATE_EN);
> - mask = gen9_dc_mask(dev_priv);
> + mask = gen9_dc_mask(i915);
> DRM_DEBUG_KMS("Setting DC state from %02x to %02x\n",
> val & mask, state);
>
> /* Check if DMC is ignoring our DC state requests */
> - if ((val & mask) != dev_priv->csr.dc_state)
> + if ((val & mask) != i915->csr.dc_state)
> DRM_ERROR("DC state mismatch (0x%x -> 0x%x)\n",
> - dev_priv->csr.dc_state, val & mask);
> + i915->csr.dc_state, val & mask);
>
> val &= ~mask;
> val |= state;
>
> - gen9_write_dc_state(dev_priv, val);
> + gen9_write_dc_state(i915, val);
>
> - dev_priv->csr.dc_state = val & mask;
> + i915->csr.dc_state = val & mask;
> }
>
> -void bxt_enable_dc9(struct drm_i915_private *dev_priv)
> +void bxt_enable_dc9(struct drm_i915_private *i915)
> {
> - assert_can_enable_dc9(dev_priv);
> + assert_can_enable_dc9(i915);
>
> DRM_DEBUG_KMS("Enabling DC9\n");
> /*
> @@ -642,23 +642,23 @@ void bxt_enable_dc9(struct drm_i915_private *dev_priv)
> * platforms with South Display Engine on PCH,
> * because PPS registers are always on.
> */
> - if (!HAS_PCH_SPLIT(dev_priv))
> - intel_power_sequencer_reset(dev_priv);
> - gen9_set_dc_state(dev_priv, DC_STATE_EN_DC9);
> + if (!HAS_PCH_SPLIT(i915))
> + intel_power_sequencer_reset(i915);
> + gen9_set_dc_state(i915, DC_STATE_EN_DC9);
> }
>
> -void bxt_disable_dc9(struct drm_i915_private *dev_priv)
> +void bxt_disable_dc9(struct drm_i915_private *i915)
> {
> - assert_can_disable_dc9(dev_priv);
> + assert_can_disable_dc9(i915);
>
> DRM_DEBUG_KMS("Disabling DC9\n");
>
> - gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
> + gen9_set_dc_state(i915, DC_STATE_DISABLE);
>
> - intel_pps_unlock_regs_wa(dev_priv);
> + intel_pps_unlock_regs_wa(i915);
> }
>
> -static void assert_csr_loaded(struct drm_i915_private *dev_priv)
> +static void assert_csr_loaded(struct drm_i915_private *i915)
> {
> WARN_ONCE(!I915_READ(CSR_PROGRAM(0)),
> "CSR program storage start is NULL\n");
> @@ -667,12 +667,12 @@ static void assert_csr_loaded(struct drm_i915_private *dev_priv)
> }
>
> static struct i915_power_well *
> -lookup_power_well(struct drm_i915_private *dev_priv,
> +lookup_power_well(struct drm_i915_private *i915,
> enum i915_power_well_id power_well_id)
> {
> struct i915_power_well *power_well;
>
> - for_each_power_well(dev_priv, power_well)
> + for_each_power_well(i915, power_well)
> if (power_well->desc->id == power_well_id)
> return power_well;
>
> @@ -684,62 +684,62 @@ lookup_power_well(struct drm_i915_private *dev_priv,
> * our driver.
> */
> WARN(1, "Power well %d not defined for this platform\n", power_well_id);
> - return &dev_priv->power_domains.power_wells[0];
> + return &i915->power_domains.power_wells[0];
> }
>
> -static void assert_can_enable_dc5(struct drm_i915_private *dev_priv)
> +static void assert_can_enable_dc5(struct drm_i915_private *i915)
> {
> - bool pg2_enabled = intel_display_power_well_is_enabled(dev_priv,
> + bool pg2_enabled = intel_display_power_well_is_enabled(i915,
> SKL_DISP_PW_2);
>
> WARN_ONCE(pg2_enabled, "PG2 not disabled to enable DC5.\n");
>
> WARN_ONCE((I915_READ(DC_STATE_EN) & DC_STATE_EN_UPTO_DC5),
> "DC5 already programmed to be enabled.\n");
> - assert_rpm_wakelock_held(dev_priv);
> + assert_rpm_wakelock_held(i915);
>
> - assert_csr_loaded(dev_priv);
> + assert_csr_loaded(i915);
> }
>
> -void gen9_enable_dc5(struct drm_i915_private *dev_priv)
> +void gen9_enable_dc5(struct drm_i915_private *i915)
> {
> - assert_can_enable_dc5(dev_priv);
> + assert_can_enable_dc5(i915);
>
> DRM_DEBUG_KMS("Enabling DC5\n");
>
> /* Wa Display #1183: skl,kbl,cfl */
> - if (IS_GEN9_BC(dev_priv))
> + if (IS_GEN9_BC(i915))
> I915_WRITE(GEN8_CHICKEN_DCPR_1, I915_READ(GEN8_CHICKEN_DCPR_1) |
> SKL_SELECT_ALTERNATE_DC_EXIT);
>
> - gen9_set_dc_state(dev_priv, DC_STATE_EN_UPTO_DC5);
> + gen9_set_dc_state(i915, DC_STATE_EN_UPTO_DC5);
> }
>
> -static void assert_can_enable_dc6(struct drm_i915_private *dev_priv)
> +static void assert_can_enable_dc6(struct drm_i915_private *i915)
> {
> WARN_ONCE(I915_READ(UTIL_PIN_CTL) & UTIL_PIN_ENABLE,
> "Backlight is not disabled.\n");
> WARN_ONCE((I915_READ(DC_STATE_EN) & DC_STATE_EN_UPTO_DC6),
> "DC6 already programmed to be enabled.\n");
>
> - assert_csr_loaded(dev_priv);
> + assert_csr_loaded(i915);
> }
>
> -void skl_enable_dc6(struct drm_i915_private *dev_priv)
> +void skl_enable_dc6(struct drm_i915_private *i915)
> {
> - assert_can_enable_dc6(dev_priv);
> + assert_can_enable_dc6(i915);
>
> DRM_DEBUG_KMS("Enabling DC6\n");
>
> /* Wa Display #1183: skl,kbl,cfl */
> - if (IS_GEN9_BC(dev_priv))
> + if (IS_GEN9_BC(i915))
> I915_WRITE(GEN8_CHICKEN_DCPR_1, I915_READ(GEN8_CHICKEN_DCPR_1) |
> SKL_SELECT_ALTERNATE_DC_EXIT);
>
> - gen9_set_dc_state(dev_priv, DC_STATE_EN_UPTO_DC6);
> + gen9_set_dc_state(i915, DC_STATE_EN_UPTO_DC6);
> }
>
> -static void hsw_power_well_sync_hw(struct drm_i915_private *dev_priv,
> +static void hsw_power_well_sync_hw(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> const struct i915_power_well_regs *regs = power_well->desc->hsw.regs;
> @@ -757,52 +757,52 @@ static void hsw_power_well_sync_hw(struct drm_i915_private *dev_priv,
> }
> }
>
> -static void bxt_dpio_cmn_power_well_enable(struct drm_i915_private *dev_priv,
> +static void bxt_dpio_cmn_power_well_enable(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> - bxt_ddi_phy_init(dev_priv, power_well->desc->bxt.phy);
> + bxt_ddi_phy_init(i915, power_well->desc->bxt.phy);
> }
>
> -static void bxt_dpio_cmn_power_well_disable(struct drm_i915_private *dev_priv,
> +static void bxt_dpio_cmn_power_well_disable(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> - bxt_ddi_phy_uninit(dev_priv, power_well->desc->bxt.phy);
> + bxt_ddi_phy_uninit(i915, power_well->desc->bxt.phy);
> }
>
> -static bool bxt_dpio_cmn_power_well_enabled(struct drm_i915_private *dev_priv,
> +static bool bxt_dpio_cmn_power_well_enabled(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> - return bxt_ddi_phy_is_enabled(dev_priv, power_well->desc->bxt.phy);
> + return bxt_ddi_phy_is_enabled(i915, power_well->desc->bxt.phy);
> }
>
> -static void bxt_verify_ddi_phy_power_wells(struct drm_i915_private *dev_priv)
> +static void bxt_verify_ddi_phy_power_wells(struct drm_i915_private *i915)
> {
> struct i915_power_well *power_well;
>
> - power_well = lookup_power_well(dev_priv, BXT_DISP_PW_DPIO_CMN_A);
> + power_well = lookup_power_well(i915, BXT_DISP_PW_DPIO_CMN_A);
> if (power_well->count > 0)
> - bxt_ddi_phy_verify_state(dev_priv, power_well->desc->bxt.phy);
> + bxt_ddi_phy_verify_state(i915, power_well->desc->bxt.phy);
>
> - power_well = lookup_power_well(dev_priv, VLV_DISP_PW_DPIO_CMN_BC);
> + power_well = lookup_power_well(i915, VLV_DISP_PW_DPIO_CMN_BC);
> if (power_well->count > 0)
> - bxt_ddi_phy_verify_state(dev_priv, power_well->desc->bxt.phy);
> + bxt_ddi_phy_verify_state(i915, power_well->desc->bxt.phy);
>
> - if (IS_GEMINILAKE(dev_priv)) {
> - power_well = lookup_power_well(dev_priv,
> + if (IS_GEMINILAKE(i915)) {
> + power_well = lookup_power_well(i915,
> GLK_DISP_PW_DPIO_CMN_C);
> if (power_well->count > 0)
> - bxt_ddi_phy_verify_state(dev_priv,
> + bxt_ddi_phy_verify_state(i915,
> power_well->desc->bxt.phy);
> }
> }
>
> -static bool gen9_dc_off_power_well_enabled(struct drm_i915_private *dev_priv,
> +static bool gen9_dc_off_power_well_enabled(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> return (I915_READ(DC_STATE_EN) & DC_STATE_EN_UPTO_DC5_DC6_MASK) == 0;
> }
>
> -static void gen9_assert_dbuf_enabled(struct drm_i915_private *dev_priv)
> +static void gen9_assert_dbuf_enabled(struct drm_i915_private *i915)
> {
> u32 tmp = I915_READ(DBUF_CTL);
>
> @@ -811,92 +811,92 @@ static void gen9_assert_dbuf_enabled(struct drm_i915_private *dev_priv)
> "Unexpected DBuf power power state (0x%08x)\n", tmp);
> }
>
> -static void gen9_dc_off_power_well_enable(struct drm_i915_private *dev_priv,
> +static void gen9_dc_off_power_well_enable(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> struct intel_cdclk_state cdclk_state = {};
>
> - gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
> + gen9_set_dc_state(i915, DC_STATE_DISABLE);
>
> - dev_priv->display.get_cdclk(dev_priv, &cdclk_state);
> + i915->display.get_cdclk(i915, &cdclk_state);
> /* Can't read out voltage_level so can't use intel_cdclk_changed() */
> - WARN_ON(intel_cdclk_needs_modeset(&dev_priv->cdclk.hw, &cdclk_state));
> + WARN_ON(intel_cdclk_needs_modeset(&i915->cdclk.hw, &cdclk_state));
>
> - gen9_assert_dbuf_enabled(dev_priv);
> + gen9_assert_dbuf_enabled(i915);
>
> - if (IS_GEN9_LP(dev_priv))
> - bxt_verify_ddi_phy_power_wells(dev_priv);
> + if (IS_GEN9_LP(i915))
> + bxt_verify_ddi_phy_power_wells(i915);
>
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> /*
> * DMC retains HW context only for port A, the other combo
> * PHY's HW context for port B is lost after DC transitions,
> * so we need to restore it manually.
> */
> - intel_combo_phy_init(dev_priv);
> + intel_combo_phy_init(i915);
> }
>
> -static void gen9_dc_off_power_well_disable(struct drm_i915_private *dev_priv,
> +static void gen9_dc_off_power_well_disable(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> - if (!dev_priv->csr.dmc_payload)
> + if (!i915->csr.dmc_payload)
> return;
>
> - if (dev_priv->csr.allowed_dc_mask & DC_STATE_EN_UPTO_DC6)
> - skl_enable_dc6(dev_priv);
> - else if (dev_priv->csr.allowed_dc_mask & DC_STATE_EN_UPTO_DC5)
> - gen9_enable_dc5(dev_priv);
> + if (i915->csr.allowed_dc_mask & DC_STATE_EN_UPTO_DC6)
> + skl_enable_dc6(i915);
> + else if (i915->csr.allowed_dc_mask & DC_STATE_EN_UPTO_DC5)
> + gen9_enable_dc5(i915);
> }
>
> -static void i9xx_power_well_sync_hw_noop(struct drm_i915_private *dev_priv,
> +static void i9xx_power_well_sync_hw_noop(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> }
>
> -static void i9xx_always_on_power_well_noop(struct drm_i915_private *dev_priv,
> +static void i9xx_always_on_power_well_noop(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> }
>
> -static bool i9xx_always_on_power_well_enabled(struct drm_i915_private *dev_priv,
> +static bool i9xx_always_on_power_well_enabled(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> return true;
> }
>
> -static void i830_pipes_power_well_enable(struct drm_i915_private *dev_priv,
> +static void i830_pipes_power_well_enable(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> if ((I915_READ(PIPECONF(PIPE_A)) & PIPECONF_ENABLE) == 0)
> - i830_enable_pipe(dev_priv, PIPE_A);
> + i830_enable_pipe(i915, PIPE_A);
> if ((I915_READ(PIPECONF(PIPE_B)) & PIPECONF_ENABLE) == 0)
> - i830_enable_pipe(dev_priv, PIPE_B);
> + i830_enable_pipe(i915, PIPE_B);
> }
>
> -static void i830_pipes_power_well_disable(struct drm_i915_private *dev_priv,
> +static void i830_pipes_power_well_disable(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> - i830_disable_pipe(dev_priv, PIPE_B);
> - i830_disable_pipe(dev_priv, PIPE_A);
> + i830_disable_pipe(i915, PIPE_B);
> + i830_disable_pipe(i915, PIPE_A);
> }
>
> -static bool i830_pipes_power_well_enabled(struct drm_i915_private *dev_priv,
> +static bool i830_pipes_power_well_enabled(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> return I915_READ(PIPECONF(PIPE_A)) & PIPECONF_ENABLE &&
> I915_READ(PIPECONF(PIPE_B)) & PIPECONF_ENABLE;
> }
>
> -static void i830_pipes_power_well_sync_hw(struct drm_i915_private *dev_priv,
> +static void i830_pipes_power_well_sync_hw(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> if (power_well->count > 0)
> - i830_pipes_power_well_enable(dev_priv, power_well);
> + i830_pipes_power_well_enable(i915, power_well);
> else
> - i830_pipes_power_well_disable(dev_priv, power_well);
> + i830_pipes_power_well_disable(i915, power_well);
> }
>
> -static void vlv_set_power_well(struct drm_i915_private *dev_priv,
> +static void vlv_set_power_well(struct drm_i915_private *i915,
> struct i915_power_well *power_well, bool enable)
> {
> int pw_idx = power_well->desc->vlv.idx;
> @@ -908,43 +908,43 @@ static void vlv_set_power_well(struct drm_i915_private *dev_priv,
> state = enable ? PUNIT_PWRGT_PWR_ON(pw_idx) :
> PUNIT_PWRGT_PWR_GATE(pw_idx);
>
> - vlv_punit_get(dev_priv);
> + vlv_punit_get(i915);
>
> #define COND \
> - ((vlv_punit_read(dev_priv, PUNIT_REG_PWRGT_STATUS) & mask) == state)
> + ((vlv_punit_read(i915, PUNIT_REG_PWRGT_STATUS) & mask) == state)
>
> if (COND)
> goto out;
>
> - ctrl = vlv_punit_read(dev_priv, PUNIT_REG_PWRGT_CTRL);
> + ctrl = vlv_punit_read(i915, PUNIT_REG_PWRGT_CTRL);
> ctrl &= ~mask;
> ctrl |= state;
> - vlv_punit_write(dev_priv, PUNIT_REG_PWRGT_CTRL, ctrl);
> + vlv_punit_write(i915, PUNIT_REG_PWRGT_CTRL, ctrl);
>
> if (wait_for(COND, 100))
> DRM_ERROR("timeout setting power well state %08x (%08x)\n",
> state,
> - vlv_punit_read(dev_priv, PUNIT_REG_PWRGT_CTRL));
> + vlv_punit_read(i915, PUNIT_REG_PWRGT_CTRL));
>
> #undef COND
>
> out:
> - vlv_punit_put(dev_priv);
> + vlv_punit_put(i915);
> }
>
> -static void vlv_power_well_enable(struct drm_i915_private *dev_priv,
> +static void vlv_power_well_enable(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> - vlv_set_power_well(dev_priv, power_well, true);
> + vlv_set_power_well(i915, power_well, true);
> }
>
> -static void vlv_power_well_disable(struct drm_i915_private *dev_priv,
> +static void vlv_power_well_disable(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> - vlv_set_power_well(dev_priv, power_well, false);
> + vlv_set_power_well(i915, power_well, false);
> }
>
> -static bool vlv_power_well_enabled(struct drm_i915_private *dev_priv,
> +static bool vlv_power_well_enabled(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> int pw_idx = power_well->desc->vlv.idx;
> @@ -956,9 +956,9 @@ static bool vlv_power_well_enabled(struct drm_i915_private *dev_priv,
> mask = PUNIT_PWRGT_MASK(pw_idx);
> ctrl = PUNIT_PWRGT_PWR_ON(pw_idx);
>
> - vlv_punit_get(dev_priv);
> + vlv_punit_get(i915);
>
> - state = vlv_punit_read(dev_priv, PUNIT_REG_PWRGT_STATUS) & mask;
> + state = vlv_punit_read(i915, PUNIT_REG_PWRGT_STATUS) & mask;
> /*
> * We only ever set the power-on and power-gate states, anything
> * else is unexpected.
> @@ -972,15 +972,15 @@ static bool vlv_power_well_enabled(struct drm_i915_private *dev_priv,
> * A transient state at this point would mean some unexpected party
> * is poking at the power controls too.
> */
> - ctrl = vlv_punit_read(dev_priv, PUNIT_REG_PWRGT_CTRL) & mask;
> + ctrl = vlv_punit_read(i915, PUNIT_REG_PWRGT_CTRL) & mask;
> WARN_ON(ctrl != state);
>
> - vlv_punit_put(dev_priv);
> + vlv_punit_put(i915);
>
> return enabled;
> }
>
> -static void vlv_init_display_clock_gating(struct drm_i915_private *dev_priv)
> +static void vlv_init_display_clock_gating(struct drm_i915_private *i915)
> {
> u32 val;
>
> @@ -1001,13 +1001,13 @@ static void vlv_init_display_clock_gating(struct drm_i915_private *dev_priv)
> I915_WRITE(MI_ARB_VLV, MI_ARB_DISPLAY_TRICKLE_FEED_DISABLE);
> I915_WRITE(CBR1_VLV, 0);
>
> - WARN_ON(dev_priv->rawclk_freq == 0);
> + WARN_ON(i915->rawclk_freq == 0);
>
> I915_WRITE(RAWCLK_FREQ_VLV,
> - DIV_ROUND_CLOSEST(dev_priv->rawclk_freq, 1000));
> + DIV_ROUND_CLOSEST(i915->rawclk_freq, 1000));
> }
>
> -static void vlv_display_power_well_init(struct drm_i915_private *dev_priv)
> +static void vlv_display_power_well_init(struct drm_i915_private *i915)
> {
> struct intel_encoder *encoder;
> enum pipe pipe;
> @@ -1020,7 +1020,7 @@ static void vlv_display_power_well_init(struct drm_i915_private *dev_priv)
> *
> * CHV DPLL B/C have some issues if VGA mode is enabled.
> */
> - for_each_pipe(dev_priv, pipe) {
> + for_each_pipe(i915, pipe) {
> u32 val = I915_READ(DPLL(pipe));
>
> val |= DPLL_REF_CLK_ENABLE_VLV | DPLL_VGA_MODE_DIS;
> @@ -1030,71 +1030,71 @@ static void vlv_display_power_well_init(struct drm_i915_private *dev_priv)
> I915_WRITE(DPLL(pipe), val);
> }
>
> - vlv_init_display_clock_gating(dev_priv);
> + vlv_init_display_clock_gating(i915);
>
> - spin_lock_irq(&dev_priv->irq_lock);
> - valleyview_enable_display_irqs(dev_priv);
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
> + valleyview_enable_display_irqs(i915);
> + spin_unlock_irq(&i915->irq_lock);
>
> /*
> * During driver initialization/resume we can avoid restoring the
> * part of the HW/SW state that will be inited anyway explicitly.
> */
> - if (dev_priv->power_domains.initializing)
> + if (i915->power_domains.initializing)
> return;
>
> - intel_hpd_init(dev_priv);
> + intel_hpd_init(i915);
>
> /* Re-enable the ADPA, if we have one */
> - for_each_intel_encoder(&dev_priv->drm, encoder) {
> + for_each_intel_encoder(&i915->drm, encoder) {
> if (encoder->type == INTEL_OUTPUT_ANALOG)
> intel_crt_reset(&encoder->base);
> }
>
> - i915_redisable_vga_power_on(dev_priv);
> + i915_redisable_vga_power_on(i915);
>
> - intel_pps_unlock_regs_wa(dev_priv);
> + intel_pps_unlock_regs_wa(i915);
> }
>
> -static void vlv_display_power_well_deinit(struct drm_i915_private *dev_priv)
> +static void vlv_display_power_well_deinit(struct drm_i915_private *i915)
> {
> - spin_lock_irq(&dev_priv->irq_lock);
> - valleyview_disable_display_irqs(dev_priv);
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
> + valleyview_disable_display_irqs(i915);
> + spin_unlock_irq(&i915->irq_lock);
>
> /* make sure we're done processing display irqs */
> - synchronize_irq(dev_priv->drm.irq);
> + synchronize_irq(i915->drm.irq);
>
> - intel_power_sequencer_reset(dev_priv);
> + intel_power_sequencer_reset(i915);
>
> /* Prevent us from re-enabling polling on accident in late suspend */
> - if (!dev_priv->drm.dev->power.is_suspended)
> - intel_hpd_poll_init(dev_priv);
> + if (!i915->drm.dev->power.is_suspended)
> + intel_hpd_poll_init(i915);
> }
>
> -static void vlv_display_power_well_enable(struct drm_i915_private *dev_priv,
> +static void vlv_display_power_well_enable(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> - vlv_set_power_well(dev_priv, power_well, true);
> + vlv_set_power_well(i915, power_well, true);
>
> - vlv_display_power_well_init(dev_priv);
> + vlv_display_power_well_init(i915);
> }
>
> -static void vlv_display_power_well_disable(struct drm_i915_private *dev_priv,
> +static void vlv_display_power_well_disable(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> - vlv_display_power_well_deinit(dev_priv);
> + vlv_display_power_well_deinit(i915);
>
> - vlv_set_power_well(dev_priv, power_well, false);
> + vlv_set_power_well(i915, power_well, false);
> }
>
> -static void vlv_dpio_cmn_power_well_enable(struct drm_i915_private *dev_priv,
> +static void vlv_dpio_cmn_power_well_enable(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> /* since ref/cri clock was enabled */
> udelay(1); /* >10ns for cmnreset, >0ns for sidereset */
>
> - vlv_set_power_well(dev_priv, power_well, true);
> + vlv_set_power_well(i915, power_well, true);
>
> /*
> * From VLV2A0_DP_eDP_DPIO_driver_vbios_notes_10.docx -
> @@ -1110,31 +1110,31 @@ static void vlv_dpio_cmn_power_well_enable(struct drm_i915_private *dev_priv,
> I915_WRITE(DPIO_CTL, I915_READ(DPIO_CTL) | DPIO_CMNRST);
> }
>
> -static void vlv_dpio_cmn_power_well_disable(struct drm_i915_private *dev_priv,
> +static void vlv_dpio_cmn_power_well_disable(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> enum pipe pipe;
>
> - for_each_pipe(dev_priv, pipe)
> - assert_pll_disabled(dev_priv, pipe);
> + for_each_pipe(i915, pipe)
> + assert_pll_disabled(i915, pipe);
>
> /* Assert common reset */
> I915_WRITE(DPIO_CTL, I915_READ(DPIO_CTL) & ~DPIO_CMNRST);
>
> - vlv_set_power_well(dev_priv, power_well, false);
> + vlv_set_power_well(i915, power_well, false);
> }
>
> #define POWER_DOMAIN_MASK (GENMASK_ULL(POWER_DOMAIN_NUM - 1, 0))
>
> #define BITS_SET(val, bits) (((val) & (bits)) == (bits))
>
> -static void assert_chv_phy_status(struct drm_i915_private *dev_priv)
> +static void assert_chv_phy_status(struct drm_i915_private *i915)
> {
> struct i915_power_well *cmn_bc =
> - lookup_power_well(dev_priv, VLV_DISP_PW_DPIO_CMN_BC);
> + lookup_power_well(i915, VLV_DISP_PW_DPIO_CMN_BC);
> struct i915_power_well *cmn_d =
> - lookup_power_well(dev_priv, CHV_DISP_PW_DPIO_CMN_D);
> - u32 phy_control = dev_priv->chv_phy_control;
> + lookup_power_well(i915, CHV_DISP_PW_DPIO_CMN_D);
> + u32 phy_control = i915->chv_phy_control;
> u32 phy_status = 0;
> u32 phy_status_mask = 0xffffffff;
>
> @@ -1145,7 +1145,7 @@ static void assert_chv_phy_status(struct drm_i915_private *dev_priv)
> * reset (ie. the power well has been disabled at
> * least once).
> */
> - if (!dev_priv->chv_phy_assert[DPIO_PHY0])
> + if (!i915->chv_phy_assert[DPIO_PHY0])
> phy_status_mask &= ~(PHY_STATUS_CMN_LDO(DPIO_PHY0, DPIO_CH0) |
> PHY_STATUS_SPLINE_LDO(DPIO_PHY0, DPIO_CH0, 0) |
> PHY_STATUS_SPLINE_LDO(DPIO_PHY0, DPIO_CH0, 1) |
> @@ -1153,12 +1153,12 @@ static void assert_chv_phy_status(struct drm_i915_private *dev_priv)
> PHY_STATUS_SPLINE_LDO(DPIO_PHY0, DPIO_CH1, 0) |
> PHY_STATUS_SPLINE_LDO(DPIO_PHY0, DPIO_CH1, 1));
>
> - if (!dev_priv->chv_phy_assert[DPIO_PHY1])
> + if (!i915->chv_phy_assert[DPIO_PHY1])
> phy_status_mask &= ~(PHY_STATUS_CMN_LDO(DPIO_PHY1, DPIO_CH0) |
> PHY_STATUS_SPLINE_LDO(DPIO_PHY1, DPIO_CH0, 0) |
> PHY_STATUS_SPLINE_LDO(DPIO_PHY1, DPIO_CH0, 1));
>
> - if (cmn_bc->desc->ops->is_enabled(dev_priv, cmn_bc)) {
> + if (cmn_bc->desc->ops->is_enabled(i915, cmn_bc)) {
> phy_status |= PHY_POWERGOOD(DPIO_PHY0);
>
> /* this assumes override is only used to enable lanes */
> @@ -1199,7 +1199,7 @@ static void assert_chv_phy_status(struct drm_i915_private *dev_priv)
> phy_status |= PHY_STATUS_SPLINE_LDO(DPIO_PHY0, DPIO_CH1, 1);
> }
>
> - if (cmn_d->desc->ops->is_enabled(dev_priv, cmn_d)) {
> + if (cmn_d->desc->ops->is_enabled(i915, cmn_d)) {
> phy_status |= PHY_POWERGOOD(DPIO_PHY1);
>
> /* this assumes override is only used to enable lanes */
> @@ -1224,19 +1224,19 @@ static void assert_chv_phy_status(struct drm_i915_private *dev_priv)
> * The PHY may be busy with some initial calibration and whatnot,
> * so the power state can take a while to actually change.
> */
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> DISPLAY_PHY_STATUS,
> phy_status_mask,
> phy_status,
> 10))
> DRM_ERROR("Unexpected PHY_STATUS 0x%08x, expected 0x%08x (PHY_CONTROL=0x%08x)\n",
> I915_READ(DISPLAY_PHY_STATUS) & phy_status_mask,
> - phy_status, dev_priv->chv_phy_control);
> + phy_status, i915->chv_phy_control);
> }
>
> #undef BITS_SET
>
> -static void chv_dpio_cmn_power_well_enable(struct drm_i915_private *dev_priv,
> +static void chv_dpio_cmn_power_well_enable(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> enum dpio_phy phy;
> @@ -1256,51 +1256,51 @@ static void chv_dpio_cmn_power_well_enable(struct drm_i915_private *dev_priv,
>
> /* since ref/cri clock was enabled */
> udelay(1); /* >10ns for cmnreset, >0ns for sidereset */
> - vlv_set_power_well(dev_priv, power_well, true);
> + vlv_set_power_well(i915, power_well, true);
>
> /* Poll for phypwrgood signal */
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> DISPLAY_PHY_STATUS,
> PHY_POWERGOOD(phy),
> PHY_POWERGOOD(phy),
> 1))
> DRM_ERROR("Display PHY %d is not power up\n", phy);
>
> - vlv_dpio_get(dev_priv);
> + vlv_dpio_get(i915);
>
> /* Enable dynamic power down */
> - tmp = vlv_dpio_read(dev_priv, pipe, CHV_CMN_DW28);
> + tmp = vlv_dpio_read(i915, pipe, CHV_CMN_DW28);
> tmp |= DPIO_DYNPWRDOWNEN_CH0 | DPIO_CL1POWERDOWNEN |
> DPIO_SUS_CLK_CONFIG_GATE_CLKREQ;
> - vlv_dpio_write(dev_priv, pipe, CHV_CMN_DW28, tmp);
> + vlv_dpio_write(i915, pipe, CHV_CMN_DW28, tmp);
>
> if (power_well->desc->id == VLV_DISP_PW_DPIO_CMN_BC) {
> - tmp = vlv_dpio_read(dev_priv, pipe, _CHV_CMN_DW6_CH1);
> + tmp = vlv_dpio_read(i915, pipe, _CHV_CMN_DW6_CH1);
> tmp |= DPIO_DYNPWRDOWNEN_CH1;
> - vlv_dpio_write(dev_priv, pipe, _CHV_CMN_DW6_CH1, tmp);
> + vlv_dpio_write(i915, pipe, _CHV_CMN_DW6_CH1, tmp);
> } else {
> /*
> * Force the non-existing CL2 off. BXT does this
> * too, so maybe it saves some power even though
> * CL2 doesn't exist?
> */
> - tmp = vlv_dpio_read(dev_priv, pipe, CHV_CMN_DW30);
> + tmp = vlv_dpio_read(i915, pipe, CHV_CMN_DW30);
> tmp |= DPIO_CL2_LDOFUSE_PWRENB;
> - vlv_dpio_write(dev_priv, pipe, CHV_CMN_DW30, tmp);
> + vlv_dpio_write(i915, pipe, CHV_CMN_DW30, tmp);
> }
>
> - vlv_dpio_put(dev_priv);
> + vlv_dpio_put(i915);
>
> - dev_priv->chv_phy_control |= PHY_COM_LANE_RESET_DEASSERT(phy);
> - I915_WRITE(DISPLAY_PHY_CONTROL, dev_priv->chv_phy_control);
> + i915->chv_phy_control |= PHY_COM_LANE_RESET_DEASSERT(phy);
> + I915_WRITE(DISPLAY_PHY_CONTROL, i915->chv_phy_control);
>
> DRM_DEBUG_KMS("Enabled DPIO PHY%d (PHY_CONTROL=0x%08x)\n",
> - phy, dev_priv->chv_phy_control);
> + phy, i915->chv_phy_control);
>
> - assert_chv_phy_status(dev_priv);
> + assert_chv_phy_status(i915);
> }
>
> -static void chv_dpio_cmn_power_well_disable(struct drm_i915_private *dev_priv,
> +static void chv_dpio_cmn_power_well_disable(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> enum dpio_phy phy;
> @@ -1310,28 +1310,28 @@ static void chv_dpio_cmn_power_well_disable(struct drm_i915_private *dev_priv,
>
> if (power_well->desc->id == VLV_DISP_PW_DPIO_CMN_BC) {
> phy = DPIO_PHY0;
> - assert_pll_disabled(dev_priv, PIPE_A);
> - assert_pll_disabled(dev_priv, PIPE_B);
> + assert_pll_disabled(i915, PIPE_A);
> + assert_pll_disabled(i915, PIPE_B);
> } else {
> phy = DPIO_PHY1;
> - assert_pll_disabled(dev_priv, PIPE_C);
> + assert_pll_disabled(i915, PIPE_C);
> }
>
> - dev_priv->chv_phy_control &= ~PHY_COM_LANE_RESET_DEASSERT(phy);
> - I915_WRITE(DISPLAY_PHY_CONTROL, dev_priv->chv_phy_control);
> + i915->chv_phy_control &= ~PHY_COM_LANE_RESET_DEASSERT(phy);
> + I915_WRITE(DISPLAY_PHY_CONTROL, i915->chv_phy_control);
>
> - vlv_set_power_well(dev_priv, power_well, false);
> + vlv_set_power_well(i915, power_well, false);
>
> DRM_DEBUG_KMS("Disabled DPIO PHY%d (PHY_CONTROL=0x%08x)\n",
> - phy, dev_priv->chv_phy_control);
> + phy, i915->chv_phy_control);
>
> /* PHY is fully reset now, so we can enable the PHY state asserts */
> - dev_priv->chv_phy_assert[phy] = true;
> + i915->chv_phy_assert[phy] = true;
>
> - assert_chv_phy_status(dev_priv);
> + assert_chv_phy_status(i915);
> }
>
> -static void assert_chv_phy_powergate(struct drm_i915_private *dev_priv, enum dpio_phy phy,
> +static void assert_chv_phy_powergate(struct drm_i915_private *i915, enum dpio_phy phy,
> enum dpio_channel ch, bool override, unsigned int mask)
> {
> enum pipe pipe = phy == DPIO_PHY0 ? PIPE_A : PIPE_C;
> @@ -1344,7 +1344,7 @@ static void assert_chv_phy_powergate(struct drm_i915_private *dev_priv, enum dpi
> * reset (ie. the power well has been disabled at
> * least once).
> */
> - if (!dev_priv->chv_phy_assert[phy])
> + if (!i915->chv_phy_assert[phy])
> return;
>
> if (ch == DPIO_CH0)
> @@ -1352,9 +1352,9 @@ static void assert_chv_phy_powergate(struct drm_i915_private *dev_priv, enum dpi
> else
> reg = _CHV_CMN_DW6_CH1;
>
> - vlv_dpio_get(dev_priv);
> - val = vlv_dpio_read(dev_priv, pipe, reg);
> - vlv_dpio_put(dev_priv);
> + vlv_dpio_get(i915);
> + val = vlv_dpio_read(i915, pipe, reg);
> + vlv_dpio_put(i915);
>
> /*
> * This assumes !override is only used when the port is disabled.
> @@ -1393,30 +1393,30 @@ static void assert_chv_phy_powergate(struct drm_i915_private *dev_priv, enum dpi
> reg, val);
> }
>
> -bool chv_phy_powergate_ch(struct drm_i915_private *dev_priv, enum dpio_phy phy,
> +bool chv_phy_powergate_ch(struct drm_i915_private *i915, enum dpio_phy phy,
> enum dpio_channel ch, bool override)
> {
> - struct i915_power_domains *power_domains = &dev_priv->power_domains;
> + struct i915_power_domains *power_domains = &i915->power_domains;
> bool was_override;
>
> mutex_lock(&power_domains->lock);
>
> - was_override = dev_priv->chv_phy_control & PHY_CH_POWER_DOWN_OVRD_EN(phy, ch);
> + was_override = i915->chv_phy_control & PHY_CH_POWER_DOWN_OVRD_EN(phy, ch);
>
> if (override == was_override)
> goto out;
>
> if (override)
> - dev_priv->chv_phy_control |= PHY_CH_POWER_DOWN_OVRD_EN(phy, ch);
> + i915->chv_phy_control |= PHY_CH_POWER_DOWN_OVRD_EN(phy, ch);
> else
> - dev_priv->chv_phy_control &= ~PHY_CH_POWER_DOWN_OVRD_EN(phy, ch);
> + i915->chv_phy_control &= ~PHY_CH_POWER_DOWN_OVRD_EN(phy, ch);
>
> - I915_WRITE(DISPLAY_PHY_CONTROL, dev_priv->chv_phy_control);
> + I915_WRITE(DISPLAY_PHY_CONTROL, i915->chv_phy_control);
>
> DRM_DEBUG_KMS("Power gating DPIO PHY%d CH%d (DPIO_PHY_CONTROL=0x%08x)\n",
> - phy, ch, dev_priv->chv_phy_control);
> + phy, ch, i915->chv_phy_control);
>
> - assert_chv_phy_status(dev_priv);
> + assert_chv_phy_status(i915);
>
> out:
> mutex_unlock(&power_domains->lock);
> @@ -1427,43 +1427,43 @@ bool chv_phy_powergate_ch(struct drm_i915_private *dev_priv, enum dpio_phy phy,
> void chv_phy_powergate_lanes(struct intel_encoder *encoder,
> bool override, unsigned int mask)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> - struct i915_power_domains *power_domains = &dev_priv->power_domains;
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> + struct i915_power_domains *power_domains = &i915->power_domains;
> enum dpio_phy phy = vlv_dport_to_phy(enc_to_dig_port(&encoder->base));
> enum dpio_channel ch = vlv_dport_to_channel(enc_to_dig_port(&encoder->base));
>
> mutex_lock(&power_domains->lock);
>
> - dev_priv->chv_phy_control &= ~PHY_CH_POWER_DOWN_OVRD(0xf, phy, ch);
> - dev_priv->chv_phy_control |= PHY_CH_POWER_DOWN_OVRD(mask, phy, ch);
> + i915->chv_phy_control &= ~PHY_CH_POWER_DOWN_OVRD(0xf, phy, ch);
> + i915->chv_phy_control |= PHY_CH_POWER_DOWN_OVRD(mask, phy, ch);
>
> if (override)
> - dev_priv->chv_phy_control |= PHY_CH_POWER_DOWN_OVRD_EN(phy, ch);
> + i915->chv_phy_control |= PHY_CH_POWER_DOWN_OVRD_EN(phy, ch);
> else
> - dev_priv->chv_phy_control &= ~PHY_CH_POWER_DOWN_OVRD_EN(phy, ch);
> + i915->chv_phy_control &= ~PHY_CH_POWER_DOWN_OVRD_EN(phy, ch);
>
> - I915_WRITE(DISPLAY_PHY_CONTROL, dev_priv->chv_phy_control);
> + I915_WRITE(DISPLAY_PHY_CONTROL, i915->chv_phy_control);
>
> DRM_DEBUG_KMS("Power gating DPIO PHY%d CH%d lanes 0x%x (PHY_CONTROL=0x%08x)\n",
> - phy, ch, mask, dev_priv->chv_phy_control);
> + phy, ch, mask, i915->chv_phy_control);
>
> - assert_chv_phy_status(dev_priv);
> + assert_chv_phy_status(i915);
>
> - assert_chv_phy_powergate(dev_priv, phy, ch, override, mask);
> + assert_chv_phy_powergate(i915, phy, ch, override, mask);
>
> mutex_unlock(&power_domains->lock);
> }
>
> -static bool chv_pipe_power_well_enabled(struct drm_i915_private *dev_priv,
> +static bool chv_pipe_power_well_enabled(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> enum pipe pipe = PIPE_A;
> bool enabled;
> u32 state, ctrl;
>
> - vlv_punit_get(dev_priv);
> + vlv_punit_get(i915);
>
> - state = vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM) & DP_SSS_MASK(pipe);
> + state = vlv_punit_read(i915, PUNIT_REG_DSPSSPM) & DP_SSS_MASK(pipe);
> /*
> * We only ever set the power-on and power-gate states, anything
> * else is unexpected.
> @@ -1475,15 +1475,15 @@ static bool chv_pipe_power_well_enabled(struct drm_i915_private *dev_priv,
> * A transient state at this point would mean some unexpected party
> * is poking at the power controls too.
> */
> - ctrl = vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM) & DP_SSC_MASK(pipe);
> + ctrl = vlv_punit_read(i915, PUNIT_REG_DSPSSPM) & DP_SSC_MASK(pipe);
> WARN_ON(ctrl << 16 != state);
>
> - vlv_punit_put(dev_priv);
> + vlv_punit_put(i915);
>
> return enabled;
> }
>
> -static void chv_set_pipe_power_well(struct drm_i915_private *dev_priv,
> +static void chv_set_pipe_power_well(struct drm_i915_private *i915,
> struct i915_power_well *power_well,
> bool enable)
> {
> @@ -1493,44 +1493,44 @@ static void chv_set_pipe_power_well(struct drm_i915_private *dev_priv,
>
> state = enable ? DP_SSS_PWR_ON(pipe) : DP_SSS_PWR_GATE(pipe);
>
> - vlv_punit_get(dev_priv);
> + vlv_punit_get(i915);
>
> #define COND \
> - ((vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM) & DP_SSS_MASK(pipe)) == state)
> + ((vlv_punit_read(i915, PUNIT_REG_DSPSSPM) & DP_SSS_MASK(pipe)) == state)
>
> if (COND)
> goto out;
>
> - ctrl = vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM);
> + ctrl = vlv_punit_read(i915, PUNIT_REG_DSPSSPM);
> ctrl &= ~DP_SSC_MASK(pipe);
> ctrl |= enable ? DP_SSC_PWR_ON(pipe) : DP_SSC_PWR_GATE(pipe);
> - vlv_punit_write(dev_priv, PUNIT_REG_DSPSSPM, ctrl);
> + vlv_punit_write(i915, PUNIT_REG_DSPSSPM, ctrl);
>
> if (wait_for(COND, 100))
> DRM_ERROR("timeout setting power well state %08x (%08x)\n",
> state,
> - vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM));
> + vlv_punit_read(i915, PUNIT_REG_DSPSSPM));
>
> #undef COND
>
> out:
> - vlv_punit_put(dev_priv);
> + vlv_punit_put(i915);
> }
>
> -static void chv_pipe_power_well_enable(struct drm_i915_private *dev_priv,
> +static void chv_pipe_power_well_enable(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> - chv_set_pipe_power_well(dev_priv, power_well, true);
> + chv_set_pipe_power_well(i915, power_well, true);
>
> - vlv_display_power_well_init(dev_priv);
> + vlv_display_power_well_init(i915);
> }
>
> -static void chv_pipe_power_well_disable(struct drm_i915_private *dev_priv,
> +static void chv_pipe_power_well_disable(struct drm_i915_private *i915,
> struct i915_power_well *power_well)
> {
> - vlv_display_power_well_deinit(dev_priv);
> + vlv_display_power_well_deinit(i915);
>
> - chv_set_pipe_power_well(dev_priv, power_well, false);
> + chv_set_pipe_power_well(i915, power_well, false);
> }
>
> static u64 __async_put_domains_mask(struct i915_power_domains *power_domains)
> @@ -1627,10 +1627,10 @@ async_put_domains_clear_domain(struct i915_power_domains *power_domains,
> }
>
> static bool
> -intel_display_power_grab_async_put_ref(struct drm_i915_private *dev_priv,
> +intel_display_power_grab_async_put_ref(struct drm_i915_private *i915,
> enum intel_display_power_domain domain)
> {
> - struct i915_power_domains *power_domains = &dev_priv->power_domains;
> + struct i915_power_domains *power_domains = &i915->power_domains;
> bool ret = false;
>
> if (!(async_put_domains_mask(power_domains) & BIT_ULL(domain)))
> @@ -1644,7 +1644,7 @@ intel_display_power_grab_async_put_ref(struct drm_i915_private *dev_priv,
> goto out_verify;
>
> cancel_delayed_work(&power_domains->async_put_work);
> - intel_runtime_pm_put_raw(dev_priv,
> + intel_runtime_pm_put_raw(i915,
> fetch_and_zero(&power_domains->async_put_wakeref));
> out_verify:
> verify_async_put_domains_state(power_domains);
> @@ -1653,24 +1653,24 @@ intel_display_power_grab_async_put_ref(struct drm_i915_private *dev_priv,
> }
>
> static void
> -__intel_display_power_get_domain(struct drm_i915_private *dev_priv,
> +__intel_display_power_get_domain(struct drm_i915_private *i915,
> enum intel_display_power_domain domain)
> {
> - struct i915_power_domains *power_domains = &dev_priv->power_domains;
> + struct i915_power_domains *power_domains = &i915->power_domains;
> struct i915_power_well *power_well;
>
> - if (intel_display_power_grab_async_put_ref(dev_priv, domain))
> + if (intel_display_power_grab_async_put_ref(i915, domain))
> return;
>
> - for_each_power_domain_well(dev_priv, power_well, BIT_ULL(domain))
> - intel_power_well_get(dev_priv, power_well);
> + for_each_power_domain_well(i915, power_well, BIT_ULL(domain))
> + intel_power_well_get(i915, power_well);
>
> power_domains->domain_use_count[domain]++;
> }
>
> /**
> * intel_display_power_get - grab a power domain reference
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> * @domain: power domain to reference
> *
> * This function grabs a power domain reference for @domain and ensures that the
> @@ -1680,14 +1680,14 @@ __intel_display_power_get_domain(struct drm_i915_private *dev_priv,
> * Any power domain reference obtained by this function must have a symmetric
> * call to intel_display_power_put() to release the reference again.
> */
> -intel_wakeref_t intel_display_power_get(struct drm_i915_private *dev_priv,
> +intel_wakeref_t intel_display_power_get(struct drm_i915_private *i915,
> enum intel_display_power_domain domain)
> {
> - struct i915_power_domains *power_domains = &dev_priv->power_domains;
> - intel_wakeref_t wakeref = intel_runtime_pm_get(dev_priv);
> + struct i915_power_domains *power_domains = &i915->power_domains;
> + intel_wakeref_t wakeref = intel_runtime_pm_get(i915);
>
> mutex_lock(&power_domains->lock);
> - __intel_display_power_get_domain(dev_priv, domain);
> + __intel_display_power_get_domain(i915, domain);
> mutex_unlock(&power_domains->lock);
>
> return wakeref;
> @@ -1695,7 +1695,7 @@ intel_wakeref_t intel_display_power_get(struct drm_i915_private *dev_priv,
>
> /**
> * intel_display_power_get_if_enabled - grab a reference for an enabled display power domain
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> * @domain: power domain to reference
> *
> * This function grabs a power domain reference for @domain and ensures that the
> @@ -1706,21 +1706,21 @@ intel_wakeref_t intel_display_power_get(struct drm_i915_private *dev_priv,
> * call to intel_display_power_put() to release the reference again.
> */
> intel_wakeref_t
> -intel_display_power_get_if_enabled(struct drm_i915_private *dev_priv,
> +intel_display_power_get_if_enabled(struct drm_i915_private *i915,
> enum intel_display_power_domain domain)
> {
> - struct i915_power_domains *power_domains = &dev_priv->power_domains;
> + struct i915_power_domains *power_domains = &i915->power_domains;
> intel_wakeref_t wakeref;
> bool is_enabled;
>
> - wakeref = intel_runtime_pm_get_if_in_use(dev_priv);
> + wakeref = intel_runtime_pm_get_if_in_use(i915);
> if (!wakeref)
> return false;
>
> mutex_lock(&power_domains->lock);
>
> - if (__intel_display_power_is_enabled(dev_priv, domain)) {
> - __intel_display_power_get_domain(dev_priv, domain);
> + if (__intel_display_power_is_enabled(i915, domain)) {
> + __intel_display_power_get_domain(i915, domain);
> is_enabled = true;
> } else {
> is_enabled = false;
> @@ -1729,7 +1729,7 @@ intel_display_power_get_if_enabled(struct drm_i915_private *dev_priv,
> mutex_unlock(&power_domains->lock);
>
> if (!is_enabled) {
> - intel_runtime_pm_put(dev_priv, wakeref);
> + intel_runtime_pm_put(i915, wakeref);
> wakeref = 0;
> }
>
> @@ -1737,14 +1737,14 @@ intel_display_power_get_if_enabled(struct drm_i915_private *dev_priv,
> }
>
> static void
> -__intel_display_power_put_domain(struct drm_i915_private *dev_priv,
> +__intel_display_power_put_domain(struct drm_i915_private *i915,
> enum intel_display_power_domain domain)
> {
> struct i915_power_domains *power_domains;
> struct i915_power_well *power_well;
> const char *name = intel_display_power_domain_str(domain);
>
> - power_domains = &dev_priv->power_domains;
> + power_domains = &i915->power_domains;
>
> WARN(!power_domains->domain_use_count[domain],
> "Use count on domain %s is already zero\n",
> @@ -1755,23 +1755,23 @@ __intel_display_power_put_domain(struct drm_i915_private *dev_priv,
>
> power_domains->domain_use_count[domain]--;
>
> - for_each_power_domain_well_reverse(dev_priv, power_well, BIT_ULL(domain))
> - intel_power_well_put(dev_priv, power_well);
> + for_each_power_domain_well_reverse(i915, power_well, BIT_ULL(domain))
> + intel_power_well_put(i915, power_well);
> }
>
> -static void __intel_display_power_put(struct drm_i915_private *dev_priv,
> +static void __intel_display_power_put(struct drm_i915_private *i915,
> enum intel_display_power_domain domain)
> {
> - struct i915_power_domains *power_domains = &dev_priv->power_domains;
> + struct i915_power_domains *power_domains = &i915->power_domains;
>
> mutex_lock(&power_domains->lock);
> - __intel_display_power_put_domain(dev_priv, domain);
> + __intel_display_power_put_domain(i915, domain);
> mutex_unlock(&power_domains->lock);
> }
>
> /**
> * intel_display_power_put_unchecked - release an unchecked power domain reference
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> * @domain: power domain to reference
> *
> * This function drops the power domain reference obtained by
> @@ -1782,11 +1782,11 @@ static void __intel_display_power_put(struct drm_i915_private *dev_priv,
> * new code, as the correctness of its use cannot be checked. Always use
> * intel_display_power_put() instead.
> */
> -void intel_display_power_put_unchecked(struct drm_i915_private *dev_priv,
> +void intel_display_power_put_unchecked(struct drm_i915_private *i915,
> enum intel_display_power_domain domain)
> {
> - __intel_display_power_put(dev_priv, domain);
> - intel_runtime_pm_put_unchecked(dev_priv);
> + __intel_display_power_put(i915, domain);
> + intel_runtime_pm_put_unchecked(i915);
> }
>
> static void
> @@ -1803,7 +1803,7 @@ queue_async_put_domains_work(struct i915_power_domains *power_domains,
> static void
> release_async_put_domains(struct i915_power_domains *power_domains, u64 mask)
> {
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> container_of(power_domains, struct drm_i915_private,
> power_domains);
> enum intel_display_power_domain domain;
> @@ -1814,26 +1814,26 @@ release_async_put_domains(struct i915_power_domains *power_domains, u64 mask)
> * wakeref to make the state checker happy about the HW access during
> * power well disabling.
> */
> - assert_rpm_raw_wakeref_held(dev_priv);
> - wakeref = intel_runtime_pm_get(dev_priv);
> + assert_rpm_raw_wakeref_held(i915);
> + wakeref = intel_runtime_pm_get(i915);
>
> for_each_power_domain(domain, mask) {
> /* Clear before put, so put's sanity check is happy. */
> async_put_domains_clear_domain(power_domains, domain);
> - __intel_display_power_put_domain(dev_priv, domain);
> + __intel_display_power_put_domain(i915, domain);
> }
>
> - intel_runtime_pm_put(dev_priv, wakeref);
> + intel_runtime_pm_put(i915, wakeref);
> }
>
> static void
> intel_display_power_put_async_work(struct work_struct *work)
> {
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> container_of(work, struct drm_i915_private,
> power_domains.async_put_work.work);
> - struct i915_power_domains *power_domains = &dev_priv->power_domains;
> - intel_wakeref_t new_work_wakeref = intel_runtime_pm_get_raw(dev_priv);
> + struct i915_power_domains *power_domains = &i915->power_domains;
> + intel_wakeref_t new_work_wakeref = intel_runtime_pm_get_raw(i915);
> intel_wakeref_t old_work_wakeref = 0;
>
> mutex_lock(&power_domains->lock);
> @@ -1863,9 +1863,9 @@ intel_display_power_put_async_work(struct work_struct *work)
> mutex_unlock(&power_domains->lock);
>
> if (old_work_wakeref)
> - intel_runtime_pm_put_raw(dev_priv, old_work_wakeref);
> + intel_runtime_pm_put_raw(i915, old_work_wakeref);
> if (new_work_wakeref)
> - intel_runtime_pm_put_raw(dev_priv, new_work_wakeref);
> + intel_runtime_pm_put_raw(i915, new_work_wakeref);
> }
>
> /**
> @@ -1974,7 +1974,7 @@ intel_display_power_flush_work_sync(struct drm_i915_private *i915)
> #if IS_ENABLED(CONFIG_DRM_I915_DEBUG_RUNTIME_PM)
> /**
> * intel_display_power_put - release a power domain reference
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> * @domain: power domain to reference
> * @wakeref: wakeref acquired for the reference that is being released
> *
> @@ -1982,12 +1982,12 @@ intel_display_power_flush_work_sync(struct drm_i915_private *i915)
> * intel_display_power_get() and might power down the corresponding hardware
> * block right away if this is the last reference.
> */
> -void intel_display_power_put(struct drm_i915_private *dev_priv,
> +void intel_display_power_put(struct drm_i915_private *i915,
> enum intel_display_power_domain domain,
> intel_wakeref_t wakeref)
> {
> - __intel_display_power_put(dev_priv, domain);
> - intel_runtime_pm_put(dev_priv, wakeref);
> + __intel_display_power_put(i915, domain);
> + intel_runtime_pm_put(i915, wakeref);
> }
> #endif
>
> @@ -2655,14 +2655,14 @@ static const struct i915_power_well_desc chv_power_wells[] = {
> },
> };
>
> -bool intel_display_power_well_is_enabled(struct drm_i915_private *dev_priv,
> +bool intel_display_power_well_is_enabled(struct drm_i915_private *i915,
> enum i915_power_well_id power_well_id)
> {
> struct i915_power_well *power_well;
> bool ret;
>
> - power_well = lookup_power_well(dev_priv, power_well_id);
> - ret = power_well->desc->ops->is_enabled(dev_priv, power_well);
> + power_well = lookup_power_well(i915, power_well_id);
> + ret = power_well->desc->ops->is_enabled(i915, power_well);
>
> return ret;
> }
> @@ -3352,7 +3352,7 @@ static const struct i915_power_well_desc icl_power_wells[] = {
> };
>
> static int
> -sanitize_disable_power_well_option(const struct drm_i915_private *dev_priv,
> +sanitize_disable_power_well_option(const struct drm_i915_private *i915,
> int disable_power_well)
> {
> if (disable_power_well >= 0)
> @@ -3361,14 +3361,14 @@ sanitize_disable_power_well_option(const struct drm_i915_private *dev_priv,
> return 1;
> }
>
> -static u32 get_allowed_dc_mask(const struct drm_i915_private *dev_priv,
> +static u32 get_allowed_dc_mask(const struct drm_i915_private *i915,
> int enable_dc)
> {
> u32 mask;
> int requested_dc;
> int max_dc;
>
> - if (INTEL_GEN(dev_priv) >= 11) {
> + if (INTEL_GEN(i915) >= 11) {
> max_dc = 2;
> /*
> * DC9 has a separate HW flow from the rest of the DC states,
> @@ -3376,10 +3376,10 @@ static u32 get_allowed_dc_mask(const struct drm_i915_private *dev_priv,
> * suspend/resume, so allow it unconditionally.
> */
> mask = DC_STATE_EN_DC9;
> - } else if (IS_GEN(dev_priv, 10) || IS_GEN9_BC(dev_priv)) {
> + } else if (IS_GEN(i915, 10) || IS_GEN9_BC(i915)) {
> max_dc = 2;
> mask = 0;
> - } else if (IS_GEN9_LP(dev_priv)) {
> + } else if (IS_GEN9_LP(i915)) {
> max_dc = 1;
> mask = DC_STATE_EN_DC9;
> } else {
> @@ -3451,21 +3451,21 @@ __set_power_wells(struct i915_power_domains *power_domains,
>
> /**
> * intel_power_domains_init - initializes the power domain structures
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> *
> - * Initializes the power domain structures for @dev_priv depending upon the
> + * Initializes the power domain structures for @i915 depending upon the
> * supported platform.
> */
> -int intel_power_domains_init(struct drm_i915_private *dev_priv)
> +int intel_power_domains_init(struct drm_i915_private *i915)
> {
> - struct i915_power_domains *power_domains = &dev_priv->power_domains;
> + struct i915_power_domains *power_domains = &i915->power_domains;
> int err;
>
> i915_modparams.disable_power_well =
> - sanitize_disable_power_well_option(dev_priv,
> + sanitize_disable_power_well_option(i915,
> i915_modparams.disable_power_well);
> - dev_priv->csr.allowed_dc_mask =
> - get_allowed_dc_mask(dev_priv, i915_modparams.enable_dc);
> + i915->csr.allowed_dc_mask =
> + get_allowed_dc_mask(i915, i915_modparams.enable_dc);
>
> BUILD_BUG_ON(POWER_DOMAIN_NUM > 64);
>
> @@ -3478,9 +3478,9 @@ int intel_power_domains_init(struct drm_i915_private *dev_priv)
> * The enabling order will be from lower to higher indexed wells,
> * the disabling order is reversed.
> */
> - if (IS_GEN(dev_priv, 11)) {
> + if (IS_GEN(i915, 11)) {
> err = set_power_wells(power_domains, icl_power_wells);
> - } else if (IS_CANNONLAKE(dev_priv)) {
> + } else if (IS_CANNONLAKE(i915)) {
> err = set_power_wells(power_domains, cnl_power_wells);
>
> /*
> @@ -3489,23 +3489,23 @@ int intel_power_domains_init(struct drm_i915_private *dev_priv)
> * timeouts, lets remove them from the list
> * for the SKUs without port F.
> */
> - if (!IS_CNL_WITH_PORT_F(dev_priv))
> + if (!IS_CNL_WITH_PORT_F(i915))
> power_domains->power_well_count -= 2;
> - } else if (IS_GEMINILAKE(dev_priv)) {
> + } else if (IS_GEMINILAKE(i915)) {
> err = set_power_wells(power_domains, glk_power_wells);
> - } else if (IS_BROXTON(dev_priv)) {
> + } else if (IS_BROXTON(i915)) {
> err = set_power_wells(power_domains, bxt_power_wells);
> - } else if (IS_GEN9_BC(dev_priv)) {
> + } else if (IS_GEN9_BC(i915)) {
> err = set_power_wells(power_domains, skl_power_wells);
> - } else if (IS_CHERRYVIEW(dev_priv)) {
> + } else if (IS_CHERRYVIEW(i915)) {
> err = set_power_wells(power_domains, chv_power_wells);
> - } else if (IS_BROADWELL(dev_priv)) {
> + } else if (IS_BROADWELL(i915)) {
> err = set_power_wells(power_domains, bdw_power_wells);
> - } else if (IS_HASWELL(dev_priv)) {
> + } else if (IS_HASWELL(i915)) {
> err = set_power_wells(power_domains, hsw_power_wells);
> - } else if (IS_VALLEYVIEW(dev_priv)) {
> + } else if (IS_VALLEYVIEW(i915)) {
> err = set_power_wells(power_domains, vlv_power_wells);
> - } else if (IS_I830(dev_priv)) {
> + } else if (IS_I830(i915)) {
> err = set_power_wells(power_domains, i830_power_wells);
> } else {
> err = set_power_wells(power_domains, i9xx_always_on_power_well);
> @@ -3516,31 +3516,31 @@ int intel_power_domains_init(struct drm_i915_private *dev_priv)
>
> /**
> * intel_power_domains_cleanup - clean up power domains resources
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> *
> * Release any resources acquired by intel_power_domains_init()
> */
> -void intel_power_domains_cleanup(struct drm_i915_private *dev_priv)
> +void intel_power_domains_cleanup(struct drm_i915_private *i915)
> {
> - kfree(dev_priv->power_domains.power_wells);
> + kfree(i915->power_domains.power_wells);
> }
>
> -static void intel_power_domains_sync_hw(struct drm_i915_private *dev_priv)
> +static void intel_power_domains_sync_hw(struct drm_i915_private *i915)
> {
> - struct i915_power_domains *power_domains = &dev_priv->power_domains;
> + struct i915_power_domains *power_domains = &i915->power_domains;
> struct i915_power_well *power_well;
>
> mutex_lock(&power_domains->lock);
> - for_each_power_well(dev_priv, power_well) {
> - power_well->desc->ops->sync_hw(dev_priv, power_well);
> + for_each_power_well(i915, power_well) {
> + power_well->desc->ops->sync_hw(i915, power_well);
> power_well->hw_enabled =
> - power_well->desc->ops->is_enabled(dev_priv, power_well);
> + power_well->desc->ops->is_enabled(i915, power_well);
> }
> mutex_unlock(&power_domains->lock);
> }
>
> static inline
> -bool intel_dbuf_slice_set(struct drm_i915_private *dev_priv,
> +bool intel_dbuf_slice_set(struct drm_i915_private *i915,
> i915_reg_t reg, bool enable)
> {
> u32 val, status;
> @@ -3560,30 +3560,30 @@ bool intel_dbuf_slice_set(struct drm_i915_private *dev_priv,
> return true;
> }
>
> -static void gen9_dbuf_enable(struct drm_i915_private *dev_priv)
> +static void gen9_dbuf_enable(struct drm_i915_private *i915)
> {
> - intel_dbuf_slice_set(dev_priv, DBUF_CTL, true);
> + intel_dbuf_slice_set(i915, DBUF_CTL, true);
> }
>
> -static void gen9_dbuf_disable(struct drm_i915_private *dev_priv)
> +static void gen9_dbuf_disable(struct drm_i915_private *i915)
> {
> - intel_dbuf_slice_set(dev_priv, DBUF_CTL, false);
> + intel_dbuf_slice_set(i915, DBUF_CTL, false);
> }
>
> -static u8 intel_dbuf_max_slices(struct drm_i915_private *dev_priv)
> +static u8 intel_dbuf_max_slices(struct drm_i915_private *i915)
> {
> - if (INTEL_GEN(dev_priv) < 11)
> + if (INTEL_GEN(i915) < 11)
> return 1;
> return 2;
> }
>
> -void icl_dbuf_slices_update(struct drm_i915_private *dev_priv,
> +void icl_dbuf_slices_update(struct drm_i915_private *i915,
> u8 req_slices)
> {
> - const u8 hw_enabled_slices = dev_priv->wm.skl_hw.ddb.enabled_slices;
> + const u8 hw_enabled_slices = i915->wm.skl_hw.ddb.enabled_slices;
> bool ret;
>
> - if (req_slices > intel_dbuf_max_slices(dev_priv)) {
> + if (req_slices > intel_dbuf_max_slices(i915)) {
> DRM_ERROR("Invalid number of dbuf slices requested\n");
> return;
> }
> @@ -3592,15 +3592,15 @@ void icl_dbuf_slices_update(struct drm_i915_private *dev_priv,
> return;
>
> if (req_slices > hw_enabled_slices)
> - ret = intel_dbuf_slice_set(dev_priv, DBUF_CTL_S2, true);
> + ret = intel_dbuf_slice_set(i915, DBUF_CTL_S2, true);
> else
> - ret = intel_dbuf_slice_set(dev_priv, DBUF_CTL_S2, false);
> + ret = intel_dbuf_slice_set(i915, DBUF_CTL_S2, false);
>
> if (ret)
> - dev_priv->wm.skl_hw.ddb.enabled_slices = req_slices;
> + i915->wm.skl_hw.ddb.enabled_slices = req_slices;
> }
>
> -static void icl_dbuf_enable(struct drm_i915_private *dev_priv)
> +static void icl_dbuf_enable(struct drm_i915_private *i915)
> {
> I915_WRITE(DBUF_CTL_S1, I915_READ(DBUF_CTL_S1) | DBUF_POWER_REQUEST);
> I915_WRITE(DBUF_CTL_S2, I915_READ(DBUF_CTL_S2) | DBUF_POWER_REQUEST);
> @@ -3616,10 +3616,10 @@ static void icl_dbuf_enable(struct drm_i915_private *dev_priv)
> * FIXME: for now pretend that we only have 1 slice, see
> * intel_enabled_dbuf_slices_num().
> */
> - dev_priv->wm.skl_hw.ddb.enabled_slices = 1;
> + i915->wm.skl_hw.ddb.enabled_slices = 1;
> }
>
> -static void icl_dbuf_disable(struct drm_i915_private *dev_priv)
> +static void icl_dbuf_disable(struct drm_i915_private *i915)
> {
> I915_WRITE(DBUF_CTL_S1, I915_READ(DBUF_CTL_S1) & ~DBUF_POWER_REQUEST);
> I915_WRITE(DBUF_CTL_S2, I915_READ(DBUF_CTL_S2) & ~DBUF_POWER_REQUEST);
> @@ -3635,10 +3635,10 @@ static void icl_dbuf_disable(struct drm_i915_private *dev_priv)
> * FIXME: for now pretend that the first slice is always
> * enabled, see intel_enabled_dbuf_slices_num().
> */
> - dev_priv->wm.skl_hw.ddb.enabled_slices = 1;
> + i915->wm.skl_hw.ddb.enabled_slices = 1;
> }
>
> -static void icl_mbus_init(struct drm_i915_private *dev_priv)
> +static void icl_mbus_init(struct drm_i915_private *i915)
> {
> u32 val;
>
> @@ -3650,7 +3650,7 @@ static void icl_mbus_init(struct drm_i915_private *dev_priv)
> I915_WRITE(MBUS_ABOX_CTL, val);
> }
>
> -static void hsw_assert_cdclk(struct drm_i915_private *dev_priv)
> +static void hsw_assert_cdclk(struct drm_i915_private *i915)
> {
> u32 val = I915_READ(LCPLL_CTL);
>
> @@ -3670,9 +3670,9 @@ static void hsw_assert_cdclk(struct drm_i915_private *dev_priv)
> DRM_ERROR("LCPLL not using non-SSC reference\n");
> }
>
> -static void assert_can_disable_lcpll(struct drm_i915_private *dev_priv)
> +static void assert_can_disable_lcpll(struct drm_i915_private *i915)
> {
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_device *dev = &i915->drm;
> struct intel_crtc *crtc;
>
> for_each_intel_crtc(dev, crtc)
> @@ -3691,7 +3691,7 @@ static void assert_can_disable_lcpll(struct drm_i915_private *dev_priv)
> "Panel power on\n");
> I915_STATE_WARN(I915_READ(BLC_PWM_CPU_CTL2) & BLM_PWM_ENABLE,
> "CPU PWM1 enabled\n");
> - if (IS_HASWELL(dev_priv))
> + if (IS_HASWELL(i915))
> I915_STATE_WARN(I915_READ(HSW_BLC_PWM2_CTL) & BLM_PWM_ENABLE,
> "CPU PWM2 enabled\n");
> I915_STATE_WARN(I915_READ(BLC_PWM_PCH_CTL1) & BLM_PCH_PWM_ENABLE,
> @@ -3707,21 +3707,21 @@ static void assert_can_disable_lcpll(struct drm_i915_private *dev_priv)
> * gen-specific and since we only disable LCPLL after we fully disable
> * the interrupts, the check below should be enough.
> */
> - I915_STATE_WARN(intel_irqs_enabled(dev_priv), "IRQs enabled\n");
> + I915_STATE_WARN(intel_irqs_enabled(i915), "IRQs enabled\n");
> }
>
> -static u32 hsw_read_dcomp(struct drm_i915_private *dev_priv)
> +static u32 hsw_read_dcomp(struct drm_i915_private *i915)
> {
> - if (IS_HASWELL(dev_priv))
> + if (IS_HASWELL(i915))
> return I915_READ(D_COMP_HSW);
> else
> return I915_READ(D_COMP_BDW);
> }
>
> -static void hsw_write_dcomp(struct drm_i915_private *dev_priv, u32 val)
> +static void hsw_write_dcomp(struct drm_i915_private *i915, u32 val)
> {
> - if (IS_HASWELL(dev_priv)) {
> - if (sandybridge_pcode_write(dev_priv,
> + if (IS_HASWELL(i915)) {
> + if (sandybridge_pcode_write(i915,
> GEN6_PCODE_WRITE_D_COMP, val))
> DRM_DEBUG_KMS("Failed to write to D_COMP\n");
> } else {
> @@ -3738,12 +3738,12 @@ static void hsw_write_dcomp(struct drm_i915_private *dev_priv, u32 val)
> * register. Callers should take care of disabling all the display engine
> * functions, doing the mode unset, fixing interrupts, etc.
> */
> -static void hsw_disable_lcpll(struct drm_i915_private *dev_priv,
> +static void hsw_disable_lcpll(struct drm_i915_private *i915,
> bool switch_to_fclk, bool allow_power_down)
> {
> u32 val;
>
> - assert_can_disable_lcpll(dev_priv);
> + assert_can_disable_lcpll(i915);
>
> val = I915_READ(LCPLL_CTL);
>
> @@ -3762,16 +3762,16 @@ static void hsw_disable_lcpll(struct drm_i915_private *dev_priv,
> I915_WRITE(LCPLL_CTL, val);
> POSTING_READ(LCPLL_CTL);
>
> - if (intel_wait_for_register(&dev_priv->uncore, LCPLL_CTL,
> + if (intel_wait_for_register(&i915->uncore, LCPLL_CTL,
> LCPLL_PLL_LOCK, 0, 1))
> DRM_ERROR("LCPLL still locked\n");
>
> - val = hsw_read_dcomp(dev_priv);
> + val = hsw_read_dcomp(i915);
> val |= D_COMP_COMP_DISABLE;
> - hsw_write_dcomp(dev_priv, val);
> + hsw_write_dcomp(i915, val);
> ndelay(100);
>
> - if (wait_for((hsw_read_dcomp(dev_priv) &
> + if (wait_for((hsw_read_dcomp(i915) &
> D_COMP_RCOMP_IN_PROGRESS) == 0, 1))
> DRM_ERROR("D_COMP RCOMP still in progress\n");
>
> @@ -3787,7 +3787,7 @@ static void hsw_disable_lcpll(struct drm_i915_private *dev_priv,
> * Fully restores LCPLL, disallowing power down and switching back to LCPLL
> * source.
> */
> -static void hsw_restore_lcpll(struct drm_i915_private *dev_priv)
> +static void hsw_restore_lcpll(struct drm_i915_private *i915)
> {
> u32 val;
>
> @@ -3801,7 +3801,7 @@ static void hsw_restore_lcpll(struct drm_i915_private *dev_priv)
> * Make sure we're not on PC8 state before disabling PC8, otherwise
> * we'll hang the machine. To prevent PC8 state, just enable force_wake.
> */
> - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
>
> if (val & LCPLL_POWER_DOWN_ALLOW) {
> val &= ~LCPLL_POWER_DOWN_ALLOW;
> @@ -3809,16 +3809,16 @@ static void hsw_restore_lcpll(struct drm_i915_private *dev_priv)
> POSTING_READ(LCPLL_CTL);
> }
>
> - val = hsw_read_dcomp(dev_priv);
> + val = hsw_read_dcomp(i915);
> val |= D_COMP_COMP_FORCE;
> val &= ~D_COMP_COMP_DISABLE;
> - hsw_write_dcomp(dev_priv, val);
> + hsw_write_dcomp(i915, val);
>
> val = I915_READ(LCPLL_CTL);
> val &= ~LCPLL_PLL_DISABLE;
> I915_WRITE(LCPLL_CTL, val);
>
> - if (intel_wait_for_register(&dev_priv->uncore, LCPLL_CTL,
> + if (intel_wait_for_register(&i915->uncore, LCPLL_CTL,
> LCPLL_PLL_LOCK, LCPLL_PLL_LOCK, 5))
> DRM_ERROR("LCPLL not locked yet\n");
>
> @@ -3832,10 +3832,10 @@ static void hsw_restore_lcpll(struct drm_i915_private *dev_priv)
> DRM_ERROR("Switching back to LCPLL failed\n");
> }
>
> - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
>
> - intel_update_cdclk(dev_priv);
> - intel_dump_cdclk_state(&dev_priv->cdclk.hw, "Current CDCLK");
> + intel_update_cdclk(i915);
> + intel_dump_cdclk_state(&i915->cdclk.hw, "Current CDCLK");
> }
>
> /*
> @@ -3861,45 +3861,45 @@ static void hsw_restore_lcpll(struct drm_i915_private *dev_priv)
> * For more, read "Display Sequences for Package C8" on the hardware
> * documentation.
> */
> -void hsw_enable_pc8(struct drm_i915_private *dev_priv)
> +void hsw_enable_pc8(struct drm_i915_private *i915)
> {
> u32 val;
>
> DRM_DEBUG_KMS("Enabling package C8+\n");
>
> - if (HAS_PCH_LPT_LP(dev_priv)) {
> + if (HAS_PCH_LPT_LP(i915)) {
> val = I915_READ(SOUTH_DSPCLK_GATE_D);
> val &= ~PCH_LP_PARTITION_LEVEL_DISABLE;
> I915_WRITE(SOUTH_DSPCLK_GATE_D, val);
> }
>
> - lpt_disable_clkout_dp(dev_priv);
> - hsw_disable_lcpll(dev_priv, true, true);
> + lpt_disable_clkout_dp(i915);
> + hsw_disable_lcpll(i915, true, true);
> }
>
> -void hsw_disable_pc8(struct drm_i915_private *dev_priv)
> +void hsw_disable_pc8(struct drm_i915_private *i915)
> {
> u32 val;
>
> DRM_DEBUG_KMS("Disabling package C8+\n");
>
> - hsw_restore_lcpll(dev_priv);
> - intel_init_pch_refclk(dev_priv);
> + hsw_restore_lcpll(i915);
> + intel_init_pch_refclk(i915);
>
> - if (HAS_PCH_LPT_LP(dev_priv)) {
> + if (HAS_PCH_LPT_LP(i915)) {
> val = I915_READ(SOUTH_DSPCLK_GATE_D);
> val |= PCH_LP_PARTITION_LEVEL_DISABLE;
> I915_WRITE(SOUTH_DSPCLK_GATE_D, val);
> }
> }
>
> -static void intel_pch_reset_handshake(struct drm_i915_private *dev_priv,
> +static void intel_pch_reset_handshake(struct drm_i915_private *i915,
> bool enable)
> {
> i915_reg_t reg;
> u32 reset_bits, val;
>
> - if (IS_IVYBRIDGE(dev_priv)) {
> + if (IS_IVYBRIDGE(i915)) {
> reg = GEN7_MSG_CTL;
> reset_bits = WAIT_FOR_PCH_FLR_ACK | WAIT_FOR_PCH_RESET_ACK;
> } else {
> @@ -3917,46 +3917,46 @@ static void intel_pch_reset_handshake(struct drm_i915_private *dev_priv,
> I915_WRITE(reg, val);
> }
>
> -static void skl_display_core_init(struct drm_i915_private *dev_priv,
> +static void skl_display_core_init(struct drm_i915_private *i915,
> bool resume)
> {
> - struct i915_power_domains *power_domains = &dev_priv->power_domains;
> + struct i915_power_domains *power_domains = &i915->power_domains;
> struct i915_power_well *well;
>
> - gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
> + gen9_set_dc_state(i915, DC_STATE_DISABLE);
>
> /* enable PCH reset handshake */
> - intel_pch_reset_handshake(dev_priv, !HAS_PCH_NOP(dev_priv));
> + intel_pch_reset_handshake(i915, !HAS_PCH_NOP(i915));
>
> /* enable PG1 and Misc I/O */
> mutex_lock(&power_domains->lock);
>
> - well = lookup_power_well(dev_priv, SKL_DISP_PW_1);
> - intel_power_well_enable(dev_priv, well);
> + well = lookup_power_well(i915, SKL_DISP_PW_1);
> + intel_power_well_enable(i915, well);
>
> - well = lookup_power_well(dev_priv, SKL_DISP_PW_MISC_IO);
> - intel_power_well_enable(dev_priv, well);
> + well = lookup_power_well(i915, SKL_DISP_PW_MISC_IO);
> + intel_power_well_enable(i915, well);
>
> mutex_unlock(&power_domains->lock);
>
> - intel_cdclk_init(dev_priv);
> + intel_cdclk_init(i915);
>
> - gen9_dbuf_enable(dev_priv);
> + gen9_dbuf_enable(i915);
>
> - if (resume && dev_priv->csr.dmc_payload)
> - intel_csr_load_program(dev_priv);
> + if (resume && i915->csr.dmc_payload)
> + intel_csr_load_program(i915);
> }
>
> -static void skl_display_core_uninit(struct drm_i915_private *dev_priv)
> +static void skl_display_core_uninit(struct drm_i915_private *i915)
> {
> - struct i915_power_domains *power_domains = &dev_priv->power_domains;
> + struct i915_power_domains *power_domains = &i915->power_domains;
> struct i915_power_well *well;
>
> - gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
> + gen9_set_dc_state(i915, DC_STATE_DISABLE);
>
> - gen9_dbuf_disable(dev_priv);
> + gen9_dbuf_disable(i915);
>
> - intel_cdclk_uninit(dev_priv);
> + intel_cdclk_uninit(i915);
>
> /* The spec doesn't call for removing the reset handshake flag */
> /* disable PG1 and Misc I/O */
> @@ -3969,21 +3969,21 @@ static void skl_display_core_uninit(struct drm_i915_private *dev_priv)
> * Note that even though the driver's request is removed power well 1
> * may stay enabled after this due to DMC's own request on it.
> */
> - well = lookup_power_well(dev_priv, SKL_DISP_PW_1);
> - intel_power_well_disable(dev_priv, well);
> + well = lookup_power_well(i915, SKL_DISP_PW_1);
> + intel_power_well_disable(i915, well);
>
> mutex_unlock(&power_domains->lock);
>
> usleep_range(10, 30); /* 10 us delay per Bspec */
> }
>
> -void bxt_display_core_init(struct drm_i915_private *dev_priv,
> +void bxt_display_core_init(struct drm_i915_private *i915,
> bool resume)
> {
> - struct i915_power_domains *power_domains = &dev_priv->power_domains;
> + struct i915_power_domains *power_domains = &i915->power_domains;
> struct i915_power_well *well;
>
> - gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
> + gen9_set_dc_state(i915, DC_STATE_DISABLE);
>
> /*
> * NDE_RSTWRN_OPT RST PCH Handshake En must always be 0b on BXT
> @@ -3991,34 +3991,34 @@ void bxt_display_core_init(struct drm_i915_private *dev_priv,
> * Move the handshake programming to initialization sequence.
> * Previously was left up to BIOS.
> */
> - intel_pch_reset_handshake(dev_priv, false);
> + intel_pch_reset_handshake(i915, false);
>
> /* Enable PG1 */
> mutex_lock(&power_domains->lock);
>
> - well = lookup_power_well(dev_priv, SKL_DISP_PW_1);
> - intel_power_well_enable(dev_priv, well);
> + well = lookup_power_well(i915, SKL_DISP_PW_1);
> + intel_power_well_enable(i915, well);
>
> mutex_unlock(&power_domains->lock);
>
> - intel_cdclk_init(dev_priv);
> + intel_cdclk_init(i915);
>
> - gen9_dbuf_enable(dev_priv);
> + gen9_dbuf_enable(i915);
>
> - if (resume && dev_priv->csr.dmc_payload)
> - intel_csr_load_program(dev_priv);
> + if (resume && i915->csr.dmc_payload)
> + intel_csr_load_program(i915);
> }
>
> -void bxt_display_core_uninit(struct drm_i915_private *dev_priv)
> +void bxt_display_core_uninit(struct drm_i915_private *i915)
> {
> - struct i915_power_domains *power_domains = &dev_priv->power_domains;
> + struct i915_power_domains *power_domains = &i915->power_domains;
> struct i915_power_well *well;
>
> - gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
> + gen9_set_dc_state(i915, DC_STATE_DISABLE);
>
> - gen9_dbuf_disable(dev_priv);
> + gen9_dbuf_disable(i915);
>
> - intel_cdclk_uninit(dev_priv);
> + intel_cdclk_uninit(i915);
>
> /* The spec doesn't call for removing the reset handshake flag */
>
> @@ -4029,60 +4029,60 @@ void bxt_display_core_uninit(struct drm_i915_private *dev_priv)
> */
> mutex_lock(&power_domains->lock);
>
> - well = lookup_power_well(dev_priv, SKL_DISP_PW_1);
> - intel_power_well_disable(dev_priv, well);
> + well = lookup_power_well(i915, SKL_DISP_PW_1);
> + intel_power_well_disable(i915, well);
>
> mutex_unlock(&power_domains->lock);
>
> usleep_range(10, 30); /* 10 us delay per Bspec */
> }
>
> -static void cnl_display_core_init(struct drm_i915_private *dev_priv, bool resume)
> +static void cnl_display_core_init(struct drm_i915_private *i915, bool resume)
> {
> - struct i915_power_domains *power_domains = &dev_priv->power_domains;
> + struct i915_power_domains *power_domains = &i915->power_domains;
> struct i915_power_well *well;
>
> - gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
> + gen9_set_dc_state(i915, DC_STATE_DISABLE);
>
> /* 1. Enable PCH Reset Handshake */
> - intel_pch_reset_handshake(dev_priv, !HAS_PCH_NOP(dev_priv));
> + intel_pch_reset_handshake(i915, !HAS_PCH_NOP(i915));
>
> /* 2-3. */
> - intel_combo_phy_init(dev_priv);
> + intel_combo_phy_init(i915);
>
> /*
> * 4. Enable Power Well 1 (PG1).
> * The AUX IO power wells will be enabled on demand.
> */
> mutex_lock(&power_domains->lock);
> - well = lookup_power_well(dev_priv, SKL_DISP_PW_1);
> - intel_power_well_enable(dev_priv, well);
> + well = lookup_power_well(i915, SKL_DISP_PW_1);
> + intel_power_well_enable(i915, well);
> mutex_unlock(&power_domains->lock);
>
> /* 5. Enable CD clock */
> - intel_cdclk_init(dev_priv);
> + intel_cdclk_init(i915);
>
> /* 6. Enable DBUF */
> - gen9_dbuf_enable(dev_priv);
> + gen9_dbuf_enable(i915);
>
> - if (resume && dev_priv->csr.dmc_payload)
> - intel_csr_load_program(dev_priv);
> + if (resume && i915->csr.dmc_payload)
> + intel_csr_load_program(i915);
> }
>
> -static void cnl_display_core_uninit(struct drm_i915_private *dev_priv)
> +static void cnl_display_core_uninit(struct drm_i915_private *i915)
> {
> - struct i915_power_domains *power_domains = &dev_priv->power_domains;
> + struct i915_power_domains *power_domains = &i915->power_domains;
> struct i915_power_well *well;
>
> - gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
> + gen9_set_dc_state(i915, DC_STATE_DISABLE);
>
> /* 1. Disable all display engine functions -> aready done */
>
> /* 2. Disable DBUF */
> - gen9_dbuf_disable(dev_priv);
> + gen9_dbuf_disable(i915);
>
> /* 3. Disable CD clock */
> - intel_cdclk_uninit(dev_priv);
> + intel_cdclk_uninit(i915);
>
> /*
> * 4. Disable Power Well 1 (PG1).
> @@ -4090,66 +4090,66 @@ static void cnl_display_core_uninit(struct drm_i915_private *dev_priv)
> * disabled at this point.
> */
> mutex_lock(&power_domains->lock);
> - well = lookup_power_well(dev_priv, SKL_DISP_PW_1);
> - intel_power_well_disable(dev_priv, well);
> + well = lookup_power_well(i915, SKL_DISP_PW_1);
> + intel_power_well_disable(i915, well);
> mutex_unlock(&power_domains->lock);
>
> usleep_range(10, 30); /* 10 us delay per Bspec */
>
> /* 5. */
> - intel_combo_phy_uninit(dev_priv);
> + intel_combo_phy_uninit(i915);
> }
>
> -void icl_display_core_init(struct drm_i915_private *dev_priv,
> +void icl_display_core_init(struct drm_i915_private *i915,
> bool resume)
> {
> - struct i915_power_domains *power_domains = &dev_priv->power_domains;
> + struct i915_power_domains *power_domains = &i915->power_domains;
> struct i915_power_well *well;
>
> - gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
> + gen9_set_dc_state(i915, DC_STATE_DISABLE);
>
> /* 1. Enable PCH reset handshake. */
> - intel_pch_reset_handshake(dev_priv, !HAS_PCH_NOP(dev_priv));
> + intel_pch_reset_handshake(i915, !HAS_PCH_NOP(i915));
>
> /* 2. Initialize all combo phys */
> - intel_combo_phy_init(dev_priv);
> + intel_combo_phy_init(i915);
>
> /*
> * 3. Enable Power Well 1 (PG1).
> * The AUX IO power wells will be enabled on demand.
> */
> mutex_lock(&power_domains->lock);
> - well = lookup_power_well(dev_priv, SKL_DISP_PW_1);
> - intel_power_well_enable(dev_priv, well);
> + well = lookup_power_well(i915, SKL_DISP_PW_1);
> + intel_power_well_enable(i915, well);
> mutex_unlock(&power_domains->lock);
>
> /* 4. Enable CDCLK. */
> - intel_cdclk_init(dev_priv);
> + intel_cdclk_init(i915);
>
> /* 5. Enable DBUF. */
> - icl_dbuf_enable(dev_priv);
> + icl_dbuf_enable(i915);
>
> /* 6. Setup MBUS. */
> - icl_mbus_init(dev_priv);
> + icl_mbus_init(i915);
>
> - if (resume && dev_priv->csr.dmc_payload)
> - intel_csr_load_program(dev_priv);
> + if (resume && i915->csr.dmc_payload)
> + intel_csr_load_program(i915);
> }
>
> -void icl_display_core_uninit(struct drm_i915_private *dev_priv)
> +void icl_display_core_uninit(struct drm_i915_private *i915)
> {
> - struct i915_power_domains *power_domains = &dev_priv->power_domains;
> + struct i915_power_domains *power_domains = &i915->power_domains;
> struct i915_power_well *well;
>
> - gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
> + gen9_set_dc_state(i915, DC_STATE_DISABLE);
>
> /* 1. Disable all display engine functions -> aready done */
>
> /* 2. Disable DBUF */
> - icl_dbuf_disable(dev_priv);
> + icl_dbuf_disable(i915);
>
> /* 3. Disable CD clock */
> - intel_cdclk_uninit(dev_priv);
> + intel_cdclk_uninit(i915);
>
> /*
> * 4. Disable Power Well 1 (PG1).
> @@ -4157,20 +4157,20 @@ void icl_display_core_uninit(struct drm_i915_private *dev_priv)
> * disabled at this point.
> */
> mutex_lock(&power_domains->lock);
> - well = lookup_power_well(dev_priv, SKL_DISP_PW_1);
> - intel_power_well_disable(dev_priv, well);
> + well = lookup_power_well(i915, SKL_DISP_PW_1);
> + intel_power_well_disable(i915, well);
> mutex_unlock(&power_domains->lock);
>
> /* 5. */
> - intel_combo_phy_uninit(dev_priv);
> + intel_combo_phy_uninit(i915);
> }
>
> -static void chv_phy_control_init(struct drm_i915_private *dev_priv)
> +static void chv_phy_control_init(struct drm_i915_private *i915)
> {
> struct i915_power_well *cmn_bc =
> - lookup_power_well(dev_priv, VLV_DISP_PW_DPIO_CMN_BC);
> + lookup_power_well(i915, VLV_DISP_PW_DPIO_CMN_BC);
> struct i915_power_well *cmn_d =
> - lookup_power_well(dev_priv, CHV_DISP_PW_DPIO_CMN_D);
> + lookup_power_well(i915, CHV_DISP_PW_DPIO_CMN_D);
>
> /*
> * DISPLAY_PHY_CONTROL can get corrupted if read. As a
> @@ -4179,7 +4179,7 @@ static void chv_phy_control_init(struct drm_i915_private *dev_priv)
> * power well state and lane status to reconstruct the
> * expected initial value.
> */
> - dev_priv->chv_phy_control =
> + i915->chv_phy_control =
> PHY_LDO_SEQ_DELAY(PHY_LDO_DELAY_600NS, DPIO_PHY0) |
> PHY_LDO_SEQ_DELAY(PHY_LDO_DELAY_600NS, DPIO_PHY1) |
> PHY_CH_POWER_MODE(PHY_CH_DEEP_PSR, DPIO_PHY0, DPIO_CH0) |
> @@ -4193,7 +4193,7 @@ static void chv_phy_control_init(struct drm_i915_private *dev_priv)
> * override and set the lane powerdown bits accding to the
> * current lane status.
> */
> - if (cmn_bc->desc->ops->is_enabled(dev_priv, cmn_bc)) {
> + if (cmn_bc->desc->ops->is_enabled(i915, cmn_bc)) {
> u32 status = I915_READ(DPLL(PIPE_A));
> unsigned int mask;
>
> @@ -4201,30 +4201,30 @@ static void chv_phy_control_init(struct drm_i915_private *dev_priv)
> if (mask == 0xf)
> mask = 0x0;
> else
> - dev_priv->chv_phy_control |=
> + i915->chv_phy_control |=
> PHY_CH_POWER_DOWN_OVRD_EN(DPIO_PHY0, DPIO_CH0);
>
> - dev_priv->chv_phy_control |=
> + i915->chv_phy_control |=
> PHY_CH_POWER_DOWN_OVRD(mask, DPIO_PHY0, DPIO_CH0);
>
> mask = (status & DPLL_PORTC_READY_MASK) >> 4;
> if (mask == 0xf)
> mask = 0x0;
> else
> - dev_priv->chv_phy_control |=
> + i915->chv_phy_control |=
> PHY_CH_POWER_DOWN_OVRD_EN(DPIO_PHY0, DPIO_CH1);
>
> - dev_priv->chv_phy_control |=
> + i915->chv_phy_control |=
> PHY_CH_POWER_DOWN_OVRD(mask, DPIO_PHY0, DPIO_CH1);
>
> - dev_priv->chv_phy_control |= PHY_COM_LANE_RESET_DEASSERT(DPIO_PHY0);
> + i915->chv_phy_control |= PHY_COM_LANE_RESET_DEASSERT(DPIO_PHY0);
>
> - dev_priv->chv_phy_assert[DPIO_PHY0] = false;
> + i915->chv_phy_assert[DPIO_PHY0] = false;
> } else {
> - dev_priv->chv_phy_assert[DPIO_PHY0] = true;
> + i915->chv_phy_assert[DPIO_PHY0] = true;
> }
>
> - if (cmn_d->desc->ops->is_enabled(dev_priv, cmn_d)) {
> + if (cmn_d->desc->ops->is_enabled(i915, cmn_d)) {
> u32 status = I915_READ(DPIO_PHY_STATUS);
> unsigned int mask;
>
> @@ -4233,42 +4233,42 @@ static void chv_phy_control_init(struct drm_i915_private *dev_priv)
> if (mask == 0xf)
> mask = 0x0;
> else
> - dev_priv->chv_phy_control |=
> + i915->chv_phy_control |=
> PHY_CH_POWER_DOWN_OVRD_EN(DPIO_PHY1, DPIO_CH0);
>
> - dev_priv->chv_phy_control |=
> + i915->chv_phy_control |=
> PHY_CH_POWER_DOWN_OVRD(mask, DPIO_PHY1, DPIO_CH0);
>
> - dev_priv->chv_phy_control |= PHY_COM_LANE_RESET_DEASSERT(DPIO_PHY1);
> + i915->chv_phy_control |= PHY_COM_LANE_RESET_DEASSERT(DPIO_PHY1);
>
> - dev_priv->chv_phy_assert[DPIO_PHY1] = false;
> + i915->chv_phy_assert[DPIO_PHY1] = false;
> } else {
> - dev_priv->chv_phy_assert[DPIO_PHY1] = true;
> + i915->chv_phy_assert[DPIO_PHY1] = true;
> }
>
> - I915_WRITE(DISPLAY_PHY_CONTROL, dev_priv->chv_phy_control);
> + I915_WRITE(DISPLAY_PHY_CONTROL, i915->chv_phy_control);
>
> DRM_DEBUG_KMS("Initial PHY_CONTROL=0x%08x\n",
> - dev_priv->chv_phy_control);
> + i915->chv_phy_control);
> }
>
> -static void vlv_cmnlane_wa(struct drm_i915_private *dev_priv)
> +static void vlv_cmnlane_wa(struct drm_i915_private *i915)
> {
> struct i915_power_well *cmn =
> - lookup_power_well(dev_priv, VLV_DISP_PW_DPIO_CMN_BC);
> + lookup_power_well(i915, VLV_DISP_PW_DPIO_CMN_BC);
> struct i915_power_well *disp2d =
> - lookup_power_well(dev_priv, VLV_DISP_PW_DISP2D);
> + lookup_power_well(i915, VLV_DISP_PW_DISP2D);
>
> /* If the display might be already active skip this */
> - if (cmn->desc->ops->is_enabled(dev_priv, cmn) &&
> - disp2d->desc->ops->is_enabled(dev_priv, disp2d) &&
> + if (cmn->desc->ops->is_enabled(i915, cmn) &&
> + disp2d->desc->ops->is_enabled(i915, disp2d) &&
> I915_READ(DPIO_CTL) & DPIO_CMNRST)
> return;
>
> DRM_DEBUG_KMS("toggling display PHY side reset\n");
>
> /* cmnlane needs DPLL registers */
> - disp2d->desc->ops->enable(dev_priv, disp2d);
> + disp2d->desc->ops->enable(i915, disp2d);
>
> /*
> * From VLV2A0_DP_eDP_HDMI_DPIO_driver_vbios_notes_11.docx:
> @@ -4277,27 +4277,27 @@ static void vlv_cmnlane_wa(struct drm_i915_private *dev_priv)
> * Simply ungating isn't enough to reset the PHY enough to get
> * ports and lanes running.
> */
> - cmn->desc->ops->disable(dev_priv, cmn);
> + cmn->desc->ops->disable(i915, cmn);
> }
>
> -static bool vlv_punit_is_power_gated(struct drm_i915_private *dev_priv, u32 reg0)
> +static bool vlv_punit_is_power_gated(struct drm_i915_private *i915, u32 reg0)
> {
> bool ret;
>
> - vlv_punit_get(dev_priv);
> - ret = (vlv_punit_read(dev_priv, reg0) & SSPM0_SSC_MASK) == SSPM0_SSC_PWR_GATE;
> - vlv_punit_put(dev_priv);
> + vlv_punit_get(i915);
> + ret = (vlv_punit_read(i915, reg0) & SSPM0_SSC_MASK) == SSPM0_SSC_PWR_GATE;
> + vlv_punit_put(i915);
>
> return ret;
> }
>
> -static void assert_ved_power_gated(struct drm_i915_private *dev_priv)
> +static void assert_ved_power_gated(struct drm_i915_private *i915)
> {
> - WARN(!vlv_punit_is_power_gated(dev_priv, PUNIT_REG_VEDSSPM0),
> + WARN(!vlv_punit_is_power_gated(i915, PUNIT_REG_VEDSSPM0),
> "VED not power gated\n");
> }
>
> -static void assert_isp_power_gated(struct drm_i915_private *dev_priv)
> +static void assert_isp_power_gated(struct drm_i915_private *i915)
> {
> static const struct pci_device_id isp_ids[] = {
> {PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x0f38)},
> @@ -4306,11 +4306,11 @@ static void assert_isp_power_gated(struct drm_i915_private *dev_priv)
> };
>
> WARN(!pci_dev_present(isp_ids) &&
> - !vlv_punit_is_power_gated(dev_priv, PUNIT_REG_ISPSSPM0),
> + !vlv_punit_is_power_gated(i915, PUNIT_REG_ISPSSPM0),
> "ISP not power gated\n");
> }
>
> -static void intel_power_domains_verify_state(struct drm_i915_private *dev_priv);
> +static void intel_power_domains_verify_state(struct drm_i915_private *i915);
>
> /**
> * intel_power_domains_init_hw - initialize hardware power domain state
> diff --git a/drivers/gpu/drm/i915/intel_display_power.h b/drivers/gpu/drm/i915/intel_display_power.h
> index ff57b0a7fe59..0310fe114ecc 100644
> --- a/drivers/gpu/drm/i915/intel_display_power.h
> +++ b/drivers/gpu/drm/i915/intel_display_power.h
> @@ -80,23 +80,23 @@ struct i915_power_well_ops {
> * during driver init and resume time, possibly after first calling
> * the enable/disable handlers.
> */
> - void (*sync_hw)(struct drm_i915_private *dev_priv,
> + void (*sync_hw)(struct drm_i915_private *i915,
> struct i915_power_well *power_well);
> /*
> * Enable the well and resources that depend on it (for example
> * interrupts located on the well). Called after the 0->1 refcount
> * transition.
> */
> - void (*enable)(struct drm_i915_private *dev_priv,
> + void (*enable)(struct drm_i915_private *i915,
> struct i915_power_well *power_well);
> /*
> * Disable the well and resources that depend on it. Called after
> * the 1->0 refcount transition.
> */
> - void (*disable)(struct drm_i915_private *dev_priv,
> + void (*disable)(struct drm_i915_private *i915,
> struct i915_power_well *power_well);
> /* Returns the hw enabled state. */
> - bool (*is_enabled)(struct drm_i915_private *dev_priv,
> + bool (*is_enabled)(struct drm_i915_private *i915,
> struct i915_power_well *power_well);
> };
>
> @@ -204,48 +204,48 @@ struct i915_power_domains {
> for_each_power_well_reverse(__dev_priv, __power_well) \
> for_each_if((__power_well)->desc->domains & (__domain_mask))
>
> -void skl_enable_dc6(struct drm_i915_private *dev_priv);
> -void gen9_sanitize_dc_state(struct drm_i915_private *dev_priv);
> -void bxt_enable_dc9(struct drm_i915_private *dev_priv);
> -void bxt_disable_dc9(struct drm_i915_private *dev_priv);
> -void gen9_enable_dc5(struct drm_i915_private *dev_priv);
> +void skl_enable_dc6(struct drm_i915_private *i915);
> +void gen9_sanitize_dc_state(struct drm_i915_private *i915);
> +void bxt_enable_dc9(struct drm_i915_private *i915);
> +void bxt_disable_dc9(struct drm_i915_private *i915);
> +void gen9_enable_dc5(struct drm_i915_private *i915);
>
> -int intel_power_domains_init(struct drm_i915_private *dev_priv);
> -void intel_power_domains_cleanup(struct drm_i915_private *dev_priv);
> -void intel_power_domains_init_hw(struct drm_i915_private *dev_priv, bool resume);
> -void intel_power_domains_fini_hw(struct drm_i915_private *dev_priv);
> -void icl_display_core_init(struct drm_i915_private *dev_priv, bool resume);
> -void icl_display_core_uninit(struct drm_i915_private *dev_priv);
> -void intel_power_domains_enable(struct drm_i915_private *dev_priv);
> -void intel_power_domains_disable(struct drm_i915_private *dev_priv);
> -void intel_power_domains_suspend(struct drm_i915_private *dev_priv,
> +int intel_power_domains_init(struct drm_i915_private *i915);
> +void intel_power_domains_cleanup(struct drm_i915_private *i915);
> +void intel_power_domains_init_hw(struct drm_i915_private *i915, bool resume);
> +void intel_power_domains_fini_hw(struct drm_i915_private *i915);
> +void icl_display_core_init(struct drm_i915_private *i915, bool resume);
> +void icl_display_core_uninit(struct drm_i915_private *i915);
> +void intel_power_domains_enable(struct drm_i915_private *i915);
> +void intel_power_domains_disable(struct drm_i915_private *i915);
> +void intel_power_domains_suspend(struct drm_i915_private *i915,
> enum i915_drm_suspend_mode);
> -void intel_power_domains_resume(struct drm_i915_private *dev_priv);
> -void hsw_enable_pc8(struct drm_i915_private *dev_priv);
> -void hsw_disable_pc8(struct drm_i915_private *dev_priv);
> -void bxt_display_core_init(struct drm_i915_private *dev_priv, bool resume);
> -void bxt_display_core_uninit(struct drm_i915_private *dev_priv);
> +void intel_power_domains_resume(struct drm_i915_private *i915);
> +void hsw_enable_pc8(struct drm_i915_private *i915);
> +void hsw_disable_pc8(struct drm_i915_private *i915);
> +void bxt_display_core_init(struct drm_i915_private *i915, bool resume);
> +void bxt_display_core_uninit(struct drm_i915_private *i915);
>
> const char *
> intel_display_power_domain_str(enum intel_display_power_domain domain);
>
> -bool intel_display_power_is_enabled(struct drm_i915_private *dev_priv,
> +bool intel_display_power_is_enabled(struct drm_i915_private *i915,
> enum intel_display_power_domain domain);
> -bool __intel_display_power_is_enabled(struct drm_i915_private *dev_priv,
> +bool __intel_display_power_is_enabled(struct drm_i915_private *i915,
> enum intel_display_power_domain domain);
> -intel_wakeref_t intel_display_power_get(struct drm_i915_private *dev_priv,
> +intel_wakeref_t intel_display_power_get(struct drm_i915_private *i915,
> enum intel_display_power_domain domain);
> intel_wakeref_t
> -intel_display_power_get_if_enabled(struct drm_i915_private *dev_priv,
> +intel_display_power_get_if_enabled(struct drm_i915_private *i915,
> enum intel_display_power_domain domain);
> -void intel_display_power_put_unchecked(struct drm_i915_private *dev_priv,
> +void intel_display_power_put_unchecked(struct drm_i915_private *i915,
> enum intel_display_power_domain domain);
> void __intel_display_power_put_async(struct drm_i915_private *i915,
> enum intel_display_power_domain domain,
> intel_wakeref_t wakeref);
> void intel_display_power_flush_work(struct drm_i915_private *i915);
> #if IS_ENABLED(CONFIG_DRM_I915_DEBUG_RUNTIME_PM)
> -void intel_display_power_put(struct drm_i915_private *dev_priv,
> +void intel_display_power_put(struct drm_i915_private *i915,
> enum intel_display_power_domain domain,
> intel_wakeref_t wakeref);
> static inline void
> @@ -277,12 +277,12 @@ intel_display_power_put_async(struct drm_i915_private *i915,
> for ((wf) = intel_display_power_get((i915), (domain)); (wf); \
> intel_display_power_put_async((i915), (domain), (wf)), (wf) = 0)
>
> -void icl_dbuf_slices_update(struct drm_i915_private *dev_priv,
> +void icl_dbuf_slices_update(struct drm_i915_private *i915,
> u8 req_slices);
>
> void chv_phy_powergate_lanes(struct intel_encoder *encoder,
> bool override, unsigned int mask);
> -bool chv_phy_powergate_ch(struct drm_i915_private *dev_priv, enum dpio_phy phy,
> +bool chv_phy_powergate_ch(struct drm_i915_private *i915, enum dpio_phy phy,
> enum dpio_channel ch, bool override);
>
> #endif /* __INTEL_DISPLAY_POWER_H__ */
> diff --git a/drivers/gpu/drm/i915/intel_dp.c b/drivers/gpu/drm/i915/intel_dp.c
> index 4336df46fe78..d4dc2b9506e5 100644
> --- a/drivers/gpu/drm/i915/intel_dp.c
> +++ b/drivers/gpu/drm/i915/intel_dp.c
> @@ -160,7 +160,7 @@ static bool edp_panel_vdd_on(struct intel_dp *intel_dp);
> static void edp_panel_vdd_off(struct intel_dp *intel_dp, bool sync);
> static void vlv_init_panel_power_sequencer(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state);
> -static void vlv_steal_power_sequencer(struct drm_i915_private *dev_priv,
> +static void vlv_steal_power_sequencer(struct drm_i915_private *i915,
> enum pipe pipe);
> static void intel_dp_unset_edid(struct intel_dp *intel_dp);
>
> @@ -214,8 +214,8 @@ static int intel_dp_max_common_rate(struct intel_dp *intel_dp)
> static int intel_dp_get_fia_supported_lane_count(struct intel_dp *intel_dp)
> {
> struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> - struct drm_i915_private *dev_priv = to_i915(dig_port->base.base.dev);
> - enum tc_port tc_port = intel_port_to_tc(dev_priv, dig_port->base.port);
> + struct drm_i915_private *i915 = to_i915(dig_port->base.base.dev);
> + enum tc_port tc_port = intel_port_to_tc(i915, dig_port->base.port);
> intel_wakeref_t wakeref;
> u32 lane_info;
>
> @@ -223,7 +223,7 @@ static int intel_dp_get_fia_supported_lane_count(struct intel_dp *intel_dp)
> return 4;
>
> lane_info = 0;
> - with_intel_display_power(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref)
> + with_intel_display_power(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref)
> lane_info = (I915_READ(PORT_TX_DFLEXDPSP) &
> DP_LANE_ASSIGNMENT_MASK(tc_port)) >>
> DP_LANE_ASSIGNMENT_SHIFT(tc_port);
> @@ -284,8 +284,8 @@ intel_dp_downstream_max_dotclock(struct intel_dp *intel_dp)
> {
> struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp);
> struct intel_encoder *encoder = &intel_dig_port->base;
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> - int max_dotclk = dev_priv->max_dotclk_freq;
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> + int max_dotclk = i915->max_dotclk_freq;
> int ds_max_dotclk;
>
> int type = intel_dp->downstream_ports[0] & DP_DS_PORT_TYPE_MASK;
> @@ -305,7 +305,7 @@ intel_dp_downstream_max_dotclock(struct intel_dp *intel_dp)
> static int cnl_max_source_rate(struct intel_dp *intel_dp)
> {
> struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> - struct drm_i915_private *dev_priv = to_i915(dig_port->base.base.dev);
> + struct drm_i915_private *i915 = to_i915(dig_port->base.base.dev);
> enum port port = dig_port->base.port;
>
> u32 voltage = I915_READ(CNL_PORT_COMP_DW3) & VOLTAGE_INFO_MASK;
> @@ -315,7 +315,7 @@ static int cnl_max_source_rate(struct intel_dp *intel_dp)
> return 540000;
>
> /* For this SKU 8.1G is supported in all ports */
> - if (IS_CNL_WITH_PORT_F(dev_priv))
> + if (IS_CNL_WITH_PORT_F(i915))
> return 810000;
>
> /* For other SKUs, max rate on ports A and D is 5.4G */
> @@ -328,11 +328,11 @@ static int cnl_max_source_rate(struct intel_dp *intel_dp)
> static int icl_max_source_rate(struct intel_dp *intel_dp)
> {
> struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> - struct drm_i915_private *dev_priv = to_i915(dig_port->base.base.dev);
> + struct drm_i915_private *i915 = to_i915(dig_port->base.base.dev);
> enum port port = dig_port->base.port;
>
> - if (intel_port_is_combophy(dev_priv, port) &&
> - !IS_ELKHARTLAKE(dev_priv) &&
> + if (intel_port_is_combophy(i915, port) &&
> + !IS_ELKHARTLAKE(i915) &&
> !intel_dp_is_edp(intel_dp))
> return 540000;
>
> @@ -359,30 +359,30 @@ intel_dp_set_source_rates(struct intel_dp *intel_dp)
> 162000, 270000
> };
> struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> - struct drm_i915_private *dev_priv = to_i915(dig_port->base.base.dev);
> + struct drm_i915_private *i915 = to_i915(dig_port->base.base.dev);
> const struct ddi_vbt_port_info *info =
> - &dev_priv->vbt.ddi_port_info[dig_port->base.port];
> + &i915->vbt.ddi_port_info[dig_port->base.port];
> const int *source_rates;
> int size, max_rate = 0, vbt_max_rate = info->dp_max_link_rate;
>
> /* This should only be done once */
> WARN_ON(intel_dp->source_rates || intel_dp->num_source_rates);
>
> - if (INTEL_GEN(dev_priv) >= 10) {
> + if (INTEL_GEN(i915) >= 10) {
> source_rates = cnl_rates;
> size = ARRAY_SIZE(cnl_rates);
> - if (IS_GEN(dev_priv, 10))
> + if (IS_GEN(i915, 10))
> max_rate = cnl_max_source_rate(intel_dp);
> else
> max_rate = icl_max_source_rate(intel_dp);
> - } else if (IS_GEN9_LP(dev_priv)) {
> + } else if (IS_GEN9_LP(i915)) {
> source_rates = bxt_rates;
> size = ARRAY_SIZE(bxt_rates);
> - } else if (IS_GEN9_BC(dev_priv)) {
> + } else if (IS_GEN9_BC(i915)) {
> source_rates = skl_rates;
> size = ARRAY_SIZE(skl_rates);
> - } else if ((IS_HASWELL(dev_priv) && !IS_HSW_ULX(dev_priv)) ||
> - IS_BROADWELL(dev_priv)) {
> + } else if ((IS_HASWELL(i915) && !IS_HSW_ULX(i915)) ||
> + IS_BROADWELL(i915)) {
> source_rates = hsw_rates;
> size = ARRAY_SIZE(hsw_rates);
> } else {
> @@ -532,7 +532,7 @@ intel_dp_mode_valid(struct drm_connector *connector,
> struct intel_dp *intel_dp = intel_attached_dp(connector);
> struct intel_connector *intel_connector = to_intel_connector(connector);
> struct drm_display_mode *fixed_mode = intel_connector->panel.fixed_mode;
> - struct drm_i915_private *dev_priv = to_i915(connector->dev);
> + struct drm_i915_private *i915 = to_i915(connector->dev);
> int target_clock = mode->clock;
> int max_rate, mode_rate, max_lanes, max_link_clock;
> int max_dotclk;
> @@ -564,7 +564,7 @@ intel_dp_mode_valid(struct drm_connector *connector,
> * Output bpp is stored in 6.4 format so right shift by 4 to get the
> * integer value since we support only integer values of bpp.
> */
> - if ((INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) &&
> + if ((INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) &&
> drm_dp_sink_supports_dsc(intel_dp->dsc_dpcd)) {
> if (intel_dp_is_edp(intel_dp)) {
> dsc_max_output_bpp =
> @@ -630,17 +630,17 @@ intel_dp_pps_init(struct intel_dp *intel_dp);
> static intel_wakeref_t
> pps_lock(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> intel_wakeref_t wakeref;
>
> /*
> * See intel_power_sequencer_reset() why we need
> * a power domain reference here.
> */
> - wakeref = intel_display_power_get(dev_priv,
> + wakeref = intel_display_power_get(i915,
> intel_aux_power_domain(dp_to_dig_port(intel_dp)));
>
> - mutex_lock(&dev_priv->pps_mutex);
> + mutex_lock(&i915->pps_mutex);
>
> return wakeref;
> }
> @@ -648,10 +648,10 @@ pps_lock(struct intel_dp *intel_dp)
> static intel_wakeref_t
> pps_unlock(struct intel_dp *intel_dp, intel_wakeref_t wakeref)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
>
> - mutex_unlock(&dev_priv->pps_mutex);
> - intel_display_power_put(dev_priv,
> + mutex_unlock(&i915->pps_mutex);
> + intel_display_power_put(i915,
> intel_aux_power_domain(dp_to_dig_port(intel_dp)),
> wakeref);
> return 0;
> @@ -663,7 +663,7 @@ pps_unlock(struct intel_dp *intel_dp, intel_wakeref_t wakeref)
> static void
> vlv_power_sequencer_kick(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp);
> enum pipe pipe = intel_dp->pps_pipe;
> bool pll_enabled, release_cl_override = false;
> @@ -687,7 +687,7 @@ vlv_power_sequencer_kick(struct intel_dp *intel_dp)
> DP |= DP_PORT_WIDTH(1);
> DP |= DP_LINK_TRAIN_PAT_1;
>
> - if (IS_CHERRYVIEW(dev_priv))
> + if (IS_CHERRYVIEW(i915))
> DP |= DP_PIPE_SEL_CHV(pipe);
> else
> DP |= DP_PIPE_SEL(pipe);
> @@ -699,10 +699,10 @@ vlv_power_sequencer_kick(struct intel_dp *intel_dp)
> * So enable temporarily it if it's not already enabled.
> */
> if (!pll_enabled) {
> - release_cl_override = IS_CHERRYVIEW(dev_priv) &&
> - !chv_phy_powergate_ch(dev_priv, phy, ch, true);
> + release_cl_override = IS_CHERRYVIEW(i915) &&
> + !chv_phy_powergate_ch(i915, phy, ch, true);
>
> - if (vlv_force_pll_on(dev_priv, pipe, IS_CHERRYVIEW(dev_priv) ?
> + if (vlv_force_pll_on(i915, pipe, IS_CHERRYVIEW(i915) ?
> &chv_dpll[0].dpll : &vlv_dpll[0].dpll)) {
> DRM_ERROR("Failed to force on pll for pipe %c!\n",
> pipe_name(pipe));
> @@ -726,14 +726,14 @@ vlv_power_sequencer_kick(struct intel_dp *intel_dp)
> POSTING_READ(intel_dp->output_reg);
>
> if (!pll_enabled) {
> - vlv_force_pll_off(dev_priv, pipe);
> + vlv_force_pll_off(i915, pipe);
>
> if (release_cl_override)
> - chv_phy_powergate_ch(dev_priv, phy, ch, false);
> + chv_phy_powergate_ch(i915, phy, ch, false);
> }
> }
>
> -static enum pipe vlv_find_free_pps(struct drm_i915_private *dev_priv)
> +static enum pipe vlv_find_free_pps(struct drm_i915_private *i915)
> {
> struct intel_encoder *encoder;
> unsigned int pipes = (1 << PIPE_A) | (1 << PIPE_B);
> @@ -742,7 +742,7 @@ static enum pipe vlv_find_free_pps(struct drm_i915_private *dev_priv)
> * We don't have power sequencer currently.
> * Pick one that's not used by other ports.
> */
> - for_each_intel_dp(&dev_priv->drm, encoder) {
> + for_each_intel_dp(&i915->drm, encoder) {
> struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
>
> if (encoder->type == INTEL_OUTPUT_EDP) {
> @@ -768,11 +768,11 @@ static enum pipe vlv_find_free_pps(struct drm_i915_private *dev_priv)
> static enum pipe
> vlv_power_sequencer_pipe(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp);
> enum pipe pipe;
>
> - lockdep_assert_held(&dev_priv->pps_mutex);
> + lockdep_assert_held(&i915->pps_mutex);
>
> /* We should never land here with regular DP ports */
> WARN_ON(!intel_dp_is_edp(intel_dp));
> @@ -783,7 +783,7 @@ vlv_power_sequencer_pipe(struct intel_dp *intel_dp)
> if (intel_dp->pps_pipe != INVALID_PIPE)
> return intel_dp->pps_pipe;
>
> - pipe = vlv_find_free_pps(dev_priv);
> + pipe = vlv_find_free_pps(i915);
>
> /*
> * Didn't find one. This should not happen since there
> @@ -792,7 +792,7 @@ vlv_power_sequencer_pipe(struct intel_dp *intel_dp)
> if (WARN_ON(pipe == INVALID_PIPE))
> pipe = PIPE_A;
>
> - vlv_steal_power_sequencer(dev_priv, pipe);
> + vlv_steal_power_sequencer(i915, pipe);
> intel_dp->pps_pipe = pipe;
>
> DRM_DEBUG_KMS("picked pipe %c power sequencer for port %c\n",
> @@ -815,10 +815,10 @@ vlv_power_sequencer_pipe(struct intel_dp *intel_dp)
> static int
> bxt_power_sequencer_idx(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> - int backlight_controller = dev_priv->vbt.backlight.controller;
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> + int backlight_controller = i915->vbt.backlight.controller;
>
> - lockdep_assert_held(&dev_priv->pps_mutex);
> + lockdep_assert_held(&i915->pps_mutex);
>
> /* We should never land here with regular DP ports */
> WARN_ON(!intel_dp_is_edp(intel_dp));
> @@ -837,29 +837,29 @@ bxt_power_sequencer_idx(struct intel_dp *intel_dp)
> return backlight_controller;
> }
>
> -typedef bool (*vlv_pipe_check)(struct drm_i915_private *dev_priv,
> +typedef bool (*vlv_pipe_check)(struct drm_i915_private *i915,
> enum pipe pipe);
>
> -static bool vlv_pipe_has_pp_on(struct drm_i915_private *dev_priv,
> +static bool vlv_pipe_has_pp_on(struct drm_i915_private *i915,
> enum pipe pipe)
> {
> return I915_READ(PP_STATUS(pipe)) & PP_ON;
> }
>
> -static bool vlv_pipe_has_vdd_on(struct drm_i915_private *dev_priv,
> +static bool vlv_pipe_has_vdd_on(struct drm_i915_private *i915,
> enum pipe pipe)
> {
> return I915_READ(PP_CONTROL(pipe)) & EDP_FORCE_VDD;
> }
>
> -static bool vlv_pipe_any(struct drm_i915_private *dev_priv,
> +static bool vlv_pipe_any(struct drm_i915_private *i915,
> enum pipe pipe)
> {
> return true;
> }
>
> static enum pipe
> -vlv_initial_pps_pipe(struct drm_i915_private *dev_priv,
> +vlv_initial_pps_pipe(struct drm_i915_private *i915,
> enum port port,
> vlv_pipe_check pipe_check)
> {
> @@ -872,7 +872,7 @@ vlv_initial_pps_pipe(struct drm_i915_private *dev_priv,
> if (port_sel != PANEL_PORT_SELECT_VLV(port))
> continue;
>
> - if (!pipe_check(dev_priv, pipe))
> + if (!pipe_check(i915, pipe))
> continue;
>
> return pipe;
> @@ -884,23 +884,23 @@ vlv_initial_pps_pipe(struct drm_i915_private *dev_priv,
> static void
> vlv_initial_power_sequencer_setup(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp);
> enum port port = intel_dig_port->base.port;
>
> - lockdep_assert_held(&dev_priv->pps_mutex);
> + lockdep_assert_held(&i915->pps_mutex);
>
> /* try to find a pipe with this port selected */
> /* first pick one where the panel is on */
> - intel_dp->pps_pipe = vlv_initial_pps_pipe(dev_priv, port,
> + intel_dp->pps_pipe = vlv_initial_pps_pipe(i915, port,
> vlv_pipe_has_pp_on);
> /* didn't find one? pick one where vdd is on */
> if (intel_dp->pps_pipe == INVALID_PIPE)
> - intel_dp->pps_pipe = vlv_initial_pps_pipe(dev_priv, port,
> + intel_dp->pps_pipe = vlv_initial_pps_pipe(i915, port,
> vlv_pipe_has_vdd_on);
> /* didn't find one? pick one with just the correct port */
> if (intel_dp->pps_pipe == INVALID_PIPE)
> - intel_dp->pps_pipe = vlv_initial_pps_pipe(dev_priv, port,
> + intel_dp->pps_pipe = vlv_initial_pps_pipe(i915, port,
> vlv_pipe_any);
>
> /* didn't find one? just let vlv_power_sequencer_pipe() pick one when needed */
> @@ -917,12 +917,12 @@ vlv_initial_power_sequencer_setup(struct intel_dp *intel_dp)
> intel_dp_init_panel_power_sequencer_registers(intel_dp, false);
> }
>
> -void intel_power_sequencer_reset(struct drm_i915_private *dev_priv)
> +void intel_power_sequencer_reset(struct drm_i915_private *i915)
> {
> struct intel_encoder *encoder;
>
> - if (WARN_ON(!IS_VALLEYVIEW(dev_priv) && !IS_CHERRYVIEW(dev_priv) &&
> - !IS_GEN9_LP(dev_priv)))
> + if (WARN_ON(!IS_VALLEYVIEW(i915) && !IS_CHERRYVIEW(i915) &&
> + !IS_GEN9_LP(i915)))
> return;
>
> /*
> @@ -935,7 +935,7 @@ void intel_power_sequencer_reset(struct drm_i915_private *dev_priv)
> * should use them always.
> */
>
> - for_each_intel_dp(&dev_priv->drm, encoder) {
> + for_each_intel_dp(&i915->drm, encoder) {
> struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
>
> WARN_ON(intel_dp->active_pipe != INVALID_PIPE);
> @@ -943,7 +943,7 @@ void intel_power_sequencer_reset(struct drm_i915_private *dev_priv)
> if (encoder->type != INTEL_OUTPUT_EDP)
> continue;
>
> - if (IS_GEN9_LP(dev_priv))
> + if (IS_GEN9_LP(i915))
> intel_dp->pps_reset = true;
> else
> intel_dp->pps_pipe = INVALID_PIPE;
> @@ -961,14 +961,14 @@ struct pps_registers {
> static void intel_pps_get_registers(struct intel_dp *intel_dp,
> struct pps_registers *regs)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> int pps_idx = 0;
>
> memset(regs, 0, sizeof(*regs));
>
> - if (IS_GEN9_LP(dev_priv))
> + if (IS_GEN9_LP(i915))
> pps_idx = bxt_power_sequencer_idx(intel_dp);
> - else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> + else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> pps_idx = vlv_power_sequencer_pipe(intel_dp);
>
> regs->pp_ctrl = PP_CONTROL(pps_idx);
> @@ -977,7 +977,7 @@ static void intel_pps_get_registers(struct intel_dp *intel_dp,
> regs->pp_off = PP_OFF_DELAYS(pps_idx);
>
> /* Cycle delay moved from PP_DIVISOR to PP_CONTROL */
> - if (IS_GEN9_LP(dev_priv) || INTEL_PCH_TYPE(dev_priv) >= PCH_CNP)
> + if (IS_GEN9_LP(i915) || INTEL_PCH_TYPE(i915) >= PCH_CNP)
> regs->pp_div = INVALID_MMIO_REG;
> else
> regs->pp_div = PP_DIVISOR(pps_idx);
> @@ -1010,14 +1010,14 @@ static int edp_notify_handler(struct notifier_block *this, unsigned long code,
> {
> struct intel_dp *intel_dp = container_of(this, typeof(* intel_dp),
> edp_notifier);
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> intel_wakeref_t wakeref;
>
> if (!intel_dp_is_edp(intel_dp) || code != SYS_RESTART)
> return 0;
>
> with_pps_lock(intel_dp, wakeref) {
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> enum pipe pipe = vlv_power_sequencer_pipe(intel_dp);
> i915_reg_t pp_ctrl_reg, pp_div_reg;
> u32 pp_div;
> @@ -1039,11 +1039,11 @@ static int edp_notify_handler(struct notifier_block *this, unsigned long code,
>
> static bool edp_have_panel_power(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
>
> - lockdep_assert_held(&dev_priv->pps_mutex);
> + lockdep_assert_held(&i915->pps_mutex);
>
> - if ((IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) &&
> + if ((IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) &&
> intel_dp->pps_pipe == INVALID_PIPE)
> return false;
>
> @@ -1052,11 +1052,11 @@ static bool edp_have_panel_power(struct intel_dp *intel_dp)
>
> static bool edp_have_panel_vdd(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
>
> - lockdep_assert_held(&dev_priv->pps_mutex);
> + lockdep_assert_held(&i915->pps_mutex);
>
> - if ((IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) &&
> + if ((IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) &&
> intel_dp->pps_pipe == INVALID_PIPE)
> return false;
>
> @@ -1066,7 +1066,7 @@ static bool edp_have_panel_vdd(struct intel_dp *intel_dp)
> static void
> intel_dp_check_edp(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
>
> if (!intel_dp_is_edp(intel_dp))
> return;
> @@ -1103,7 +1103,7 @@ intel_dp_aux_wait_done(struct intel_dp *intel_dp)
>
> static u32 g4x_get_aux_clock_divider(struct intel_dp *intel_dp, int index)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
>
> if (index)
> return 0;
> @@ -1112,12 +1112,12 @@ static u32 g4x_get_aux_clock_divider(struct intel_dp *intel_dp, int index)
> * The clock divider is based off the hrawclk, and would like to run at
> * 2MHz. So, take the hrawclk value and divide by 2000 and use that
> */
> - return DIV_ROUND_CLOSEST(dev_priv->rawclk_freq, 2000);
> + return DIV_ROUND_CLOSEST(i915->rawclk_freq, 2000);
> }
>
> static u32 ilk_get_aux_clock_divider(struct intel_dp *intel_dp, int index)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
>
> if (index)
> @@ -1129,17 +1129,17 @@ static u32 ilk_get_aux_clock_divider(struct intel_dp *intel_dp, int index)
> * divide by 2000 and use that
> */
> if (dig_port->aux_ch == AUX_CH_A)
> - return DIV_ROUND_CLOSEST(dev_priv->cdclk.hw.cdclk, 2000);
> + return DIV_ROUND_CLOSEST(i915->cdclk.hw.cdclk, 2000);
> else
> - return DIV_ROUND_CLOSEST(dev_priv->rawclk_freq, 2000);
> + return DIV_ROUND_CLOSEST(i915->rawclk_freq, 2000);
> }
>
> static u32 hsw_get_aux_clock_divider(struct intel_dp *intel_dp, int index)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
>
> - if (dig_port->aux_ch != AUX_CH_A && HAS_PCH_LPT_H(dev_priv)) {
> + if (dig_port->aux_ch != AUX_CH_A && HAS_PCH_LPT_H(i915)) {
> /* Workaround for non-ULT HSW */
> switch (index) {
> case 0: return 63;
> @@ -1166,16 +1166,16 @@ static u32 g4x_get_aux_send_ctl(struct intel_dp *intel_dp,
> u32 aux_clock_divider)
> {
> struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp);
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> to_i915(intel_dig_port->base.base.dev);
> u32 precharge, timeout;
>
> - if (IS_GEN(dev_priv, 6))
> + if (IS_GEN(i915, 6))
> precharge = 3;
> else
> precharge = 5;
>
> - if (IS_BROADWELL(dev_priv))
> + if (IS_BROADWELL(i915))
> timeout = DP_AUX_CH_CTL_TIME_OUT_600us;
> else
> timeout = DP_AUX_CH_CTL_TIME_OUT_400us;
> @@ -1481,7 +1481,7 @@ intel_dp_aux_transfer(struct drm_dp_aux *aux, struct drm_dp_aux_msg *msg)
>
> static i915_reg_t g4x_aux_ctl_reg(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> enum aux_ch aux_ch = dig_port->aux_ch;
>
> @@ -1498,7 +1498,7 @@ static i915_reg_t g4x_aux_ctl_reg(struct intel_dp *intel_dp)
>
> static i915_reg_t g4x_aux_data_reg(struct intel_dp *intel_dp, int index)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> enum aux_ch aux_ch = dig_port->aux_ch;
>
> @@ -1515,7 +1515,7 @@ static i915_reg_t g4x_aux_data_reg(struct intel_dp *intel_dp, int index)
>
> static i915_reg_t ilk_aux_ctl_reg(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> enum aux_ch aux_ch = dig_port->aux_ch;
>
> @@ -1534,7 +1534,7 @@ static i915_reg_t ilk_aux_ctl_reg(struct intel_dp *intel_dp)
>
> static i915_reg_t ilk_aux_data_reg(struct intel_dp *intel_dp, int index)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> enum aux_ch aux_ch = dig_port->aux_ch;
>
> @@ -1553,7 +1553,7 @@ static i915_reg_t ilk_aux_data_reg(struct intel_dp *intel_dp, int index)
>
> static i915_reg_t skl_aux_ctl_reg(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> enum aux_ch aux_ch = dig_port->aux_ch;
>
> @@ -1573,7 +1573,7 @@ static i915_reg_t skl_aux_ctl_reg(struct intel_dp *intel_dp)
>
> static i915_reg_t skl_aux_data_reg(struct intel_dp *intel_dp, int index)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> enum aux_ch aux_ch = dig_port->aux_ch;
>
> @@ -1600,14 +1600,14 @@ intel_dp_aux_fini(struct intel_dp *intel_dp)
> static void
> intel_dp_aux_init(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> struct intel_encoder *encoder = &dig_port->base;
>
> - if (INTEL_GEN(dev_priv) >= 9) {
> + if (INTEL_GEN(i915) >= 9) {
> intel_dp->aux_ch_ctl_reg = skl_aux_ctl_reg;
> intel_dp->aux_ch_data_reg = skl_aux_data_reg;
> - } else if (HAS_PCH_SPLIT(dev_priv)) {
> + } else if (HAS_PCH_SPLIT(i915)) {
> intel_dp->aux_ch_ctl_reg = ilk_aux_ctl_reg;
> intel_dp->aux_ch_data_reg = ilk_aux_data_reg;
> } else {
> @@ -1615,16 +1615,16 @@ intel_dp_aux_init(struct intel_dp *intel_dp)
> intel_dp->aux_ch_data_reg = g4x_aux_data_reg;
> }
>
> - if (INTEL_GEN(dev_priv) >= 9)
> + if (INTEL_GEN(i915) >= 9)
> intel_dp->get_aux_clock_divider = skl_get_aux_clock_divider;
> - else if (IS_BROADWELL(dev_priv) || IS_HASWELL(dev_priv))
> + else if (IS_BROADWELL(i915) || IS_HASWELL(i915))
> intel_dp->get_aux_clock_divider = hsw_get_aux_clock_divider;
> - else if (HAS_PCH_SPLIT(dev_priv))
> + else if (HAS_PCH_SPLIT(i915))
> intel_dp->get_aux_clock_divider = ilk_get_aux_clock_divider;
> else
> intel_dp->get_aux_clock_divider = g4x_get_aux_clock_divider;
>
> - if (INTEL_GEN(dev_priv) >= 9)
> + if (INTEL_GEN(i915) >= 9)
> intel_dp->get_aux_send_ctl = skl_get_aux_send_ctl;
> else
> intel_dp->get_aux_send_ctl = g4x_get_aux_send_ctl;
> @@ -1655,20 +1655,20 @@ static void
> intel_dp_set_clock(struct intel_encoder *encoder,
> struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> const struct dp_link_dpll *divisor = NULL;
> int i, count = 0;
>
> - if (IS_G4X(dev_priv)) {
> + if (IS_G4X(i915)) {
> divisor = g4x_dpll;
> count = ARRAY_SIZE(g4x_dpll);
> - } else if (HAS_PCH_SPLIT(dev_priv)) {
> + } else if (HAS_PCH_SPLIT(i915)) {
> divisor = pch_dpll;
> count = ARRAY_SIZE(pch_dpll);
> - } else if (IS_CHERRYVIEW(dev_priv)) {
> + } else if (IS_CHERRYVIEW(i915)) {
> divisor = chv_dpll;
> count = ARRAY_SIZE(chv_dpll);
> - } else if (IS_VALLEYVIEW(dev_priv)) {
> + } else if (IS_VALLEYVIEW(i915)) {
> divisor = vlv_dpll;
> count = ARRAY_SIZE(vlv_dpll);
> }
> @@ -1760,9 +1760,9 @@ void intel_dp_compute_rate(struct intel_dp *intel_dp, int port_clock,
> static bool intel_dp_source_supports_fec(struct intel_dp *intel_dp,
> const struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
>
> - return INTEL_GEN(dev_priv) >= 11 &&
> + return INTEL_GEN(i915) >= 11 &&
> pipe_config->cpu_transcoder != TRANSCODER_A;
> }
>
> @@ -1776,9 +1776,9 @@ static bool intel_dp_supports_fec(struct intel_dp *intel_dp,
> static bool intel_dp_source_supports_dsc(struct intel_dp *intel_dp,
> const struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
>
> - return INTEL_GEN(dev_priv) >= 10 &&
> + return INTEL_GEN(i915) >= 10 &&
> pipe_config->cpu_transcoder != TRANSCODER_A;
> }
>
> @@ -1795,7 +1795,7 @@ static bool intel_dp_supports_dsc(struct intel_dp *intel_dp,
> static int intel_dp_compute_bpp(struct intel_dp *intel_dp,
> struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> struct intel_connector *intel_connector = intel_dp->attached_connector;
> int bpp, bpc;
>
> @@ -1808,10 +1808,10 @@ static int intel_dp_compute_bpp(struct intel_dp *intel_dp,
> if (intel_dp_is_edp(intel_dp)) {
> /* Get bpp from vbt only for panels that dont have bpp in edid */
> if (intel_connector->base.display_info.bpc == 0 &&
> - dev_priv->vbt.edp.bpp && dev_priv->vbt.edp.bpp < bpp) {
> + i915->vbt.edp.bpp && i915->vbt.edp.bpp < bpp) {
> DRM_DEBUG_KMS("clamping bpp for eDP panel to BIOS-provided %i\n",
> - dev_priv->vbt.edp.bpp);
> - bpp = dev_priv->vbt.edp.bpp;
> + i915->vbt.edp.bpp);
> + bpp = i915->vbt.edp.bpp;
> }
> }
>
> @@ -1924,7 +1924,7 @@ static int intel_dp_dsc_compute_config(struct intel_dp *intel_dp,
> struct link_config_limits *limits)
> {
> struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> - struct drm_i915_private *dev_priv = to_i915(dig_port->base.base.dev);
> + struct drm_i915_private *i915 = to_i915(dig_port->base.base.dev);
> struct drm_display_mode *adjusted_mode = &pipe_config->base.adjusted_mode;
> u8 dsc_max_bpc;
> int pipe_bpp;
> @@ -1988,7 +1988,7 @@ static int intel_dp_dsc_compute_config(struct intel_dp *intel_dp,
> * is greater than the maximum Cdclock and if slice count is even
> * then we need to use 2 VDSC instances.
> */
> - if (adjusted_mode->crtc_clock > dev_priv->max_cdclk_freq) {
> + if (adjusted_mode->crtc_clock > i915->max_cdclk_freq) {
> if (pipe_config->dsc_params.slice_count > 1) {
> pipe_config->dsc_params.dsc_split = true;
> } else {
> @@ -2168,7 +2168,7 @@ intel_dp_compute_config(struct intel_encoder *encoder,
> struct intel_crtc_state *pipe_config,
> struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct drm_display_mode *adjusted_mode = &pipe_config->base.adjusted_mode;
> struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> struct intel_lspcon *lspcon = enc_to_intel_lspcon(&encoder->base);
> @@ -2181,7 +2181,7 @@ intel_dp_compute_config(struct intel_encoder *encoder,
> DP_DPCD_QUIRK_CONSTANT_N);
> int ret = 0, output_bpp;
>
> - if (HAS_PCH_SPLIT(dev_priv) && !HAS_DDI(dev_priv) && port != PORT_A)
> + if (HAS_PCH_SPLIT(i915) && !HAS_DDI(i915) && port != PORT_A)
> pipe_config->has_pch_encoder = true;
>
> pipe_config->output_format = INTEL_OUTPUT_FORMAT_RGB;
> @@ -2195,7 +2195,7 @@ intel_dp_compute_config(struct intel_encoder *encoder,
> return ret;
>
> pipe_config->has_drrs = false;
> - if (IS_G4X(dev_priv) || port == PORT_A)
> + if (IS_G4X(i915) || port == PORT_A)
> pipe_config->has_audio = false;
> else if (intel_conn_state->force_audio == HDMI_AUDIO_AUTO)
> pipe_config->has_audio = intel_dp->has_audio;
> @@ -2206,13 +2206,13 @@ intel_dp_compute_config(struct intel_encoder *encoder,
> intel_fixed_panel_mode(intel_connector->panel.fixed_mode,
> adjusted_mode);
>
> - if (INTEL_GEN(dev_priv) >= 9) {
> + if (INTEL_GEN(i915) >= 9) {
> ret = skl_update_scaler_crtc(pipe_config);
> if (ret)
> return ret;
> }
>
> - if (HAS_GMCH(dev_priv))
> + if (HAS_GMCH(i915))
> intel_gmch_panel_fitting(intel_crtc, pipe_config,
> conn_state->scaling_mode);
> else
> @@ -2223,7 +2223,7 @@ intel_dp_compute_config(struct intel_encoder *encoder,
> if (adjusted_mode->flags & DRM_MODE_FLAG_DBLSCAN)
> return -EINVAL;
>
> - if (HAS_GMCH(dev_priv) &&
> + if (HAS_GMCH(i915) &&
> adjusted_mode->flags & DRM_MODE_FLAG_INTERLACE)
> return -EINVAL;
>
> @@ -2250,7 +2250,7 @@ intel_dp_compute_config(struct intel_encoder *encoder,
> constant_n);
>
> if (intel_connector->panel.downclock_mode != NULL &&
> - dev_priv->drrs.type == SEAMLESS_DRRS_SUPPORT) {
> + i915->drrs.type == SEAMLESS_DRRS_SUPPORT) {
> pipe_config->has_drrs = true;
> intel_link_compute_m_n(output_bpp,
> pipe_config->lane_count,
> @@ -2260,7 +2260,7 @@ intel_dp_compute_config(struct intel_encoder *encoder,
> constant_n);
> }
>
> - if (!HAS_DDI(dev_priv))
> + if (!HAS_DDI(i915))
> intel_dp_set_clock(encoder, pipe_config);
>
> intel_psr_compute_config(intel_dp, pipe_config);
> @@ -2281,7 +2281,7 @@ void intel_dp_set_link_params(struct intel_dp *intel_dp,
> static void intel_dp_prepare(struct intel_encoder *encoder,
> const struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> enum port port = encoder->port;
> struct intel_crtc *crtc = to_intel_crtc(pipe_config->base.crtc);
> @@ -2320,7 +2320,7 @@ static void intel_dp_prepare(struct intel_encoder *encoder,
>
> /* Split out the IBX/CPU vs CPT settings */
>
> - if (IS_IVYBRIDGE(dev_priv) && port == PORT_A) {
> + if (IS_IVYBRIDGE(i915) && port == PORT_A) {
> if (adjusted_mode->flags & DRM_MODE_FLAG_PHSYNC)
> intel_dp->DP |= DP_SYNC_HS_HIGH;
> if (adjusted_mode->flags & DRM_MODE_FLAG_PVSYNC)
> @@ -2331,7 +2331,7 @@ static void intel_dp_prepare(struct intel_encoder *encoder,
> intel_dp->DP |= DP_ENHANCED_FRAMING;
>
> intel_dp->DP |= DP_PIPE_SEL_IVB(crtc->pipe);
> - } else if (HAS_PCH_CPT(dev_priv) && port != PORT_A) {
> + } else if (HAS_PCH_CPT(i915) && port != PORT_A) {
> u32 trans_dp;
>
> intel_dp->DP |= DP_LINK_TRAIN_OFF_CPT;
> @@ -2343,7 +2343,7 @@ static void intel_dp_prepare(struct intel_encoder *encoder,
> trans_dp &= ~TRANS_DP_ENH_FRAMING;
> I915_WRITE(TRANS_DP_CTL(crtc->pipe), trans_dp);
> } else {
> - if (IS_G4X(dev_priv) && pipe_config->limited_color_range)
> + if (IS_G4X(i915) && pipe_config->limited_color_range)
> intel_dp->DP |= DP_COLOR_RANGE_16_235;
>
> if (adjusted_mode->flags & DRM_MODE_FLAG_PHSYNC)
> @@ -2355,7 +2355,7 @@ static void intel_dp_prepare(struct intel_encoder *encoder,
> if (drm_dp_enhanced_frame_cap(intel_dp->dpcd))
> intel_dp->DP |= DP_ENHANCED_FRAMING;
>
> - if (IS_CHERRYVIEW(dev_priv))
> + if (IS_CHERRYVIEW(i915))
> intel_dp->DP |= DP_PIPE_SEL_CHV(crtc->pipe);
> else
> intel_dp->DP |= DP_PIPE_SEL(crtc->pipe);
> @@ -2377,10 +2377,10 @@ static void wait_panel_status(struct intel_dp *intel_dp,
> u32 mask,
> u32 value)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> i915_reg_t pp_stat_reg, pp_ctrl_reg;
>
> - lockdep_assert_held(&dev_priv->pps_mutex);
> + lockdep_assert_held(&i915->pps_mutex);
>
> intel_pps_verify_state(intel_dp);
>
> @@ -2392,7 +2392,7 @@ static void wait_panel_status(struct intel_dp *intel_dp,
> I915_READ(pp_stat_reg),
> I915_READ(pp_ctrl_reg));
>
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> pp_stat_reg, mask, value,
> 5000))
> DRM_ERROR("Panel status timeout: status %08x control %08x\n",
> @@ -2453,13 +2453,13 @@ static void edp_wait_backlight_off(struct intel_dp *intel_dp)
>
> static u32 ironlake_get_pp_control(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> u32 control;
>
> - lockdep_assert_held(&dev_priv->pps_mutex);
> + lockdep_assert_held(&i915->pps_mutex);
>
> control = I915_READ(_pp_ctrl_reg(intel_dp));
> - if (WARN_ON(!HAS_DDI(dev_priv) &&
> + if (WARN_ON(!HAS_DDI(i915) &&
> (control & PANEL_UNLOCK_MASK) != PANEL_UNLOCK_REGS)) {
> control &= ~PANEL_UNLOCK_MASK;
> control |= PANEL_UNLOCK_REGS;
> @@ -2474,13 +2474,13 @@ static u32 ironlake_get_pp_control(struct intel_dp *intel_dp)
> */
> static bool edp_panel_vdd_on(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp);
> u32 pp;
> i915_reg_t pp_stat_reg, pp_ctrl_reg;
> bool need_to_disable = !intel_dp->want_panel_vdd;
>
> - lockdep_assert_held(&dev_priv->pps_mutex);
> + lockdep_assert_held(&i915->pps_mutex);
>
> if (!intel_dp_is_edp(intel_dp))
> return false;
> @@ -2491,7 +2491,7 @@ static bool edp_panel_vdd_on(struct intel_dp *intel_dp)
> if (edp_have_panel_vdd(intel_dp))
> return need_to_disable;
>
> - intel_display_power_get(dev_priv,
> + intel_display_power_get(i915,
> intel_aux_power_domain(intel_dig_port));
>
> DRM_DEBUG_KMS("Turning eDP port %c VDD on\n",
> @@ -2546,13 +2546,13 @@ void intel_edp_panel_vdd_on(struct intel_dp *intel_dp)
>
> static void edp_panel_vdd_off_sync(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> struct intel_digital_port *intel_dig_port =
> dp_to_dig_port(intel_dp);
> u32 pp;
> i915_reg_t pp_stat_reg, pp_ctrl_reg;
>
> - lockdep_assert_held(&dev_priv->pps_mutex);
> + lockdep_assert_held(&i915->pps_mutex);
>
> WARN_ON(intel_dp->want_panel_vdd);
>
> @@ -2578,7 +2578,7 @@ static void edp_panel_vdd_off_sync(struct intel_dp *intel_dp)
> if ((pp & PANEL_POWER_ON) == 0)
> intel_dp->panel_power_off_time = ktime_get_boottime();
>
> - intel_display_power_put_unchecked(dev_priv,
> + intel_display_power_put_unchecked(i915,
> intel_aux_power_domain(intel_dig_port));
> }
>
> @@ -2615,9 +2615,9 @@ static void edp_panel_vdd_schedule_off(struct intel_dp *intel_dp)
> */
> static void edp_panel_vdd_off(struct intel_dp *intel_dp, bool sync)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
>
> - lockdep_assert_held(&dev_priv->pps_mutex);
> + lockdep_assert_held(&i915->pps_mutex);
>
> if (!intel_dp_is_edp(intel_dp))
> return;
> @@ -2635,11 +2635,11 @@ static void edp_panel_vdd_off(struct intel_dp *intel_dp, bool sync)
>
> static void edp_panel_on(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> u32 pp;
> i915_reg_t pp_ctrl_reg;
>
> - lockdep_assert_held(&dev_priv->pps_mutex);
> + lockdep_assert_held(&i915->pps_mutex);
>
> if (!intel_dp_is_edp(intel_dp))
> return;
> @@ -2656,7 +2656,7 @@ static void edp_panel_on(struct intel_dp *intel_dp)
>
> pp_ctrl_reg = _pp_ctrl_reg(intel_dp);
> pp = ironlake_get_pp_control(intel_dp);
> - if (IS_GEN(dev_priv, 5)) {
> + if (IS_GEN(i915, 5)) {
> /* ILK workaround: disable reset around power sequence */
> pp &= ~PANEL_POWER_RESET;
> I915_WRITE(pp_ctrl_reg, pp);
> @@ -2664,7 +2664,7 @@ static void edp_panel_on(struct intel_dp *intel_dp)
> }
>
> pp |= PANEL_POWER_ON;
> - if (!IS_GEN(dev_priv, 5))
> + if (!IS_GEN(i915, 5))
> pp |= PANEL_POWER_RESET;
>
> I915_WRITE(pp_ctrl_reg, pp);
> @@ -2673,7 +2673,7 @@ static void edp_panel_on(struct intel_dp *intel_dp)
> wait_panel_on(intel_dp);
> intel_dp->last_power_on = jiffies;
>
> - if (IS_GEN(dev_priv, 5)) {
> + if (IS_GEN(i915, 5)) {
> pp |= PANEL_POWER_RESET; /* restore panel reset bit */
> I915_WRITE(pp_ctrl_reg, pp);
> POSTING_READ(pp_ctrl_reg);
> @@ -2694,12 +2694,12 @@ void intel_edp_panel_on(struct intel_dp *intel_dp)
>
> static void edp_panel_off(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> u32 pp;
> i915_reg_t pp_ctrl_reg;
>
> - lockdep_assert_held(&dev_priv->pps_mutex);
> + lockdep_assert_held(&i915->pps_mutex);
>
> if (!intel_dp_is_edp(intel_dp))
> return;
> @@ -2727,7 +2727,7 @@ static void edp_panel_off(struct intel_dp *intel_dp)
> intel_dp->panel_power_off_time = ktime_get_boottime();
>
> /* We got a reference when we enabled the VDD. */
> - intel_display_power_put_unchecked(dev_priv, intel_aux_power_domain(dig_port));
> + intel_display_power_put_unchecked(i915, intel_aux_power_domain(dig_port));
> }
>
> void intel_edp_panel_off(struct intel_dp *intel_dp)
> @@ -2744,7 +2744,7 @@ void intel_edp_panel_off(struct intel_dp *intel_dp)
> /* Enable backlight in the panel power control. */
> static void _intel_edp_backlight_on(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> intel_wakeref_t wakeref;
>
> /*
> @@ -2785,7 +2785,7 @@ void intel_edp_backlight_on(const struct intel_crtc_state *crtc_state,
> /* Disable backlight in the panel power control. */
> static void _intel_edp_backlight_off(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> intel_wakeref_t wakeref;
>
> if (!intel_dp_is_edp(intel_dp))
> @@ -2849,7 +2849,7 @@ static void intel_edp_backlight_power(struct intel_connector *connector,
> static void assert_dp_port(struct intel_dp *intel_dp, bool state)
> {
> struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> - struct drm_i915_private *dev_priv = to_i915(dig_port->base.base.dev);
> + struct drm_i915_private *i915 = to_i915(dig_port->base.base.dev);
> bool cur_state = I915_READ(intel_dp->output_reg) & DP_PORT_EN;
>
> I915_STATE_WARN(cur_state != state,
> @@ -2859,7 +2859,7 @@ static void assert_dp_port(struct intel_dp *intel_dp, bool state)
> }
> #define assert_dp_port_disabled(d) assert_dp_port((d), false)
>
> -static void assert_edp_pll(struct drm_i915_private *dev_priv, bool state)
> +static void assert_edp_pll(struct drm_i915_private *i915, bool state)
> {
> bool cur_state = I915_READ(DP_A) & DP_PLL_ENABLE;
>
> @@ -2874,11 +2874,11 @@ static void ironlake_edp_pll_on(struct intel_dp *intel_dp,
> const struct intel_crtc_state *pipe_config)
> {
> struct intel_crtc *crtc = to_intel_crtc(pipe_config->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
>
> - assert_pipe_disabled(dev_priv, crtc->pipe);
> + assert_pipe_disabled(i915, crtc->pipe);
> assert_dp_port_disabled(intel_dp);
> - assert_edp_pll_disabled(dev_priv);
> + assert_edp_pll_disabled(i915);
>
> DRM_DEBUG_KMS("enabling eDP PLL for clock %d\n",
> pipe_config->port_clock);
> @@ -2900,8 +2900,8 @@ static void ironlake_edp_pll_on(struct intel_dp *intel_dp,
> * 1. Wait for the start of vertical blank on the enabled pipe going to FDI
> * 2. Program DP PLL enable
> */
> - if (IS_GEN(dev_priv, 5))
> - intel_wait_for_vblank_if_active(dev_priv, !crtc->pipe);
> + if (IS_GEN(i915, 5))
> + intel_wait_for_vblank_if_active(i915, !crtc->pipe);
>
> intel_dp->DP |= DP_PLL_ENABLE;
>
> @@ -2914,11 +2914,11 @@ static void ironlake_edp_pll_off(struct intel_dp *intel_dp,
> const struct intel_crtc_state *old_crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
>
> - assert_pipe_disabled(dev_priv, crtc->pipe);
> + assert_pipe_disabled(i915, crtc->pipe);
> assert_dp_port_disabled(intel_dp);
> - assert_edp_pll_enabled(dev_priv);
> + assert_edp_pll_enabled(i915);
>
> DRM_DEBUG_KMS("disabling eDP PLL\n");
>
> @@ -2999,12 +2999,12 @@ void intel_dp_sink_dpms(struct intel_dp *intel_dp, int mode)
> mode == DRM_MODE_DPMS_ON ? "enable" : "disable");
> }
>
> -static bool cpt_dp_port_selected(struct drm_i915_private *dev_priv,
> +static bool cpt_dp_port_selected(struct drm_i915_private *i915,
> enum port port, enum pipe *pipe)
> {
> enum pipe p;
>
> - for_each_pipe(dev_priv, p) {
> + for_each_pipe(i915, p) {
> u32 val = I915_READ(TRANS_DP_CTL(p));
>
> if ((val & TRANS_DP_PORT_SEL_MASK) == TRANS_DP_PORT_SEL(port)) {
> @@ -3021,7 +3021,7 @@ static bool cpt_dp_port_selected(struct drm_i915_private *dev_priv,
> return false;
> }
>
> -bool intel_dp_port_enabled(struct drm_i915_private *dev_priv,
> +bool intel_dp_port_enabled(struct drm_i915_private *i915,
> i915_reg_t dp_reg, enum port port,
> enum pipe *pipe)
> {
> @@ -3033,11 +3033,11 @@ bool intel_dp_port_enabled(struct drm_i915_private *dev_priv,
> ret = val & DP_PORT_EN;
>
> /* asserts want to know the pipe even if the port is disabled */
> - if (IS_IVYBRIDGE(dev_priv) && port == PORT_A)
> + if (IS_IVYBRIDGE(i915) && port == PORT_A)
> *pipe = (val & DP_PIPE_SEL_MASK_IVB) >> DP_PIPE_SEL_SHIFT_IVB;
> - else if (HAS_PCH_CPT(dev_priv) && port != PORT_A)
> - ret &= cpt_dp_port_selected(dev_priv, port, pipe);
> - else if (IS_CHERRYVIEW(dev_priv))
> + else if (HAS_PCH_CPT(i915) && port != PORT_A)
> + ret &= cpt_dp_port_selected(i915, port, pipe);
> + else if (IS_CHERRYVIEW(i915))
> *pipe = (val & DP_PIPE_SEL_MASK_CHV) >> DP_PIPE_SEL_SHIFT_CHV;
> else
> *pipe = (val & DP_PIPE_SEL_MASK) >> DP_PIPE_SEL_SHIFT;
> @@ -3048,20 +3048,20 @@ bool intel_dp_port_enabled(struct drm_i915_private *dev_priv,
> static bool intel_dp_get_hw_state(struct intel_encoder *encoder,
> enum pipe *pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> intel_wakeref_t wakeref;
> bool ret;
>
> - wakeref = intel_display_power_get_if_enabled(dev_priv,
> + wakeref = intel_display_power_get_if_enabled(i915,
> encoder->power_domain);
> if (!wakeref)
> return false;
>
> - ret = intel_dp_port_enabled(dev_priv, intel_dp->output_reg,
> + ret = intel_dp_port_enabled(i915, intel_dp->output_reg,
> encoder->port, pipe);
>
> - intel_display_power_put(dev_priv, encoder->power_domain, wakeref);
> + intel_display_power_put(i915, encoder->power_domain, wakeref);
>
> return ret;
> }
> @@ -3069,7 +3069,7 @@ static bool intel_dp_get_hw_state(struct intel_encoder *encoder,
> static void intel_dp_get_config(struct intel_encoder *encoder,
> struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> u32 tmp, flags = 0;
> enum port port = encoder->port;
> @@ -3084,7 +3084,7 @@ static void intel_dp_get_config(struct intel_encoder *encoder,
>
> pipe_config->has_audio = tmp & DP_AUDIO_OUTPUT_ENABLE && port != PORT_A;
>
> - if (HAS_PCH_CPT(dev_priv) && port != PORT_A) {
> + if (HAS_PCH_CPT(i915) && port != PORT_A) {
> u32 trans_dp = I915_READ(TRANS_DP_CTL(crtc->pipe));
>
> if (trans_dp & TRANS_DP_HSYNC_ACTIVE_HIGH)
> @@ -3110,7 +3110,7 @@ static void intel_dp_get_config(struct intel_encoder *encoder,
>
> pipe_config->base.adjusted_mode.flags |= flags;
>
> - if (IS_G4X(dev_priv) && tmp & DP_COLOR_RANGE_16_235)
> + if (IS_G4X(i915) && tmp & DP_COLOR_RANGE_16_235)
> pipe_config->limited_color_range = true;
>
> pipe_config->lane_count =
> @@ -3129,8 +3129,8 @@ static void intel_dp_get_config(struct intel_encoder *encoder,
> intel_dotclock_calculate(pipe_config->port_clock,
> &pipe_config->dp_m_n);
>
> - if (intel_dp_is_edp(intel_dp) && dev_priv->vbt.edp.bpp &&
> - pipe_config->pipe_bpp > dev_priv->vbt.edp.bpp) {
> + if (intel_dp_is_edp(intel_dp) && i915->vbt.edp.bpp &&
> + pipe_config->pipe_bpp > i915->vbt.edp.bpp) {
> /*
> * This is a big fat ugly hack.
> *
> @@ -3145,8 +3145,8 @@ static void intel_dp_get_config(struct intel_encoder *encoder,
> * load.
> */
> DRM_DEBUG_KMS("pipe has %d bpp for eDP panel, overriding BIOS-provided max %d bpp\n",
> - pipe_config->pipe_bpp, dev_priv->vbt.edp.bpp);
> - dev_priv->vbt.edp.bpp = pipe_config->pipe_bpp;
> + pipe_config->pipe_bpp, i915->vbt.edp.bpp);
> + i915->vbt.edp.bpp = pipe_config->pipe_bpp;
> }
> }
>
> @@ -3215,16 +3215,16 @@ static void chv_post_disable_dp(struct intel_encoder *encoder,
> const struct intel_crtc_state *old_crtc_state,
> const struct drm_connector_state *old_conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
>
> intel_dp_link_down(encoder, old_crtc_state);
>
> - vlv_dpio_get(dev_priv);
> + vlv_dpio_get(i915);
>
> /* Assert data lane reset */
> chv_data_lane_soft_reset(encoder, old_crtc_state, true);
>
> - vlv_dpio_put(dev_priv);
> + vlv_dpio_put(i915);
> }
>
> static void
> @@ -3232,7 +3232,7 @@ _intel_dp_set_link_train(struct intel_dp *intel_dp,
> u32 *DP,
> u8 dp_train_pat)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp);
> enum port port = intel_dig_port->base.port;
> u8 train_pat_mask = drm_dp_training_pattern_mask(intel_dp->dpcd);
> @@ -3241,7 +3241,7 @@ _intel_dp_set_link_train(struct intel_dp *intel_dp,
> DRM_DEBUG_KMS("Using DP training pattern TPS%d\n",
> dp_train_pat & train_pat_mask);
>
> - if (HAS_DDI(dev_priv)) {
> + if (HAS_DDI(i915)) {
> u32 temp = I915_READ(DP_TP_CTL(port));
>
> if (dp_train_pat & DP_LINK_SCRAMBLING_DISABLE)
> @@ -3270,8 +3270,8 @@ _intel_dp_set_link_train(struct intel_dp *intel_dp,
> }
> I915_WRITE(DP_TP_CTL(port), temp);
>
> - } else if ((IS_IVYBRIDGE(dev_priv) && port == PORT_A) ||
> - (HAS_PCH_CPT(dev_priv) && port != PORT_A)) {
> + } else if ((IS_IVYBRIDGE(i915) && port == PORT_A) ||
> + (HAS_PCH_CPT(i915) && port != PORT_A)) {
> *DP &= ~DP_LINK_TRAIN_MASK_CPT;
>
> switch (dp_train_pat & DP_TRAINING_PATTERN_MASK) {
> @@ -3314,7 +3314,7 @@ _intel_dp_set_link_train(struct intel_dp *intel_dp,
> static void intel_dp_enable_port(struct intel_dp *intel_dp,
> const struct intel_crtc_state *old_crtc_state)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
>
> /* enable with pattern 1 (as per spec) */
>
> @@ -3338,7 +3338,7 @@ static void intel_enable_dp(struct intel_encoder *encoder,
> const struct intel_crtc_state *pipe_config,
> const struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> struct intel_crtc *crtc = to_intel_crtc(pipe_config->base.crtc);
> u32 dp_reg = I915_READ(intel_dp->output_reg);
> @@ -3349,7 +3349,7 @@ static void intel_enable_dp(struct intel_encoder *encoder,
> return;
>
> with_pps_lock(intel_dp, wakeref) {
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> vlv_init_panel_power_sequencer(encoder, pipe_config);
>
> intel_dp_enable_port(intel_dp, pipe_config);
> @@ -3359,13 +3359,13 @@ static void intel_enable_dp(struct intel_encoder *encoder,
> edp_panel_vdd_off(intel_dp, true);
> }
>
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> unsigned int lane_mask = 0x0;
>
> - if (IS_CHERRYVIEW(dev_priv))
> + if (IS_CHERRYVIEW(i915))
> lane_mask = intel_dp_unused_lane_mask(pipe_config->lane_count);
>
> - vlv_wait_port_ready(dev_priv, dp_to_dig_port(intel_dp),
> + vlv_wait_port_ready(i915, dp_to_dig_port(intel_dp),
> lane_mask);
> }
>
> @@ -3412,7 +3412,7 @@ static void g4x_pre_enable_dp(struct intel_encoder *encoder,
> static void vlv_detach_power_sequencer(struct intel_dp *intel_dp)
> {
> struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp);
> - struct drm_i915_private *dev_priv = to_i915(intel_dig_port->base.base.dev);
> + struct drm_i915_private *i915 = to_i915(intel_dig_port->base.base.dev);
> enum pipe pipe = intel_dp->pps_pipe;
> i915_reg_t pp_on_reg = PP_ON_DELAYS(pipe);
>
> @@ -3440,14 +3440,14 @@ static void vlv_detach_power_sequencer(struct intel_dp *intel_dp)
> intel_dp->pps_pipe = INVALID_PIPE;
> }
>
> -static void vlv_steal_power_sequencer(struct drm_i915_private *dev_priv,
> +static void vlv_steal_power_sequencer(struct drm_i915_private *i915,
> enum pipe pipe)
> {
> struct intel_encoder *encoder;
>
> - lockdep_assert_held(&dev_priv->pps_mutex);
> + lockdep_assert_held(&i915->pps_mutex);
>
> - for_each_intel_dp(&dev_priv->drm, encoder) {
> + for_each_intel_dp(&i915->drm, encoder) {
> struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> enum port port = encoder->port;
>
> @@ -3469,11 +3469,11 @@ static void vlv_steal_power_sequencer(struct drm_i915_private *dev_priv,
> static void vlv_init_panel_power_sequencer(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
>
> - lockdep_assert_held(&dev_priv->pps_mutex);
> + lockdep_assert_held(&i915->pps_mutex);
>
> WARN_ON(intel_dp->active_pipe != INVALID_PIPE);
>
> @@ -3491,7 +3491,7 @@ static void vlv_init_panel_power_sequencer(struct intel_encoder *encoder,
> * We may be stealing the power
> * sequencer from another port.
> */
> - vlv_steal_power_sequencer(dev_priv, crtc->pipe);
> + vlv_steal_power_sequencer(i915, crtc->pipe);
>
> intel_dp->active_pipe = crtc->pipe;
>
> @@ -3570,17 +3570,17 @@ intel_dp_get_link_status(struct intel_dp *intel_dp, u8 link_status[DP_LINK_STATU
> u8
> intel_dp_voltage_max(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> struct intel_encoder *encoder = &dp_to_dig_port(intel_dp)->base;
> enum port port = encoder->port;
>
> - if (HAS_DDI(dev_priv))
> + if (HAS_DDI(i915))
> return intel_ddi_dp_voltage_max(encoder);
> - else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> + else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> return DP_TRAIN_VOLTAGE_SWING_LEVEL_3;
> - else if (IS_IVYBRIDGE(dev_priv) && port == PORT_A)
> + else if (IS_IVYBRIDGE(i915) && port == PORT_A)
> return DP_TRAIN_VOLTAGE_SWING_LEVEL_2;
> - else if (HAS_PCH_CPT(dev_priv) && port != PORT_A)
> + else if (HAS_PCH_CPT(i915) && port != PORT_A)
> return DP_TRAIN_VOLTAGE_SWING_LEVEL_3;
> else
> return DP_TRAIN_VOLTAGE_SWING_LEVEL_2;
> @@ -3589,13 +3589,13 @@ intel_dp_voltage_max(struct intel_dp *intel_dp)
> u8
> intel_dp_pre_emphasis_max(struct intel_dp *intel_dp, u8 voltage_swing)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> struct intel_encoder *encoder = &dp_to_dig_port(intel_dp)->base;
> enum port port = encoder->port;
>
> - if (HAS_DDI(dev_priv)) {
> + if (HAS_DDI(i915)) {
> return intel_ddi_dp_pre_emphasis_max(encoder, voltage_swing);
> - } else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> + } else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> switch (voltage_swing & DP_TRAIN_VOLTAGE_SWING_MASK) {
> case DP_TRAIN_VOLTAGE_SWING_LEVEL_0:
> return DP_TRAIN_PRE_EMPH_LEVEL_3;
> @@ -3607,7 +3607,7 @@ intel_dp_pre_emphasis_max(struct intel_dp *intel_dp, u8 voltage_swing)
> default:
> return DP_TRAIN_PRE_EMPH_LEVEL_0;
> }
> - } else if (IS_IVYBRIDGE(dev_priv) && port == PORT_A) {
> + } else if (IS_IVYBRIDGE(i915) && port == PORT_A) {
> switch (voltage_swing & DP_TRAIN_VOLTAGE_SWING_MASK) {
> case DP_TRAIN_VOLTAGE_SWING_LEVEL_0:
> return DP_TRAIN_PRE_EMPH_LEVEL_2;
> @@ -3901,25 +3901,25 @@ ivb_cpu_edp_signal_levels(u8 train_set)
> void
> intel_dp_set_signal_levels(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp);
> enum port port = intel_dig_port->base.port;
> u32 signal_levels, mask = 0;
> u8 train_set = intel_dp->train_set[0];
>
> - if (IS_GEN9_LP(dev_priv) || INTEL_GEN(dev_priv) >= 10) {
> + if (IS_GEN9_LP(i915) || INTEL_GEN(i915) >= 10) {
> signal_levels = bxt_signal_levels(intel_dp);
> - } else if (HAS_DDI(dev_priv)) {
> + } else if (HAS_DDI(i915)) {
> signal_levels = ddi_signal_levels(intel_dp);
> mask = DDI_BUF_EMP_MASK;
> - } else if (IS_CHERRYVIEW(dev_priv)) {
> + } else if (IS_CHERRYVIEW(i915)) {
> signal_levels = chv_signal_levels(intel_dp);
> - } else if (IS_VALLEYVIEW(dev_priv)) {
> + } else if (IS_VALLEYVIEW(i915)) {
> signal_levels = vlv_signal_levels(intel_dp);
> - } else if (IS_IVYBRIDGE(dev_priv) && port == PORT_A) {
> + } else if (IS_IVYBRIDGE(i915) && port == PORT_A) {
> signal_levels = ivb_cpu_edp_signal_levels(train_set);
> mask = EDP_LINK_TRAIN_VOL_EMP_MASK_IVB;
> - } else if (IS_GEN(dev_priv, 6) && port == PORT_A) {
> + } else if (IS_GEN(i915, 6) && port == PORT_A) {
> signal_levels = snb_cpu_edp_signal_levels(train_set);
> mask = EDP_LINK_TRAIN_VOL_EMP_MASK_SNB;
> } else {
> @@ -3947,7 +3947,7 @@ intel_dp_program_link_training_pattern(struct intel_dp *intel_dp,
> u8 dp_train_pat)
> {
> struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp);
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> to_i915(intel_dig_port->base.base.dev);
>
> _intel_dp_set_link_train(intel_dp, &intel_dp->DP, dp_train_pat);
> @@ -3958,12 +3958,12 @@ intel_dp_program_link_training_pattern(struct intel_dp *intel_dp,
>
> void intel_dp_set_idle_link_train(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp);
> enum port port = intel_dig_port->base.port;
> u32 val;
>
> - if (!HAS_DDI(dev_priv))
> + if (!HAS_DDI(i915))
> return;
>
> val = I915_READ(DP_TP_CTL(port));
> @@ -3981,7 +3981,7 @@ void intel_dp_set_idle_link_train(struct intel_dp *intel_dp)
> if (port == PORT_A)
> return;
>
> - if (intel_wait_for_register(&dev_priv->uncore, DP_TP_STATUS(port),
> + if (intel_wait_for_register(&i915->uncore, DP_TP_STATUS(port),
> DP_TP_STATUS_IDLE_DONE,
> DP_TP_STATUS_IDLE_DONE,
> 1))
> @@ -3992,7 +3992,7 @@ static void
> intel_dp_link_down(struct intel_encoder *encoder,
> const struct intel_crtc_state *old_crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc);
> enum port port = encoder->port;
> @@ -4003,8 +4003,8 @@ intel_dp_link_down(struct intel_encoder *encoder,
>
> DRM_DEBUG_KMS("\n");
>
> - if ((IS_IVYBRIDGE(dev_priv) && port == PORT_A) ||
> - (HAS_PCH_CPT(dev_priv) && port != PORT_A)) {
> + if ((IS_IVYBRIDGE(i915) && port == PORT_A) ||
> + (HAS_PCH_CPT(i915) && port != PORT_A)) {
> DP &= ~DP_LINK_TRAIN_MASK_CPT;
> DP |= DP_LINK_TRAIN_PAT_IDLE_CPT;
> } else {
> @@ -4023,13 +4023,13 @@ intel_dp_link_down(struct intel_encoder *encoder,
> * to transcoder A after disabling it to allow the
> * matching HDMI port to be enabled on transcoder A.
> */
> - if (HAS_PCH_IBX(dev_priv) && crtc->pipe == PIPE_B && port != PORT_A) {
> + if (HAS_PCH_IBX(i915) && crtc->pipe == PIPE_B && port != PORT_A) {
> /*
> * We get CPU/PCH FIFO underruns on the other pipe when
> * doing the workaround. Sweep them under the rug.
> */
> - intel_set_cpu_fifo_underrun_reporting(dev_priv, PIPE_A, false);
> - intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, false);
> + intel_set_cpu_fifo_underrun_reporting(i915, PIPE_A, false);
> + intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, false);
>
> /* always enable with pattern 1 (as per spec) */
> DP &= ~(DP_PIPE_SEL_MASK | DP_LINK_TRAIN_MASK);
> @@ -4042,16 +4042,16 @@ intel_dp_link_down(struct intel_encoder *encoder,
> I915_WRITE(intel_dp->output_reg, DP);
> POSTING_READ(intel_dp->output_reg);
>
> - intel_wait_for_vblank_if_active(dev_priv, PIPE_A);
> - intel_set_cpu_fifo_underrun_reporting(dev_priv, PIPE_A, true);
> - intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, true);
> + intel_wait_for_vblank_if_active(i915, PIPE_A);
> + intel_set_cpu_fifo_underrun_reporting(i915, PIPE_A, true);
> + intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, true);
> }
>
> msleep(intel_dp->panel_power_down_delay);
>
> intel_dp->DP = DP;
>
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> intel_wakeref_t wakeref;
>
> with_pps_lock(intel_dp, wakeref)
> @@ -4156,7 +4156,7 @@ static void intel_dp_get_dsc_sink_cap(struct intel_dp *intel_dp)
> static bool
> intel_edp_init_dpcd(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> to_i915(dp_to_dig_port(intel_dp)->base.base.dev);
>
> /* this function is meant to be called only once */
> @@ -4169,7 +4169,7 @@ intel_edp_init_dpcd(struct intel_dp *intel_dp)
> drm_dp_is_branch(intel_dp->dpcd));
>
> if (intel_dp->dpcd[DP_DPCD_REV] >= 0x11)
> - dev_priv->no_aux_handshake = intel_dp->dpcd[DP_MAX_DOWNSPREAD] &
> + i915->no_aux_handshake = intel_dp->dpcd[DP_MAX_DOWNSPREAD] &
> DP_NO_AUX_HANDSHAKE_LINK_TRAINING;
>
> /*
> @@ -4230,7 +4230,7 @@ intel_edp_init_dpcd(struct intel_dp *intel_dp)
> intel_dp_set_common_rates(intel_dp);
>
> /* Read the eDP DSC DPCD registers */
> - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> intel_dp_get_dsc_sink_cap(intel_dp);
>
> return true;
> @@ -4804,7 +4804,7 @@ intel_dp_needs_link_retrain(struct intel_dp *intel_dp)
> int intel_dp_retrain_link(struct intel_encoder *encoder,
> struct drm_modeset_acquire_ctx *ctx)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> struct intel_connector *connector = intel_dp->attached_connector;
> struct drm_connector_state *conn_state;
> @@ -4817,7 +4817,7 @@ int intel_dp_retrain_link(struct intel_encoder *encoder,
> if (!connector || connector->base.status != connector_status_connected)
> return 0;
>
> - ret = drm_modeset_lock(&dev_priv->drm.mode_config.connection_mutex,
> + ret = drm_modeset_lock(&i915->drm.mode_config.connection_mutex,
> ctx);
> if (ret)
> return ret;
> @@ -4847,20 +4847,20 @@ int intel_dp_retrain_link(struct intel_encoder *encoder,
> return 0;
>
> /* Suppress underruns caused by re-training */
> - intel_set_cpu_fifo_underrun_reporting(dev_priv, crtc->pipe, false);
> + intel_set_cpu_fifo_underrun_reporting(i915, crtc->pipe, false);
> if (crtc_state->has_pch_encoder)
> - intel_set_pch_fifo_underrun_reporting(dev_priv,
> + intel_set_pch_fifo_underrun_reporting(i915,
> intel_crtc_pch_transcoder(crtc), false);
>
> intel_dp_start_link_train(intel_dp);
> intel_dp_stop_link_train(intel_dp);
>
> /* Keep underrun reporting disabled until things are stable */
> - intel_wait_for_vblank(dev_priv, crtc->pipe);
> + intel_wait_for_vblank(i915, crtc->pipe);
>
> - intel_set_cpu_fifo_underrun_reporting(dev_priv, crtc->pipe, true);
> + intel_set_cpu_fifo_underrun_reporting(i915, crtc->pipe, true);
> if (crtc_state->has_pch_encoder)
> - intel_set_pch_fifo_underrun_reporting(dev_priv,
> + intel_set_pch_fifo_underrun_reporting(i915,
> intel_crtc_pch_transcoder(crtc), true);
>
> return 0;
> @@ -4946,7 +4946,7 @@ static void intel_dp_check_service_irq(struct intel_dp *intel_dp)
> static bool
> intel_dp_short_pulse(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> u8 old_sink_count = intel_dp->sink_count;
> bool ret;
>
> @@ -4983,7 +4983,7 @@ intel_dp_short_pulse(struct intel_dp *intel_dp)
> if (intel_dp->compliance.test_type == DP_TEST_LINK_TRAINING) {
> DRM_DEBUG_KMS("Link Training Compliance Test requested\n");
> /* Send a Hotplug Uevent to userspace to start modeset */
> - drm_kms_helper_hotplug_event(&dev_priv->drm);
> + drm_kms_helper_hotplug_event(&i915->drm);
> }
>
> return true;
> @@ -5052,7 +5052,7 @@ edp_detect(struct intel_dp *intel_dp)
>
> static bool ibx_digital_port_connected(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> u32 bit;
>
> switch (encoder->hpd_pin) {
> @@ -5075,7 +5075,7 @@ static bool ibx_digital_port_connected(struct intel_encoder *encoder)
>
> static bool cpt_digital_port_connected(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> u32 bit;
>
> switch (encoder->hpd_pin) {
> @@ -5098,7 +5098,7 @@ static bool cpt_digital_port_connected(struct intel_encoder *encoder)
>
> static bool spt_digital_port_connected(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> u32 bit;
>
> switch (encoder->hpd_pin) {
> @@ -5117,7 +5117,7 @@ static bool spt_digital_port_connected(struct intel_encoder *encoder)
>
> static bool g4x_digital_port_connected(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> u32 bit;
>
> switch (encoder->hpd_pin) {
> @@ -5140,7 +5140,7 @@ static bool g4x_digital_port_connected(struct intel_encoder *encoder)
>
> static bool gm45_digital_port_connected(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> u32 bit;
>
> switch (encoder->hpd_pin) {
> @@ -5163,7 +5163,7 @@ static bool gm45_digital_port_connected(struct intel_encoder *encoder)
>
> static bool ilk_digital_port_connected(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
>
> if (encoder->hpd_pin == HPD_PORT_A)
> return I915_READ(DEISR) & DE_DP_A_HOTPLUG;
> @@ -5173,7 +5173,7 @@ static bool ilk_digital_port_connected(struct intel_encoder *encoder)
>
> static bool snb_digital_port_connected(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
>
> if (encoder->hpd_pin == HPD_PORT_A)
> return I915_READ(DEISR) & DE_DP_A_HOTPLUG;
> @@ -5183,7 +5183,7 @@ static bool snb_digital_port_connected(struct intel_encoder *encoder)
>
> static bool ivb_digital_port_connected(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
>
> if (encoder->hpd_pin == HPD_PORT_A)
> return I915_READ(DEISR) & DE_DP_A_HOTPLUG_IVB;
> @@ -5193,7 +5193,7 @@ static bool ivb_digital_port_connected(struct intel_encoder *encoder)
>
> static bool bdw_digital_port_connected(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
>
> if (encoder->hpd_pin == HPD_PORT_A)
> return I915_READ(GEN8_DE_PORT_ISR) & GEN8_PORT_DP_A_HOTPLUG;
> @@ -5203,7 +5203,7 @@ static bool bdw_digital_port_connected(struct intel_encoder *encoder)
>
> static bool bxt_digital_port_connected(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> u32 bit;
>
> switch (encoder->hpd_pin) {
> @@ -5224,7 +5224,7 @@ static bool bxt_digital_port_connected(struct intel_encoder *encoder)
> return I915_READ(GEN8_DE_PORT_ISR) & bit;
> }
>
> -static bool icl_combo_port_connected(struct drm_i915_private *dev_priv,
> +static bool icl_combo_port_connected(struct drm_i915_private *i915,
> struct intel_digital_port *intel_dig_port)
> {
> enum port port = intel_dig_port->base.port;
> @@ -5247,7 +5247,7 @@ static const char *tc_type_name(enum tc_port_type type)
> return names[type];
> }
>
> -static void icl_update_tc_port_type(struct drm_i915_private *dev_priv,
> +static void icl_update_tc_port_type(struct drm_i915_private *i915,
> struct intel_digital_port *intel_dig_port,
> bool is_legacy, bool is_typec, bool is_tbt)
> {
> @@ -5295,10 +5295,10 @@ static void icl_update_tc_port_type(struct drm_i915_private *dev_priv,
> * will require a lot of coordination with user space and thorough testing for
> * the extra possible cases.
> */
> -static bool icl_tc_phy_connect(struct drm_i915_private *dev_priv,
> +static bool icl_tc_phy_connect(struct drm_i915_private *i915,
> struct intel_digital_port *dig_port)
> {
> - enum tc_port tc_port = intel_port_to_tc(dev_priv, dig_port->base.port);
> + enum tc_port tc_port = intel_port_to_tc(i915, dig_port->base.port);
> u32 val;
>
> if (dig_port->tc_type != TC_PORT_LEGACY &&
> @@ -5329,7 +5329,7 @@ static bool icl_tc_phy_connect(struct drm_i915_private *dev_priv,
> if (dig_port->tc_type == TC_PORT_TYPEC &&
> !(I915_READ(PORT_TX_DFLEXDPSP) & TC_LIVE_STATE_TC(tc_port))) {
> DRM_DEBUG_KMS("TC PHY %d sudden disconnect.\n", tc_port);
> - icl_tc_phy_disconnect(dev_priv, dig_port);
> + icl_tc_phy_disconnect(i915, dig_port);
> return false;
> }
>
> @@ -5340,10 +5340,10 @@ static bool icl_tc_phy_connect(struct drm_i915_private *dev_priv,
> * See the comment at the connect function. This implements the Disconnect
> * Flow.
> */
> -void icl_tc_phy_disconnect(struct drm_i915_private *dev_priv,
> +void icl_tc_phy_disconnect(struct drm_i915_private *i915,
> struct intel_digital_port *dig_port)
> {
> - enum tc_port tc_port = intel_port_to_tc(dev_priv, dig_port->base.port);
> + enum tc_port tc_port = intel_port_to_tc(i915, dig_port->base.port);
>
> if (dig_port->tc_type == TC_PORT_UNKNOWN)
> return;
> @@ -5378,11 +5378,11 @@ void icl_tc_phy_disconnect(struct drm_i915_private *dev_priv,
> * connected ports are usable, and avoids exposing to the users objects they
> * can't really use.
> */
> -static bool icl_tc_port_connected(struct drm_i915_private *dev_priv,
> +static bool icl_tc_port_connected(struct drm_i915_private *i915,
> struct intel_digital_port *intel_dig_port)
> {
> enum port port = intel_dig_port->base.port;
> - enum tc_port tc_port = intel_port_to_tc(dev_priv, port);
> + enum tc_port tc_port = intel_port_to_tc(i915, port);
> bool is_legacy, is_typec, is_tbt;
> u32 dpsp;
>
> @@ -5407,15 +5407,15 @@ static bool icl_tc_port_connected(struct drm_i915_private *dev_priv,
> is_tbt = dpsp & TC_LIVE_STATE_TBT(tc_port);
>
> if (!is_legacy && !is_typec && !is_tbt) {
> - icl_tc_phy_disconnect(dev_priv, intel_dig_port);
> + icl_tc_phy_disconnect(i915, intel_dig_port);
>
> return false;
> }
>
> - icl_update_tc_port_type(dev_priv, intel_dig_port, is_legacy, is_typec,
> + icl_update_tc_port_type(i915, intel_dig_port, is_legacy, is_typec,
> is_tbt);
>
> - if (!icl_tc_phy_connect(dev_priv, intel_dig_port))
> + if (!icl_tc_phy_connect(i915, intel_dig_port))
> return false;
>
> return true;
> @@ -5423,13 +5423,13 @@ static bool icl_tc_port_connected(struct drm_i915_private *dev_priv,
>
> static bool icl_digital_port_connected(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base);
>
> - if (intel_port_is_combophy(dev_priv, encoder->port))
> - return icl_combo_port_connected(dev_priv, dig_port);
> - else if (intel_port_is_tc(dev_priv, encoder->port))
> - return icl_tc_port_connected(dev_priv, dig_port);
> + if (intel_port_is_combophy(i915, encoder->port))
> + return icl_combo_port_connected(i915, dig_port);
> + else if (intel_port_is_tc(i915, encoder->port))
> + return icl_tc_port_connected(i915, dig_port);
> else
> MISSING_CASE(encoder->hpd_pin);
>
> @@ -5449,41 +5449,41 @@ static bool icl_digital_port_connected(struct intel_encoder *encoder)
> */
> static bool __intel_digital_port_connected(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
>
> - if (HAS_GMCH(dev_priv)) {
> - if (IS_GM45(dev_priv))
> + if (HAS_GMCH(i915)) {
> + if (IS_GM45(i915))
> return gm45_digital_port_connected(encoder);
> else
> return g4x_digital_port_connected(encoder);
> }
>
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> return icl_digital_port_connected(encoder);
> - else if (IS_GEN(dev_priv, 10) || IS_GEN9_BC(dev_priv))
> + else if (IS_GEN(i915, 10) || IS_GEN9_BC(i915))
> return spt_digital_port_connected(encoder);
> - else if (IS_GEN9_LP(dev_priv))
> + else if (IS_GEN9_LP(i915))
> return bxt_digital_port_connected(encoder);
> - else if (IS_GEN(dev_priv, 8))
> + else if (IS_GEN(i915, 8))
> return bdw_digital_port_connected(encoder);
> - else if (IS_GEN(dev_priv, 7))
> + else if (IS_GEN(i915, 7))
> return ivb_digital_port_connected(encoder);
> - else if (IS_GEN(dev_priv, 6))
> + else if (IS_GEN(i915, 6))
> return snb_digital_port_connected(encoder);
> - else if (IS_GEN(dev_priv, 5))
> + else if (IS_GEN(i915, 5))
> return ilk_digital_port_connected(encoder);
>
> - MISSING_CASE(INTEL_GEN(dev_priv));
> + MISSING_CASE(INTEL_GEN(i915));
> return false;
> }
>
> bool intel_digital_port_connected(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> bool is_connected = false;
> intel_wakeref_t wakeref;
>
> - with_intel_display_power(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref)
> + with_intel_display_power(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref)
> is_connected = __intel_digital_port_connected(encoder);
>
> return is_connected;
> @@ -5537,7 +5537,7 @@ intel_dp_detect(struct drm_connector *connector,
> struct drm_modeset_acquire_ctx *ctx,
> bool force)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->dev);
> + struct drm_i915_private *i915 = to_i915(connector->dev);
> struct intel_dp *intel_dp = intel_attached_dp(connector);
> struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> struct intel_encoder *encoder = &dig_port->base;
> @@ -5545,7 +5545,7 @@ intel_dp_detect(struct drm_connector *connector,
>
> DRM_DEBUG_KMS("[CONNECTOR:%d:%s]\n",
> connector->base.id, connector->name);
> - WARN_ON(!drm_modeset_is_locked(&dev_priv->drm.mode_config.connection_mutex));
> + WARN_ON(!drm_modeset_is_locked(&i915->drm.mode_config.connection_mutex));
>
> /* Can't disconnect eDP */
> if (intel_dp_is_edp(intel_dp))
> @@ -5584,7 +5584,7 @@ intel_dp_detect(struct drm_connector *connector,
> intel_dp_print_rates(intel_dp);
>
> /* Read DP Sink DSC Cap DPCD regs for DP v1.4 */
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> intel_dp_get_dsc_sink_cap(intel_dp);
>
> drm_dp_read_desc(&intel_dp->aux, &intel_dp->desc,
> @@ -5642,7 +5642,7 @@ intel_dp_force(struct drm_connector *connector)
> struct intel_dp *intel_dp = intel_attached_dp(connector);
> struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> struct intel_encoder *intel_encoder = &dig_port->base;
> - struct drm_i915_private *dev_priv = to_i915(intel_encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(intel_encoder->base.dev);
> enum intel_display_power_domain aux_domain =
> intel_aux_power_domain(dig_port);
> intel_wakeref_t wakeref;
> @@ -5654,11 +5654,11 @@ intel_dp_force(struct drm_connector *connector)
> if (connector->status != connector_status_connected)
> return;
>
> - wakeref = intel_display_power_get(dev_priv, aux_domain);
> + wakeref = intel_display_power_get(i915, aux_domain);
>
> intel_dp_set_edid(intel_dp);
>
> - intel_display_power_put(dev_priv, aux_domain, wakeref);
> + intel_display_power_put(i915, aux_domain, wakeref);
> }
>
> static int intel_dp_get_modes(struct drm_connector *connector)
> @@ -6362,10 +6362,10 @@ static const struct intel_hdcp_shim intel_dp_hdcp_shim = {
>
> static void intel_edp_panel_vdd_sanitize(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
>
> - lockdep_assert_held(&dev_priv->pps_mutex);
> + lockdep_assert_held(&i915->pps_mutex);
>
> if (!edp_have_panel_vdd(intel_dp))
> return;
> @@ -6377,18 +6377,18 @@ static void intel_edp_panel_vdd_sanitize(struct intel_dp *intel_dp)
> * indefinitely.
> */
> DRM_DEBUG_KMS("VDD left on by BIOS, adjusting state tracking\n");
> - intel_display_power_get(dev_priv, intel_aux_power_domain(dig_port));
> + intel_display_power_get(i915, intel_aux_power_domain(dig_port));
>
> edp_panel_vdd_schedule_off(intel_dp);
> }
>
> static enum pipe vlv_active_pipe(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> struct intel_encoder *encoder = &dp_to_dig_port(intel_dp)->base;
> enum pipe pipe;
>
> - if (intel_dp_port_enabled(dev_priv, intel_dp->output_reg,
> + if (intel_dp_port_enabled(i915, intel_dp->output_reg,
> encoder->port, &pipe))
> return pipe;
>
> @@ -6397,12 +6397,12 @@ static enum pipe vlv_active_pipe(struct intel_dp *intel_dp)
>
> void intel_dp_encoder_reset(struct drm_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->dev);
> + struct drm_i915_private *i915 = to_i915(encoder->dev);
> struct intel_dp *intel_dp = enc_to_intel_dp(encoder);
> struct intel_lspcon *lspcon = dp_to_lspcon(intel_dp);
> intel_wakeref_t wakeref;
>
> - if (!HAS_DDI(dev_priv))
> + if (!HAS_DDI(i915))
> intel_dp->DP = I915_READ(intel_dp->output_reg);
>
> if (lspcon->active)
> @@ -6410,12 +6410,12 @@ void intel_dp_encoder_reset(struct drm_encoder *encoder)
>
> intel_dp->reset_link_params = true;
>
> - if (!IS_VALLEYVIEW(dev_priv) && !IS_CHERRYVIEW(dev_priv) &&
> + if (!IS_VALLEYVIEW(i915) && !IS_CHERRYVIEW(i915) &&
> !intel_dp_is_edp(intel_dp))
> return;
>
> with_pps_lock(intel_dp, wakeref) {
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> intel_dp->active_pipe = vlv_active_pipe(intel_dp);
>
> if (intel_dp_is_edp(intel_dp)) {
> @@ -6508,41 +6508,41 @@ intel_dp_hpd_pulse(struct intel_digital_port *intel_dig_port, bool long_hpd)
> }
>
> /* check the VBT to see whether the eDP is on another port */
> -bool intel_dp_is_port_edp(struct drm_i915_private *dev_priv, enum port port)
> +bool intel_dp_is_port_edp(struct drm_i915_private *i915, enum port port)
> {
> /*
> * eDP not supported on g4x. so bail out early just
> * for a bit extra safety in case the VBT is bonkers.
> */
> - if (INTEL_GEN(dev_priv) < 5)
> + if (INTEL_GEN(i915) < 5)
> return false;
>
> - if (INTEL_GEN(dev_priv) < 9 && port == PORT_A)
> + if (INTEL_GEN(i915) < 9 && port == PORT_A)
> return true;
>
> - return intel_bios_is_port_edp(dev_priv, port);
> + return intel_bios_is_port_edp(i915, port);
> }
>
> static void
> intel_dp_add_properties(struct intel_dp *intel_dp, struct drm_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->dev);
> + struct drm_i915_private *i915 = to_i915(connector->dev);
> enum port port = dp_to_dig_port(intel_dp)->base.port;
>
> - if (!IS_G4X(dev_priv) && port != PORT_A)
> + if (!IS_G4X(i915) && port != PORT_A)
> intel_attach_force_audio_property(connector);
>
> intel_attach_broadcast_rgb_property(connector);
> - if (HAS_GMCH(dev_priv))
> + if (HAS_GMCH(i915))
> drm_connector_attach_max_bpc_property(connector, 6, 10);
> - else if (INTEL_GEN(dev_priv) >= 5)
> + else if (INTEL_GEN(i915) >= 5)
> drm_connector_attach_max_bpc_property(connector, 6, 12);
>
> if (intel_dp_is_edp(intel_dp)) {
> u32 allowed_scalers;
>
> allowed_scalers = BIT(DRM_MODE_SCALE_ASPECT) | BIT(DRM_MODE_SCALE_FULLSCREEN);
> - if (!HAS_GMCH(dev_priv))
> + if (!HAS_GMCH(i915))
> allowed_scalers |= BIT(DRM_MODE_SCALE_CENTER);
>
> drm_connector_attach_scaling_mode_property(connector, allowed_scalers);
> @@ -6562,7 +6562,7 @@ static void intel_dp_init_panel_power_timestamps(struct intel_dp *intel_dp)
> static void
> intel_pps_readout_hw_state(struct intel_dp *intel_dp, struct edp_power_seq *seq)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> u32 pp_on, pp_off, pp_ctl;
> struct pps_registers regs;
>
> @@ -6571,7 +6571,7 @@ intel_pps_readout_hw_state(struct intel_dp *intel_dp, struct edp_power_seq *seq)
> pp_ctl = ironlake_get_pp_control(intel_dp);
>
> /* Ensure PPS is unlocked */
> - if (!HAS_DDI(dev_priv))
> + if (!HAS_DDI(i915))
> I915_WRITE(regs.pp_ctrl, pp_ctl);
>
> pp_on = I915_READ(regs.pp_on);
> @@ -6621,11 +6621,11 @@ intel_pps_verify_state(struct intel_dp *intel_dp)
> static void
> intel_dp_init_panel_power_sequencer(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> struct edp_power_seq cur, vbt, spec,
> *final = &intel_dp->pps_delays;
>
> - lockdep_assert_held(&dev_priv->pps_mutex);
> + lockdep_assert_held(&i915->pps_mutex);
>
> /* already initialized? */
> if (final->t11_t12 != 0)
> @@ -6635,13 +6635,13 @@ intel_dp_init_panel_power_sequencer(struct intel_dp *intel_dp)
>
> intel_pps_dump_state("cur", &cur);
>
> - vbt = dev_priv->vbt.edp.pps;
> + vbt = i915->vbt.edp.pps;
> /* On Toshiba Satellite P50-C-18C system the VBT T12 delay
> * of 500ms appears to be too short. Ocassionally the panel
> * just fails to power back on. Increasing the delay to 800ms
> * seems sufficient to avoid this problem.
> */
> - if (dev_priv->quirks & QUIRK_INCREASE_T12_DELAY) {
> + if (i915->quirks & QUIRK_INCREASE_T12_DELAY) {
> vbt.t11_t12 = max_t(u16, vbt.t11_t12, 1300 * 10);
> DRM_DEBUG_KMS("Increasing T12 panel delay as per the quirk to %d\n",
> vbt.t11_t12);
> @@ -6714,14 +6714,14 @@ static void
> intel_dp_init_panel_power_sequencer_registers(struct intel_dp *intel_dp,
> bool force_disable_vdd)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> u32 pp_on, pp_off, port_sel = 0;
> - int div = dev_priv->rawclk_freq / 1000;
> + int div = i915->rawclk_freq / 1000;
> struct pps_registers regs;
> enum port port = dp_to_dig_port(intel_dp)->base.port;
> const struct edp_power_seq *seq = &intel_dp->pps_delays;
>
> - lockdep_assert_held(&dev_priv->pps_mutex);
> + lockdep_assert_held(&i915->pps_mutex);
>
> intel_pps_get_registers(intel_dp, ®s);
>
> @@ -6757,9 +6757,9 @@ intel_dp_init_panel_power_sequencer_registers(struct intel_dp *intel_dp,
>
> /* Haswell doesn't have any port selection bits for the panel
> * power sequencer any more. */
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> port_sel = PANEL_PORT_SELECT_VLV(port);
> - } else if (HAS_PCH_IBX(dev_priv) || HAS_PCH_CPT(dev_priv)) {
> + } else if (HAS_PCH_IBX(i915) || HAS_PCH_CPT(i915)) {
> switch (port) {
> case PORT_A:
> port_sel = PANEL_PORT_SELECT_DPA;
> @@ -6807,9 +6807,9 @@ intel_dp_init_panel_power_sequencer_registers(struct intel_dp *intel_dp,
>
> static void intel_dp_pps_init(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
>
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> vlv_initial_power_sequencer_setup(intel_dp);
> } else {
> intel_dp_init_panel_power_sequencer(intel_dp);
> @@ -6819,7 +6819,7 @@ static void intel_dp_pps_init(struct intel_dp *intel_dp)
>
> /**
> * intel_dp_set_drrs_state - program registers for RR switch to take effect
> - * @dev_priv: i915 device
> + * @i915: i915 device
> * @crtc_state: a pointer to the active intel_crtc_state
> * @refresh_rate: RR to be programmed
> *
> @@ -6828,15 +6828,15 @@ static void intel_dp_pps_init(struct intel_dp *intel_dp)
> * supported by the panel or to any other RR based on media playback (in
> * this case, RR value needs to be passed from user space).
> *
> - * The caller of this function needs to take a lock on dev_priv->drrs.
> + * The caller of this function needs to take a lock on i915->drrs.
> */
> -static void intel_dp_set_drrs_state(struct drm_i915_private *dev_priv,
> +static void intel_dp_set_drrs_state(struct drm_i915_private *i915,
> const struct intel_crtc_state *crtc_state,
> int refresh_rate)
> {
> struct intel_encoder *encoder;
> struct intel_digital_port *dig_port = NULL;
> - struct intel_dp *intel_dp = dev_priv->drrs.dp;
> + struct intel_dp *intel_dp = i915->drrs.dp;
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc);
> enum drrs_refresh_rate_type index = DRRS_HIGH_RR;
>
> @@ -6858,7 +6858,7 @@ static void intel_dp_set_drrs_state(struct drm_i915_private *dev_priv,
> return;
> }
>
> - if (dev_priv->drrs.type < SEAMLESS_DRRS_SUPPORT) {
> + if (i915->drrs.type < SEAMLESS_DRRS_SUPPORT) {
> DRM_DEBUG_KMS("Only Seamless DRRS supported.\n");
> return;
> }
> @@ -6867,7 +6867,7 @@ static void intel_dp_set_drrs_state(struct drm_i915_private *dev_priv,
> refresh_rate)
> index = DRRS_LOW_RR;
>
> - if (index == dev_priv->drrs.refresh_rate_type) {
> + if (index == i915->drrs.refresh_rate_type) {
> DRM_DEBUG_KMS(
> "DRRS requested for previously set RR...ignoring\n");
> return;
> @@ -6878,7 +6878,7 @@ static void intel_dp_set_drrs_state(struct drm_i915_private *dev_priv,
> return;
> }
>
> - if (INTEL_GEN(dev_priv) >= 8 && !IS_CHERRYVIEW(dev_priv)) {
> + if (INTEL_GEN(i915) >= 8 && !IS_CHERRYVIEW(i915)) {
> switch (index) {
> case DRRS_HIGH_RR:
> intel_dp_set_m_n(crtc_state, M1_N1);
> @@ -6890,18 +6890,18 @@ static void intel_dp_set_drrs_state(struct drm_i915_private *dev_priv,
> default:
> DRM_ERROR("Unsupported refreshrate type\n");
> }
> - } else if (INTEL_GEN(dev_priv) > 6) {
> + } else if (INTEL_GEN(i915) > 6) {
> i915_reg_t reg = PIPECONF(crtc_state->cpu_transcoder);
> u32 val;
>
> val = I915_READ(reg);
> if (index > DRRS_HIGH_RR) {
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> val |= PIPECONF_EDP_RR_MODE_SWITCH_VLV;
> else
> val |= PIPECONF_EDP_RR_MODE_SWITCH;
> } else {
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> val &= ~PIPECONF_EDP_RR_MODE_SWITCH_VLV;
> else
> val &= ~PIPECONF_EDP_RR_MODE_SWITCH;
> @@ -6909,7 +6909,7 @@ static void intel_dp_set_drrs_state(struct drm_i915_private *dev_priv,
> I915_WRITE(reg, val);
> }
>
> - dev_priv->drrs.refresh_rate_type = index;
> + i915->drrs.refresh_rate_type = index;
>
> DRM_DEBUG_KMS("eDP Refresh Rate set to : %dHz\n", refresh_rate);
> }
> @@ -6924,30 +6924,30 @@ static void intel_dp_set_drrs_state(struct drm_i915_private *dev_priv,
> void intel_edp_drrs_enable(struct intel_dp *intel_dp,
> const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
>
> if (!crtc_state->has_drrs) {
> DRM_DEBUG_KMS("Panel doesn't support DRRS\n");
> return;
> }
>
> - if (dev_priv->psr.enabled) {
> + if (i915->psr.enabled) {
> DRM_DEBUG_KMS("PSR enabled. Not enabling DRRS.\n");
> return;
> }
>
> - mutex_lock(&dev_priv->drrs.mutex);
> - if (dev_priv->drrs.dp) {
> + mutex_lock(&i915->drrs.mutex);
> + if (i915->drrs.dp) {
> DRM_DEBUG_KMS("DRRS already enabled\n");
> goto unlock;
> }
>
> - dev_priv->drrs.busy_frontbuffer_bits = 0;
> + i915->drrs.busy_frontbuffer_bits = 0;
>
> - dev_priv->drrs.dp = intel_dp;
> + i915->drrs.dp = intel_dp;
>
> unlock:
> - mutex_unlock(&dev_priv->drrs.mutex);
> + mutex_unlock(&i915->drrs.mutex);
> }
>
> /**
> @@ -6959,36 +6959,36 @@ void intel_edp_drrs_enable(struct intel_dp *intel_dp,
> void intel_edp_drrs_disable(struct intel_dp *intel_dp,
> const struct intel_crtc_state *old_crtc_state)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
>
> if (!old_crtc_state->has_drrs)
> return;
>
> - mutex_lock(&dev_priv->drrs.mutex);
> - if (!dev_priv->drrs.dp) {
> - mutex_unlock(&dev_priv->drrs.mutex);
> + mutex_lock(&i915->drrs.mutex);
> + if (!i915->drrs.dp) {
> + mutex_unlock(&i915->drrs.mutex);
> return;
> }
>
> - if (dev_priv->drrs.refresh_rate_type == DRRS_LOW_RR)
> - intel_dp_set_drrs_state(dev_priv, old_crtc_state,
> + if (i915->drrs.refresh_rate_type == DRRS_LOW_RR)
> + intel_dp_set_drrs_state(i915, old_crtc_state,
> intel_dp->attached_connector->panel.fixed_mode->vrefresh);
>
> - dev_priv->drrs.dp = NULL;
> - mutex_unlock(&dev_priv->drrs.mutex);
> + i915->drrs.dp = NULL;
> + mutex_unlock(&i915->drrs.mutex);
>
> - cancel_delayed_work_sync(&dev_priv->drrs.work);
> + cancel_delayed_work_sync(&i915->drrs.work);
> }
>
> static void intel_edp_drrs_downclock_work(struct work_struct *work)
> {
> - struct drm_i915_private *dev_priv =
> - container_of(work, typeof(*dev_priv), drrs.work.work);
> + struct drm_i915_private *i915 =
> + container_of(work, typeof(*i915), drrs.work.work);
> struct intel_dp *intel_dp;
>
> - mutex_lock(&dev_priv->drrs.mutex);
> + mutex_lock(&i915->drrs.mutex);
>
> - intel_dp = dev_priv->drrs.dp;
> + intel_dp = i915->drrs.dp;
>
> if (!intel_dp)
> goto unlock;
> @@ -6998,23 +6998,23 @@ static void intel_edp_drrs_downclock_work(struct work_struct *work)
> * recheck.
> */
>
> - if (dev_priv->drrs.busy_frontbuffer_bits)
> + if (i915->drrs.busy_frontbuffer_bits)
> goto unlock;
>
> - if (dev_priv->drrs.refresh_rate_type != DRRS_LOW_RR) {
> + if (i915->drrs.refresh_rate_type != DRRS_LOW_RR) {
> struct drm_crtc *crtc = dp_to_dig_port(intel_dp)->base.base.crtc;
>
> - intel_dp_set_drrs_state(dev_priv, to_intel_crtc(crtc)->config,
> + intel_dp_set_drrs_state(i915, to_intel_crtc(crtc)->config,
> intel_dp->attached_connector->panel.downclock_mode->vrefresh);
> }
>
> unlock:
> - mutex_unlock(&dev_priv->drrs.mutex);
> + mutex_unlock(&i915->drrs.mutex);
> }
>
> /**
> * intel_edp_drrs_invalidate - Disable Idleness DRRS
> - * @dev_priv: i915 device
> + * @i915: i915 device
> * @frontbuffer_bits: frontbuffer plane tracking bits
> *
> * This function gets called everytime rendering on the given planes start.
> @@ -7022,40 +7022,40 @@ static void intel_edp_drrs_downclock_work(struct work_struct *work)
> *
> * Dirty frontbuffers relevant to DRRS are tracked in busy_frontbuffer_bits.
> */
> -void intel_edp_drrs_invalidate(struct drm_i915_private *dev_priv,
> +void intel_edp_drrs_invalidate(struct drm_i915_private *i915,
> unsigned int frontbuffer_bits)
> {
> struct drm_crtc *crtc;
> enum pipe pipe;
>
> - if (dev_priv->drrs.type == DRRS_NOT_SUPPORTED)
> + if (i915->drrs.type == DRRS_NOT_SUPPORTED)
> return;
>
> - cancel_delayed_work(&dev_priv->drrs.work);
> + cancel_delayed_work(&i915->drrs.work);
>
> - mutex_lock(&dev_priv->drrs.mutex);
> - if (!dev_priv->drrs.dp) {
> - mutex_unlock(&dev_priv->drrs.mutex);
> + mutex_lock(&i915->drrs.mutex);
> + if (!i915->drrs.dp) {
> + mutex_unlock(&i915->drrs.mutex);
> return;
> }
>
> - crtc = dp_to_dig_port(dev_priv->drrs.dp)->base.base.crtc;
> + crtc = dp_to_dig_port(i915->drrs.dp)->base.base.crtc;
> pipe = to_intel_crtc(crtc)->pipe;
>
> frontbuffer_bits &= INTEL_FRONTBUFFER_ALL_MASK(pipe);
> - dev_priv->drrs.busy_frontbuffer_bits |= frontbuffer_bits;
> + i915->drrs.busy_frontbuffer_bits |= frontbuffer_bits;
>
> /* invalidate means busy screen hence upclock */
> - if (frontbuffer_bits && dev_priv->drrs.refresh_rate_type == DRRS_LOW_RR)
> - intel_dp_set_drrs_state(dev_priv, to_intel_crtc(crtc)->config,
> - dev_priv->drrs.dp->attached_connector->panel.fixed_mode->vrefresh);
> + if (frontbuffer_bits && i915->drrs.refresh_rate_type == DRRS_LOW_RR)
> + intel_dp_set_drrs_state(i915, to_intel_crtc(crtc)->config,
> + i915->drrs.dp->attached_connector->panel.fixed_mode->vrefresh);
>
> - mutex_unlock(&dev_priv->drrs.mutex);
> + mutex_unlock(&i915->drrs.mutex);
> }
>
> /**
> * intel_edp_drrs_flush - Restart Idleness DRRS
> - * @dev_priv: i915 device
> + * @i915: i915 device
> * @frontbuffer_bits: frontbuffer plane tracking bits
> *
> * This function gets called every time rendering on the given planes has
> @@ -7065,42 +7065,42 @@ void intel_edp_drrs_invalidate(struct drm_i915_private *dev_priv,
> *
> * Dirty frontbuffers relevant to DRRS are tracked in busy_frontbuffer_bits.
> */
> -void intel_edp_drrs_flush(struct drm_i915_private *dev_priv,
> +void intel_edp_drrs_flush(struct drm_i915_private *i915,
> unsigned int frontbuffer_bits)
> {
> struct drm_crtc *crtc;
> enum pipe pipe;
>
> - if (dev_priv->drrs.type == DRRS_NOT_SUPPORTED)
> + if (i915->drrs.type == DRRS_NOT_SUPPORTED)
> return;
>
> - cancel_delayed_work(&dev_priv->drrs.work);
> + cancel_delayed_work(&i915->drrs.work);
>
> - mutex_lock(&dev_priv->drrs.mutex);
> - if (!dev_priv->drrs.dp) {
> - mutex_unlock(&dev_priv->drrs.mutex);
> + mutex_lock(&i915->drrs.mutex);
> + if (!i915->drrs.dp) {
> + mutex_unlock(&i915->drrs.mutex);
> return;
> }
>
> - crtc = dp_to_dig_port(dev_priv->drrs.dp)->base.base.crtc;
> + crtc = dp_to_dig_port(i915->drrs.dp)->base.base.crtc;
> pipe = to_intel_crtc(crtc)->pipe;
>
> frontbuffer_bits &= INTEL_FRONTBUFFER_ALL_MASK(pipe);
> - dev_priv->drrs.busy_frontbuffer_bits &= ~frontbuffer_bits;
> + i915->drrs.busy_frontbuffer_bits &= ~frontbuffer_bits;
>
> /* flush means busy screen hence upclock */
> - if (frontbuffer_bits && dev_priv->drrs.refresh_rate_type == DRRS_LOW_RR)
> - intel_dp_set_drrs_state(dev_priv, to_intel_crtc(crtc)->config,
> - dev_priv->drrs.dp->attached_connector->panel.fixed_mode->vrefresh);
> + if (frontbuffer_bits && i915->drrs.refresh_rate_type == DRRS_LOW_RR)
> + intel_dp_set_drrs_state(i915, to_intel_crtc(crtc)->config,
> + i915->drrs.dp->attached_connector->panel.fixed_mode->vrefresh);
>
> /*
> * flush also means no more activity hence schedule downclock, if all
> * other fbs are quiescent too
> */
> - if (!dev_priv->drrs.busy_frontbuffer_bits)
> - schedule_delayed_work(&dev_priv->drrs.work,
> + if (!i915->drrs.busy_frontbuffer_bits)
> + schedule_delayed_work(&i915->drrs.work,
> msecs_to_jiffies(1000));
> - mutex_unlock(&dev_priv->drrs.mutex);
> + mutex_unlock(&i915->drrs.mutex);
> }
>
> /**
> @@ -7157,18 +7157,18 @@ static struct drm_display_mode *
> intel_dp_drrs_init(struct intel_connector *connector,
> struct drm_display_mode *fixed_mode)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct drm_display_mode *downclock_mode = NULL;
>
> - INIT_DELAYED_WORK(&dev_priv->drrs.work, intel_edp_drrs_downclock_work);
> - mutex_init(&dev_priv->drrs.mutex);
> + INIT_DELAYED_WORK(&i915->drrs.work, intel_edp_drrs_downclock_work);
> + mutex_init(&i915->drrs.mutex);
>
> - if (INTEL_GEN(dev_priv) <= 6) {
> + if (INTEL_GEN(i915) <= 6) {
> DRM_DEBUG_KMS("DRRS supported for Gen7 and above\n");
> return NULL;
> }
>
> - if (dev_priv->vbt.drrs_type != SEAMLESS_DRRS_SUPPORT) {
> + if (i915->vbt.drrs_type != SEAMLESS_DRRS_SUPPORT) {
> DRM_DEBUG_KMS("VBT doesn't support DRRS\n");
> return NULL;
> }
> @@ -7179,9 +7179,9 @@ intel_dp_drrs_init(struct intel_connector *connector,
> return NULL;
> }
>
> - dev_priv->drrs.type = dev_priv->vbt.drrs_type;
> + i915->drrs.type = i915->vbt.drrs_type;
>
> - dev_priv->drrs.refresh_rate_type = DRRS_HIGH_RR;
> + i915->drrs.refresh_rate_type = DRRS_HIGH_RR;
> DRM_DEBUG_KMS("seamless DRRS supported for eDP panel.\n");
> return downclock_mode;
> }
> @@ -7189,8 +7189,8 @@ intel_dp_drrs_init(struct intel_connector *connector,
> static bool intel_edp_init_connector(struct intel_dp *intel_dp,
> struct intel_connector *intel_connector)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> + struct drm_device *dev = &i915->drm;
> struct drm_connector *connector = &intel_connector->base;
> struct drm_display_mode *fixed_mode = NULL;
> struct drm_display_mode *downclock_mode = NULL;
> @@ -7210,8 +7210,8 @@ static bool intel_edp_init_connector(struct intel_dp *intel_dp,
> * eDP and LVDS bail out early in this case to prevent interfering
> * with an already powered-on LVDS power sequencer.
> */
> - if (intel_get_lvds_encoder(dev_priv)) {
> - WARN_ON(!(HAS_PCH_IBX(dev_priv) || HAS_PCH_CPT(dev_priv)));
> + if (intel_get_lvds_encoder(i915)) {
> + WARN_ON(!(HAS_PCH_IBX(i915) || HAS_PCH_CPT(i915)));
> DRM_INFO("LVDS was detected, not registering eDP\n");
>
> return false;
> @@ -7256,7 +7256,7 @@ static bool intel_edp_init_connector(struct intel_dp *intel_dp,
> fixed_mode = intel_panel_vbt_fixed_mode(intel_connector);
> mutex_unlock(&dev->mode_config.mutex);
>
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> intel_dp->edp_notifier.notifier_call = edp_notify_handler;
> register_reboot_notifier(&intel_dp->edp_notifier);
>
> @@ -7330,7 +7330,7 @@ intel_dp_init_connector(struct intel_digital_port *intel_dig_port,
> struct intel_dp *intel_dp = &intel_dig_port->dp;
> struct intel_encoder *intel_encoder = &intel_dig_port->base;
> struct drm_device *dev = intel_encoder->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> enum port port = intel_encoder->port;
> int type;
>
> @@ -7353,18 +7353,18 @@ intel_dp_init_connector(struct intel_digital_port *intel_dig_port,
> intel_dp->DP = I915_READ(intel_dp->output_reg);
> intel_dp->attached_connector = intel_connector;
>
> - if (intel_dp_is_port_edp(dev_priv, port)) {
> + if (intel_dp_is_port_edp(i915, port)) {
> /*
> * Currently we don't support eDP on TypeC ports, although in
> * theory it could work on TypeC legacy ports.
> */
> - WARN_ON(intel_port_is_tc(dev_priv, port));
> + WARN_ON(intel_port_is_tc(i915, port));
> type = DRM_MODE_CONNECTOR_eDP;
> } else {
> type = DRM_MODE_CONNECTOR_DisplayPort;
> }
>
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> intel_dp->active_pipe = vlv_active_pipe(intel_dp);
>
> /*
> @@ -7376,7 +7376,7 @@ intel_dp_init_connector(struct intel_digital_port *intel_dig_port,
> intel_encoder->type = INTEL_OUTPUT_EDP;
>
> /* eDP only on port B and/or C on vlv/chv */
> - if (WARN_ON((IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) &&
> + if (WARN_ON((IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) &&
> intel_dp_is_edp(intel_dp) &&
> port != PORT_B && port != PORT_C))
> return false;
> @@ -7388,26 +7388,26 @@ intel_dp_init_connector(struct intel_digital_port *intel_dig_port,
> drm_connector_init(dev, connector, &intel_dp_connector_funcs, type);
> drm_connector_helper_add(connector, &intel_dp_connector_helper_funcs);
>
> - if (!HAS_GMCH(dev_priv))
> + if (!HAS_GMCH(i915))
> connector->interlace_allowed = true;
> connector->doublescan_allowed = 0;
>
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> connector->ycbcr_420_allowed = true;
>
> - intel_encoder->hpd_pin = intel_hpd_pin_default(dev_priv, port);
> + intel_encoder->hpd_pin = intel_hpd_pin_default(i915, port);
>
> intel_dp_aux_init(intel_dp);
>
> intel_connector_attach_encoder(intel_connector, intel_encoder);
>
> - if (HAS_DDI(dev_priv))
> + if (HAS_DDI(i915))
> intel_connector->get_hw_state = intel_ddi_connector_get_hw_state;
> else
> intel_connector->get_hw_state = intel_connector_get_hw_state;
>
> /* init MST on ports that can support it */
> - if (HAS_DP_MST(dev_priv) && !intel_dp_is_edp(intel_dp) &&
> + if (HAS_DP_MST(i915) && !intel_dp_is_edp(intel_dp) &&
> (port == PORT_B || port == PORT_C ||
> port == PORT_D || port == PORT_F))
> intel_dp_mst_encoder_init(intel_dig_port,
> @@ -7421,7 +7421,7 @@ intel_dp_init_connector(struct intel_digital_port *intel_dig_port,
>
> intel_dp_add_properties(intel_dp, connector);
>
> - if (is_hdcp_supported(dev_priv, port) && !intel_dp_is_edp(intel_dp)) {
> + if (is_hdcp_supported(i915, port) && !intel_dp_is_edp(intel_dp)) {
> int ret = intel_hdcp_init(intel_connector, &intel_dp_hdcp_shim);
> if (ret)
> DRM_DEBUG_KMS("HDCP init failed, skipping.\n");
> @@ -7431,7 +7431,7 @@ intel_dp_init_connector(struct intel_digital_port *intel_dig_port,
> * 0xd. Failure to do so will result in spurious interrupts being
> * generated on the port when a cable is not attached.
> */
> - if (IS_G45(dev_priv)) {
> + if (IS_G45(i915)) {
> u32 temp = I915_READ(PEG_BAND_GAP_DATA);
> I915_WRITE(PEG_BAND_GAP_DATA, (temp & ~0xf) | 0xd);
> }
> @@ -7444,7 +7444,7 @@ intel_dp_init_connector(struct intel_digital_port *intel_dig_port,
> return false;
> }
>
> -bool intel_dp_init(struct drm_i915_private *dev_priv,
> +bool intel_dp_init(struct drm_i915_private *i915,
> i915_reg_t output_reg,
> enum port port)
> {
> @@ -7464,7 +7464,7 @@ bool intel_dp_init(struct drm_i915_private *dev_priv,
> intel_encoder = &intel_dig_port->base;
> encoder = &intel_encoder->base;
>
> - if (drm_encoder_init(&dev_priv->drm, &intel_encoder->base,
> + if (drm_encoder_init(&i915->drm, &intel_encoder->base,
> &intel_dp_enc_funcs, DRM_MODE_ENCODER_TMDS,
> "DP %c", port_name(port)))
> goto err_encoder_init;
> @@ -7475,14 +7475,14 @@ bool intel_dp_init(struct drm_i915_private *dev_priv,
> intel_encoder->get_config = intel_dp_get_config;
> intel_encoder->update_pipe = intel_panel_update_backlight;
> intel_encoder->suspend = intel_dp_encoder_suspend;
> - if (IS_CHERRYVIEW(dev_priv)) {
> + if (IS_CHERRYVIEW(i915)) {
> intel_encoder->pre_pll_enable = chv_dp_pre_pll_enable;
> intel_encoder->pre_enable = chv_pre_enable_dp;
> intel_encoder->enable = vlv_enable_dp;
> intel_encoder->disable = vlv_disable_dp;
> intel_encoder->post_disable = chv_post_disable_dp;
> intel_encoder->post_pll_disable = chv_dp_post_pll_disable;
> - } else if (IS_VALLEYVIEW(dev_priv)) {
> + } else if (IS_VALLEYVIEW(i915)) {
> intel_encoder->pre_pll_enable = vlv_dp_pre_pll_enable;
> intel_encoder->pre_enable = vlv_pre_enable_dp;
> intel_encoder->enable = vlv_enable_dp;
> @@ -7500,7 +7500,7 @@ bool intel_dp_init(struct drm_i915_private *dev_priv,
>
> intel_encoder->type = INTEL_OUTPUT_DP;
> intel_encoder->power_domain = intel_port_to_power_domain(port);
> - if (IS_CHERRYVIEW(dev_priv)) {
> + if (IS_CHERRYVIEW(i915)) {
> if (port == PORT_D)
> intel_encoder->crtc_mask = 1 << 2;
> else
> @@ -7516,7 +7516,7 @@ bool intel_dp_init(struct drm_i915_private *dev_priv,
> if (port != PORT_A)
> intel_infoframe_init(intel_dig_port);
>
> - intel_dig_port->aux_ch = intel_bios_port_aux_ch(dev_priv, port);
> + intel_dig_port->aux_ch = intel_bios_port_aux_ch(i915, port);
> if (!intel_dp_init_connector(intel_dig_port, intel_connector))
> goto err_init_connector;
>
> @@ -7531,11 +7531,11 @@ bool intel_dp_init(struct drm_i915_private *dev_priv,
> return false;
> }
>
> -void intel_dp_mst_suspend(struct drm_i915_private *dev_priv)
> +void intel_dp_mst_suspend(struct drm_i915_private *i915)
> {
> struct intel_encoder *encoder;
>
> - for_each_intel_encoder(&dev_priv->drm, encoder) {
> + for_each_intel_encoder(&i915->drm, encoder) {
> struct intel_dp *intel_dp;
>
> if (encoder->type != INTEL_OUTPUT_DDI)
> @@ -7551,11 +7551,11 @@ void intel_dp_mst_suspend(struct drm_i915_private *dev_priv)
> }
> }
>
> -void intel_dp_mst_resume(struct drm_i915_private *dev_priv)
> +void intel_dp_mst_resume(struct drm_i915_private *i915)
> {
> struct intel_encoder *encoder;
>
> - for_each_intel_encoder(&dev_priv->drm, encoder) {
> + for_each_intel_encoder(&i915->drm, encoder) {
> struct intel_dp *intel_dp;
> int ret;
>
> diff --git a/drivers/gpu/drm/i915/intel_dp.h b/drivers/gpu/drm/i915/intel_dp.h
> index da70b1a41c83..b6fecb185b23 100644
> --- a/drivers/gpu/drm/i915/intel_dp.h
> +++ b/drivers/gpu/drm/i915/intel_dp.h
> @@ -35,10 +35,10 @@ void intel_dp_adjust_compliance_config(struct intel_dp *intel_dp,
> bool intel_dp_limited_color_range(const struct intel_crtc_state *crtc_state,
> const struct drm_connector_state *conn_state);
> int intel_dp_min_bpp(const struct intel_crtc_state *crtc_state);
> -bool intel_dp_port_enabled(struct drm_i915_private *dev_priv,
> +bool intel_dp_port_enabled(struct drm_i915_private *i915,
> i915_reg_t dp_reg, enum port port,
> enum pipe *pipe);
> -bool intel_dp_init(struct drm_i915_private *dev_priv, i915_reg_t output_reg,
> +bool intel_dp_init(struct drm_i915_private *i915, i915_reg_t output_reg,
> enum port port);
> bool intel_dp_init_connector(struct intel_digital_port *intel_dig_port,
> struct intel_connector *intel_connector);
> @@ -60,7 +60,7 @@ int intel_dp_compute_config(struct intel_encoder *encoder,
> struct intel_crtc_state *pipe_config,
> struct drm_connector_state *conn_state);
> bool intel_dp_is_edp(struct intel_dp *intel_dp);
> -bool intel_dp_is_port_edp(struct drm_i915_private *dev_priv, enum port port);
> +bool intel_dp_is_port_edp(struct drm_i915_private *i915, enum port port);
> enum irqreturn intel_dp_hpd_pulse(struct intel_digital_port *intel_dig_port,
> bool long_hpd);
> void intel_edp_backlight_on(const struct intel_crtc_state *crtc_state,
> @@ -69,21 +69,21 @@ void intel_edp_backlight_off(const struct drm_connector_state *conn_state);
> void intel_edp_panel_vdd_on(struct intel_dp *intel_dp);
> void intel_edp_panel_on(struct intel_dp *intel_dp);
> void intel_edp_panel_off(struct intel_dp *intel_dp);
> -void intel_dp_mst_suspend(struct drm_i915_private *dev_priv);
> -void intel_dp_mst_resume(struct drm_i915_private *dev_priv);
> +void intel_dp_mst_suspend(struct drm_i915_private *i915);
> +void intel_dp_mst_resume(struct drm_i915_private *i915);
> int intel_dp_max_link_rate(struct intel_dp *intel_dp);
> int intel_dp_max_lane_count(struct intel_dp *intel_dp);
> int intel_dp_rate_select(struct intel_dp *intel_dp, int rate);
> -void intel_power_sequencer_reset(struct drm_i915_private *dev_priv);
> +void intel_power_sequencer_reset(struct drm_i915_private *i915);
> u32 intel_dp_pack_aux(const u8 *src, int src_bytes);
>
> void intel_edp_drrs_enable(struct intel_dp *intel_dp,
> const struct intel_crtc_state *crtc_state);
> void intel_edp_drrs_disable(struct intel_dp *intel_dp,
> const struct intel_crtc_state *crtc_state);
> -void intel_edp_drrs_invalidate(struct drm_i915_private *dev_priv,
> +void intel_edp_drrs_invalidate(struct drm_i915_private *i915,
> unsigned int frontbuffer_bits);
> -void intel_edp_drrs_flush(struct drm_i915_private *dev_priv,
> +void intel_edp_drrs_flush(struct drm_i915_private *i915,
> unsigned int frontbuffer_bits);
>
> void
> @@ -112,7 +112,7 @@ bool intel_dp_get_colorimetry_status(struct intel_dp *intel_dp);
> int intel_dp_link_required(int pixel_clock, int bpp);
> int intel_dp_max_data_rate(int max_link_clock, int max_lanes);
> bool intel_digital_port_connected(struct intel_encoder *encoder);
> -void icl_tc_phy_disconnect(struct drm_i915_private *dev_priv,
> +void icl_tc_phy_disconnect(struct drm_i915_private *i915,
> struct intel_digital_port *dig_port);
>
> static inline unsigned int intel_dp_unused_lane_mask(int lane_count)
> diff --git a/drivers/gpu/drm/i915/intel_dp_aux_backlight.c b/drivers/gpu/drm/i915/intel_dp_aux_backlight.c
> index 7ded95a334db..7a3bd29dcf0b 100644
> --- a/drivers/gpu/drm/i915/intel_dp_aux_backlight.c
> +++ b/drivers/gpu/drm/i915/intel_dp_aux_backlight.c
> @@ -109,7 +109,7 @@ intel_dp_aux_set_backlight(const struct drm_connector_state *conn_state, u32 lev
> */
> static bool intel_dp_aux_set_pwm_freq(struct intel_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_dp *intel_dp = enc_to_intel_dp(&connector->encoder->base);
> int freq, fxp, fxp_min, fxp_max, fxp_actual, f = 1;
> u8 pn, pn_min, pn_max;
> @@ -118,7 +118,7 @@ static bool intel_dp_aux_set_pwm_freq(struct intel_connector *connector)
> * Note that, if F x P is out of supported range, the maximum value or
> * minimum value will applied automatically. So no need to check that.
> */
> - freq = dev_priv->vbt.backlight.pwm_freq_hz;
> + freq = i915->vbt.backlight.pwm_freq_hz;
> DRM_DEBUG_KMS("VBT defined backlight frequency %u Hz\n", freq);
> if (!freq) {
> DRM_DEBUG_KMS("Use panel default backlight frequency\n");
> diff --git a/drivers/gpu/drm/i915/intel_dp_mst.c b/drivers/gpu/drm/i915/intel_dp_mst.c
> index 0caf645fbbb8..ab21dc690c9f 100644
> --- a/drivers/gpu/drm/i915/intel_dp_mst.c
> +++ b/drivers/gpu/drm/i915/intel_dp_mst.c
> @@ -91,7 +91,7 @@ static int intel_dp_mst_compute_config(struct intel_encoder *encoder,
> struct intel_crtc_state *pipe_config,
> struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dp_mst_encoder *intel_mst = enc_to_mst(&encoder->base);
> struct intel_dp *intel_dp = &intel_mst->primary->dp;
> struct intel_connector *connector =
> @@ -140,11 +140,11 @@ static int intel_dp_mst_compute_config(struct intel_encoder *encoder,
> pipe_config->limited_color_range =
> intel_dp_limited_color_range(pipe_config, conn_state);
>
> - if (IS_GEN9_LP(dev_priv))
> + if (IS_GEN9_LP(i915))
> pipe_config->lane_lat_optim_mask =
> bxt_ddi_phy_calc_lane_lat_optim_mask(pipe_config->lane_count);
>
> - intel_ddi_compute_min_voltage_level(dev_priv, pipe_config);
> + intel_ddi_compute_min_voltage_level(i915, pipe_config);
>
> return 0;
> }
> @@ -285,7 +285,7 @@ static void intel_mst_pre_enable_dp(struct intel_encoder *encoder,
> struct intel_dp_mst_encoder *intel_mst = enc_to_mst(&encoder->base);
> struct intel_digital_port *intel_dig_port = intel_mst->primary;
> struct intel_dp *intel_dp = &intel_dig_port->dp;
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> enum port port = intel_dig_port->base.port;
> struct intel_connector *connector =
> to_intel_connector(conn_state->connector);
> @@ -332,12 +332,12 @@ static void intel_mst_enable_dp(struct intel_encoder *encoder,
> struct intel_dp_mst_encoder *intel_mst = enc_to_mst(&encoder->base);
> struct intel_digital_port *intel_dig_port = intel_mst->primary;
> struct intel_dp *intel_dp = &intel_dig_port->dp;
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> enum port port = intel_dig_port->base.port;
>
> DRM_DEBUG_KMS("active links %d\n", intel_dp->active_mst_links);
>
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> DP_TP_STATUS(port),
> DP_TP_STATUS_ACT_SENT,
> DP_TP_STATUS_ACT_SENT,
> @@ -495,7 +495,7 @@ static struct drm_connector *intel_dp_add_mst_connector(struct drm_dp_mst_topolo
> struct intel_dp *intel_dp = container_of(mgr, struct intel_dp, mst_mgr);
> struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp);
> struct drm_device *dev = intel_dig_port->base.base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_connector *intel_connector;
> struct drm_connector *connector;
> enum pipe pipe;
> @@ -520,7 +520,7 @@ static struct drm_connector *intel_dp_add_mst_connector(struct drm_dp_mst_topolo
>
> drm_connector_helper_add(connector, &intel_dp_mst_connector_helper_funcs);
>
> - for_each_pipe(dev_priv, pipe) {
> + for_each_pipe(i915, pipe) {
> struct drm_encoder *enc =
> &intel_dp->mst_encoders[pipe]->base.base;
>
> @@ -549,10 +549,10 @@ static struct drm_connector *intel_dp_add_mst_connector(struct drm_dp_mst_topolo
>
> static void intel_dp_register_mst_connector(struct drm_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->dev);
> + struct drm_i915_private *i915 = to_i915(connector->dev);
>
> - if (dev_priv->fbdev)
> - drm_fb_helper_add_one_connector(&dev_priv->fbdev->helper,
> + if (i915->fbdev)
> + drm_fb_helper_add_one_connector(&i915->fbdev->helper,
> connector);
>
> drm_connector_register(connector);
> @@ -561,13 +561,13 @@ static void intel_dp_register_mst_connector(struct drm_connector *connector)
> static void intel_dp_destroy_mst_connector(struct drm_dp_mst_topology_mgr *mgr,
> struct drm_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->dev);
> + struct drm_i915_private *i915 = to_i915(connector->dev);
>
> DRM_DEBUG_KMS("[CONNECTOR:%d:%s]\n", connector->base.id, connector->name);
> drm_connector_unregister(connector);
>
> - if (dev_priv->fbdev)
> - drm_fb_helper_remove_one_connector(&dev_priv->fbdev->helper,
> + if (i915->fbdev)
> + drm_fb_helper_remove_one_connector(&i915->fbdev->helper,
> connector);
>
> drm_connector_put(connector);
> @@ -622,10 +622,10 @@ static bool
> intel_dp_create_fake_mst_encoders(struct intel_digital_port *intel_dig_port)
> {
> struct intel_dp *intel_dp = &intel_dig_port->dp;
> - struct drm_i915_private *dev_priv = to_i915(intel_dig_port->base.base.dev);
> + struct drm_i915_private *i915 = to_i915(intel_dig_port->base.base.dev);
> enum pipe pipe;
>
> - for_each_pipe(dev_priv, pipe)
> + for_each_pipe(i915, pipe)
> intel_dp->mst_encoders[pipe] = intel_dp_create_fake_mst_encoder(intel_dig_port, pipe);
> return true;
> }
> diff --git a/drivers/gpu/drm/i915/intel_dpio_phy.c b/drivers/gpu/drm/i915/intel_dpio_phy.c
> index bdbe41759827..c85672bb1a7d 100644
> --- a/drivers/gpu/drm/i915/intel_dpio_phy.c
> +++ b/drivers/gpu/drm/i915/intel_dpio_phy.c
> @@ -212,9 +212,9 @@ static const struct bxt_ddi_phy_info glk_ddi_phy_info[] = {
> };
>
> static const struct bxt_ddi_phy_info *
> -bxt_get_phy_list(struct drm_i915_private *dev_priv, int *count)
> +bxt_get_phy_list(struct drm_i915_private *i915, int *count)
> {
> - if (IS_GEMINILAKE(dev_priv)) {
> + if (IS_GEMINILAKE(i915)) {
> *count = ARRAY_SIZE(glk_ddi_phy_info);
> return glk_ddi_phy_info;
> } else {
> @@ -224,22 +224,22 @@ bxt_get_phy_list(struct drm_i915_private *dev_priv, int *count)
> }
>
> static const struct bxt_ddi_phy_info *
> -bxt_get_phy_info(struct drm_i915_private *dev_priv, enum dpio_phy phy)
> +bxt_get_phy_info(struct drm_i915_private *i915, enum dpio_phy phy)
> {
> int count;
> const struct bxt_ddi_phy_info *phy_list =
> - bxt_get_phy_list(dev_priv, &count);
> + bxt_get_phy_list(i915, &count);
>
> return &phy_list[phy];
> }
>
> -void bxt_port_to_phy_channel(struct drm_i915_private *dev_priv, enum port port,
> +void bxt_port_to_phy_channel(struct drm_i915_private *i915, enum port port,
> enum dpio_phy *phy, enum dpio_channel *ch)
> {
> const struct bxt_ddi_phy_info *phy_info, *phys;
> int i, count;
>
> - phys = bxt_get_phy_list(dev_priv, &count);
> + phys = bxt_get_phy_list(i915, &count);
>
> for (i = 0; i < count; i++) {
> phy_info = &phys[i];
> @@ -263,7 +263,7 @@ void bxt_port_to_phy_channel(struct drm_i915_private *dev_priv, enum port port,
> *ch = DPIO_CH0;
> }
>
> -void bxt_ddi_phy_set_signal_level(struct drm_i915_private *dev_priv,
> +void bxt_ddi_phy_set_signal_level(struct drm_i915_private *i915,
> enum port port, u32 margin, u32 scale,
> u32 enable, u32 deemphasis)
> {
> @@ -271,7 +271,7 @@ void bxt_ddi_phy_set_signal_level(struct drm_i915_private *dev_priv,
> enum dpio_phy phy;
> enum dpio_channel ch;
>
> - bxt_port_to_phy_channel(dev_priv, port, &phy, &ch);
> + bxt_port_to_phy_channel(i915, port, &phy, &ch);
>
> /*
> * While we write to the group register to program all lanes at once we
> @@ -306,12 +306,12 @@ void bxt_ddi_phy_set_signal_level(struct drm_i915_private *dev_priv,
> I915_WRITE(BXT_PORT_PCS_DW10_GRP(phy, ch), val);
> }
>
> -bool bxt_ddi_phy_is_enabled(struct drm_i915_private *dev_priv,
> +bool bxt_ddi_phy_is_enabled(struct drm_i915_private *i915,
> enum dpio_phy phy)
> {
> const struct bxt_ddi_phy_info *phy_info;
>
> - phy_info = bxt_get_phy_info(dev_priv, phy);
> + phy_info = bxt_get_phy_info(i915, phy);
>
> if (!(I915_READ(BXT_P_CR_GT_DISP_PWRON) & phy_info->pwron_mask))
> return false;
> @@ -334,37 +334,37 @@ bool bxt_ddi_phy_is_enabled(struct drm_i915_private *dev_priv,
> return true;
> }
>
> -static u32 bxt_get_grc(struct drm_i915_private *dev_priv, enum dpio_phy phy)
> +static u32 bxt_get_grc(struct drm_i915_private *i915, enum dpio_phy phy)
> {
> u32 val = I915_READ(BXT_PORT_REF_DW6(phy));
>
> return (val & GRC_CODE_MASK) >> GRC_CODE_SHIFT;
> }
>
> -static void bxt_phy_wait_grc_done(struct drm_i915_private *dev_priv,
> +static void bxt_phy_wait_grc_done(struct drm_i915_private *i915,
> enum dpio_phy phy)
> {
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> BXT_PORT_REF_DW3(phy),
> GRC_DONE, GRC_DONE,
> 10))
> DRM_ERROR("timeout waiting for PHY%d GRC\n", phy);
> }
>
> -static void _bxt_ddi_phy_init(struct drm_i915_private *dev_priv,
> +static void _bxt_ddi_phy_init(struct drm_i915_private *i915,
> enum dpio_phy phy)
> {
> const struct bxt_ddi_phy_info *phy_info;
> u32 val;
>
> - phy_info = bxt_get_phy_info(dev_priv, phy);
> + phy_info = bxt_get_phy_info(i915, phy);
>
> - if (bxt_ddi_phy_is_enabled(dev_priv, phy)) {
> + if (bxt_ddi_phy_is_enabled(i915, phy)) {
> /* Still read out the GRC value for state verification */
> if (phy_info->rcomp_phy != -1)
> - dev_priv->bxt_phy_grc = bxt_get_grc(dev_priv, phy);
> + i915->bxt_phy_grc = bxt_get_grc(i915, phy);
>
> - if (bxt_ddi_phy_verify_state(dev_priv, phy)) {
> + if (bxt_ddi_phy_verify_state(i915, phy)) {
> DRM_DEBUG_DRIVER("DDI PHY %d already enabled, "
> "won't reprogram it\n", phy);
> return;
> @@ -386,7 +386,7 @@ static void _bxt_ddi_phy_init(struct drm_i915_private *dev_priv,
> * The flag should get set in 100us according to the HW team, but
> * use 1ms due to occasional timeouts observed with that.
> */
> - if (intel_wait_for_register_fw(&dev_priv->uncore,
> + if (intel_wait_for_register_fw(&i915->uncore,
> BXT_PORT_CL1CM_DW0(phy),
> PHY_RESERVED | PHY_POWER_GOOD,
> PHY_POWER_GOOD,
> @@ -419,14 +419,14 @@ static void _bxt_ddi_phy_init(struct drm_i915_private *dev_priv,
> if (phy_info->rcomp_phy != -1) {
> u32 grc_code;
>
> - bxt_phy_wait_grc_done(dev_priv, phy_info->rcomp_phy);
> + bxt_phy_wait_grc_done(i915, phy_info->rcomp_phy);
>
> /*
> * PHY0 isn't connected to an RCOMP resistor so copy over
> * the corresponding calibrated value from PHY1, and disable
> * the automatic calibration on PHY0.
> */
> - val = dev_priv->bxt_phy_grc = bxt_get_grc(dev_priv,
> + val = i915->bxt_phy_grc = bxt_get_grc(i915,
> phy_info->rcomp_phy);
> grc_code = val << GRC_CODE_FAST_SHIFT |
> val << GRC_CODE_SLOW_SHIFT |
> @@ -446,12 +446,12 @@ static void _bxt_ddi_phy_init(struct drm_i915_private *dev_priv,
> I915_WRITE(BXT_PHY_CTL_FAMILY(phy), val);
> }
>
> -void bxt_ddi_phy_uninit(struct drm_i915_private *dev_priv, enum dpio_phy phy)
> +void bxt_ddi_phy_uninit(struct drm_i915_private *i915, enum dpio_phy phy)
> {
> const struct bxt_ddi_phy_info *phy_info;
> u32 val;
>
> - phy_info = bxt_get_phy_info(dev_priv, phy);
> + phy_info = bxt_get_phy_info(i915, phy);
>
> val = I915_READ(BXT_PHY_CTL_FAMILY(phy));
> val &= ~COMMON_RESET_DIS;
> @@ -462,34 +462,34 @@ void bxt_ddi_phy_uninit(struct drm_i915_private *dev_priv, enum dpio_phy phy)
> I915_WRITE(BXT_P_CR_GT_DISP_PWRON, val);
> }
>
> -void bxt_ddi_phy_init(struct drm_i915_private *dev_priv, enum dpio_phy phy)
> +void bxt_ddi_phy_init(struct drm_i915_private *i915, enum dpio_phy phy)
> {
> const struct bxt_ddi_phy_info *phy_info =
> - bxt_get_phy_info(dev_priv, phy);
> + bxt_get_phy_info(i915, phy);
> enum dpio_phy rcomp_phy = phy_info->rcomp_phy;
> bool was_enabled;
>
> - lockdep_assert_held(&dev_priv->power_domains.lock);
> + lockdep_assert_held(&i915->power_domains.lock);
>
> was_enabled = true;
> if (rcomp_phy != -1)
> - was_enabled = bxt_ddi_phy_is_enabled(dev_priv, rcomp_phy);
> + was_enabled = bxt_ddi_phy_is_enabled(i915, rcomp_phy);
>
> /*
> * We need to copy the GRC calibration value from rcomp_phy,
> * so make sure it's powered up.
> */
> if (!was_enabled)
> - _bxt_ddi_phy_init(dev_priv, rcomp_phy);
> + _bxt_ddi_phy_init(i915, rcomp_phy);
>
> - _bxt_ddi_phy_init(dev_priv, phy);
> + _bxt_ddi_phy_init(i915, phy);
>
> if (!was_enabled)
> - bxt_ddi_phy_uninit(dev_priv, rcomp_phy);
> + bxt_ddi_phy_uninit(i915, rcomp_phy);
> }
>
> static bool __printf(6, 7)
> -__phy_reg_verify_state(struct drm_i915_private *dev_priv, enum dpio_phy phy,
> +__phy_reg_verify_state(struct drm_i915_private *i915, enum dpio_phy phy,
> i915_reg_t reg, u32 mask, u32 expected,
> const char *reg_fmt, ...)
> {
> @@ -515,20 +515,20 @@ __phy_reg_verify_state(struct drm_i915_private *dev_priv, enum dpio_phy phy,
> return false;
> }
>
> -bool bxt_ddi_phy_verify_state(struct drm_i915_private *dev_priv,
> +bool bxt_ddi_phy_verify_state(struct drm_i915_private *i915,
> enum dpio_phy phy)
> {
> const struct bxt_ddi_phy_info *phy_info;
> u32 mask;
> bool ok;
>
> - phy_info = bxt_get_phy_info(dev_priv, phy);
> + phy_info = bxt_get_phy_info(i915, phy);
>
> #define _CHK(reg, mask, exp, fmt, ...) \
> - __phy_reg_verify_state(dev_priv, phy, reg, mask, exp, fmt, \
> + __phy_reg_verify_state(i915, phy, reg, mask, exp, fmt, \
> ## __VA_ARGS__)
>
> - if (!bxt_ddi_phy_is_enabled(dev_priv, phy))
> + if (!bxt_ddi_phy_is_enabled(i915, phy))
> return false;
>
> ok = true;
> @@ -552,7 +552,7 @@ bool bxt_ddi_phy_verify_state(struct drm_i915_private *dev_priv,
> "BXT_PORT_CL2CM_DW6(%d)", phy);
>
> if (phy_info->rcomp_phy != -1) {
> - u32 grc_code = dev_priv->bxt_phy_grc;
> + u32 grc_code = i915->bxt_phy_grc;
>
> grc_code = grc_code << GRC_CODE_FAST_SHIFT |
> grc_code << GRC_CODE_SLOW_SHIFT |
> @@ -591,13 +591,13 @@ bxt_ddi_phy_calc_lane_lat_optim_mask(u8 lane_count)
> void bxt_ddi_phy_set_lane_optim_mask(struct intel_encoder *encoder,
> u8 lane_lat_optim_mask)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> enum port port = encoder->port;
> enum dpio_phy phy;
> enum dpio_channel ch;
> int lane;
>
> - bxt_port_to_phy_channel(dev_priv, port, &phy, &ch);
> + bxt_port_to_phy_channel(i915, port, &phy, &ch);
>
> for (lane = 0; lane < 4; lane++) {
> u32 val = I915_READ(BXT_PORT_TX_DW14_LN(phy, ch, lane));
> @@ -617,14 +617,14 @@ void bxt_ddi_phy_set_lane_optim_mask(struct intel_encoder *encoder,
> u8
> bxt_ddi_phy_get_lane_lat_optim_mask(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> enum port port = encoder->port;
> enum dpio_phy phy;
> enum dpio_channel ch;
> int lane;
> u8 mask;
>
> - bxt_port_to_phy_channel(dev_priv, port, &phy, &ch);
> + bxt_port_to_phy_channel(i915, port, &phy, &ch);
>
> mask = 0;
> for (lane = 0; lane < 4; lane++) {
> @@ -642,7 +642,7 @@ void chv_set_phy_signal_level(struct intel_encoder *encoder,
> u32 deemph_reg_value, u32 margin_reg_value,
> bool uniq_trans_scale)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_digital_port *dport = enc_to_dig_port(&encoder->base);
> struct intel_crtc *intel_crtc = to_intel_crtc(encoder->base.crtc);
> enum dpio_channel ch = vlv_dport_to_channel(dport);
> @@ -650,46 +650,46 @@ void chv_set_phy_signal_level(struct intel_encoder *encoder,
> u32 val;
> int i;
>
> - vlv_dpio_get(dev_priv);
> + vlv_dpio_get(i915);
>
> /* Clear calc init */
> - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS01_DW10(ch));
> + val = vlv_dpio_read(i915, pipe, VLV_PCS01_DW10(ch));
> val &= ~(DPIO_PCS_SWING_CALC_TX0_TX2 | DPIO_PCS_SWING_CALC_TX1_TX3);
> val &= ~(DPIO_PCS_TX1DEEMP_MASK | DPIO_PCS_TX2DEEMP_MASK);
> val |= DPIO_PCS_TX1DEEMP_9P5 | DPIO_PCS_TX2DEEMP_9P5;
> - vlv_dpio_write(dev_priv, pipe, VLV_PCS01_DW10(ch), val);
> + vlv_dpio_write(i915, pipe, VLV_PCS01_DW10(ch), val);
>
> if (intel_crtc->config->lane_count > 2) {
> - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS23_DW10(ch));
> + val = vlv_dpio_read(i915, pipe, VLV_PCS23_DW10(ch));
> val &= ~(DPIO_PCS_SWING_CALC_TX0_TX2 | DPIO_PCS_SWING_CALC_TX1_TX3);
> val &= ~(DPIO_PCS_TX1DEEMP_MASK | DPIO_PCS_TX2DEEMP_MASK);
> val |= DPIO_PCS_TX1DEEMP_9P5 | DPIO_PCS_TX2DEEMP_9P5;
> - vlv_dpio_write(dev_priv, pipe, VLV_PCS23_DW10(ch), val);
> + vlv_dpio_write(i915, pipe, VLV_PCS23_DW10(ch), val);
> }
>
> - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS01_DW9(ch));
> + val = vlv_dpio_read(i915, pipe, VLV_PCS01_DW9(ch));
> val &= ~(DPIO_PCS_TX1MARGIN_MASK | DPIO_PCS_TX2MARGIN_MASK);
> val |= DPIO_PCS_TX1MARGIN_000 | DPIO_PCS_TX2MARGIN_000;
> - vlv_dpio_write(dev_priv, pipe, VLV_PCS01_DW9(ch), val);
> + vlv_dpio_write(i915, pipe, VLV_PCS01_DW9(ch), val);
>
> if (intel_crtc->config->lane_count > 2) {
> - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS23_DW9(ch));
> + val = vlv_dpio_read(i915, pipe, VLV_PCS23_DW9(ch));
> val &= ~(DPIO_PCS_TX1MARGIN_MASK | DPIO_PCS_TX2MARGIN_MASK);
> val |= DPIO_PCS_TX1MARGIN_000 | DPIO_PCS_TX2MARGIN_000;
> - vlv_dpio_write(dev_priv, pipe, VLV_PCS23_DW9(ch), val);
> + vlv_dpio_write(i915, pipe, VLV_PCS23_DW9(ch), val);
> }
>
> /* Program swing deemph */
> for (i = 0; i < intel_crtc->config->lane_count; i++) {
> - val = vlv_dpio_read(dev_priv, pipe, CHV_TX_DW4(ch, i));
> + val = vlv_dpio_read(i915, pipe, CHV_TX_DW4(ch, i));
> val &= ~DPIO_SWING_DEEMPH9P5_MASK;
> val |= deemph_reg_value << DPIO_SWING_DEEMPH9P5_SHIFT;
> - vlv_dpio_write(dev_priv, pipe, CHV_TX_DW4(ch, i), val);
> + vlv_dpio_write(i915, pipe, CHV_TX_DW4(ch, i), val);
> }
>
> /* Program swing margin */
> for (i = 0; i < intel_crtc->config->lane_count; i++) {
> - val = vlv_dpio_read(dev_priv, pipe, CHV_TX_DW2(ch, i));
> + val = vlv_dpio_read(i915, pipe, CHV_TX_DW2(ch, i));
>
> val &= ~DPIO_SWING_MARGIN000_MASK;
> val |= margin_reg_value << DPIO_SWING_MARGIN000_SHIFT;
> @@ -702,7 +702,7 @@ void chv_set_phy_signal_level(struct intel_encoder *encoder,
> val &= ~(0xff << DPIO_UNIQ_TRANS_SCALE_SHIFT);
> val |= 0x9a << DPIO_UNIQ_TRANS_SCALE_SHIFT;
>
> - vlv_dpio_write(dev_priv, pipe, CHV_TX_DW2(ch, i), val);
> + vlv_dpio_write(i915, pipe, CHV_TX_DW2(ch, i), val);
> }
>
> /*
> @@ -712,70 +712,70 @@ void chv_set_phy_signal_level(struct intel_encoder *encoder,
> * 27 for ch0 and ch1.
> */
> for (i = 0; i < intel_crtc->config->lane_count; i++) {
> - val = vlv_dpio_read(dev_priv, pipe, CHV_TX_DW3(ch, i));
> + val = vlv_dpio_read(i915, pipe, CHV_TX_DW3(ch, i));
> if (uniq_trans_scale)
> val |= DPIO_TX_UNIQ_TRANS_SCALE_EN;
> else
> val &= ~DPIO_TX_UNIQ_TRANS_SCALE_EN;
> - vlv_dpio_write(dev_priv, pipe, CHV_TX_DW3(ch, i), val);
> + vlv_dpio_write(i915, pipe, CHV_TX_DW3(ch, i), val);
> }
>
> /* Start swing calculation */
> - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS01_DW10(ch));
> + val = vlv_dpio_read(i915, pipe, VLV_PCS01_DW10(ch));
> val |= DPIO_PCS_SWING_CALC_TX0_TX2 | DPIO_PCS_SWING_CALC_TX1_TX3;
> - vlv_dpio_write(dev_priv, pipe, VLV_PCS01_DW10(ch), val);
> + vlv_dpio_write(i915, pipe, VLV_PCS01_DW10(ch), val);
>
> if (intel_crtc->config->lane_count > 2) {
> - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS23_DW10(ch));
> + val = vlv_dpio_read(i915, pipe, VLV_PCS23_DW10(ch));
> val |= DPIO_PCS_SWING_CALC_TX0_TX2 | DPIO_PCS_SWING_CALC_TX1_TX3;
> - vlv_dpio_write(dev_priv, pipe, VLV_PCS23_DW10(ch), val);
> + vlv_dpio_write(i915, pipe, VLV_PCS23_DW10(ch), val);
> }
>
> - vlv_dpio_put(dev_priv);
> + vlv_dpio_put(i915);
> }
>
> void chv_data_lane_soft_reset(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state,
> bool reset)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> enum dpio_channel ch = vlv_dport_to_channel(enc_to_dig_port(&encoder->base));
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> enum pipe pipe = crtc->pipe;
> u32 val;
>
> - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS01_DW0(ch));
> + val = vlv_dpio_read(i915, pipe, VLV_PCS01_DW0(ch));
> if (reset)
> val &= ~(DPIO_PCS_TX_LANE2_RESET | DPIO_PCS_TX_LANE1_RESET);
> else
> val |= DPIO_PCS_TX_LANE2_RESET | DPIO_PCS_TX_LANE1_RESET;
> - vlv_dpio_write(dev_priv, pipe, VLV_PCS01_DW0(ch), val);
> + vlv_dpio_write(i915, pipe, VLV_PCS01_DW0(ch), val);
>
> if (crtc_state->lane_count > 2) {
> - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS23_DW0(ch));
> + val = vlv_dpio_read(i915, pipe, VLV_PCS23_DW0(ch));
> if (reset)
> val &= ~(DPIO_PCS_TX_LANE2_RESET | DPIO_PCS_TX_LANE1_RESET);
> else
> val |= DPIO_PCS_TX_LANE2_RESET | DPIO_PCS_TX_LANE1_RESET;
> - vlv_dpio_write(dev_priv, pipe, VLV_PCS23_DW0(ch), val);
> + vlv_dpio_write(i915, pipe, VLV_PCS23_DW0(ch), val);
> }
>
> - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS01_DW1(ch));
> + val = vlv_dpio_read(i915, pipe, VLV_PCS01_DW1(ch));
> val |= CHV_PCS_REQ_SOFTRESET_EN;
> if (reset)
> val &= ~DPIO_PCS_CLK_SOFT_RESET;
> else
> val |= DPIO_PCS_CLK_SOFT_RESET;
> - vlv_dpio_write(dev_priv, pipe, VLV_PCS01_DW1(ch), val);
> + vlv_dpio_write(i915, pipe, VLV_PCS01_DW1(ch), val);
>
> if (crtc_state->lane_count > 2) {
> - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS23_DW1(ch));
> + val = vlv_dpio_read(i915, pipe, VLV_PCS23_DW1(ch));
> val |= CHV_PCS_REQ_SOFTRESET_EN;
> if (reset)
> val &= ~DPIO_PCS_CLK_SOFT_RESET;
> else
> val |= DPIO_PCS_CLK_SOFT_RESET;
> - vlv_dpio_write(dev_priv, pipe, VLV_PCS23_DW1(ch), val);
> + vlv_dpio_write(i915, pipe, VLV_PCS23_DW1(ch), val);
> }
> }
>
> @@ -783,7 +783,7 @@ void chv_phy_pre_pll_enable(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state)
> {
> struct intel_digital_port *dport = enc_to_dig_port(&encoder->base);
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> enum dpio_channel ch = vlv_dport_to_channel(dport);
> enum pipe pipe = crtc->pipe;
> @@ -797,51 +797,51 @@ void chv_phy_pre_pll_enable(struct intel_encoder *encoder,
> */
> if (ch == DPIO_CH0 && pipe == PIPE_B)
> dport->release_cl2_override =
> - !chv_phy_powergate_ch(dev_priv, DPIO_PHY0, DPIO_CH1, true);
> + !chv_phy_powergate_ch(i915, DPIO_PHY0, DPIO_CH1, true);
>
> chv_phy_powergate_lanes(encoder, true, lane_mask);
>
> - vlv_dpio_get(dev_priv);
> + vlv_dpio_get(i915);
>
> /* Assert data lane reset */
> chv_data_lane_soft_reset(encoder, crtc_state, true);
>
> /* program left/right clock distribution */
> if (pipe != PIPE_B) {
> - val = vlv_dpio_read(dev_priv, pipe, _CHV_CMN_DW5_CH0);
> + val = vlv_dpio_read(i915, pipe, _CHV_CMN_DW5_CH0);
> val &= ~(CHV_BUFLEFTENA1_MASK | CHV_BUFRIGHTENA1_MASK);
> if (ch == DPIO_CH0)
> val |= CHV_BUFLEFTENA1_FORCE;
> if (ch == DPIO_CH1)
> val |= CHV_BUFRIGHTENA1_FORCE;
> - vlv_dpio_write(dev_priv, pipe, _CHV_CMN_DW5_CH0, val);
> + vlv_dpio_write(i915, pipe, _CHV_CMN_DW5_CH0, val);
> } else {
> - val = vlv_dpio_read(dev_priv, pipe, _CHV_CMN_DW1_CH1);
> + val = vlv_dpio_read(i915, pipe, _CHV_CMN_DW1_CH1);
> val &= ~(CHV_BUFLEFTENA2_MASK | CHV_BUFRIGHTENA2_MASK);
> if (ch == DPIO_CH0)
> val |= CHV_BUFLEFTENA2_FORCE;
> if (ch == DPIO_CH1)
> val |= CHV_BUFRIGHTENA2_FORCE;
> - vlv_dpio_write(dev_priv, pipe, _CHV_CMN_DW1_CH1, val);
> + vlv_dpio_write(i915, pipe, _CHV_CMN_DW1_CH1, val);
> }
>
> /* program clock channel usage */
> - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS01_DW8(ch));
> + val = vlv_dpio_read(i915, pipe, VLV_PCS01_DW8(ch));
> val |= CHV_PCS_USEDCLKCHANNEL_OVRRIDE;
> if (pipe != PIPE_B)
> val &= ~CHV_PCS_USEDCLKCHANNEL;
> else
> val |= CHV_PCS_USEDCLKCHANNEL;
> - vlv_dpio_write(dev_priv, pipe, VLV_PCS01_DW8(ch), val);
> + vlv_dpio_write(i915, pipe, VLV_PCS01_DW8(ch), val);
>
> if (crtc_state->lane_count > 2) {
> - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS23_DW8(ch));
> + val = vlv_dpio_read(i915, pipe, VLV_PCS23_DW8(ch));
> val |= CHV_PCS_USEDCLKCHANNEL_OVRRIDE;
> if (pipe != PIPE_B)
> val &= ~CHV_PCS_USEDCLKCHANNEL;
> else
> val |= CHV_PCS_USEDCLKCHANNEL;
> - vlv_dpio_write(dev_priv, pipe, VLV_PCS23_DW8(ch), val);
> + vlv_dpio_write(i915, pipe, VLV_PCS23_DW8(ch), val);
> }
>
> /*
> @@ -849,14 +849,14 @@ void chv_phy_pre_pll_enable(struct intel_encoder *encoder,
> * matches the pipe, but here we need to
> * pick the CL based on the port.
> */
> - val = vlv_dpio_read(dev_priv, pipe, CHV_CMN_DW19(ch));
> + val = vlv_dpio_read(i915, pipe, CHV_CMN_DW19(ch));
> if (pipe != PIPE_B)
> val &= ~CHV_CMN_USEDCLKCHANNEL;
> else
> val |= CHV_CMN_USEDCLKCHANNEL;
> - vlv_dpio_write(dev_priv, pipe, CHV_CMN_DW19(ch), val);
> + vlv_dpio_write(i915, pipe, CHV_CMN_DW19(ch), val);
>
> - vlv_dpio_put(dev_priv);
> + vlv_dpio_put(i915);
> }
>
> void chv_phy_pre_encoder_enable(struct intel_encoder *encoder,
> @@ -864,24 +864,24 @@ void chv_phy_pre_encoder_enable(struct intel_encoder *encoder,
> {
> struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> struct intel_digital_port *dport = dp_to_dig_port(intel_dp);
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> enum dpio_channel ch = vlv_dport_to_channel(dport);
> enum pipe pipe = crtc->pipe;
> int data, i, stagger;
> u32 val;
>
> - vlv_dpio_get(dev_priv);
> + vlv_dpio_get(i915);
>
> /* allow hardware to manage TX FIFO reset source */
> - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS01_DW11(ch));
> + val = vlv_dpio_read(i915, pipe, VLV_PCS01_DW11(ch));
> val &= ~DPIO_LANEDESKEW_STRAP_OVRD;
> - vlv_dpio_write(dev_priv, pipe, VLV_PCS01_DW11(ch), val);
> + vlv_dpio_write(i915, pipe, VLV_PCS01_DW11(ch), val);
>
> if (crtc_state->lane_count > 2) {
> - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS23_DW11(ch));
> + val = vlv_dpio_read(i915, pipe, VLV_PCS23_DW11(ch));
> val &= ~DPIO_LANEDESKEW_STRAP_OVRD;
> - vlv_dpio_write(dev_priv, pipe, VLV_PCS23_DW11(ch), val);
> + vlv_dpio_write(i915, pipe, VLV_PCS23_DW11(ch), val);
> }
>
> /* Program Tx lane latency optimal setting*/
> @@ -891,7 +891,7 @@ void chv_phy_pre_encoder_enable(struct intel_encoder *encoder,
> data = 0x0;
> else
> data = (i == 1) ? 0x0 : 0x1;
> - vlv_dpio_write(dev_priv, pipe, CHV_TX_DW14(ch, i),
> + vlv_dpio_write(i915, pipe, CHV_TX_DW14(ch, i),
> data << DPIO_UPAR_SHIFT);
> }
>
> @@ -907,17 +907,17 @@ void chv_phy_pre_encoder_enable(struct intel_encoder *encoder,
> else
> stagger = 0x2;
>
> - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS01_DW11(ch));
> + val = vlv_dpio_read(i915, pipe, VLV_PCS01_DW11(ch));
> val |= DPIO_TX2_STAGGER_MASK(0x1f);
> - vlv_dpio_write(dev_priv, pipe, VLV_PCS01_DW11(ch), val);
> + vlv_dpio_write(i915, pipe, VLV_PCS01_DW11(ch), val);
>
> if (crtc_state->lane_count > 2) {
> - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS23_DW11(ch));
> + val = vlv_dpio_read(i915, pipe, VLV_PCS23_DW11(ch));
> val |= DPIO_TX2_STAGGER_MASK(0x1f);
> - vlv_dpio_write(dev_priv, pipe, VLV_PCS23_DW11(ch), val);
> + vlv_dpio_write(i915, pipe, VLV_PCS23_DW11(ch), val);
> }
>
> - vlv_dpio_write(dev_priv, pipe, VLV_PCS01_DW12(ch),
> + vlv_dpio_write(i915, pipe, VLV_PCS01_DW12(ch),
> DPIO_LANESTAGGER_STRAP(stagger) |
> DPIO_LANESTAGGER_STRAP_OVRD |
> DPIO_TX1_STAGGER_MASK(0x1f) |
> @@ -925,7 +925,7 @@ void chv_phy_pre_encoder_enable(struct intel_encoder *encoder,
> DPIO_TX2_STAGGER_MULT(0));
>
> if (crtc_state->lane_count > 2) {
> - vlv_dpio_write(dev_priv, pipe, VLV_PCS23_DW12(ch),
> + vlv_dpio_write(i915, pipe, VLV_PCS23_DW12(ch),
> DPIO_LANESTAGGER_STRAP(stagger) |
> DPIO_LANESTAGGER_STRAP_OVRD |
> DPIO_TX1_STAGGER_MASK(0x1f) |
> @@ -936,16 +936,16 @@ void chv_phy_pre_encoder_enable(struct intel_encoder *encoder,
> /* Deassert data lane reset */
> chv_data_lane_soft_reset(encoder, crtc_state, false);
>
> - vlv_dpio_put(dev_priv);
> + vlv_dpio_put(i915);
> }
>
> void chv_phy_release_cl2_override(struct intel_encoder *encoder)
> {
> struct intel_digital_port *dport = enc_to_dig_port(&encoder->base);
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
>
> if (dport->release_cl2_override) {
> - chv_phy_powergate_ch(dev_priv, DPIO_PHY0, DPIO_CH1, false);
> + chv_phy_powergate_ch(i915, DPIO_PHY0, DPIO_CH1, false);
> dport->release_cl2_override = false;
> }
> }
> @@ -953,24 +953,24 @@ void chv_phy_release_cl2_override(struct intel_encoder *encoder)
> void chv_phy_post_pll_disable(struct intel_encoder *encoder,
> const struct intel_crtc_state *old_crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> enum pipe pipe = to_intel_crtc(old_crtc_state->base.crtc)->pipe;
> u32 val;
>
> - vlv_dpio_get(dev_priv);
> + vlv_dpio_get(i915);
>
> /* disable left/right clock distribution */
> if (pipe != PIPE_B) {
> - val = vlv_dpio_read(dev_priv, pipe, _CHV_CMN_DW5_CH0);
> + val = vlv_dpio_read(i915, pipe, _CHV_CMN_DW5_CH0);
> val &= ~(CHV_BUFLEFTENA1_MASK | CHV_BUFRIGHTENA1_MASK);
> - vlv_dpio_write(dev_priv, pipe, _CHV_CMN_DW5_CH0, val);
> + vlv_dpio_write(i915, pipe, _CHV_CMN_DW5_CH0, val);
> } else {
> - val = vlv_dpio_read(dev_priv, pipe, _CHV_CMN_DW1_CH1);
> + val = vlv_dpio_read(i915, pipe, _CHV_CMN_DW1_CH1);
> val &= ~(CHV_BUFLEFTENA2_MASK | CHV_BUFRIGHTENA2_MASK);
> - vlv_dpio_write(dev_priv, pipe, _CHV_CMN_DW1_CH1, val);
> + vlv_dpio_write(i915, pipe, _CHV_CMN_DW1_CH1, val);
> }
>
> - vlv_dpio_put(dev_priv);
> + vlv_dpio_put(i915);
>
> /*
> * Leave the power down bit cleared for at least one
> @@ -988,57 +988,57 @@ void vlv_set_phy_signal_level(struct intel_encoder *encoder,
> u32 demph_reg_value, u32 preemph_reg_value,
> u32 uniqtranscale_reg_value, u32 tx3_demph)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(encoder->base.crtc);
> struct intel_digital_port *dport = enc_to_dig_port(&encoder->base);
> enum dpio_channel port = vlv_dport_to_channel(dport);
> enum pipe pipe = intel_crtc->pipe;
>
> - vlv_dpio_get(dev_priv);
> + vlv_dpio_get(i915);
>
> - vlv_dpio_write(dev_priv, pipe, VLV_TX_DW5(port), 0x00000000);
> - vlv_dpio_write(dev_priv, pipe, VLV_TX_DW4(port), demph_reg_value);
> - vlv_dpio_write(dev_priv, pipe, VLV_TX_DW2(port),
> + vlv_dpio_write(i915, pipe, VLV_TX_DW5(port), 0x00000000);
> + vlv_dpio_write(i915, pipe, VLV_TX_DW4(port), demph_reg_value);
> + vlv_dpio_write(i915, pipe, VLV_TX_DW2(port),
> uniqtranscale_reg_value);
> - vlv_dpio_write(dev_priv, pipe, VLV_TX_DW3(port), 0x0C782040);
> + vlv_dpio_write(i915, pipe, VLV_TX_DW3(port), 0x0C782040);
>
> if (tx3_demph)
> - vlv_dpio_write(dev_priv, pipe, VLV_TX3_DW4(port), tx3_demph);
> + vlv_dpio_write(i915, pipe, VLV_TX3_DW4(port), tx3_demph);
>
> - vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW11(port), 0x00030000);
> - vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW9(port), preemph_reg_value);
> - vlv_dpio_write(dev_priv, pipe, VLV_TX_DW5(port), DPIO_TX_OCALINIT_EN);
> + vlv_dpio_write(i915, pipe, VLV_PCS_DW11(port), 0x00030000);
> + vlv_dpio_write(i915, pipe, VLV_PCS_DW9(port), preemph_reg_value);
> + vlv_dpio_write(i915, pipe, VLV_TX_DW5(port), DPIO_TX_OCALINIT_EN);
>
> - vlv_dpio_put(dev_priv);
> + vlv_dpio_put(i915);
> }
>
> void vlv_phy_pre_pll_enable(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state)
> {
> struct intel_digital_port *dport = enc_to_dig_port(&encoder->base);
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> enum dpio_channel port = vlv_dport_to_channel(dport);
> enum pipe pipe = crtc->pipe;
>
> /* Program Tx lane resets to default */
> - vlv_dpio_get(dev_priv);
> + vlv_dpio_get(i915);
>
> - vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW0(port),
> + vlv_dpio_write(i915, pipe, VLV_PCS_DW0(port),
> DPIO_PCS_TX_LANE2_RESET |
> DPIO_PCS_TX_LANE1_RESET);
> - vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW1(port),
> + vlv_dpio_write(i915, pipe, VLV_PCS_DW1(port),
> DPIO_PCS_CLK_CRI_RXEB_EIOS_EN |
> DPIO_PCS_CLK_CRI_RXDIGFILTSG_EN |
> (1<<DPIO_PCS_CLK_DATAWIDTH_SHIFT) |
> DPIO_PCS_CLK_SOFT_RESET);
>
> /* Fix up inter-pair skew failure */
> - vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW12(port), 0x00750f00);
> - vlv_dpio_write(dev_priv, pipe, VLV_TX_DW11(port), 0x00001500);
> - vlv_dpio_write(dev_priv, pipe, VLV_TX_DW14(port), 0x40400000);
> + vlv_dpio_write(i915, pipe, VLV_PCS_DW12(port), 0x00750f00);
> + vlv_dpio_write(i915, pipe, VLV_TX_DW11(port), 0x00001500);
> + vlv_dpio_write(i915, pipe, VLV_TX_DW14(port), 0x40400000);
>
> - vlv_dpio_put(dev_priv);
> + vlv_dpio_put(i915);
> }
>
> void vlv_phy_pre_encoder_enable(struct intel_encoder *encoder,
> @@ -1046,42 +1046,42 @@ void vlv_phy_pre_encoder_enable(struct intel_encoder *encoder,
> {
> struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> struct intel_digital_port *dport = dp_to_dig_port(intel_dp);
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> enum dpio_channel port = vlv_dport_to_channel(dport);
> enum pipe pipe = crtc->pipe;
> u32 val;
>
> - vlv_dpio_get(dev_priv);
> + vlv_dpio_get(i915);
>
> /* Enable clock channels for this port */
> - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS01_DW8(port));
> + val = vlv_dpio_read(i915, pipe, VLV_PCS01_DW8(port));
> val = 0;
> if (pipe)
> val |= (1<<21);
> else
> val &= ~(1<<21);
> val |= 0x001000c4;
> - vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW8(port), val);
> + vlv_dpio_write(i915, pipe, VLV_PCS_DW8(port), val);
>
> /* Program lane clock */
> - vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW14(port), 0x00760018);
> - vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW23(port), 0x00400888);
> + vlv_dpio_write(i915, pipe, VLV_PCS_DW14(port), 0x00760018);
> + vlv_dpio_write(i915, pipe, VLV_PCS_DW23(port), 0x00400888);
>
> - vlv_dpio_put(dev_priv);
> + vlv_dpio_put(i915);
> }
>
> void vlv_phy_reset_lanes(struct intel_encoder *encoder,
> const struct intel_crtc_state *old_crtc_state)
> {
> struct intel_digital_port *dport = enc_to_dig_port(&encoder->base);
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc);
> enum dpio_channel port = vlv_dport_to_channel(dport);
> enum pipe pipe = crtc->pipe;
>
> - vlv_dpio_get(dev_priv);
> - vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW0(port), 0x00000000);
> - vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW1(port), 0x00e00060);
> - vlv_dpio_put(dev_priv);
> + vlv_dpio_get(i915);
> + vlv_dpio_write(i915, pipe, VLV_PCS_DW0(port), 0x00000000);
> + vlv_dpio_write(i915, pipe, VLV_PCS_DW1(port), 0x00e00060);
> + vlv_dpio_put(i915);
> }
> diff --git a/drivers/gpu/drm/i915/intel_dpio_phy.h b/drivers/gpu/drm/i915/intel_dpio_phy.h
> index f418aab90b7e..d4c9393bebce 100644
> --- a/drivers/gpu/drm/i915/intel_dpio_phy.h
> +++ b/drivers/gpu/drm/i915/intel_dpio_phy.h
> @@ -15,16 +15,16 @@ struct drm_i915_private;
> struct intel_crtc_state;
> struct intel_encoder;
>
> -void bxt_port_to_phy_channel(struct drm_i915_private *dev_priv, enum port port,
> +void bxt_port_to_phy_channel(struct drm_i915_private *i915, enum port port,
> enum dpio_phy *phy, enum dpio_channel *ch);
> -void bxt_ddi_phy_set_signal_level(struct drm_i915_private *dev_priv,
> +void bxt_ddi_phy_set_signal_level(struct drm_i915_private *i915,
> enum port port, u32 margin, u32 scale,
> u32 enable, u32 deemphasis);
> -void bxt_ddi_phy_init(struct drm_i915_private *dev_priv, enum dpio_phy phy);
> -void bxt_ddi_phy_uninit(struct drm_i915_private *dev_priv, enum dpio_phy phy);
> -bool bxt_ddi_phy_is_enabled(struct drm_i915_private *dev_priv,
> +void bxt_ddi_phy_init(struct drm_i915_private *i915, enum dpio_phy phy);
> +void bxt_ddi_phy_uninit(struct drm_i915_private *i915, enum dpio_phy phy);
> +bool bxt_ddi_phy_is_enabled(struct drm_i915_private *i915,
> enum dpio_phy phy);
> -bool bxt_ddi_phy_verify_state(struct drm_i915_private *dev_priv,
> +bool bxt_ddi_phy_verify_state(struct drm_i915_private *i915,
> enum dpio_phy phy);
> u8 bxt_ddi_phy_calc_lane_lat_optim_mask(u8 lane_count);
> void bxt_ddi_phy_set_lane_optim_mask(struct intel_encoder *encoder,
> diff --git a/drivers/gpu/drm/i915/intel_dpll_mgr.c b/drivers/gpu/drm/i915/intel_dpll_mgr.c
> index 2d4e7b9a7b9d..2dabcdf27df4 100644
> --- a/drivers/gpu/drm/i915/intel_dpll_mgr.c
> +++ b/drivers/gpu/drm/i915/intel_dpll_mgr.c
> @@ -45,14 +45,14 @@
> */
>
> static void
> -intel_atomic_duplicate_dpll_state(struct drm_i915_private *dev_priv,
> +intel_atomic_duplicate_dpll_state(struct drm_i915_private *i915,
> struct intel_shared_dpll_state *shared_dpll)
> {
> enum intel_dpll_id i;
>
> /* Copy shared dpll state */
> - for (i = 0; i < dev_priv->num_shared_dpll; i++) {
> - struct intel_shared_dpll *pll = &dev_priv->shared_dplls[i];
> + for (i = 0; i < i915->num_shared_dpll; i++) {
> + struct intel_shared_dpll *pll = &i915->shared_dplls[i];
>
> shared_dpll[i] = pll->state;
> }
> @@ -77,40 +77,40 @@ intel_atomic_get_shared_dpll_state(struct drm_atomic_state *s)
>
> /**
> * intel_get_shared_dpll_by_id - get a DPLL given its id
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> * @id: pll id
> *
> * Returns:
> * A pointer to the DPLL with @id
> */
> struct intel_shared_dpll *
> -intel_get_shared_dpll_by_id(struct drm_i915_private *dev_priv,
> +intel_get_shared_dpll_by_id(struct drm_i915_private *i915,
> enum intel_dpll_id id)
> {
> - return &dev_priv->shared_dplls[id];
> + return &i915->shared_dplls[id];
> }
>
> /**
> * intel_get_shared_dpll_id - get the id of a DPLL
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> * @pll: the DPLL
> *
> * Returns:
> * The id of @pll
> */
> enum intel_dpll_id
> -intel_get_shared_dpll_id(struct drm_i915_private *dev_priv,
> +intel_get_shared_dpll_id(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll)
> {
> - if (WARN_ON(pll < dev_priv->shared_dplls||
> - pll > &dev_priv->shared_dplls[dev_priv->num_shared_dpll]))
> + if (WARN_ON(pll < i915->shared_dplls||
> + pll > &i915->shared_dplls[i915->num_shared_dpll]))
> return -1;
>
> - return (enum intel_dpll_id) (pll - dev_priv->shared_dplls);
> + return (enum intel_dpll_id) (pll - i915->shared_dplls);
> }
>
> /* For ILK+ */
> -void assert_shared_dpll(struct drm_i915_private *dev_priv,
> +void assert_shared_dpll(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll,
> bool state)
> {
> @@ -120,7 +120,7 @@ void assert_shared_dpll(struct drm_i915_private *dev_priv,
> if (WARN(!pll, "asserting DPLL %s with no DPLL\n", onoff(state)))
> return;
>
> - cur_state = pll->info->funcs->get_hw_state(dev_priv, pll, &hw_state);
> + cur_state = pll->info->funcs->get_hw_state(i915, pll, &hw_state);
> I915_STATE_WARN(cur_state != state,
> "%s assertion failure (expected %s, current %s)\n",
> pll->info->name, onoff(state), onoff(cur_state));
> @@ -136,22 +136,22 @@ void assert_shared_dpll(struct drm_i915_private *dev_priv,
> void intel_prepare_shared_dpll(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> struct intel_shared_dpll *pll = crtc_state->shared_dpll;
>
> if (WARN_ON(pll == NULL))
> return;
>
> - mutex_lock(&dev_priv->dpll_lock);
> + mutex_lock(&i915->dpll_lock);
> WARN_ON(!pll->state.crtc_mask);
> if (!pll->active_mask) {
> DRM_DEBUG_DRIVER("setting up %s\n", pll->info->name);
> WARN_ON(pll->on);
> - assert_shared_dpll_disabled(dev_priv, pll);
> + assert_shared_dpll_disabled(i915, pll);
>
> - pll->info->funcs->prepare(dev_priv, pll);
> + pll->info->funcs->prepare(i915, pll);
> }
> - mutex_unlock(&dev_priv->dpll_lock);
> + mutex_unlock(&i915->dpll_lock);
> }
>
> /**
> @@ -163,7 +163,7 @@ void intel_prepare_shared_dpll(const struct intel_crtc_state *crtc_state)
> void intel_enable_shared_dpll(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> struct intel_shared_dpll *pll = crtc_state->shared_dpll;
> unsigned int crtc_mask = drm_crtc_mask(&crtc->base);
> unsigned int old_mask;
> @@ -171,7 +171,7 @@ void intel_enable_shared_dpll(const struct intel_crtc_state *crtc_state)
> if (WARN_ON(pll == NULL))
> return;
>
> - mutex_lock(&dev_priv->dpll_lock);
> + mutex_lock(&i915->dpll_lock);
> old_mask = pll->active_mask;
>
> if (WARN_ON(!(pll->state.crtc_mask & crtc_mask)) ||
> @@ -186,17 +186,17 @@ void intel_enable_shared_dpll(const struct intel_crtc_state *crtc_state)
>
> if (old_mask) {
> WARN_ON(!pll->on);
> - assert_shared_dpll_enabled(dev_priv, pll);
> + assert_shared_dpll_enabled(i915, pll);
> goto out;
> }
> WARN_ON(pll->on);
>
> DRM_DEBUG_KMS("enabling %s\n", pll->info->name);
> - pll->info->funcs->enable(dev_priv, pll);
> + pll->info->funcs->enable(i915, pll);
> pll->on = true;
>
> out:
> - mutex_unlock(&dev_priv->dpll_lock);
> + mutex_unlock(&i915->dpll_lock);
> }
>
> /**
> @@ -208,18 +208,18 @@ void intel_enable_shared_dpll(const struct intel_crtc_state *crtc_state)
> void intel_disable_shared_dpll(const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> struct intel_shared_dpll *pll = crtc_state->shared_dpll;
> unsigned int crtc_mask = drm_crtc_mask(&crtc->base);
>
> /* PCH only available on ILK+ */
> - if (INTEL_GEN(dev_priv) < 5)
> + if (INTEL_GEN(i915) < 5)
> return;
>
> if (pll == NULL)
> return;
>
> - mutex_lock(&dev_priv->dpll_lock);
> + mutex_lock(&i915->dpll_lock);
> if (WARN_ON(!(pll->active_mask & crtc_mask)))
> goto out;
>
> @@ -227,7 +227,7 @@ void intel_disable_shared_dpll(const struct intel_crtc_state *crtc_state)
> pll->info->name, pll->active_mask, pll->on,
> crtc->base.base.id);
>
> - assert_shared_dpll_enabled(dev_priv, pll);
> + assert_shared_dpll_enabled(i915, pll);
> WARN_ON(!pll->on);
>
> pll->active_mask &= ~crtc_mask;
> @@ -235,11 +235,11 @@ void intel_disable_shared_dpll(const struct intel_crtc_state *crtc_state)
> goto out;
>
> DRM_DEBUG_KMS("disabling %s\n", pll->info->name);
> - pll->info->funcs->disable(dev_priv, pll);
> + pll->info->funcs->disable(i915, pll);
> pll->on = false;
>
> out:
> - mutex_unlock(&dev_priv->dpll_lock);
> + mutex_unlock(&i915->dpll_lock);
> }
>
> static struct intel_shared_dpll *
> @@ -248,7 +248,7 @@ intel_find_shared_dpll(struct intel_crtc_state *crtc_state,
> enum intel_dpll_id range_max)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> struct intel_shared_dpll *pll, *unused_pll = NULL;
> struct intel_shared_dpll_state *shared_dpll;
> enum intel_dpll_id i;
> @@ -256,7 +256,7 @@ intel_find_shared_dpll(struct intel_crtc_state *crtc_state,
> shared_dpll = intel_atomic_get_shared_dpll_state(crtc_state->base.state);
>
> for (i = range_min; i <= range_max; i++) {
> - pll = &dev_priv->shared_dplls[i];
> + pll = &i915->shared_dplls[i];
>
> /* Only want to check enabled timings first */
> if (shared_dpll[i].crtc_mask == 0) {
> @@ -322,7 +322,7 @@ intel_reference_shared_dpll(struct intel_shared_dpll *pll,
> */
> void intel_shared_dpll_swap_state(struct drm_atomic_state *state)
> {
> - struct drm_i915_private *dev_priv = to_i915(state->dev);
> + struct drm_i915_private *i915 = to_i915(state->dev);
> struct intel_shared_dpll_state *shared_dpll;
> struct intel_shared_dpll *pll;
> enum intel_dpll_id i;
> @@ -331,10 +331,10 @@ void intel_shared_dpll_swap_state(struct drm_atomic_state *state)
> return;
>
> shared_dpll = to_intel_atomic_state(state)->shared_dpll;
> - for (i = 0; i < dev_priv->num_shared_dpll; i++) {
> + for (i = 0; i < i915->num_shared_dpll; i++) {
> struct intel_shared_dpll_state tmp;
>
> - pll = &dev_priv->shared_dplls[i];
> + pll = &i915->shared_dplls[i];
>
> tmp = pll->state;
> pll->state = shared_dpll[i];
> @@ -342,7 +342,7 @@ void intel_shared_dpll_swap_state(struct drm_atomic_state *state)
> }
> }
>
> -static bool ibx_pch_dpll_get_hw_state(struct drm_i915_private *dev_priv,
> +static bool ibx_pch_dpll_get_hw_state(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll,
> struct intel_dpll_hw_state *hw_state)
> {
> @@ -350,7 +350,7 @@ static bool ibx_pch_dpll_get_hw_state(struct drm_i915_private *dev_priv,
> intel_wakeref_t wakeref;
> u32 val;
>
> - wakeref = intel_display_power_get_if_enabled(dev_priv,
> + wakeref = intel_display_power_get_if_enabled(i915,
> POWER_DOMAIN_DISPLAY_CORE);
> if (!wakeref)
> return false;
> @@ -360,12 +360,12 @@ static bool ibx_pch_dpll_get_hw_state(struct drm_i915_private *dev_priv,
> hw_state->fp0 = I915_READ(PCH_FP0(id));
> hw_state->fp1 = I915_READ(PCH_FP1(id));
>
> - intel_display_power_put(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> + intel_display_power_put(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref);
>
> return val & DPLL_VCO_ENABLE;
> }
>
> -static void ibx_pch_dpll_prepare(struct drm_i915_private *dev_priv,
> +static void ibx_pch_dpll_prepare(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll)
> {
> const enum intel_dpll_id id = pll->info->id;
> @@ -374,12 +374,12 @@ static void ibx_pch_dpll_prepare(struct drm_i915_private *dev_priv,
> I915_WRITE(PCH_FP1(id), pll->state.hw_state.fp1);
> }
>
> -static void ibx_assert_pch_refclk_enabled(struct drm_i915_private *dev_priv)
> +static void ibx_assert_pch_refclk_enabled(struct drm_i915_private *i915)
> {
> u32 val;
> bool enabled;
>
> - I915_STATE_WARN_ON(!(HAS_PCH_IBX(dev_priv) || HAS_PCH_CPT(dev_priv)));
> + I915_STATE_WARN_ON(!(HAS_PCH_IBX(i915) || HAS_PCH_CPT(i915)));
>
> val = I915_READ(PCH_DREF_CONTROL);
> enabled = !!(val & (DREF_SSC_SOURCE_MASK | DREF_NONSPREAD_SOURCE_MASK |
> @@ -387,13 +387,13 @@ static void ibx_assert_pch_refclk_enabled(struct drm_i915_private *dev_priv)
> I915_STATE_WARN(!enabled, "PCH refclk assertion failure, should be active but is disabled\n");
> }
>
> -static void ibx_pch_dpll_enable(struct drm_i915_private *dev_priv,
> +static void ibx_pch_dpll_enable(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll)
> {
> const enum intel_dpll_id id = pll->info->id;
>
> /* PCH refclock must be enabled first */
> - ibx_assert_pch_refclk_enabled(dev_priv);
> + ibx_assert_pch_refclk_enabled(i915);
>
> I915_WRITE(PCH_DPLL(id), pll->state.hw_state.dpll);
>
> @@ -411,7 +411,7 @@ static void ibx_pch_dpll_enable(struct drm_i915_private *dev_priv,
> udelay(200);
> }
>
> -static void ibx_pch_dpll_disable(struct drm_i915_private *dev_priv,
> +static void ibx_pch_dpll_disable(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll)
> {
> const enum intel_dpll_id id = pll->info->id;
> @@ -426,14 +426,14 @@ ibx_get_dpll(struct intel_crtc_state *crtc_state,
> struct intel_encoder *encoder)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> struct intel_shared_dpll *pll;
> enum intel_dpll_id i;
>
> - if (HAS_PCH_IBX(dev_priv)) {
> + if (HAS_PCH_IBX(i915)) {
> /* Ironlake PCH has a fixed PLL->PCH pipe mapping. */
> i = (enum intel_dpll_id) crtc->pipe;
> - pll = &dev_priv->shared_dplls[i];
> + pll = &i915->shared_dplls[i];
>
> DRM_DEBUG_KMS("[CRTC:%d:%s] using pre-allocated %s\n",
> crtc->base.base.id, crtc->base.name,
> @@ -453,7 +453,7 @@ ibx_get_dpll(struct intel_crtc_state *crtc_state,
> return pll;
> }
>
> -static void ibx_dump_hw_state(struct drm_i915_private *dev_priv,
> +static void ibx_dump_hw_state(struct drm_i915_private *i915,
> const struct intel_dpll_hw_state *hw_state)
> {
> DRM_DEBUG_KMS("dpll_hw_state: dpll: 0x%x, dpll_md: 0x%x, "
> @@ -471,7 +471,7 @@ static const struct intel_shared_dpll_funcs ibx_pch_dpll_funcs = {
> .get_hw_state = ibx_pch_dpll_get_hw_state,
> };
>
> -static void hsw_ddi_wrpll_enable(struct drm_i915_private *dev_priv,
> +static void hsw_ddi_wrpll_enable(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll)
> {
> const enum intel_dpll_id id = pll->info->id;
> @@ -481,7 +481,7 @@ static void hsw_ddi_wrpll_enable(struct drm_i915_private *dev_priv,
> udelay(20);
> }
>
> -static void hsw_ddi_spll_enable(struct drm_i915_private *dev_priv,
> +static void hsw_ddi_spll_enable(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll)
> {
> I915_WRITE(SPLL_CTL, pll->state.hw_state.spll);
> @@ -489,7 +489,7 @@ static void hsw_ddi_spll_enable(struct drm_i915_private *dev_priv,
> udelay(20);
> }
>
> -static void hsw_ddi_wrpll_disable(struct drm_i915_private *dev_priv,
> +static void hsw_ddi_wrpll_disable(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll)
> {
> const enum intel_dpll_id id = pll->info->id;
> @@ -500,7 +500,7 @@ static void hsw_ddi_wrpll_disable(struct drm_i915_private *dev_priv,
> POSTING_READ(WRPLL_CTL(id));
> }
>
> -static void hsw_ddi_spll_disable(struct drm_i915_private *dev_priv,
> +static void hsw_ddi_spll_disable(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll)
> {
> u32 val;
> @@ -510,7 +510,7 @@ static void hsw_ddi_spll_disable(struct drm_i915_private *dev_priv,
> POSTING_READ(SPLL_CTL);
> }
>
> -static bool hsw_ddi_wrpll_get_hw_state(struct drm_i915_private *dev_priv,
> +static bool hsw_ddi_wrpll_get_hw_state(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll,
> struct intel_dpll_hw_state *hw_state)
> {
> @@ -518,7 +518,7 @@ static bool hsw_ddi_wrpll_get_hw_state(struct drm_i915_private *dev_priv,
> intel_wakeref_t wakeref;
> u32 val;
>
> - wakeref = intel_display_power_get_if_enabled(dev_priv,
> + wakeref = intel_display_power_get_if_enabled(i915,
> POWER_DOMAIN_DISPLAY_CORE);
> if (!wakeref)
> return false;
> @@ -526,19 +526,19 @@ static bool hsw_ddi_wrpll_get_hw_state(struct drm_i915_private *dev_priv,
> val = I915_READ(WRPLL_CTL(id));
> hw_state->wrpll = val;
>
> - intel_display_power_put(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> + intel_display_power_put(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref);
>
> return val & WRPLL_PLL_ENABLE;
> }
>
> -static bool hsw_ddi_spll_get_hw_state(struct drm_i915_private *dev_priv,
> +static bool hsw_ddi_spll_get_hw_state(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll,
> struct intel_dpll_hw_state *hw_state)
> {
> intel_wakeref_t wakeref;
> u32 val;
>
> - wakeref = intel_display_power_get_if_enabled(dev_priv,
> + wakeref = intel_display_power_get_if_enabled(i915,
> POWER_DOMAIN_DISPLAY_CORE);
> if (!wakeref)
> return false;
> @@ -546,7 +546,7 @@ static bool hsw_ddi_spll_get_hw_state(struct drm_i915_private *dev_priv,
> val = I915_READ(SPLL_CTL);
> hw_state->spll = val;
>
> - intel_display_power_put(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> + intel_display_power_put(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref);
>
> return val & SPLL_PLL_ENABLE;
> }
> @@ -793,7 +793,7 @@ static struct intel_shared_dpll *hsw_ddi_hdmi_get_dpll(struct intel_crtc_state *
> static struct intel_shared_dpll *
> hsw_ddi_dp_get_dpll(struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> struct intel_shared_dpll *pll;
> enum intel_dpll_id pll_id;
> int clock = crtc_state->port_clock;
> @@ -813,7 +813,7 @@ hsw_ddi_dp_get_dpll(struct intel_crtc_state *crtc_state)
> return NULL;
> }
>
> - pll = intel_get_shared_dpll_by_id(dev_priv, pll_id);
> + pll = intel_get_shared_dpll_by_id(i915, pll_id);
>
> if (!pll)
> return NULL;
> @@ -855,7 +855,7 @@ hsw_get_dpll(struct intel_crtc_state *crtc_state,
> return pll;
> }
>
> -static void hsw_dump_hw_state(struct drm_i915_private *dev_priv,
> +static void hsw_dump_hw_state(struct drm_i915_private *i915,
> const struct intel_dpll_hw_state *hw_state)
> {
> DRM_DEBUG_KMS("dpll_hw_state: wrpll: 0x%x spll: 0x%x\n",
> @@ -874,17 +874,17 @@ static const struct intel_shared_dpll_funcs hsw_ddi_spll_funcs = {
> .get_hw_state = hsw_ddi_spll_get_hw_state,
> };
>
> -static void hsw_ddi_lcpll_enable(struct drm_i915_private *dev_priv,
> +static void hsw_ddi_lcpll_enable(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll)
> {
> }
>
> -static void hsw_ddi_lcpll_disable(struct drm_i915_private *dev_priv,
> +static void hsw_ddi_lcpll_disable(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll)
> {
> }
>
> -static bool hsw_ddi_lcpll_get_hw_state(struct drm_i915_private *dev_priv,
> +static bool hsw_ddi_lcpll_get_hw_state(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll,
> struct intel_dpll_hw_state *hw_state)
> {
> @@ -928,7 +928,7 @@ static const struct skl_dpll_regs skl_dpll_regs[4] = {
> },
> };
>
> -static void skl_ddi_pll_write_ctrl1(struct drm_i915_private *dev_priv,
> +static void skl_ddi_pll_write_ctrl1(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll)
> {
> const enum intel_dpll_id id = pll->info->id;
> @@ -945,13 +945,13 @@ static void skl_ddi_pll_write_ctrl1(struct drm_i915_private *dev_priv,
> POSTING_READ(DPLL_CTRL1);
> }
>
> -static void skl_ddi_pll_enable(struct drm_i915_private *dev_priv,
> +static void skl_ddi_pll_enable(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll)
> {
> const struct skl_dpll_regs *regs = skl_dpll_regs;
> const enum intel_dpll_id id = pll->info->id;
>
> - skl_ddi_pll_write_ctrl1(dev_priv, pll);
> + skl_ddi_pll_write_ctrl1(i915, pll);
>
> I915_WRITE(regs[id].cfgcr1, pll->state.hw_state.cfgcr1);
> I915_WRITE(regs[id].cfgcr2, pll->state.hw_state.cfgcr2);
> @@ -962,7 +962,7 @@ static void skl_ddi_pll_enable(struct drm_i915_private *dev_priv,
> I915_WRITE(regs[id].ctl,
> I915_READ(regs[id].ctl) | LCPLL_PLL_ENABLE);
>
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> DPLL_STATUS,
> DPLL_LOCK(id),
> DPLL_LOCK(id),
> @@ -970,13 +970,13 @@ static void skl_ddi_pll_enable(struct drm_i915_private *dev_priv,
> DRM_ERROR("DPLL %d not locked\n", id);
> }
>
> -static void skl_ddi_dpll0_enable(struct drm_i915_private *dev_priv,
> +static void skl_ddi_dpll0_enable(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll)
> {
> - skl_ddi_pll_write_ctrl1(dev_priv, pll);
> + skl_ddi_pll_write_ctrl1(i915, pll);
> }
>
> -static void skl_ddi_pll_disable(struct drm_i915_private *dev_priv,
> +static void skl_ddi_pll_disable(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll)
> {
> const struct skl_dpll_regs *regs = skl_dpll_regs;
> @@ -988,12 +988,12 @@ static void skl_ddi_pll_disable(struct drm_i915_private *dev_priv,
> POSTING_READ(regs[id].ctl);
> }
>
> -static void skl_ddi_dpll0_disable(struct drm_i915_private *dev_priv,
> +static void skl_ddi_dpll0_disable(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll)
> {
> }
>
> -static bool skl_ddi_pll_get_hw_state(struct drm_i915_private *dev_priv,
> +static bool skl_ddi_pll_get_hw_state(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll,
> struct intel_dpll_hw_state *hw_state)
> {
> @@ -1003,7 +1003,7 @@ static bool skl_ddi_pll_get_hw_state(struct drm_i915_private *dev_priv,
> intel_wakeref_t wakeref;
> bool ret;
>
> - wakeref = intel_display_power_get_if_enabled(dev_priv,
> + wakeref = intel_display_power_get_if_enabled(i915,
> POWER_DOMAIN_DISPLAY_CORE);
> if (!wakeref)
> return false;
> @@ -1025,12 +1025,12 @@ static bool skl_ddi_pll_get_hw_state(struct drm_i915_private *dev_priv,
> ret = true;
>
> out:
> - intel_display_power_put(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> + intel_display_power_put(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref);
>
> return ret;
> }
>
> -static bool skl_ddi_dpll0_get_hw_state(struct drm_i915_private *dev_priv,
> +static bool skl_ddi_dpll0_get_hw_state(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll,
> struct intel_dpll_hw_state *hw_state)
> {
> @@ -1040,7 +1040,7 @@ static bool skl_ddi_dpll0_get_hw_state(struct drm_i915_private *dev_priv,
> u32 val;
> bool ret;
>
> - wakeref = intel_display_power_get_if_enabled(dev_priv,
> + wakeref = intel_display_power_get_if_enabled(i915,
> POWER_DOMAIN_DISPLAY_CORE);
> if (!wakeref)
> return false;
> @@ -1058,7 +1058,7 @@ static bool skl_ddi_dpll0_get_hw_state(struct drm_i915_private *dev_priv,
> ret = true;
>
> out:
> - intel_display_power_put(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> + intel_display_power_put(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref);
>
> return ret;
> }
> @@ -1424,7 +1424,7 @@ skl_get_dpll(struct intel_crtc_state *crtc_state,
> return pll;
> }
>
> -static void skl_dump_hw_state(struct drm_i915_private *dev_priv,
> +static void skl_dump_hw_state(struct drm_i915_private *i915,
> const struct intel_dpll_hw_state *hw_state)
> {
> DRM_DEBUG_KMS("dpll_hw_state: "
> @@ -1446,7 +1446,7 @@ static const struct intel_shared_dpll_funcs skl_ddi_dpll0_funcs = {
> .get_hw_state = skl_ddi_dpll0_get_hw_state,
> };
>
> -static void bxt_ddi_pll_enable(struct drm_i915_private *dev_priv,
> +static void bxt_ddi_pll_enable(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll)
> {
> u32 temp;
> @@ -1454,14 +1454,14 @@ static void bxt_ddi_pll_enable(struct drm_i915_private *dev_priv,
> enum dpio_phy phy;
> enum dpio_channel ch;
>
> - bxt_port_to_phy_channel(dev_priv, port, &phy, &ch);
> + bxt_port_to_phy_channel(i915, port, &phy, &ch);
>
> /* Non-SSC reference */
> temp = I915_READ(BXT_PORT_PLL_ENABLE(port));
> temp |= PORT_PLL_REF_SEL;
> I915_WRITE(BXT_PORT_PLL_ENABLE(port), temp);
>
> - if (IS_GEMINILAKE(dev_priv)) {
> + if (IS_GEMINILAKE(i915)) {
> temp = I915_READ(BXT_PORT_PLL_ENABLE(port));
> temp |= PORT_PLL_POWER_ENABLE;
> I915_WRITE(BXT_PORT_PLL_ENABLE(port), temp);
> @@ -1549,7 +1549,7 @@ static void bxt_ddi_pll_enable(struct drm_i915_private *dev_priv,
> 200))
> DRM_ERROR("PLL %d not locked\n", port);
>
> - if (IS_GEMINILAKE(dev_priv)) {
> + if (IS_GEMINILAKE(i915)) {
> temp = I915_READ(BXT_PORT_TX_DW5_LN0(phy, ch));
> temp |= DCC_DELAY_RANGE_2;
> I915_WRITE(BXT_PORT_TX_DW5_GRP(phy, ch), temp);
> @@ -1566,7 +1566,7 @@ static void bxt_ddi_pll_enable(struct drm_i915_private *dev_priv,
> I915_WRITE(BXT_PORT_PCS_DW12_GRP(phy, ch), temp);
> }
>
> -static void bxt_ddi_pll_disable(struct drm_i915_private *dev_priv,
> +static void bxt_ddi_pll_disable(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll)
> {
> enum port port = (enum port)pll->info->id; /* 1:1 port->PLL mapping */
> @@ -1577,7 +1577,7 @@ static void bxt_ddi_pll_disable(struct drm_i915_private *dev_priv,
> I915_WRITE(BXT_PORT_PLL_ENABLE(port), temp);
> POSTING_READ(BXT_PORT_PLL_ENABLE(port));
>
> - if (IS_GEMINILAKE(dev_priv)) {
> + if (IS_GEMINILAKE(i915)) {
> temp = I915_READ(BXT_PORT_PLL_ENABLE(port));
> temp &= ~PORT_PLL_POWER_ENABLE;
> I915_WRITE(BXT_PORT_PLL_ENABLE(port), temp);
> @@ -1588,7 +1588,7 @@ static void bxt_ddi_pll_disable(struct drm_i915_private *dev_priv,
> }
> }
>
> -static bool bxt_ddi_pll_get_hw_state(struct drm_i915_private *dev_priv,
> +static bool bxt_ddi_pll_get_hw_state(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll,
> struct intel_dpll_hw_state *hw_state)
> {
> @@ -1599,9 +1599,9 @@ static bool bxt_ddi_pll_get_hw_state(struct drm_i915_private *dev_priv,
> u32 val;
> bool ret;
>
> - bxt_port_to_phy_channel(dev_priv, port, &phy, &ch);
> + bxt_port_to_phy_channel(i915, port, &phy, &ch);
>
> - wakeref = intel_display_power_get_if_enabled(dev_priv,
> + wakeref = intel_display_power_get_if_enabled(i915,
> POWER_DOMAIN_DISPLAY_CORE);
> if (!wakeref)
> return false;
> @@ -1660,7 +1660,7 @@ static bool bxt_ddi_pll_get_hw_state(struct drm_i915_private *dev_priv,
> ret = true;
>
> out:
> - intel_display_power_put(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> + intel_display_power_put(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref);
>
> return ret;
> }
> @@ -1832,7 +1832,7 @@ bxt_get_dpll(struct intel_crtc_state *crtc_state,
> struct intel_encoder *encoder)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> struct intel_shared_dpll *pll;
> enum intel_dpll_id id;
>
> @@ -1846,7 +1846,7 @@ bxt_get_dpll(struct intel_crtc_state *crtc_state,
>
> /* 1:1 mapping between ports and PLLs */
> id = (enum intel_dpll_id) encoder->port;
> - pll = intel_get_shared_dpll_by_id(dev_priv, id);
> + pll = intel_get_shared_dpll_by_id(i915, id);
>
> DRM_DEBUG_KMS("[CRTC:%d:%s] using pre-allocated %s\n",
> crtc->base.base.id, crtc->base.name, pll->info->name);
> @@ -1856,7 +1856,7 @@ bxt_get_dpll(struct intel_crtc_state *crtc_state,
> return pll;
> }
>
> -static void bxt_dump_hw_state(struct drm_i915_private *dev_priv,
> +static void bxt_dump_hw_state(struct drm_i915_private *i915,
> const struct intel_dpll_hw_state *hw_state)
> {
> DRM_DEBUG_KMS("dpll_hw_state: ebb0: 0x%x, ebb4: 0x%x,"
> @@ -1887,7 +1887,7 @@ struct intel_dpll_mgr {
> struct intel_shared_dpll *(*get_dpll)(struct intel_crtc_state *crtc_state,
> struct intel_encoder *encoder);
>
> - void (*dump_hw_state)(struct drm_i915_private *dev_priv,
> + void (*dump_hw_state)(struct drm_i915_private *i915,
> const struct intel_dpll_hw_state *hw_state);
> };
>
> @@ -1946,7 +1946,7 @@ static const struct intel_dpll_mgr bxt_pll_mgr = {
> .dump_hw_state = bxt_dump_hw_state,
> };
>
> -static void cnl_ddi_pll_enable(struct drm_i915_private *dev_priv,
> +static void cnl_ddi_pll_enable(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll)
> {
> const enum intel_dpll_id id = pll->info->id;
> @@ -1958,7 +1958,7 @@ static void cnl_ddi_pll_enable(struct drm_i915_private *dev_priv,
> I915_WRITE(CNL_DPLL_ENABLE(id), val);
>
> /* 2. Wait for DPLL power state enabled in DPLL_ENABLE. */
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> CNL_DPLL_ENABLE(id),
> PLL_POWER_STATE,
> PLL_POWER_STATE,
> @@ -1999,7 +1999,7 @@ static void cnl_ddi_pll_enable(struct drm_i915_private *dev_priv,
> I915_WRITE(CNL_DPLL_ENABLE(id), val);
>
> /* 7. Wait for PLL lock status in DPLL_ENABLE. */
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> CNL_DPLL_ENABLE(id),
> PLL_LOCK,
> PLL_LOCK,
> @@ -2021,7 +2021,7 @@ static void cnl_ddi_pll_enable(struct drm_i915_private *dev_priv,
> */
> }
>
> -static void cnl_ddi_pll_disable(struct drm_i915_private *dev_priv,
> +static void cnl_ddi_pll_disable(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll)
> {
> const enum intel_dpll_id id = pll->info->id;
> @@ -2047,7 +2047,7 @@ static void cnl_ddi_pll_disable(struct drm_i915_private *dev_priv,
> I915_WRITE(CNL_DPLL_ENABLE(id), val);
>
> /* 4. Wait for PLL not locked status in DPLL_ENABLE. */
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> CNL_DPLL_ENABLE(id),
> PLL_LOCK,
> 0,
> @@ -2069,7 +2069,7 @@ static void cnl_ddi_pll_disable(struct drm_i915_private *dev_priv,
> I915_WRITE(CNL_DPLL_ENABLE(id), val);
>
> /* 7. Wait for DPLL power state disabled in DPLL_ENABLE. */
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> CNL_DPLL_ENABLE(id),
> PLL_POWER_STATE,
> 0,
> @@ -2077,7 +2077,7 @@ static void cnl_ddi_pll_disable(struct drm_i915_private *dev_priv,
> DRM_ERROR("PLL %d Power not disabled\n", id);
> }
>
> -static bool cnl_ddi_pll_get_hw_state(struct drm_i915_private *dev_priv,
> +static bool cnl_ddi_pll_get_hw_state(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll,
> struct intel_dpll_hw_state *hw_state)
> {
> @@ -2086,7 +2086,7 @@ static bool cnl_ddi_pll_get_hw_state(struct drm_i915_private *dev_priv,
> u32 val;
> bool ret;
>
> - wakeref = intel_display_power_get_if_enabled(dev_priv,
> + wakeref = intel_display_power_get_if_enabled(i915,
> POWER_DOMAIN_DISPLAY_CORE);
> if (!wakeref)
> return false;
> @@ -2107,7 +2107,7 @@ static bool cnl_ddi_pll_get_hw_state(struct drm_i915_private *dev_priv,
> ret = true;
>
> out:
> - intel_display_power_put(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> + intel_display_power_put(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref);
>
> return ret;
> }
> @@ -2199,15 +2199,15 @@ static void cnl_wrpll_params_populate(struct skl_wrpll_params *params,
> params->dco_fraction = dco & 0x7fff;
> }
>
> -int cnl_hdmi_pll_ref_clock(struct drm_i915_private *dev_priv)
> +int cnl_hdmi_pll_ref_clock(struct drm_i915_private *i915)
> {
> - int ref_clock = dev_priv->cdclk.hw.ref;
> + int ref_clock = i915->cdclk.hw.ref;
>
> /*
> * For ICL+, the spec states: if reference frequency is 38.4,
> * use 19.2 because the DPLL automatically divides that by 2.
> */
> - if (INTEL_GEN(dev_priv) >= 11 && ref_clock == 38400)
> + if (INTEL_GEN(i915) >= 11 && ref_clock == 38400)
> ref_clock = 19200;
>
> return ref_clock;
> @@ -2217,7 +2217,7 @@ static bool
> cnl_ddi_calculate_wrpll(struct intel_crtc_state *crtc_state,
> struct skl_wrpll_params *wrpll_params)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> u32 afe_clock = crtc_state->port_clock * 5;
> u32 ref_clock;
> u32 dco_min = 7998000;
> @@ -2252,7 +2252,7 @@ cnl_ddi_calculate_wrpll(struct intel_crtc_state *crtc_state,
>
> cnl_wrpll_get_multipliers(best_div, &pdiv, &qdiv, &kdiv);
>
> - ref_clock = cnl_hdmi_pll_ref_clock(dev_priv);
> + ref_clock = cnl_hdmi_pll_ref_clock(i915);
>
> cnl_wrpll_params_populate(wrpll_params, best_dco, ref_clock,
> pdiv, qdiv, kdiv);
> @@ -2370,7 +2370,7 @@ cnl_get_dpll(struct intel_crtc_state *crtc_state,
> return pll;
> }
>
> -static void cnl_dump_hw_state(struct drm_i915_private *dev_priv,
> +static void cnl_dump_hw_state(struct drm_i915_private *i915,
> const struct intel_dpll_hw_state *hw_state)
> {
> DRM_DEBUG_KMS("dpll_hw_state: "
> @@ -2476,9 +2476,9 @@ static const struct skl_wrpll_params icl_tbt_pll_19_2MHz_values = {
> static bool icl_calc_dp_combo_pll(struct intel_crtc_state *crtc_state,
> struct skl_wrpll_params *pll_params)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> const struct icl_combo_pll_params *params =
> - dev_priv->cdclk.hw.ref == 24000 ?
> + i915->cdclk.hw.ref == 24000 ?
> icl_dp_combo_pll_24MHz_values :
> icl_dp_combo_pll_19_2MHz_values;
> int clock = crtc_state->port_clock;
> @@ -2498,9 +2498,9 @@ static bool icl_calc_dp_combo_pll(struct intel_crtc_state *crtc_state,
> static bool icl_calc_tbt_pll(struct intel_crtc_state *crtc_state,
> struct skl_wrpll_params *pll_params)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
>
> - *pll_params = dev_priv->cdclk.hw.ref == 24000 ?
> + *pll_params = i915->cdclk.hw.ref == 24000 ?
> icl_tbt_pll_24MHz_values : icl_tbt_pll_19_2MHz_values;
> return true;
> }
> @@ -2508,12 +2508,12 @@ static bool icl_calc_tbt_pll(struct intel_crtc_state *crtc_state,
> static bool icl_calc_dpll_state(struct intel_crtc_state *crtc_state,
> struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> u32 cfgcr0, cfgcr1;
> struct skl_wrpll_params pll_params = { 0 };
> bool ret;
>
> - if (intel_port_is_tc(dev_priv, encoder->port))
> + if (intel_port_is_tc(i915, encoder->port))
> ret = icl_calc_tbt_pll(crtc_state, &pll_params);
> else if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_HDMI) ||
> intel_crtc_has_type(crtc_state, INTEL_OUTPUT_DSI))
> @@ -2629,9 +2629,9 @@ static bool icl_mg_pll_find_divisors(int clock_khz, bool is_dp, bool use_ssc,
> */
> static bool icl_calc_mg_pll_state(struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> struct intel_dpll_hw_state *pll_state = &crtc_state->dpll_hw_state;
> - int refclk_khz = dev_priv->cdclk.hw.ref;
> + int refclk_khz = i915->cdclk.hw.ref;
> int clock = crtc_state->port_clock;
> u32 dco_khz, m1div, m2div_int, m2div_rem, m2div_frac;
> u32 iref_ndiv, iref_trim, iref_pulse_w;
> @@ -2796,18 +2796,18 @@ static struct intel_shared_dpll *
> icl_get_dpll(struct intel_crtc_state *crtc_state,
> struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> struct intel_digital_port *intel_dig_port;
> struct intel_shared_dpll *pll;
> enum port port = encoder->port;
> enum intel_dpll_id min, max;
> bool ret;
>
> - if (intel_port_is_combophy(dev_priv, port)) {
> + if (intel_port_is_combophy(i915, port)) {
> min = DPLL_ID_ICL_DPLL0;
> max = DPLL_ID_ICL_DPLL1;
> ret = icl_calc_dpll_state(crtc_state, encoder);
> - } else if (intel_port_is_tc(dev_priv, port)) {
> + } else if (intel_port_is_tc(i915, port)) {
> if (encoder->type == INTEL_OUTPUT_DP_MST) {
> struct intel_dp_mst_encoder *mst_encoder;
>
> @@ -2824,7 +2824,7 @@ icl_get_dpll(struct intel_crtc_state *crtc_state,
> } else {
> enum tc_port tc_port;
>
> - tc_port = intel_port_to_tc(dev_priv, port);
> + tc_port = intel_port_to_tc(i915, port);
> min = icl_tc_port_to_pll_id(tc_port);
> max = min;
> ret = icl_calc_mg_pll_state(crtc_state);
> @@ -2851,7 +2851,7 @@ icl_get_dpll(struct intel_crtc_state *crtc_state,
> return pll;
> }
>
> -static bool mg_pll_get_hw_state(struct drm_i915_private *dev_priv,
> +static bool mg_pll_get_hw_state(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll,
> struct intel_dpll_hw_state *hw_state)
> {
> @@ -2861,7 +2861,7 @@ static bool mg_pll_get_hw_state(struct drm_i915_private *dev_priv,
> bool ret = false;
> u32 val;
>
> - wakeref = intel_display_power_get_if_enabled(dev_priv,
> + wakeref = intel_display_power_get_if_enabled(i915,
> POWER_DOMAIN_DISPLAY_CORE);
> if (!wakeref)
> return false;
> @@ -2896,7 +2896,7 @@ static bool mg_pll_get_hw_state(struct drm_i915_private *dev_priv,
> hw_state->mg_pll_tdc_coldst_bias =
> I915_READ(MG_PLL_TDC_COLDST_BIAS(tc_port));
>
> - if (dev_priv->cdclk.hw.ref == 38400) {
> + if (i915->cdclk.hw.ref == 38400) {
> hw_state->mg_pll_tdc_coldst_bias_mask = MG_PLL_TDC_COLDST_COLDSTART;
> hw_state->mg_pll_bias_mask = 0;
> } else {
> @@ -2909,11 +2909,11 @@ static bool mg_pll_get_hw_state(struct drm_i915_private *dev_priv,
>
> ret = true;
> out:
> - intel_display_power_put(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> + intel_display_power_put(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> return ret;
> }
>
> -static bool icl_pll_get_hw_state(struct drm_i915_private *dev_priv,
> +static bool icl_pll_get_hw_state(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll,
> struct intel_dpll_hw_state *hw_state,
> i915_reg_t enable_reg)
> @@ -2923,7 +2923,7 @@ static bool icl_pll_get_hw_state(struct drm_i915_private *dev_priv,
> bool ret = false;
> u32 val;
>
> - wakeref = intel_display_power_get_if_enabled(dev_priv,
> + wakeref = intel_display_power_get_if_enabled(i915,
> POWER_DOMAIN_DISPLAY_CORE);
> if (!wakeref)
> return false;
> @@ -2937,26 +2937,26 @@ static bool icl_pll_get_hw_state(struct drm_i915_private *dev_priv,
>
> ret = true;
> out:
> - intel_display_power_put(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> + intel_display_power_put(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> return ret;
> }
>
> -static bool combo_pll_get_hw_state(struct drm_i915_private *dev_priv,
> +static bool combo_pll_get_hw_state(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll,
> struct intel_dpll_hw_state *hw_state)
> {
> - return icl_pll_get_hw_state(dev_priv, pll, hw_state,
> + return icl_pll_get_hw_state(i915, pll, hw_state,
> CNL_DPLL_ENABLE(pll->info->id));
> }
>
> -static bool tbt_pll_get_hw_state(struct drm_i915_private *dev_priv,
> +static bool tbt_pll_get_hw_state(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll,
> struct intel_dpll_hw_state *hw_state)
> {
> - return icl_pll_get_hw_state(dev_priv, pll, hw_state, TBT_PLL_ENABLE);
> + return icl_pll_get_hw_state(i915, pll, hw_state, TBT_PLL_ENABLE);
> }
>
> -static void icl_dpll_write(struct drm_i915_private *dev_priv,
> +static void icl_dpll_write(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll)
> {
> struct intel_dpll_hw_state *hw_state = &pll->state.hw_state;
> @@ -2967,7 +2967,7 @@ static void icl_dpll_write(struct drm_i915_private *dev_priv,
> POSTING_READ(ICL_DPLL_CFGCR1(id));
> }
>
> -static void icl_mg_pll_write(struct drm_i915_private *dev_priv,
> +static void icl_mg_pll_write(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll)
> {
> struct intel_dpll_hw_state *hw_state = &pll->state.hw_state;
> @@ -3017,7 +3017,7 @@ static void icl_mg_pll_write(struct drm_i915_private *dev_priv,
> POSTING_READ(MG_PLL_TDC_COLDST_BIAS(tc_port));
> }
>
> -static void icl_pll_power_enable(struct drm_i915_private *dev_priv,
> +static void icl_pll_power_enable(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll,
> i915_reg_t enable_reg)
> {
> @@ -3031,12 +3031,12 @@ static void icl_pll_power_enable(struct drm_i915_private *dev_priv,
> * The spec says we need to "wait" but it also says it should be
> * immediate.
> */
> - if (intel_wait_for_register(&dev_priv->uncore, enable_reg,
> + if (intel_wait_for_register(&i915->uncore, enable_reg,
> PLL_POWER_STATE, PLL_POWER_STATE, 1))
> DRM_ERROR("PLL %d Power not enabled\n", pll->info->id);
> }
>
> -static void icl_pll_enable(struct drm_i915_private *dev_priv,
> +static void icl_pll_enable(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll,
> i915_reg_t enable_reg)
> {
> @@ -3047,19 +3047,19 @@ static void icl_pll_enable(struct drm_i915_private *dev_priv,
> I915_WRITE(enable_reg, val);
>
> /* Timeout is actually 600us. */
> - if (intel_wait_for_register(&dev_priv->uncore, enable_reg,
> + if (intel_wait_for_register(&i915->uncore, enable_reg,
> PLL_LOCK, PLL_LOCK, 1))
> DRM_ERROR("PLL %d not locked\n", pll->info->id);
> }
>
> -static void combo_pll_enable(struct drm_i915_private *dev_priv,
> +static void combo_pll_enable(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll)
> {
> i915_reg_t enable_reg = CNL_DPLL_ENABLE(pll->info->id);
>
> - icl_pll_power_enable(dev_priv, pll, enable_reg);
> + icl_pll_power_enable(i915, pll, enable_reg);
>
> - icl_dpll_write(dev_priv, pll);
> + icl_dpll_write(i915, pll);
>
> /*
> * DVFS pre sequence would be here, but in our driver the cdclk code
> @@ -3067,17 +3067,17 @@ static void combo_pll_enable(struct drm_i915_private *dev_priv,
> * nothing here.
> */
>
> - icl_pll_enable(dev_priv, pll, enable_reg);
> + icl_pll_enable(i915, pll, enable_reg);
>
> /* DVFS post sequence would be here. See the comment above. */
> }
>
> -static void tbt_pll_enable(struct drm_i915_private *dev_priv,
> +static void tbt_pll_enable(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll)
> {
> - icl_pll_power_enable(dev_priv, pll, TBT_PLL_ENABLE);
> + icl_pll_power_enable(i915, pll, TBT_PLL_ENABLE);
>
> - icl_dpll_write(dev_priv, pll);
> + icl_dpll_write(i915, pll);
>
> /*
> * DVFS pre sequence would be here, but in our driver the cdclk code
> @@ -3085,20 +3085,20 @@ static void tbt_pll_enable(struct drm_i915_private *dev_priv,
> * nothing here.
> */
>
> - icl_pll_enable(dev_priv, pll, TBT_PLL_ENABLE);
> + icl_pll_enable(i915, pll, TBT_PLL_ENABLE);
>
> /* DVFS post sequence would be here. See the comment above. */
> }
>
> -static void mg_pll_enable(struct drm_i915_private *dev_priv,
> +static void mg_pll_enable(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll)
> {
> i915_reg_t enable_reg =
> MG_PLL_ENABLE(icl_pll_id_to_tc_port(pll->info->id));
>
> - icl_pll_power_enable(dev_priv, pll, enable_reg);
> + icl_pll_power_enable(i915, pll, enable_reg);
>
> - icl_mg_pll_write(dev_priv, pll);
> + icl_mg_pll_write(i915, pll);
>
> /*
> * DVFS pre sequence would be here, but in our driver the cdclk code
> @@ -3106,12 +3106,12 @@ static void mg_pll_enable(struct drm_i915_private *dev_priv,
> * nothing here.
> */
>
> - icl_pll_enable(dev_priv, pll, enable_reg);
> + icl_pll_enable(i915, pll, enable_reg);
>
> /* DVFS post sequence would be here. See the comment above. */
> }
>
> -static void icl_pll_disable(struct drm_i915_private *dev_priv,
> +static void icl_pll_disable(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll,
> i915_reg_t enable_reg)
> {
> @@ -3130,7 +3130,7 @@ static void icl_pll_disable(struct drm_i915_private *dev_priv,
> I915_WRITE(enable_reg, val);
>
> /* Timeout is actually 1us. */
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> enable_reg, PLL_LOCK, 0, 1))
> DRM_ERROR("PLL %d locked\n", pll->info->id);
>
> @@ -3144,33 +3144,33 @@ static void icl_pll_disable(struct drm_i915_private *dev_priv,
> * The spec says we need to "wait" but it also says it should be
> * immediate.
> */
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> enable_reg, PLL_POWER_STATE, 0, 1))
> DRM_ERROR("PLL %d Power not disabled\n", pll->info->id);
> }
>
> -static void combo_pll_disable(struct drm_i915_private *dev_priv,
> +static void combo_pll_disable(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll)
> {
> - icl_pll_disable(dev_priv, pll, CNL_DPLL_ENABLE(pll->info->id));
> + icl_pll_disable(i915, pll, CNL_DPLL_ENABLE(pll->info->id));
> }
>
> -static void tbt_pll_disable(struct drm_i915_private *dev_priv,
> +static void tbt_pll_disable(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll)
> {
> - icl_pll_disable(dev_priv, pll, TBT_PLL_ENABLE);
> + icl_pll_disable(i915, pll, TBT_PLL_ENABLE);
> }
>
> -static void mg_pll_disable(struct drm_i915_private *dev_priv,
> +static void mg_pll_disable(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll)
> {
> i915_reg_t enable_reg =
> MG_PLL_ENABLE(icl_pll_id_to_tc_port(pll->info->id));
>
> - icl_pll_disable(dev_priv, pll, enable_reg);
> + icl_pll_disable(i915, pll, enable_reg);
> }
>
> -static void icl_dump_hw_state(struct drm_i915_private *dev_priv,
> +static void icl_dump_hw_state(struct drm_i915_private *i915,
> const struct intel_dpll_hw_state *hw_state)
> {
> DRM_DEBUG_KMS("dpll_hw_state: cfgcr0: 0x%x, cfgcr1: 0x%x, "
> @@ -3247,28 +3247,28 @@ static const struct intel_dpll_mgr ehl_pll_mgr = {
> */
> void intel_shared_dpll_init(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> const struct intel_dpll_mgr *dpll_mgr = NULL;
> const struct dpll_info *dpll_info;
> int i;
>
> - if (IS_ELKHARTLAKE(dev_priv))
> + if (IS_ELKHARTLAKE(i915))
> dpll_mgr = &ehl_pll_mgr;
> - else if (INTEL_GEN(dev_priv) >= 11)
> + else if (INTEL_GEN(i915) >= 11)
> dpll_mgr = &icl_pll_mgr;
> - else if (IS_CANNONLAKE(dev_priv))
> + else if (IS_CANNONLAKE(i915))
> dpll_mgr = &cnl_pll_mgr;
> - else if (IS_GEN9_BC(dev_priv))
> + else if (IS_GEN9_BC(i915))
> dpll_mgr = &skl_pll_mgr;
> - else if (IS_GEN9_LP(dev_priv))
> + else if (IS_GEN9_LP(i915))
> dpll_mgr = &bxt_pll_mgr;
> - else if (HAS_DDI(dev_priv))
> + else if (HAS_DDI(i915))
> dpll_mgr = &hsw_pll_mgr;
> - else if (HAS_PCH_IBX(dev_priv) || HAS_PCH_CPT(dev_priv))
> + else if (HAS_PCH_IBX(i915) || HAS_PCH_CPT(i915))
> dpll_mgr = &pch_pll_mgr;
>
> if (!dpll_mgr) {
> - dev_priv->num_shared_dpll = 0;
> + i915->num_shared_dpll = 0;
> return;
> }
>
> @@ -3276,14 +3276,14 @@ void intel_shared_dpll_init(struct drm_device *dev)
>
> for (i = 0; dpll_info[i].name; i++) {
> WARN_ON(i != dpll_info[i].id);
> - dev_priv->shared_dplls[i].info = &dpll_info[i];
> + i915->shared_dplls[i].info = &dpll_info[i];
> }
>
> - dev_priv->dpll_mgr = dpll_mgr;
> - dev_priv->num_shared_dpll = i;
> - mutex_init(&dev_priv->dpll_lock);
> + i915->dpll_mgr = dpll_mgr;
> + i915->num_shared_dpll = i;
> + mutex_init(&i915->dpll_lock);
>
> - BUG_ON(dev_priv->num_shared_dpll > I915_NUM_PLLS);
> + BUG_ON(i915->num_shared_dpll > I915_NUM_PLLS);
> }
>
> /**
> @@ -3304,8 +3304,8 @@ struct intel_shared_dpll *
> intel_get_shared_dpll(struct intel_crtc_state *crtc_state,
> struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> - const struct intel_dpll_mgr *dpll_mgr = dev_priv->dpll_mgr;
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> + const struct intel_dpll_mgr *dpll_mgr = i915->dpll_mgr;
>
> if (WARN_ON(!dpll_mgr))
> return NULL;
> @@ -3335,16 +3335,16 @@ void intel_release_shared_dpll(struct intel_shared_dpll *dpll,
>
> /**
> * intel_shared_dpll_dump_hw_state - write hw_state to dmesg
> - * @dev_priv: i915 drm device
> + * @i915: i915 drm device
> * @hw_state: hw state to be written to the log
> *
> * Write the relevant values in @hw_state to dmesg using DRM_DEBUG_KMS.
> */
> -void intel_dpll_dump_hw_state(struct drm_i915_private *dev_priv,
> +void intel_dpll_dump_hw_state(struct drm_i915_private *i915,
> const struct intel_dpll_hw_state *hw_state)
> {
> - if (dev_priv->dpll_mgr) {
> - dev_priv->dpll_mgr->dump_hw_state(dev_priv, hw_state);
> + if (i915->dpll_mgr) {
> + i915->dpll_mgr->dump_hw_state(i915, hw_state);
> } else {
> /* fallback for platforms that don't use the shared dpll
> * infrastructure
> diff --git a/drivers/gpu/drm/i915/intel_dpll_mgr.h b/drivers/gpu/drm/i915/intel_dpll_mgr.h
> index d0570414f3d1..b068c8441328 100644
> --- a/drivers/gpu/drm/i915/intel_dpll_mgr.h
> +++ b/drivers/gpu/drm/i915/intel_dpll_mgr.h
> @@ -221,7 +221,7 @@ struct intel_shared_dpll_funcs {
> * Called from intel_prepare_shared_dpll() function unless the PLL
> * is already enabled.
> */
> - void (*prepare)(struct drm_i915_private *dev_priv,
> + void (*prepare)(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll);
>
> /**
> @@ -230,7 +230,7 @@ struct intel_shared_dpll_funcs {
> * Hook for enabling the pll, called from intel_enable_shared_dpll()
> * if the pll is not already enabled.
> */
> - void (*enable)(struct drm_i915_private *dev_priv,
> + void (*enable)(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll);
>
> /**
> @@ -240,7 +240,7 @@ struct intel_shared_dpll_funcs {
> * only when it is safe to disable the pll, i.e., there are no more
> * tracked users for it.
> */
> - void (*disable)(struct drm_i915_private *dev_priv,
> + void (*disable)(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll);
>
> /**
> @@ -250,7 +250,7 @@ struct intel_shared_dpll_funcs {
> * registers. This is used for initial hw state readout and state
> * verification after a mode set.
> */
> - bool (*get_hw_state)(struct drm_i915_private *dev_priv,
> + bool (*get_hw_state)(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll,
> struct intel_dpll_hw_state *hw_state);
> };
> @@ -271,7 +271,7 @@ struct dpll_info {
>
> /**
> * @id: unique indentifier for this DPLL; should match the index in the
> - * dev_priv->shared_dplls array
> + * i915->shared_dplls array
> */
> enum intel_dpll_id id;
>
> @@ -321,12 +321,12 @@ struct intel_shared_dpll {
>
> /* shared dpll functions */
> struct intel_shared_dpll *
> -intel_get_shared_dpll_by_id(struct drm_i915_private *dev_priv,
> +intel_get_shared_dpll_by_id(struct drm_i915_private *i915,
> enum intel_dpll_id id);
> enum intel_dpll_id
> -intel_get_shared_dpll_id(struct drm_i915_private *dev_priv,
> +intel_get_shared_dpll_id(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll);
> -void assert_shared_dpll(struct drm_i915_private *dev_priv,
> +void assert_shared_dpll(struct drm_i915_private *i915,
> struct intel_shared_dpll *pll,
> bool state);
> #define assert_shared_dpll_enabled(d, p) assert_shared_dpll(d, p, true)
> @@ -342,9 +342,9 @@ void intel_disable_shared_dpll(const struct intel_crtc_state *crtc_state);
> void intel_shared_dpll_swap_state(struct drm_atomic_state *state);
> void intel_shared_dpll_init(struct drm_device *dev);
>
> -void intel_dpll_dump_hw_state(struct drm_i915_private *dev_priv,
> +void intel_dpll_dump_hw_state(struct drm_i915_private *i915,
> const struct intel_dpll_hw_state *hw_state);
> -int cnl_hdmi_pll_ref_clock(struct drm_i915_private *dev_priv);
> +int cnl_hdmi_pll_ref_clock(struct drm_i915_private *i915);
> enum intel_dpll_id icl_tc_port_to_pll_id(enum tc_port tc_port);
> bool intel_dpll_is_combophy(enum intel_dpll_id id);
>
> diff --git a/drivers/gpu/drm/i915/intel_drv.h b/drivers/gpu/drm/i915/intel_drv.h
> index 3e337317f77e..23065dda82d9 100644
> --- a/drivers/gpu/drm/i915/intel_drv.h
> +++ b/drivers/gpu/drm/i915/intel_drv.h
> @@ -963,7 +963,7 @@ struct intel_crtc {
>
> struct intel_crtc_state *config;
>
> - /* Access to these should be protected by dev_priv->irq_lock. */
> + /* Access to these should be protected by i915->irq_lock. */
> bool cpu_fifo_underrun_disabled;
> bool pch_fifo_underrun_disabled;
>
> @@ -1293,15 +1293,15 @@ vlv_pipe_to_channel(enum pipe pipe)
> }
>
> static inline struct intel_crtc *
> -intel_get_crtc_for_pipe(struct drm_i915_private *dev_priv, enum pipe pipe)
> +intel_get_crtc_for_pipe(struct drm_i915_private *i915, enum pipe pipe)
> {
> - return dev_priv->pipe_to_crtc_mapping[pipe];
> + return i915->pipe_to_crtc_mapping[pipe];
> }
>
> static inline struct intel_crtc *
> -intel_get_crtc_for_plane(struct drm_i915_private *dev_priv, enum i9xx_plane_id plane)
> +intel_get_crtc_for_plane(struct drm_i915_private *i915, enum i9xx_plane_id plane)
> {
> - return dev_priv->plane_to_crtc_mapping[plane];
> + return i915->plane_to_crtc_mapping[plane];
> }
>
> struct intel_load_detect_pipe {
> @@ -1447,17 +1447,17 @@ intel_atomic_get_new_crtc_state(struct intel_atomic_state *state,
>
> /* intel_display.c */
> void intel_plane_destroy(struct drm_plane *plane);
> -void i830_enable_pipe(struct drm_i915_private *dev_priv, enum pipe pipe);
> -void i830_disable_pipe(struct drm_i915_private *dev_priv, enum pipe pipe);
> +void i830_enable_pipe(struct drm_i915_private *i915, enum pipe pipe);
> +void i830_disable_pipe(struct drm_i915_private *i915, enum pipe pipe);
> enum pipe intel_crtc_pch_transcoder(struct intel_crtc *crtc);
> -int vlv_get_hpll_vco(struct drm_i915_private *dev_priv);
> -int vlv_get_cck_clock(struct drm_i915_private *dev_priv,
> +int vlv_get_hpll_vco(struct drm_i915_private *i915);
> +int vlv_get_cck_clock(struct drm_i915_private *i915,
> const char *name, u32 reg, int ref_freq);
> -int vlv_get_cck_clock_hpll(struct drm_i915_private *dev_priv,
> +int vlv_get_cck_clock_hpll(struct drm_i915_private *i915,
> const char *name, u32 reg);
> -void lpt_disable_pch_transcoder(struct drm_i915_private *dev_priv);
> -void lpt_disable_iclkip(struct drm_i915_private *dev_priv);
> -void intel_init_display_hooks(struct drm_i915_private *dev_priv);
> +void lpt_disable_pch_transcoder(struct drm_i915_private *i915);
> +void lpt_disable_iclkip(struct drm_i915_private *i915);
> +void intel_init_display_hooks(struct drm_i915_private *i915);
> unsigned int intel_fb_xy_to_linear(int x, int y,
> const struct intel_plane_state *state,
> int plane);
> @@ -1467,19 +1467,19 @@ void intel_add_fb_offsets(int *x, int *y,
> const struct intel_plane_state *state, int plane);
> unsigned int intel_rotation_info_size(const struct intel_rotation_info *rot_info);
> unsigned int intel_remapped_info_size(const struct intel_remapped_info *rem_info);
> -bool intel_has_pending_fb_unpin(struct drm_i915_private *dev_priv);
> +bool intel_has_pending_fb_unpin(struct drm_i915_private *i915);
> int intel_display_suspend(struct drm_device *dev);
> -void intel_pps_unlock_regs_wa(struct drm_i915_private *dev_priv);
> +void intel_pps_unlock_regs_wa(struct drm_i915_private *i915);
> void intel_encoder_destroy(struct drm_encoder *encoder);
> struct drm_display_mode *
> intel_encoder_current_mode(struct intel_encoder *encoder);
> -bool intel_port_is_combophy(struct drm_i915_private *dev_priv, enum port port);
> -bool intel_port_is_tc(struct drm_i915_private *dev_priv, enum port port);
> -enum tc_port intel_port_to_tc(struct drm_i915_private *dev_priv,
> +bool intel_port_is_combophy(struct drm_i915_private *i915, enum port port);
> +bool intel_port_is_tc(struct drm_i915_private *i915, enum port port);
> +enum tc_port intel_port_to_tc(struct drm_i915_private *i915,
> enum port port);
> int intel_get_pipe_from_crtc_id_ioctl(struct drm_device *dev, void *data,
> struct drm_file *file_priv);
> -enum transcoder intel_pipe_to_cpu_transcoder(struct drm_i915_private *dev_priv,
> +enum transcoder intel_pipe_to_cpu_transcoder(struct drm_i915_private *i915,
> enum pipe pipe);
> static inline bool
> intel_crtc_has_type(const struct intel_crtc_state *crtc_state,
> @@ -1496,23 +1496,23 @@ intel_crtc_has_dp_encoder(const struct intel_crtc_state *crtc_state)
> (1 << INTEL_OUTPUT_EDP));
> }
> static inline void
> -intel_wait_for_vblank(struct drm_i915_private *dev_priv, enum pipe pipe)
> +intel_wait_for_vblank(struct drm_i915_private *i915, enum pipe pipe)
> {
> - drm_wait_one_vblank(&dev_priv->drm, pipe);
> + drm_wait_one_vblank(&i915->drm, pipe);
> }
> static inline void
> -intel_wait_for_vblank_if_active(struct drm_i915_private *dev_priv, int pipe)
> +intel_wait_for_vblank_if_active(struct drm_i915_private *i915, int pipe)
> {
> - const struct intel_crtc *crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> + const struct intel_crtc *crtc = intel_get_crtc_for_pipe(i915, pipe);
>
> if (crtc->active)
> - intel_wait_for_vblank(dev_priv, pipe);
> + intel_wait_for_vblank(i915, pipe);
> }
>
> u32 intel_crtc_get_vblank_counter(struct intel_crtc *crtc);
>
> int ironlake_get_lanes_required(int target_clock, int link_bw, int bpp);
> -void vlv_wait_port_ready(struct drm_i915_private *dev_priv,
> +void vlv_wait_port_ready(struct drm_i915_private *i915,
> struct intel_digital_port *dport,
> unsigned int expected_mask);
> int intel_get_load_detect_pipe(struct drm_connector *connector,
> @@ -1536,34 +1536,34 @@ int intel_prepare_plane_fb(struct drm_plane *plane,
> void intel_cleanup_plane_fb(struct drm_plane *plane,
> struct drm_plane_state *old_state);
>
> -void assert_pch_transcoder_disabled(struct drm_i915_private *dev_priv,
> +void assert_pch_transcoder_disabled(struct drm_i915_private *i915,
> enum pipe pipe);
>
> -int vlv_force_pll_on(struct drm_i915_private *dev_priv, enum pipe pipe,
> +int vlv_force_pll_on(struct drm_i915_private *i915, enum pipe pipe,
> const struct dpll *dpll);
> -void vlv_force_pll_off(struct drm_i915_private *dev_priv, enum pipe pipe);
> -int lpt_get_iclkip(struct drm_i915_private *dev_priv);
> +void vlv_force_pll_off(struct drm_i915_private *i915, enum pipe pipe);
> +int lpt_get_iclkip(struct drm_i915_private *i915);
> bool intel_fuzzy_clock_check(int clock1, int clock2);
>
> /* modesetting asserts */
> -void assert_panel_unlocked(struct drm_i915_private *dev_priv,
> +void assert_panel_unlocked(struct drm_i915_private *i915,
> enum pipe pipe);
> -void assert_pll(struct drm_i915_private *dev_priv,
> +void assert_pll(struct drm_i915_private *i915,
> enum pipe pipe, bool state);
> #define assert_pll_enabled(d, p) assert_pll(d, p, true)
> #define assert_pll_disabled(d, p) assert_pll(d, p, false)
> -void assert_dsi_pll(struct drm_i915_private *dev_priv, bool state);
> +void assert_dsi_pll(struct drm_i915_private *i915, bool state);
> #define assert_dsi_pll_enabled(d) assert_dsi_pll(d, true)
> #define assert_dsi_pll_disabled(d) assert_dsi_pll(d, false)
> -void assert_fdi_rx_pll(struct drm_i915_private *dev_priv,
> +void assert_fdi_rx_pll(struct drm_i915_private *i915,
> enum pipe pipe, bool state);
> #define assert_fdi_rx_pll_enabled(d, p) assert_fdi_rx_pll(d, p, true)
> #define assert_fdi_rx_pll_disabled(d, p) assert_fdi_rx_pll(d, p, false)
> -void assert_pipe(struct drm_i915_private *dev_priv, enum pipe pipe, bool state);
> +void assert_pipe(struct drm_i915_private *i915, enum pipe pipe, bool state);
> #define assert_pipe_enabled(d, p) assert_pipe(d, p, true)
> #define assert_pipe_disabled(d, p) assert_pipe(d, p, false)
> -void intel_prepare_reset(struct drm_i915_private *dev_priv);
> -void intel_finish_reset(struct drm_i915_private *dev_priv);
> +void intel_prepare_reset(struct drm_i915_private *i915);
> +void intel_finish_reset(struct drm_i915_private *i915);
> void intel_dp_get_m_n(struct intel_crtc *crtc,
> struct intel_crtc_state *pipe_config);
> void intel_dp_set_m_n(const struct intel_crtc_state *crtc_state,
> diff --git a/drivers/gpu/drm/i915/intel_dsi.c b/drivers/gpu/drm/i915/intel_dsi.c
> index 5fec02aceaed..09414da3d818 100644
> --- a/drivers/gpu/drm/i915/intel_dsi.c
> +++ b/drivers/gpu/drm/i915/intel_dsi.c
> @@ -113,14 +113,14 @@ struct intel_dsi_host *intel_dsi_host_init(struct intel_dsi *intel_dsi,
> enum drm_panel_orientation
> intel_dsi_get_panel_orientation(struct intel_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> enum drm_panel_orientation orientation;
>
> - orientation = dev_priv->vbt.dsi.orientation;
> + orientation = i915->vbt.dsi.orientation;
> if (orientation != DRM_MODE_PANEL_ORIENTATION_UNKNOWN)
> return orientation;
>
> - orientation = dev_priv->vbt.orientation;
> + orientation = i915->vbt.orientation;
> if (orientation != DRM_MODE_PANEL_ORIENTATION_UNKNOWN)
> return orientation;
>
> diff --git a/drivers/gpu/drm/i915/intel_dsi.h b/drivers/gpu/drm/i915/intel_dsi.h
> index 6d20434636cd..2f477263d1f6 100644
> --- a/drivers/gpu/drm/i915/intel_dsi.h
> +++ b/drivers/gpu/drm/i915/intel_dsi.h
> @@ -155,7 +155,7 @@ static inline u16 intel_dsi_encoder_ports(struct intel_encoder *encoder)
> }
>
> /* icl_dsi.c */
> -void icl_dsi_init(struct drm_i915_private *dev_priv);
> +void icl_dsi_init(struct drm_i915_private *i915);
>
> /* intel_dsi.c */
> int intel_dsi_bitrate(const struct intel_dsi *intel_dsi);
> @@ -172,7 +172,7 @@ enum drm_mode_status intel_dsi_mode_valid(struct drm_connector *connector,
> struct intel_dsi_host *intel_dsi_host_init(struct intel_dsi *intel_dsi,
> const struct mipi_dsi_host_ops *funcs,
> enum port port);
> -void vlv_dsi_init(struct drm_i915_private *dev_priv);
> +void vlv_dsi_init(struct drm_i915_private *i915);
>
> /* vlv_dsi_pll.c */
> int vlv_dsi_pll_compute(struct intel_encoder *encoder,
> @@ -184,7 +184,7 @@ u32 vlv_dsi_get_pclk(struct intel_encoder *encoder,
> struct intel_crtc_state *config);
> void vlv_dsi_reset_clocks(struct intel_encoder *encoder, enum port port);
>
> -bool bxt_dsi_pll_is_enabled(struct drm_i915_private *dev_priv);
> +bool bxt_dsi_pll_is_enabled(struct drm_i915_private *i915);
> int bxt_dsi_pll_compute(struct intel_encoder *encoder,
> struct intel_crtc_state *config);
> void bxt_dsi_pll_enable(struct intel_encoder *encoder,
> diff --git a/drivers/gpu/drm/i915/intel_dsi_dcs_backlight.c b/drivers/gpu/drm/i915/intel_dsi_dcs_backlight.c
> index 8c33262cb0b2..e5dd4962705c 100644
> --- a/drivers/gpu/drm/i915/intel_dsi_dcs_backlight.c
> +++ b/drivers/gpu/drm/i915/intel_dsi_dcs_backlight.c
> @@ -159,11 +159,11 @@ static int dcs_setup_backlight(struct intel_connector *connector,
> int intel_dsi_dcs_init_backlight_funcs(struct intel_connector *intel_connector)
> {
> struct drm_device *dev = intel_connector->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_encoder *encoder = intel_connector->encoder;
> struct intel_panel *panel = &intel_connector->panel;
>
> - if (dev_priv->vbt.backlight.type != INTEL_BACKLIGHT_DSI_DCS)
> + if (i915->vbt.backlight.type != INTEL_BACKLIGHT_DSI_DCS)
> return -ENODEV;
>
> if (WARN_ON(encoder->type != INTEL_OUTPUT_DSI))
> diff --git a/drivers/gpu/drm/i915/intel_dsi_vbt.c b/drivers/gpu/drm/i915/intel_dsi_vbt.c
> index e5b178660408..d44ebdfd8760 100644
> --- a/drivers/gpu/drm/i915/intel_dsi_vbt.c
> +++ b/drivers/gpu/drm/i915/intel_dsi_vbt.c
> @@ -121,7 +121,7 @@ static inline enum port intel_dsi_seq_port_to_port(u8 port)
> static const u8 *mipi_exec_send_packet(struct intel_dsi *intel_dsi,
> const u8 *data)
> {
> - struct drm_i915_private *dev_priv = to_i915(intel_dsi->base.base.dev);
> + struct drm_i915_private *i915 = to_i915(intel_dsi->base.base.dev);
> struct mipi_dsi_device *dsi_device;
> u8 type, flags, seq_port;
> u16 len;
> @@ -192,7 +192,7 @@ static const u8 *mipi_exec_send_packet(struct intel_dsi *intel_dsi,
> break;
> }
>
> - if (INTEL_GEN(dev_priv) < 11)
> + if (INTEL_GEN(i915) < 11)
> vlv_dsi_wait_for_fifo_empty(intel_dsi, port);
>
> out:
> @@ -213,7 +213,7 @@ static const u8 *mipi_exec_delay(struct intel_dsi *intel_dsi, const u8 *data)
> return data;
> }
>
> -static void vlv_exec_gpio(struct drm_i915_private *dev_priv,
> +static void vlv_exec_gpio(struct drm_i915_private *i915,
> u8 gpio_source, u8 gpio_index, bool value)
> {
> struct gpio_map *map;
> @@ -228,7 +228,7 @@ static void vlv_exec_gpio(struct drm_i915_private *dev_priv,
>
> map = &vlv_gpio_table[gpio_index];
>
> - if (dev_priv->vbt.dsi.seq_version >= 3) {
> + if (i915->vbt.dsi.seq_version >= 3) {
> /* XXX: this assumes vlv_gpio_table only has NC GPIOs. */
> port = IOSF_PORT_GPIO_NC;
> } else {
> @@ -246,26 +246,26 @@ static void vlv_exec_gpio(struct drm_i915_private *dev_priv,
> pconf0 = VLV_GPIO_PCONF0(map->base_offset);
> padval = VLV_GPIO_PAD_VAL(map->base_offset);
>
> - vlv_iosf_sb_get(dev_priv, BIT(VLV_IOSF_SB_GPIO));
> + vlv_iosf_sb_get(i915, BIT(VLV_IOSF_SB_GPIO));
> if (!map->init) {
> /* FIXME: remove constant below */
> - vlv_iosf_sb_write(dev_priv, port, pconf0, 0x2000CC00);
> + vlv_iosf_sb_write(i915, port, pconf0, 0x2000CC00);
> map->init = true;
> }
>
> tmp = 0x4 | value;
> - vlv_iosf_sb_write(dev_priv, port, padval, tmp);
> - vlv_iosf_sb_put(dev_priv, BIT(VLV_IOSF_SB_GPIO));
> + vlv_iosf_sb_write(i915, port, padval, tmp);
> + vlv_iosf_sb_put(i915, BIT(VLV_IOSF_SB_GPIO));
> }
>
> -static void chv_exec_gpio(struct drm_i915_private *dev_priv,
> +static void chv_exec_gpio(struct drm_i915_private *i915,
> u8 gpio_source, u8 gpio_index, bool value)
> {
> u16 cfg0, cfg1;
> u16 family_num;
> u8 port;
>
> - if (dev_priv->vbt.dsi.seq_version >= 3) {
> + if (i915->vbt.dsi.seq_version >= 3) {
> if (gpio_index >= CHV_GPIO_IDX_START_SE) {
> /* XXX: it's unclear whether 255->57 is part of SE. */
> gpio_index -= CHV_GPIO_IDX_START_SE;
> @@ -301,15 +301,15 @@ static void chv_exec_gpio(struct drm_i915_private *dev_priv,
> cfg0 = CHV_GPIO_PAD_CFG0(family_num, gpio_index);
> cfg1 = CHV_GPIO_PAD_CFG1(family_num, gpio_index);
>
> - vlv_iosf_sb_get(dev_priv, BIT(VLV_IOSF_SB_GPIO));
> - vlv_iosf_sb_write(dev_priv, port, cfg1, 0);
> - vlv_iosf_sb_write(dev_priv, port, cfg0,
> + vlv_iosf_sb_get(i915, BIT(VLV_IOSF_SB_GPIO));
> + vlv_iosf_sb_write(i915, port, cfg1, 0);
> + vlv_iosf_sb_write(i915, port, cfg0,
> CHV_GPIO_GPIOEN | CHV_GPIO_GPIOCFG_GPO |
> CHV_GPIO_GPIOTXSTATE(value));
> - vlv_iosf_sb_put(dev_priv, BIT(VLV_IOSF_SB_GPIO));
> + vlv_iosf_sb_put(i915, BIT(VLV_IOSF_SB_GPIO));
> }
>
> -static void bxt_exec_gpio(struct drm_i915_private *dev_priv,
> +static void bxt_exec_gpio(struct drm_i915_private *i915,
> u8 gpio_source, u8 gpio_index, bool value)
> {
> /* XXX: this table is a quick ugly hack. */
> @@ -317,7 +317,7 @@ static void bxt_exec_gpio(struct drm_i915_private *dev_priv,
> struct gpio_desc *gpio_desc = bxt_gpio_table[gpio_index];
>
> if (!gpio_desc) {
> - gpio_desc = devm_gpiod_get_index(dev_priv->drm.dev,
> + gpio_desc = devm_gpiod_get_index(i915->drm.dev,
> NULL, gpio_index,
> value ? GPIOD_OUT_LOW :
> GPIOD_OUT_HIGH);
> @@ -334,7 +334,7 @@ static void bxt_exec_gpio(struct drm_i915_private *dev_priv,
> gpiod_set_value(gpio_desc, value);
> }
>
> -static void icl_exec_gpio(struct drm_i915_private *dev_priv,
> +static void icl_exec_gpio(struct drm_i915_private *i915,
> u8 gpio_source, u8 gpio_index, bool value)
> {
> DRM_DEBUG_KMS("Skipping ICL GPIO element execution\n");
> @@ -343,19 +343,19 @@ static void icl_exec_gpio(struct drm_i915_private *dev_priv,
> static const u8 *mipi_exec_gpio(struct intel_dsi *intel_dsi, const u8 *data)
> {
> struct drm_device *dev = intel_dsi->base.base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> u8 gpio_source, gpio_index = 0, gpio_number;
> bool value;
>
> DRM_DEBUG_KMS("\n");
>
> - if (dev_priv->vbt.dsi.seq_version >= 3)
> + if (i915->vbt.dsi.seq_version >= 3)
> gpio_index = *data++;
>
> gpio_number = *data++;
>
> /* gpio source in sequence v2 only */
> - if (dev_priv->vbt.dsi.seq_version == 2)
> + if (i915->vbt.dsi.seq_version == 2)
> gpio_source = (*data >> 1) & 3;
> else
> gpio_source = 0;
> @@ -363,14 +363,14 @@ static const u8 *mipi_exec_gpio(struct intel_dsi *intel_dsi, const u8 *data)
> /* pull up/down */
> value = *data++ & 1;
>
> - if (INTEL_GEN(dev_priv) >= 11)
> - icl_exec_gpio(dev_priv, gpio_source, gpio_index, value);
> - else if (IS_VALLEYVIEW(dev_priv))
> - vlv_exec_gpio(dev_priv, gpio_source, gpio_number, value);
> - else if (IS_CHERRYVIEW(dev_priv))
> - chv_exec_gpio(dev_priv, gpio_source, gpio_number, value);
> + if (INTEL_GEN(i915) >= 11)
> + icl_exec_gpio(i915, gpio_source, gpio_index, value);
> + else if (IS_VALLEYVIEW(i915))
> + vlv_exec_gpio(i915, gpio_source, gpio_number, value);
> + else if (IS_CHERRYVIEW(i915))
> + chv_exec_gpio(i915, gpio_source, gpio_number, value);
> else
> - bxt_exec_gpio(dev_priv, gpio_source, gpio_index, value);
> + bxt_exec_gpio(i915, gpio_source, gpio_index, value);
>
> return data;
> }
> @@ -456,14 +456,14 @@ static const char *sequence_name(enum mipi_seq seq_id)
> void intel_dsi_vbt_exec_sequence(struct intel_dsi *intel_dsi,
> enum mipi_seq seq_id)
> {
> - struct drm_i915_private *dev_priv = to_i915(intel_dsi->base.base.dev);
> + struct drm_i915_private *i915 = to_i915(intel_dsi->base.base.dev);
> const u8 *data;
> fn_mipi_elem_exec mipi_elem_exec;
>
> - if (WARN_ON(seq_id >= ARRAY_SIZE(dev_priv->vbt.dsi.sequence)))
> + if (WARN_ON(seq_id >= ARRAY_SIZE(i915->vbt.dsi.sequence)))
> return;
>
> - data = dev_priv->vbt.dsi.sequence[seq_id];
> + data = i915->vbt.dsi.sequence[seq_id];
> if (!data)
> return;
>
> @@ -476,7 +476,7 @@ void intel_dsi_vbt_exec_sequence(struct intel_dsi *intel_dsi,
> data++;
>
> /* Skip Size of Sequence. */
> - if (dev_priv->vbt.dsi.seq_version >= 3)
> + if (i915->vbt.dsi.seq_version >= 3)
> data += 4;
>
> while (1) {
> @@ -492,7 +492,7 @@ void intel_dsi_vbt_exec_sequence(struct intel_dsi *intel_dsi,
> mipi_elem_exec = NULL;
>
> /* Size of Operation. */
> - if (dev_priv->vbt.dsi.seq_version >= 3)
> + if (i915->vbt.dsi.seq_version >= 3)
> operation_size = *data++;
>
> if (mipi_elem_exec) {
> @@ -521,10 +521,10 @@ void intel_dsi_vbt_exec_sequence(struct intel_dsi *intel_dsi,
>
> void intel_dsi_msleep(struct intel_dsi *intel_dsi, int msec)
> {
> - struct drm_i915_private *dev_priv = to_i915(intel_dsi->base.base.dev);
> + struct drm_i915_private *i915 = to_i915(intel_dsi->base.base.dev);
>
> /* For v3 VBTs in vid-mode the delays are part of the VBT sequences */
> - if (is_vid_mode(intel_dsi) && dev_priv->vbt.dsi.seq_version >= 3)
> + if (is_vid_mode(intel_dsi) && i915->vbt.dsi.seq_version >= 3)
> return;
>
> msleep(msec);
> @@ -571,10 +571,10 @@ void intel_dsi_log_params(struct intel_dsi *intel_dsi)
> bool intel_dsi_vbt_init(struct intel_dsi *intel_dsi, u16 panel_id)
> {
> struct drm_device *dev = intel_dsi->base.base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct mipi_config *mipi_config = dev_priv->vbt.dsi.config;
> - struct mipi_pps_data *pps = dev_priv->vbt.dsi.pps;
> - struct drm_display_mode *mode = dev_priv->vbt.lfp_lvds_vbt_mode;
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct mipi_config *mipi_config = i915->vbt.dsi.config;
> + struct mipi_pps_data *pps = i915->vbt.dsi.pps;
> + struct drm_display_mode *mode = i915->vbt.lfp_lvds_vbt_mode;
> u16 burst_mode_ratio;
> enum port port;
>
> diff --git a/drivers/gpu/drm/i915/intel_dvo.c b/drivers/gpu/drm/i915/intel_dvo.c
> index 22666d28f4aa..6ced95b73538 100644
> --- a/drivers/gpu/drm/i915/intel_dvo.c
> +++ b/drivers/gpu/drm/i915/intel_dvo.c
> @@ -133,7 +133,7 @@ static struct intel_dvo *intel_attached_dvo(struct drm_connector *connector)
> static bool intel_dvo_connector_get_hw_state(struct intel_connector *connector)
> {
> struct drm_device *dev = connector->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_dvo *intel_dvo = intel_attached_dvo(&connector->base);
> u32 tmp;
>
> @@ -148,7 +148,7 @@ static bool intel_dvo_connector_get_hw_state(struct intel_connector *connector)
> static bool intel_dvo_get_hw_state(struct intel_encoder *encoder,
> enum pipe *pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dvo *intel_dvo = enc_to_dvo(encoder);
> u32 tmp;
>
> @@ -162,7 +162,7 @@ static bool intel_dvo_get_hw_state(struct intel_encoder *encoder,
> static void intel_dvo_get_config(struct intel_encoder *encoder,
> struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dvo *intel_dvo = enc_to_dvo(encoder);
> u32 tmp, flags = 0;
>
> @@ -187,7 +187,7 @@ static void intel_disable_dvo(struct intel_encoder *encoder,
> const struct intel_crtc_state *old_crtc_state,
> const struct drm_connector_state *old_conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dvo *intel_dvo = enc_to_dvo(encoder);
> i915_reg_t dvo_reg = intel_dvo->dev.dvo_reg;
> u32 temp = I915_READ(dvo_reg);
> @@ -201,7 +201,7 @@ static void intel_enable_dvo(struct intel_encoder *encoder,
> const struct intel_crtc_state *pipe_config,
> const struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dvo *intel_dvo = enc_to_dvo(encoder);
> i915_reg_t dvo_reg = intel_dvo->dev.dvo_reg;
> u32 temp = I915_READ(dvo_reg);
> @@ -276,7 +276,7 @@ static void intel_dvo_pre_enable(struct intel_encoder *encoder,
> const struct intel_crtc_state *pipe_config,
> const struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crtc *crtc = to_intel_crtc(pipe_config->base.crtc);
> const struct drm_display_mode *adjusted_mode = &pipe_config->base.adjusted_mode;
> struct intel_dvo *intel_dvo = enc_to_dvo(encoder);
> @@ -319,7 +319,7 @@ intel_dvo_detect(struct drm_connector *connector, bool force)
>
> static int intel_dvo_get_modes(struct drm_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->dev);
> + struct drm_i915_private *i915 = to_i915(connector->dev);
> const struct drm_display_mode *fixed_mode =
> to_intel_connector(connector)->panel.fixed_mode;
>
> @@ -330,7 +330,7 @@ static int intel_dvo_get_modes(struct drm_connector *connector)
> * that's not the case.
> */
> intel_ddc_get_modes(connector,
> - intel_gmbus_get_adapter(dev_priv, GMBUS_PIN_DPC));
> + intel_gmbus_get_adapter(i915, GMBUS_PIN_DPC));
> if (!list_empty(&connector->probed_modes))
> return 1;
>
> @@ -406,7 +406,7 @@ static enum port intel_dvo_port(i915_reg_t dvo_reg)
> return PORT_C;
> }
>
> -void intel_dvo_init(struct drm_i915_private *dev_priv)
> +void intel_dvo_init(struct drm_i915_private *i915)
> {
> struct intel_encoder *intel_encoder;
> struct intel_dvo *intel_dvo;
> @@ -452,7 +452,7 @@ void intel_dvo_init(struct drm_i915_private *dev_priv)
> * special cases, but otherwise default to what's defined
> * in the spec.
> */
> - if (intel_gmbus_is_valid_pin(dev_priv, dvo->gpio))
> + if (intel_gmbus_is_valid_pin(i915, dvo->gpio))
> gpio = dvo->gpio;
> else if (dvo->type == INTEL_DVO_CHIP_LVDS)
> gpio = GMBUS_PIN_SSC;
> @@ -464,7 +464,7 @@ void intel_dvo_init(struct drm_i915_private *dev_priv)
> * It appears that everything is on GPIOE except for panels
> * on i830 laptops, which are on GPIOB (DVOA).
> */
> - i2c = intel_gmbus_get_adapter(dev_priv, gpio);
> + i2c = intel_gmbus_get_adapter(i915, gpio);
>
> intel_dvo->dev = *dvo;
>
> @@ -480,7 +480,7 @@ void intel_dvo_init(struct drm_i915_private *dev_priv)
> * have the clock enabled before we attempt to
> * initialize the device.
> */
> - for_each_pipe(dev_priv, pipe) {
> + for_each_pipe(i915, pipe) {
> dpll[pipe] = I915_READ(DPLL(pipe));
> I915_WRITE(DPLL(pipe), dpll[pipe] | DPLL_DVO_2X_MODE);
> }
> @@ -488,7 +488,7 @@ void intel_dvo_init(struct drm_i915_private *dev_priv)
> dvoinit = dvo->dev_ops->init(&intel_dvo->dev, i2c);
>
> /* restore the DVO 2x clock state to original */
> - for_each_pipe(dev_priv, pipe) {
> + for_each_pipe(i915, pipe) {
> I915_WRITE(DPLL(pipe), dpll[pipe]);
> }
>
> @@ -498,7 +498,7 @@ void intel_dvo_init(struct drm_i915_private *dev_priv)
> continue;
>
> port = intel_dvo_port(dvo->dvo_reg);
> - drm_encoder_init(&dev_priv->drm, &intel_encoder->base,
> + drm_encoder_init(&i915->drm, &intel_encoder->base,
> &intel_dvo_enc_funcs, encoder_type,
> "DVO %c", port_name(port));
>
> @@ -511,14 +511,14 @@ void intel_dvo_init(struct drm_i915_private *dev_priv)
> case INTEL_DVO_CHIP_TMDS:
> intel_encoder->cloneable = (1 << INTEL_OUTPUT_ANALOG) |
> (1 << INTEL_OUTPUT_DVO);
> - drm_connector_init(&dev_priv->drm, connector,
> + drm_connector_init(&i915->drm, connector,
> &intel_dvo_connector_funcs,
> DRM_MODE_CONNECTOR_DVII);
> encoder_type = DRM_MODE_ENCODER_TMDS;
> break;
> case INTEL_DVO_CHIP_LVDS:
> intel_encoder->cloneable = 0;
> - drm_connector_init(&dev_priv->drm, connector,
> + drm_connector_init(&i915->drm, connector,
> &intel_dvo_connector_funcs,
> DRM_MODE_CONNECTOR_LVDS);
> encoder_type = DRM_MODE_ENCODER_LVDS;
> diff --git a/drivers/gpu/drm/i915/intel_dvo.h b/drivers/gpu/drm/i915/intel_dvo.h
> index 3ed0fdf8efff..d865cd439bb7 100644
> --- a/drivers/gpu/drm/i915/intel_dvo.h
> +++ b/drivers/gpu/drm/i915/intel_dvo.h
> @@ -8,6 +8,6 @@
>
> struct drm_i915_private;
>
> -void intel_dvo_init(struct drm_i915_private *dev_priv);
> +void intel_dvo_init(struct drm_i915_private *i915);
>
> #endif /* __INTEL_DVO_H__ */
> diff --git a/drivers/gpu/drm/i915/intel_dvo_dev.h b/drivers/gpu/drm/i915/intel_dvo_dev.h
> index 94a6ae1e0292..1d184dc7922b 100644
> --- a/drivers/gpu/drm/i915/intel_dvo_dev.h
> +++ b/drivers/gpu/drm/i915/intel_dvo_dev.h
> @@ -40,7 +40,7 @@ struct intel_dvo_device {
> int slave_addr;
>
> const struct intel_dvo_dev_ops *dev_ops;
> - void *dev_priv;
> + void *i915;
> struct i2c_adapter *i2c_bus;
> };
>
> diff --git a/drivers/gpu/drm/i915/intel_fbc.c b/drivers/gpu/drm/i915/intel_fbc.c
> index 5679f2fffb7c..627fed2a55d2 100644
> --- a/drivers/gpu/drm/i915/intel_fbc.c
> +++ b/drivers/gpu/drm/i915/intel_fbc.c
> @@ -45,14 +45,14 @@
> #include "intel_fbc.h"
> #include "intel_frontbuffer.h"
>
> -static inline bool fbc_supported(struct drm_i915_private *dev_priv)
> +static inline bool fbc_supported(struct drm_i915_private *i915)
> {
> - return HAS_FBC(dev_priv);
> + return HAS_FBC(i915);
> }
>
> -static inline bool no_fbc_on_multiple_pipes(struct drm_i915_private *dev_priv)
> +static inline bool no_fbc_on_multiple_pipes(struct drm_i915_private *i915)
> {
> - return INTEL_GEN(dev_priv) <= 3;
> + return INTEL_GEN(i915) <= 3;
> }
>
> /*
> @@ -82,22 +82,22 @@ static void intel_fbc_get_plane_source_size(struct intel_fbc_state_cache *cache,
> *height = cache->plane.src_h;
> }
>
> -static int intel_fbc_calculate_cfb_size(struct drm_i915_private *dev_priv,
> +static int intel_fbc_calculate_cfb_size(struct drm_i915_private *i915,
> struct intel_fbc_state_cache *cache)
> {
> int lines;
>
> intel_fbc_get_plane_source_size(cache, NULL, &lines);
> - if (IS_GEN(dev_priv, 7))
> + if (IS_GEN(i915, 7))
> lines = min(lines, 2048);
> - else if (INTEL_GEN(dev_priv) >= 8)
> + else if (INTEL_GEN(i915) >= 8)
> lines = min(lines, 2560);
>
> /* Hardware needs the full buffer stride, not just the active area. */
> return lines * cache->fb.stride;
> }
>
> -static void i8xx_fbc_deactivate(struct drm_i915_private *dev_priv)
> +static void i8xx_fbc_deactivate(struct drm_i915_private *i915)
> {
> u32 fbc_ctl;
>
> @@ -110,7 +110,7 @@ static void i8xx_fbc_deactivate(struct drm_i915_private *dev_priv)
> I915_WRITE(FBC_CONTROL, fbc_ctl);
>
> /* Wait for compressing bit to clear */
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> FBC_STATUS, FBC_STAT_COMPRESSING, 0,
> 10)) {
> DRM_DEBUG_KMS("FBC idle timed out\n");
> @@ -118,9 +118,9 @@ static void i8xx_fbc_deactivate(struct drm_i915_private *dev_priv)
> }
> }
>
> -static void i8xx_fbc_activate(struct drm_i915_private *dev_priv)
> +static void i8xx_fbc_activate(struct drm_i915_private *i915)
> {
> - struct intel_fbc_reg_params *params = &dev_priv->fbc.params;
> + struct intel_fbc_reg_params *params = &i915->fbc.params;
> int cfb_pitch;
> int i;
> u32 fbc_ctl;
> @@ -131,7 +131,7 @@ static void i8xx_fbc_activate(struct drm_i915_private *dev_priv)
> cfb_pitch = params->fb.stride;
>
> /* FBC_CTL wants 32B or 64B units */
> - if (IS_GEN(dev_priv, 2))
> + if (IS_GEN(i915, 2))
> cfb_pitch = (cfb_pitch / 32) - 1;
> else
> cfb_pitch = (cfb_pitch / 64) - 1;
> @@ -140,7 +140,7 @@ static void i8xx_fbc_activate(struct drm_i915_private *dev_priv)
> for (i = 0; i < (FBC_LL_SIZE / 32) + 1; i++)
> I915_WRITE(FBC_TAG(i), 0);
>
> - if (IS_GEN(dev_priv, 4)) {
> + if (IS_GEN(i915, 4)) {
> u32 fbc_ctl2;
>
> /* Set it up... */
> @@ -154,21 +154,21 @@ static void i8xx_fbc_activate(struct drm_i915_private *dev_priv)
> fbc_ctl = I915_READ(FBC_CONTROL);
> fbc_ctl &= 0x3fff << FBC_CTL_INTERVAL_SHIFT;
> fbc_ctl |= FBC_CTL_EN | FBC_CTL_PERIODIC;
> - if (IS_I945GM(dev_priv))
> + if (IS_I945GM(i915))
> fbc_ctl |= FBC_CTL_C3_IDLE; /* 945 needs special SR handling */
> fbc_ctl |= (cfb_pitch & 0xff) << FBC_CTL_STRIDE_SHIFT;
> fbc_ctl |= params->vma->fence->id;
> I915_WRITE(FBC_CONTROL, fbc_ctl);
> }
>
> -static bool i8xx_fbc_is_active(struct drm_i915_private *dev_priv)
> +static bool i8xx_fbc_is_active(struct drm_i915_private *i915)
> {
> return I915_READ(FBC_CONTROL) & FBC_CTL_EN;
> }
>
> -static void g4x_fbc_activate(struct drm_i915_private *dev_priv)
> +static void g4x_fbc_activate(struct drm_i915_private *i915)
> {
> - struct intel_fbc_reg_params *params = &dev_priv->fbc.params;
> + struct intel_fbc_reg_params *params = &i915->fbc.params;
> u32 dpfc_ctl;
>
> dpfc_ctl = DPFC_CTL_PLANE(params->crtc.i9xx_plane) | DPFC_SR_EN;
> @@ -188,7 +188,7 @@ static void g4x_fbc_activate(struct drm_i915_private *dev_priv)
> I915_WRITE(DPFC_CONTROL, dpfc_ctl | DPFC_CTL_EN);
> }
>
> -static void g4x_fbc_deactivate(struct drm_i915_private *dev_priv)
> +static void g4x_fbc_deactivate(struct drm_i915_private *i915)
> {
> u32 dpfc_ctl;
>
> @@ -200,23 +200,23 @@ static void g4x_fbc_deactivate(struct drm_i915_private *dev_priv)
> }
> }
>
> -static bool g4x_fbc_is_active(struct drm_i915_private *dev_priv)
> +static bool g4x_fbc_is_active(struct drm_i915_private *i915)
> {
> return I915_READ(DPFC_CONTROL) & DPFC_CTL_EN;
> }
>
> /* This function forces a CFB recompression through the nuke operation. */
> -static void intel_fbc_recompress(struct drm_i915_private *dev_priv)
> +static void intel_fbc_recompress(struct drm_i915_private *i915)
> {
> I915_WRITE(MSG_FBC_REND_STATE, FBC_REND_NUKE);
> POSTING_READ(MSG_FBC_REND_STATE);
> }
>
> -static void ilk_fbc_activate(struct drm_i915_private *dev_priv)
> +static void ilk_fbc_activate(struct drm_i915_private *i915)
> {
> - struct intel_fbc_reg_params *params = &dev_priv->fbc.params;
> + struct intel_fbc_reg_params *params = &i915->fbc.params;
> u32 dpfc_ctl;
> - int threshold = dev_priv->fbc.threshold;
> + int threshold = i915->fbc.threshold;
>
> dpfc_ctl = DPFC_CTL_PLANE(params->crtc.i9xx_plane);
> if (params->fb.format->cpp[0] == 2)
> @@ -237,9 +237,9 @@ static void ilk_fbc_activate(struct drm_i915_private *dev_priv)
>
> if (params->flags & PLANE_HAS_FENCE) {
> dpfc_ctl |= DPFC_CTL_FENCE_EN;
> - if (IS_GEN(dev_priv, 5))
> + if (IS_GEN(i915, 5))
> dpfc_ctl |= params->vma->fence->id;
> - if (IS_GEN(dev_priv, 6)) {
> + if (IS_GEN(i915, 6)) {
> I915_WRITE(SNB_DPFC_CTL_SA,
> SNB_CPU_FENCE_ENABLE |
> params->vma->fence->id);
> @@ -247,7 +247,7 @@ static void ilk_fbc_activate(struct drm_i915_private *dev_priv)
> params->crtc.fence_y_offset);
> }
> } else {
> - if (IS_GEN(dev_priv, 6)) {
> + if (IS_GEN(i915, 6)) {
> I915_WRITE(SNB_DPFC_CTL_SA, 0);
> I915_WRITE(DPFC_CPU_FENCE_OFFSET, 0);
> }
> @@ -259,10 +259,10 @@ static void ilk_fbc_activate(struct drm_i915_private *dev_priv)
> /* enable it... */
> I915_WRITE(ILK_DPFC_CONTROL, dpfc_ctl | DPFC_CTL_EN);
>
> - intel_fbc_recompress(dev_priv);
> + intel_fbc_recompress(i915);
> }
>
> -static void ilk_fbc_deactivate(struct drm_i915_private *dev_priv)
> +static void ilk_fbc_deactivate(struct drm_i915_private *i915)
> {
> u32 dpfc_ctl;
>
> @@ -274,19 +274,19 @@ static void ilk_fbc_deactivate(struct drm_i915_private *dev_priv)
> }
> }
>
> -static bool ilk_fbc_is_active(struct drm_i915_private *dev_priv)
> +static bool ilk_fbc_is_active(struct drm_i915_private *i915)
> {
> return I915_READ(ILK_DPFC_CONTROL) & DPFC_CTL_EN;
> }
>
> -static void gen7_fbc_activate(struct drm_i915_private *dev_priv)
> +static void gen7_fbc_activate(struct drm_i915_private *i915)
> {
> - struct intel_fbc_reg_params *params = &dev_priv->fbc.params;
> + struct intel_fbc_reg_params *params = &i915->fbc.params;
> u32 dpfc_ctl;
> - int threshold = dev_priv->fbc.threshold;
> + int threshold = i915->fbc.threshold;
>
> /* Display WA #0529: skl, kbl, bxt. */
> - if (IS_GEN(dev_priv, 9) && !IS_GEMINILAKE(dev_priv)) {
> + if (IS_GEN(i915, 9) && !IS_GEMINILAKE(i915)) {
> u32 val = I915_READ(CHICKEN_MISC_4);
>
> val &= ~(FBC_STRIDE_OVERRIDE | FBC_STRIDE_MASK);
> @@ -299,7 +299,7 @@ static void gen7_fbc_activate(struct drm_i915_private *dev_priv)
> }
>
> dpfc_ctl = 0;
> - if (IS_IVYBRIDGE(dev_priv))
> + if (IS_IVYBRIDGE(i915))
> dpfc_ctl |= IVB_DPFC_CTL_PLANE(params->crtc.i9xx_plane);
>
> if (params->fb.format->cpp[0] == 2)
> @@ -329,15 +329,15 @@ static void gen7_fbc_activate(struct drm_i915_private *dev_priv)
> I915_WRITE(DPFC_CPU_FENCE_OFFSET, 0);
> }
>
> - if (dev_priv->fbc.false_color)
> + if (i915->fbc.false_color)
> dpfc_ctl |= FBC_CTL_FALSE_COLOR;
>
> - if (IS_IVYBRIDGE(dev_priv)) {
> + if (IS_IVYBRIDGE(i915)) {
> /* WaFbcAsynchFlipDisableFbcQueue:ivb */
> I915_WRITE(ILK_DISPLAY_CHICKEN1,
> I915_READ(ILK_DISPLAY_CHICKEN1) |
> ILK_FBCQ_DIS);
> - } else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) {
> + } else if (IS_HASWELL(i915) || IS_BROADWELL(i915)) {
> /* WaFbcAsynchFlipDisableFbcQueue:hsw,bdw */
> I915_WRITE(CHICKEN_PIPESL_1(params->crtc.pipe),
> I915_READ(CHICKEN_PIPESL_1(params->crtc.pipe)) |
> @@ -346,72 +346,72 @@ static void gen7_fbc_activate(struct drm_i915_private *dev_priv)
>
> I915_WRITE(ILK_DPFC_CONTROL, dpfc_ctl | DPFC_CTL_EN);
>
> - intel_fbc_recompress(dev_priv);
> + intel_fbc_recompress(i915);
> }
>
> -static bool intel_fbc_hw_is_active(struct drm_i915_private *dev_priv)
> +static bool intel_fbc_hw_is_active(struct drm_i915_private *i915)
> {
> - if (INTEL_GEN(dev_priv) >= 5)
> - return ilk_fbc_is_active(dev_priv);
> - else if (IS_GM45(dev_priv))
> - return g4x_fbc_is_active(dev_priv);
> + if (INTEL_GEN(i915) >= 5)
> + return ilk_fbc_is_active(i915);
> + else if (IS_GM45(i915))
> + return g4x_fbc_is_active(i915);
> else
> - return i8xx_fbc_is_active(dev_priv);
> + return i8xx_fbc_is_active(i915);
> }
>
> -static void intel_fbc_hw_activate(struct drm_i915_private *dev_priv)
> +static void intel_fbc_hw_activate(struct drm_i915_private *i915)
> {
> - struct intel_fbc *fbc = &dev_priv->fbc;
> + struct intel_fbc *fbc = &i915->fbc;
>
> fbc->active = true;
>
> - if (INTEL_GEN(dev_priv) >= 7)
> - gen7_fbc_activate(dev_priv);
> - else if (INTEL_GEN(dev_priv) >= 5)
> - ilk_fbc_activate(dev_priv);
> - else if (IS_GM45(dev_priv))
> - g4x_fbc_activate(dev_priv);
> + if (INTEL_GEN(i915) >= 7)
> + gen7_fbc_activate(i915);
> + else if (INTEL_GEN(i915) >= 5)
> + ilk_fbc_activate(i915);
> + else if (IS_GM45(i915))
> + g4x_fbc_activate(i915);
> else
> - i8xx_fbc_activate(dev_priv);
> + i8xx_fbc_activate(i915);
> }
>
> -static void intel_fbc_hw_deactivate(struct drm_i915_private *dev_priv)
> +static void intel_fbc_hw_deactivate(struct drm_i915_private *i915)
> {
> - struct intel_fbc *fbc = &dev_priv->fbc;
> + struct intel_fbc *fbc = &i915->fbc;
>
> fbc->active = false;
>
> - if (INTEL_GEN(dev_priv) >= 5)
> - ilk_fbc_deactivate(dev_priv);
> - else if (IS_GM45(dev_priv))
> - g4x_fbc_deactivate(dev_priv);
> + if (INTEL_GEN(i915) >= 5)
> + ilk_fbc_deactivate(i915);
> + else if (IS_GM45(i915))
> + g4x_fbc_deactivate(i915);
> else
> - i8xx_fbc_deactivate(dev_priv);
> + i8xx_fbc_deactivate(i915);
> }
>
> /**
> * intel_fbc_is_active - Is FBC active?
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> *
> * This function is used to verify the current state of FBC.
> *
> * FIXME: This should be tracked in the plane config eventually
> * instead of queried at runtime for most callers.
> */
> -bool intel_fbc_is_active(struct drm_i915_private *dev_priv)
> +bool intel_fbc_is_active(struct drm_i915_private *i915)
> {
> - return dev_priv->fbc.active;
> + return i915->fbc.active;
> }
>
> -static void intel_fbc_deactivate(struct drm_i915_private *dev_priv,
> +static void intel_fbc_deactivate(struct drm_i915_private *i915,
> const char *reason)
> {
> - struct intel_fbc *fbc = &dev_priv->fbc;
> + struct intel_fbc *fbc = &i915->fbc;
>
> WARN_ON(!mutex_is_locked(&fbc->lock));
>
> if (fbc->active)
> - intel_fbc_hw_deactivate(dev_priv);
> + intel_fbc_hw_deactivate(i915);
>
> fbc->no_fbc_reason = reason;
> }
> @@ -419,12 +419,12 @@ static void intel_fbc_deactivate(struct drm_i915_private *dev_priv,
> static bool multiple_pipes_ok(struct intel_crtc *crtc,
> struct intel_plane_state *plane_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> - struct intel_fbc *fbc = &dev_priv->fbc;
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> + struct intel_fbc *fbc = &i915->fbc;
> enum pipe pipe = crtc->pipe;
>
> /* Don't even bother tracking anything we don't need. */
> - if (!no_fbc_on_multiple_pipes(dev_priv))
> + if (!no_fbc_on_multiple_pipes(i915))
> return true;
>
> if (plane_state->base.visible)
> @@ -435,7 +435,7 @@ static bool multiple_pipes_ok(struct intel_crtc *crtc,
> return (fbc->visible_pipes_mask & ~(1 << pipe)) != 0;
> }
>
> -static int find_compression_threshold(struct drm_i915_private *dev_priv,
> +static int find_compression_threshold(struct drm_i915_private *i915,
> struct drm_mm_node *node,
> int size,
> int fb_cpp)
> @@ -448,8 +448,8 @@ static int find_compression_threshold(struct drm_i915_private *dev_priv,
> * reserved range size, so it always assumes the maximum (8mb) is used.
> * If we enable FBC using a CFB on that memory range we'll get FIFO
> * underruns, even if that range is not reserved by the BIOS. */
> - if (IS_BROADWELL(dev_priv) || IS_GEN9_BC(dev_priv))
> - end = resource_size(&dev_priv->dsm) - 8 * 1024 * 1024;
> + if (IS_BROADWELL(i915) || IS_GEN9_BC(i915))
> + end = resource_size(&i915->dsm) - 8 * 1024 * 1024;
> else
> end = U64_MAX;
>
> @@ -461,7 +461,7 @@ static int find_compression_threshold(struct drm_i915_private *dev_priv,
> */
>
> /* Try to over-allocate to reduce reallocations and fragmentation. */
> - ret = i915_gem_stolen_insert_node_in_range(dev_priv, node, size <<= 1,
> + ret = i915_gem_stolen_insert_node_in_range(i915, node, size <<= 1,
> 4096, 0, end);
> if (ret == 0)
> return compression_threshold;
> @@ -472,9 +472,9 @@ static int find_compression_threshold(struct drm_i915_private *dev_priv,
> (fb_cpp == 2 && compression_threshold == 2))
> return 0;
>
> - ret = i915_gem_stolen_insert_node_in_range(dev_priv, node, size >>= 1,
> + ret = i915_gem_stolen_insert_node_in_range(i915, node, size >>= 1,
> 4096, 0, end);
> - if (ret && INTEL_GEN(dev_priv) <= 4) {
> + if (ret && INTEL_GEN(i915) <= 4) {
> return 0;
> } else if (ret) {
> compression_threshold <<= 1;
> @@ -486,17 +486,17 @@ static int find_compression_threshold(struct drm_i915_private *dev_priv,
>
> static int intel_fbc_alloc_cfb(struct intel_crtc *crtc)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> - struct intel_fbc *fbc = &dev_priv->fbc;
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> + struct intel_fbc *fbc = &i915->fbc;
> struct drm_mm_node *uninitialized_var(compressed_llb);
> int size, fb_cpp, ret;
>
> WARN_ON(drm_mm_node_allocated(&fbc->compressed_fb));
>
> - size = intel_fbc_calculate_cfb_size(dev_priv, &fbc->state_cache);
> + size = intel_fbc_calculate_cfb_size(i915, &fbc->state_cache);
> fb_cpp = fbc->state_cache.fb.format->cpp[0];
>
> - ret = find_compression_threshold(dev_priv, &fbc->compressed_fb,
> + ret = find_compression_threshold(i915, &fbc->compressed_fb,
> size, fb_cpp);
> if (!ret)
> goto err_llb;
> @@ -507,32 +507,32 @@ static int intel_fbc_alloc_cfb(struct intel_crtc *crtc)
>
> fbc->threshold = ret;
>
> - if (INTEL_GEN(dev_priv) >= 5)
> + if (INTEL_GEN(i915) >= 5)
> I915_WRITE(ILK_DPFC_CB_BASE, fbc->compressed_fb.start);
> - else if (IS_GM45(dev_priv)) {
> + else if (IS_GM45(i915)) {
> I915_WRITE(DPFC_CB_BASE, fbc->compressed_fb.start);
> } else {
> compressed_llb = kzalloc(sizeof(*compressed_llb), GFP_KERNEL);
> if (!compressed_llb)
> goto err_fb;
>
> - ret = i915_gem_stolen_insert_node(dev_priv, compressed_llb,
> + ret = i915_gem_stolen_insert_node(i915, compressed_llb,
> 4096, 4096);
> if (ret)
> goto err_fb;
>
> fbc->compressed_llb = compressed_llb;
>
> - GEM_BUG_ON(range_overflows_t(u64, dev_priv->dsm.start,
> + GEM_BUG_ON(range_overflows_t(u64, i915->dsm.start,
> fbc->compressed_fb.start,
> U32_MAX));
> - GEM_BUG_ON(range_overflows_t(u64, dev_priv->dsm.start,
> + GEM_BUG_ON(range_overflows_t(u64, i915->dsm.start,
> fbc->compressed_llb->start,
> U32_MAX));
> I915_WRITE(FBC_CFB_BASE,
> - dev_priv->dsm.start + fbc->compressed_fb.start);
> + i915->dsm.start + fbc->compressed_fb.start);
> I915_WRITE(FBC_LL_BASE,
> - dev_priv->dsm.start + compressed_llb->start);
> + i915->dsm.start + compressed_llb->start);
> }
>
> DRM_DEBUG_KMS("reserved %llu bytes of contiguous stolen space for FBC, threshold: %d\n",
> @@ -542,39 +542,39 @@ static int intel_fbc_alloc_cfb(struct intel_crtc *crtc)
>
> err_fb:
> kfree(compressed_llb);
> - i915_gem_stolen_remove_node(dev_priv, &fbc->compressed_fb);
> + i915_gem_stolen_remove_node(i915, &fbc->compressed_fb);
> err_llb:
> - if (drm_mm_initialized(&dev_priv->mm.stolen))
> + if (drm_mm_initialized(&i915->mm.stolen))
> pr_info_once("drm: not enough stolen space for compressed buffer (need %d more bytes), disabling. Hint: you may be able to increase stolen memory size in the BIOS to avoid this.\n", size);
> return -ENOSPC;
> }
>
> -static void __intel_fbc_cleanup_cfb(struct drm_i915_private *dev_priv)
> +static void __intel_fbc_cleanup_cfb(struct drm_i915_private *i915)
> {
> - struct intel_fbc *fbc = &dev_priv->fbc;
> + struct intel_fbc *fbc = &i915->fbc;
>
> if (drm_mm_node_allocated(&fbc->compressed_fb))
> - i915_gem_stolen_remove_node(dev_priv, &fbc->compressed_fb);
> + i915_gem_stolen_remove_node(i915, &fbc->compressed_fb);
>
> if (fbc->compressed_llb) {
> - i915_gem_stolen_remove_node(dev_priv, fbc->compressed_llb);
> + i915_gem_stolen_remove_node(i915, fbc->compressed_llb);
> kfree(fbc->compressed_llb);
> }
> }
>
> -void intel_fbc_cleanup_cfb(struct drm_i915_private *dev_priv)
> +void intel_fbc_cleanup_cfb(struct drm_i915_private *i915)
> {
> - struct intel_fbc *fbc = &dev_priv->fbc;
> + struct intel_fbc *fbc = &i915->fbc;
>
> - if (!fbc_supported(dev_priv))
> + if (!fbc_supported(i915))
> return;
>
> mutex_lock(&fbc->lock);
> - __intel_fbc_cleanup_cfb(dev_priv);
> + __intel_fbc_cleanup_cfb(i915);
> mutex_unlock(&fbc->lock);
> }
>
> -static bool stride_is_valid(struct drm_i915_private *dev_priv,
> +static bool stride_is_valid(struct drm_i915_private *i915,
> unsigned int stride)
> {
> /* This should have been caught earlier. */
> @@ -585,10 +585,10 @@ static bool stride_is_valid(struct drm_i915_private *dev_priv,
> if (stride < 512)
> return false;
>
> - if (IS_GEN(dev_priv, 2) || IS_GEN(dev_priv, 3))
> + if (IS_GEN(i915, 2) || IS_GEN(i915, 3))
> return stride == 4096 || stride == 8192;
>
> - if (IS_GEN(dev_priv, 4) && !IS_G4X(dev_priv) && stride < 2048)
> + if (IS_GEN(i915, 4) && !IS_G4X(i915) && stride < 2048)
> return false;
>
> if (stride > 16384)
> @@ -597,7 +597,7 @@ static bool stride_is_valid(struct drm_i915_private *dev_priv,
> return true;
> }
>
> -static bool pixel_format_is_valid(struct drm_i915_private *dev_priv,
> +static bool pixel_format_is_valid(struct drm_i915_private *i915,
> u32 pixel_format)
> {
> switch (pixel_format) {
> @@ -607,10 +607,10 @@ static bool pixel_format_is_valid(struct drm_i915_private *dev_priv,
> case DRM_FORMAT_XRGB1555:
> case DRM_FORMAT_RGB565:
> /* 16bpp not supported on gen2 */
> - if (IS_GEN(dev_priv, 2))
> + if (IS_GEN(i915, 2))
> return false;
> /* WaFbcOnly1to1Ratio:ctg */
> - if (IS_G4X(dev_priv))
> + if (IS_G4X(i915))
> return false;
> return true;
> default:
> @@ -626,17 +626,17 @@ static bool pixel_format_is_valid(struct drm_i915_private *dev_priv,
> */
> static bool intel_fbc_hw_tracking_covers_screen(struct intel_crtc *crtc)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> - struct intel_fbc *fbc = &dev_priv->fbc;
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> + struct intel_fbc *fbc = &i915->fbc;
> unsigned int effective_w, effective_h, max_w, max_h;
>
> - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) {
> + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) {
> max_w = 5120;
> max_h = 4096;
> - } else if (INTEL_GEN(dev_priv) >= 8 || IS_HASWELL(dev_priv)) {
> + } else if (INTEL_GEN(i915) >= 8 || IS_HASWELL(i915)) {
> max_w = 4096;
> max_h = 4096;
> - } else if (IS_G4X(dev_priv) || INTEL_GEN(dev_priv) >= 5) {
> + } else if (IS_G4X(i915) || INTEL_GEN(i915) >= 5) {
> max_w = 4096;
> max_h = 2048;
> } else {
> @@ -656,8 +656,8 @@ static void intel_fbc_update_state_cache(struct intel_crtc *crtc,
> struct intel_crtc_state *crtc_state,
> struct intel_plane_state *plane_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> - struct intel_fbc *fbc = &dev_priv->fbc;
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> + struct intel_fbc *fbc = &i915->fbc;
> struct intel_fbc_state_cache *cache = &fbc->state_cache;
> struct drm_framebuffer *fb = plane_state->base.fb;
>
> @@ -665,7 +665,7 @@ static void intel_fbc_update_state_cache(struct intel_crtc *crtc,
> cache->flags = 0;
>
> cache->crtc.mode_flags = crtc_state->base.adjusted_mode.flags;
> - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> + if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> cache->crtc.hsw_bdw_pixel_rate = crtc_state->pixel_rate;
>
> cache->plane.rotation = plane_state->base.rotation;
> @@ -697,8 +697,8 @@ static void intel_fbc_update_state_cache(struct intel_crtc *crtc,
>
> static bool intel_fbc_can_activate(struct intel_crtc *crtc)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> - struct intel_fbc *fbc = &dev_priv->fbc;
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> + struct intel_fbc *fbc = &i915->fbc;
> struct intel_fbc_state_cache *cache = &fbc->state_cache;
>
> /* We don't need to use a state cache here since this information is
> @@ -741,18 +741,18 @@ static bool intel_fbc_can_activate(struct intel_crtc *crtc)
> fbc->no_fbc_reason = "framebuffer not tiled or fenced";
> return false;
> }
> - if (INTEL_GEN(dev_priv) <= 4 && !IS_G4X(dev_priv) &&
> + if (INTEL_GEN(i915) <= 4 && !IS_G4X(i915) &&
> cache->plane.rotation != DRM_MODE_ROTATE_0) {
> fbc->no_fbc_reason = "rotation unsupported";
> return false;
> }
>
> - if (!stride_is_valid(dev_priv, cache->fb.stride)) {
> + if (!stride_is_valid(i915, cache->fb.stride)) {
> fbc->no_fbc_reason = "framebuffer stride not supported";
> return false;
> }
>
> - if (!pixel_format_is_valid(dev_priv, cache->fb.format->format)) {
> + if (!pixel_format_is_valid(i915, cache->fb.format->format)) {
> fbc->no_fbc_reason = "pixel format is invalid";
> return false;
> }
> @@ -764,8 +764,8 @@ static bool intel_fbc_can_activate(struct intel_crtc *crtc)
> }
>
> /* WaFbcExceedCdClockThreshold:hsw,bdw */
> - if ((IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) &&
> - cache->crtc.hsw_bdw_pixel_rate >= dev_priv->cdclk.hw.cdclk * 95 / 100) {
> + if ((IS_HASWELL(i915) || IS_BROADWELL(i915)) &&
> + cache->crtc.hsw_bdw_pixel_rate >= i915->cdclk.hw.cdclk * 95 / 100) {
> fbc->no_fbc_reason = "pixel rate is too big";
> return false;
> }
> @@ -780,7 +780,7 @@ static bool intel_fbc_can_activate(struct intel_crtc *crtc)
> * we didn't get any invalidate/deactivate calls, but this would require
> * a lot of tracking just for a specific case. If we conclude it's an
> * important case, we can implement it later. */
> - if (intel_fbc_calculate_cfb_size(dev_priv, &fbc->state_cache) >
> + if (intel_fbc_calculate_cfb_size(i915, &fbc->state_cache) >
> fbc->compressed_fb.size * fbc->threshold) {
> fbc->no_fbc_reason = "CFB requirements changed";
> return false;
> @@ -791,7 +791,7 @@ static bool intel_fbc_can_activate(struct intel_crtc *crtc)
> * having a Y offset that isn't divisible by 4 causes FIFO underrun
> * and screen flicker.
> */
> - if (IS_GEN_RANGE(dev_priv, 9, 10) &&
> + if (IS_GEN_RANGE(i915, 9, 10) &&
> (fbc->state_cache.plane.adjusted_y & 3)) {
> fbc->no_fbc_reason = "plane Y offset is misaligned";
> return false;
> @@ -800,11 +800,11 @@ static bool intel_fbc_can_activate(struct intel_crtc *crtc)
> return true;
> }
>
> -static bool intel_fbc_can_enable(struct drm_i915_private *dev_priv)
> +static bool intel_fbc_can_enable(struct drm_i915_private *i915)
> {
> - struct intel_fbc *fbc = &dev_priv->fbc;
> + struct intel_fbc *fbc = &i915->fbc;
>
> - if (intel_vgpu_active(dev_priv)) {
> + if (intel_vgpu_active(i915)) {
> fbc->no_fbc_reason = "VGPU is active";
> return false;
> }
> @@ -825,8 +825,8 @@ static bool intel_fbc_can_enable(struct drm_i915_private *dev_priv)
> static void intel_fbc_get_reg_params(struct intel_crtc *crtc,
> struct intel_fbc_reg_params *params)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> - struct intel_fbc *fbc = &dev_priv->fbc;
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> + struct intel_fbc *fbc = &i915->fbc;
> struct intel_fbc_state_cache *cache = &fbc->state_cache;
>
> /* Since all our fields are integer types, use memset here so the
> @@ -844,9 +844,9 @@ static void intel_fbc_get_reg_params(struct intel_crtc *crtc,
> params->fb.format = cache->fb.format;
> params->fb.stride = cache->fb.stride;
>
> - params->cfb_size = intel_fbc_calculate_cfb_size(dev_priv, cache);
> + params->cfb_size = intel_fbc_calculate_cfb_size(i915, cache);
>
> - if (IS_GEN(dev_priv, 9) && !IS_GEMINILAKE(dev_priv))
> + if (IS_GEN(i915, 9) && !IS_GEMINILAKE(i915))
> params->gen9_wa_cfb_stride = DIV_ROUND_UP(cache->plane.src_w,
> 32 * fbc->threshold) * 8;
> }
> @@ -855,11 +855,11 @@ void intel_fbc_pre_update(struct intel_crtc *crtc,
> struct intel_crtc_state *crtc_state,
> struct intel_plane_state *plane_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> - struct intel_fbc *fbc = &dev_priv->fbc;
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> + struct intel_fbc *fbc = &i915->fbc;
> const char *reason = "update pending";
>
> - if (!fbc_supported(dev_priv))
> + if (!fbc_supported(i915))
> return;
>
> mutex_lock(&fbc->lock);
> @@ -876,21 +876,21 @@ void intel_fbc_pre_update(struct intel_crtc *crtc,
> fbc->flip_pending = true;
>
> deactivate:
> - intel_fbc_deactivate(dev_priv, reason);
> + intel_fbc_deactivate(i915, reason);
> unlock:
> mutex_unlock(&fbc->lock);
> }
>
> /**
> * __intel_fbc_disable - disable FBC
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> *
> * This is the low level function that actually disables FBC. Callers should
> * grab the FBC lock.
> */
> -static void __intel_fbc_disable(struct drm_i915_private *dev_priv)
> +static void __intel_fbc_disable(struct drm_i915_private *i915)
> {
> - struct intel_fbc *fbc = &dev_priv->fbc;
> + struct intel_fbc *fbc = &i915->fbc;
> struct intel_crtc *crtc = fbc->crtc;
>
> WARN_ON(!mutex_is_locked(&fbc->lock));
> @@ -899,7 +899,7 @@ static void __intel_fbc_disable(struct drm_i915_private *dev_priv)
>
> DRM_DEBUG_KMS("Disabling FBC on pipe %c\n", pipe_name(crtc->pipe));
>
> - __intel_fbc_cleanup_cfb(dev_priv);
> + __intel_fbc_cleanup_cfb(i915);
>
> fbc->enabled = false;
> fbc->crtc = NULL;
> @@ -907,8 +907,8 @@ static void __intel_fbc_disable(struct drm_i915_private *dev_priv)
>
> static void __intel_fbc_post_update(struct intel_crtc *crtc)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> - struct intel_fbc *fbc = &dev_priv->fbc;
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> + struct intel_fbc *fbc = &i915->fbc;
>
> WARN_ON(!mutex_is_locked(&fbc->lock));
>
> @@ -919,8 +919,8 @@ static void __intel_fbc_post_update(struct intel_crtc *crtc)
> WARN_ON(fbc->active);
>
> if (!i915_modparams.enable_fbc) {
> - intel_fbc_deactivate(dev_priv, "disabled at runtime per module param");
> - __intel_fbc_disable(dev_priv);
> + intel_fbc_deactivate(i915, "disabled at runtime per module param");
> + __intel_fbc_disable(i915);
>
> return;
> }
> @@ -931,18 +931,18 @@ static void __intel_fbc_post_update(struct intel_crtc *crtc)
> return;
>
> if (!fbc->busy_bits) {
> - intel_fbc_deactivate(dev_priv, "FBC enabled (active or scheduled)");
> - intel_fbc_hw_activate(dev_priv);
> + intel_fbc_deactivate(i915, "FBC enabled (active or scheduled)");
> + intel_fbc_hw_activate(i915);
> } else
> - intel_fbc_deactivate(dev_priv, "frontbuffer write");
> + intel_fbc_deactivate(i915, "frontbuffer write");
> }
>
> void intel_fbc_post_update(struct intel_crtc *crtc)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> - struct intel_fbc *fbc = &dev_priv->fbc;
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> + struct intel_fbc *fbc = &i915->fbc;
>
> - if (!fbc_supported(dev_priv))
> + if (!fbc_supported(i915))
> return;
>
> mutex_lock(&fbc->lock);
> @@ -958,13 +958,13 @@ static unsigned int intel_fbc_get_frontbuffer_bit(struct intel_fbc *fbc)
> return fbc->possible_framebuffer_bits;
> }
>
> -void intel_fbc_invalidate(struct drm_i915_private *dev_priv,
> +void intel_fbc_invalidate(struct drm_i915_private *i915,
> unsigned int frontbuffer_bits,
> enum fb_op_origin origin)
> {
> - struct intel_fbc *fbc = &dev_priv->fbc;
> + struct intel_fbc *fbc = &i915->fbc;
>
> - if (!fbc_supported(dev_priv))
> + if (!fbc_supported(i915))
> return;
>
> if (origin == ORIGIN_GTT || origin == ORIGIN_FLIP)
> @@ -975,17 +975,17 @@ void intel_fbc_invalidate(struct drm_i915_private *dev_priv,
> fbc->busy_bits |= intel_fbc_get_frontbuffer_bit(fbc) & frontbuffer_bits;
>
> if (fbc->enabled && fbc->busy_bits)
> - intel_fbc_deactivate(dev_priv, "frontbuffer write");
> + intel_fbc_deactivate(i915, "frontbuffer write");
>
> mutex_unlock(&fbc->lock);
> }
>
> -void intel_fbc_flush(struct drm_i915_private *dev_priv,
> +void intel_fbc_flush(struct drm_i915_private *i915,
> unsigned int frontbuffer_bits, enum fb_op_origin origin)
> {
> - struct intel_fbc *fbc = &dev_priv->fbc;
> + struct intel_fbc *fbc = &i915->fbc;
>
> - if (!fbc_supported(dev_priv))
> + if (!fbc_supported(i915))
> return;
>
> mutex_lock(&fbc->lock);
> @@ -998,7 +998,7 @@ void intel_fbc_flush(struct drm_i915_private *dev_priv,
> if (!fbc->busy_bits && fbc->enabled &&
> (frontbuffer_bits & intel_fbc_get_frontbuffer_bit(fbc))) {
> if (fbc->active)
> - intel_fbc_recompress(dev_priv);
> + intel_fbc_recompress(i915);
> else if (!fbc->flip_pending)
> __intel_fbc_post_update(fbc->crtc);
> }
> @@ -1009,7 +1009,7 @@ void intel_fbc_flush(struct drm_i915_private *dev_priv,
>
> /**
> * intel_fbc_choose_crtc - select a CRTC to enable FBC on
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> * @state: the atomic state structure
> *
> * This function looks at the proposed state for CRTCs and planes, then chooses
> @@ -1017,12 +1017,12 @@ void intel_fbc_flush(struct drm_i915_private *dev_priv,
> * true.
> *
> * Later, intel_fbc_enable is going to look for state->enable_fbc and then maybe
> - * enable FBC for the chosen CRTC. If it does, it will set dev_priv->fbc.crtc.
> + * enable FBC for the chosen CRTC. If it does, it will set i915->fbc.crtc.
> */
> -void intel_fbc_choose_crtc(struct drm_i915_private *dev_priv,
> +void intel_fbc_choose_crtc(struct drm_i915_private *i915,
> struct intel_atomic_state *state)
> {
> - struct intel_fbc *fbc = &dev_priv->fbc;
> + struct intel_fbc *fbc = &i915->fbc;
> struct intel_plane *plane;
> struct intel_plane_state *plane_state;
> bool crtc_chosen = false;
> @@ -1035,7 +1035,7 @@ void intel_fbc_choose_crtc(struct drm_i915_private *dev_priv,
> !intel_atomic_get_new_crtc_state(state, fbc->crtc))
> goto out;
>
> - if (!intel_fbc_can_enable(dev_priv))
> + if (!intel_fbc_can_enable(i915))
> goto out;
>
> /* Simply choose the first CRTC that is compatible and has a visible
> @@ -1081,10 +1081,10 @@ void intel_fbc_enable(struct intel_crtc *crtc,
> struct intel_crtc_state *crtc_state,
> struct intel_plane_state *plane_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> - struct intel_fbc *fbc = &dev_priv->fbc;
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> + struct intel_fbc *fbc = &i915->fbc;
>
> - if (!fbc_supported(dev_priv))
> + if (!fbc_supported(i915))
> return;
>
> mutex_lock(&fbc->lock);
> @@ -1127,44 +1127,44 @@ void intel_fbc_enable(struct intel_crtc *crtc,
> */
> void intel_fbc_disable(struct intel_crtc *crtc)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> - struct intel_fbc *fbc = &dev_priv->fbc;
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> + struct intel_fbc *fbc = &i915->fbc;
>
> - if (!fbc_supported(dev_priv))
> + if (!fbc_supported(i915))
> return;
>
> mutex_lock(&fbc->lock);
> if (fbc->crtc == crtc)
> - __intel_fbc_disable(dev_priv);
> + __intel_fbc_disable(i915);
> mutex_unlock(&fbc->lock);
> }
>
> /**
> * intel_fbc_global_disable - globally disable FBC
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> *
> * This function disables FBC regardless of which CRTC is associated with it.
> */
> -void intel_fbc_global_disable(struct drm_i915_private *dev_priv)
> +void intel_fbc_global_disable(struct drm_i915_private *i915)
> {
> - struct intel_fbc *fbc = &dev_priv->fbc;
> + struct intel_fbc *fbc = &i915->fbc;
>
> - if (!fbc_supported(dev_priv))
> + if (!fbc_supported(i915))
> return;
>
> mutex_lock(&fbc->lock);
> if (fbc->enabled) {
> WARN_ON(fbc->crtc->active);
> - __intel_fbc_disable(dev_priv);
> + __intel_fbc_disable(i915);
> }
> mutex_unlock(&fbc->lock);
> }
>
> static void intel_fbc_underrun_work_fn(struct work_struct *work)
> {
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> container_of(work, struct drm_i915_private, fbc.underrun_work);
> - struct intel_fbc *fbc = &dev_priv->fbc;
> + struct intel_fbc *fbc = &i915->fbc;
>
> mutex_lock(&fbc->lock);
>
> @@ -1175,42 +1175,42 @@ static void intel_fbc_underrun_work_fn(struct work_struct *work)
> DRM_DEBUG_KMS("Disabling FBC due to FIFO underrun.\n");
> fbc->underrun_detected = true;
>
> - intel_fbc_deactivate(dev_priv, "FIFO underrun");
> + intel_fbc_deactivate(i915, "FIFO underrun");
> out:
> mutex_unlock(&fbc->lock);
> }
>
> /*
> * intel_fbc_reset_underrun - reset FBC fifo underrun status.
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> *
> * See intel_fbc_handle_fifo_underrun_irq(). For automated testing we
> * want to re-enable FBC after an underrun to increase test coverage.
> */
> -int intel_fbc_reset_underrun(struct drm_i915_private *dev_priv)
> +int intel_fbc_reset_underrun(struct drm_i915_private *i915)
> {
> int ret;
>
> - cancel_work_sync(&dev_priv->fbc.underrun_work);
> + cancel_work_sync(&i915->fbc.underrun_work);
>
> - ret = mutex_lock_interruptible(&dev_priv->fbc.lock);
> + ret = mutex_lock_interruptible(&i915->fbc.lock);
> if (ret)
> return ret;
>
> - if (dev_priv->fbc.underrun_detected) {
> + if (i915->fbc.underrun_detected) {
> DRM_DEBUG_KMS("Re-allowing FBC after fifo underrun\n");
> - dev_priv->fbc.no_fbc_reason = "FIFO underrun cleared";
> + i915->fbc.no_fbc_reason = "FIFO underrun cleared";
> }
>
> - dev_priv->fbc.underrun_detected = false;
> - mutex_unlock(&dev_priv->fbc.lock);
> + i915->fbc.underrun_detected = false;
> + mutex_unlock(&i915->fbc.lock);
>
> return 0;
> }
>
> /**
> * intel_fbc_handle_fifo_underrun_irq - disable FBC when we get a FIFO underrun
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> *
> * Without FBC, most underruns are harmless and don't really cause too many
> * problems, except for an annoying message on dmesg. With FBC, underruns can
> @@ -1222,11 +1222,11 @@ int intel_fbc_reset_underrun(struct drm_i915_private *dev_priv)
> *
> * This function is called from the IRQ handler.
> */
> -void intel_fbc_handle_fifo_underrun_irq(struct drm_i915_private *dev_priv)
> +void intel_fbc_handle_fifo_underrun_irq(struct drm_i915_private *i915)
> {
> - struct intel_fbc *fbc = &dev_priv->fbc;
> + struct intel_fbc *fbc = &i915->fbc;
>
> - if (!fbc_supported(dev_priv))
> + if (!fbc_supported(i915))
> return;
>
> /* There's no guarantee that underrun_detected won't be set to true
> @@ -1243,24 +1243,24 @@ void intel_fbc_handle_fifo_underrun_irq(struct drm_i915_private *dev_priv)
>
> /**
> * intel_fbc_init_pipe_state - initialize FBC's CRTC visibility tracking
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> *
> * The FBC code needs to track CRTC visibility since the older platforms can't
> * have FBC enabled while multiple pipes are used. This function does the
> * initial setup at driver load to make sure FBC is matching the real hardware.
> */
> -void intel_fbc_init_pipe_state(struct drm_i915_private *dev_priv)
> +void intel_fbc_init_pipe_state(struct drm_i915_private *i915)
> {
> struct intel_crtc *crtc;
>
> /* Don't even bother tracking anything if we don't need. */
> - if (!no_fbc_on_multiple_pipes(dev_priv))
> + if (!no_fbc_on_multiple_pipes(i915))
> return;
>
> - for_each_intel_crtc(&dev_priv->drm, crtc)
> + for_each_intel_crtc(&i915->drm, crtc)
> if (intel_crtc_active(crtc) &&
> crtc->base.primary->state->visible)
> - dev_priv->fbc.visible_pipes_mask |= (1 << crtc->pipe);
> + i915->fbc.visible_pipes_mask |= (1 << crtc->pipe);
> }
>
> /*
> @@ -1272,29 +1272,29 @@ void intel_fbc_init_pipe_state(struct drm_i915_private *dev_priv)
> * space to change the value during runtime without sanitizing it again. IGT
> * relies on being able to change i915.enable_fbc at runtime.
> */
> -static int intel_sanitize_fbc_option(struct drm_i915_private *dev_priv)
> +static int intel_sanitize_fbc_option(struct drm_i915_private *i915)
> {
> if (i915_modparams.enable_fbc >= 0)
> return !!i915_modparams.enable_fbc;
>
> - if (!HAS_FBC(dev_priv))
> + if (!HAS_FBC(i915))
> return 0;
>
> /* https://bugs.freedesktop.org/show_bug.cgi?id=108085 */
> - if (IS_GEMINILAKE(dev_priv))
> + if (IS_GEMINILAKE(i915))
> return 0;
>
> - if (IS_BROADWELL(dev_priv) || INTEL_GEN(dev_priv) >= 9)
> + if (IS_BROADWELL(i915) || INTEL_GEN(i915) >= 9)
> return 1;
>
> return 0;
> }
>
> -static bool need_fbc_vtd_wa(struct drm_i915_private *dev_priv)
> +static bool need_fbc_vtd_wa(struct drm_i915_private *i915)
> {
> /* WaFbcTurnOffFbcWhenHyperVisorIsUsed:skl,bxt */
> if (intel_vtd_active() &&
> - (IS_SKYLAKE(dev_priv) || IS_BROXTON(dev_priv))) {
> + (IS_SKYLAKE(i915) || IS_BROXTON(i915))) {
> DRM_INFO("Disabling framebuffer compression (FBC) to prevent screen flicker with VT-d enabled\n");
> return true;
> }
> @@ -1304,38 +1304,38 @@ static bool need_fbc_vtd_wa(struct drm_i915_private *dev_priv)
>
> /**
> * intel_fbc_init - Initialize FBC
> - * @dev_priv: the i915 device
> + * @i915: the i915 device
> *
> * This function might be called during PM init process.
> */
> -void intel_fbc_init(struct drm_i915_private *dev_priv)
> +void intel_fbc_init(struct drm_i915_private *i915)
> {
> - struct intel_fbc *fbc = &dev_priv->fbc;
> + struct intel_fbc *fbc = &i915->fbc;
>
> INIT_WORK(&fbc->underrun_work, intel_fbc_underrun_work_fn);
> mutex_init(&fbc->lock);
> fbc->enabled = false;
> fbc->active = false;
>
> - if (need_fbc_vtd_wa(dev_priv))
> - mkwrite_device_info(dev_priv)->display.has_fbc = false;
> + if (need_fbc_vtd_wa(i915))
> + mkwrite_device_info(i915)->display.has_fbc = false;
>
> - i915_modparams.enable_fbc = intel_sanitize_fbc_option(dev_priv);
> + i915_modparams.enable_fbc = intel_sanitize_fbc_option(i915);
> DRM_DEBUG_KMS("Sanitized enable_fbc value: %d\n",
> i915_modparams.enable_fbc);
>
> - if (!HAS_FBC(dev_priv)) {
> + if (!HAS_FBC(i915)) {
> fbc->no_fbc_reason = "unsupported by this chipset";
> return;
> }
>
> /* This value was pulled out of someone's hat */
> - if (INTEL_GEN(dev_priv) <= 4 && !IS_GM45(dev_priv))
> + if (INTEL_GEN(i915) <= 4 && !IS_GM45(i915))
> I915_WRITE(FBC_CONTROL, 500 << FBC_CTL_INTERVAL_SHIFT);
>
> /* We still don't have any sort of hardware state readout for FBC, so
> * deactivate it in case the BIOS activated it to make sure software
> * matches the hardware state. */
> - if (intel_fbc_hw_is_active(dev_priv))
> - intel_fbc_hw_deactivate(dev_priv);
> + if (intel_fbc_hw_is_active(i915))
> + intel_fbc_hw_deactivate(i915);
> }
> diff --git a/drivers/gpu/drm/i915/intel_fbc.h b/drivers/gpu/drm/i915/intel_fbc.h
> index 50272eda8d43..b7dc6cb43497 100644
> --- a/drivers/gpu/drm/i915/intel_fbc.h
> +++ b/drivers/gpu/drm/i915/intel_fbc.h
> @@ -16,27 +16,27 @@ struct intel_crtc;
> struct intel_crtc_state;
> struct intel_plane_state;
>
> -void intel_fbc_choose_crtc(struct drm_i915_private *dev_priv,
> +void intel_fbc_choose_crtc(struct drm_i915_private *i915,
> struct intel_atomic_state *state);
> -bool intel_fbc_is_active(struct drm_i915_private *dev_priv);
> +bool intel_fbc_is_active(struct drm_i915_private *i915);
> void intel_fbc_pre_update(struct intel_crtc *crtc,
> struct intel_crtc_state *crtc_state,
> struct intel_plane_state *plane_state);
> void intel_fbc_post_update(struct intel_crtc *crtc);
> -void intel_fbc_init(struct drm_i915_private *dev_priv);
> -void intel_fbc_init_pipe_state(struct drm_i915_private *dev_priv);
> +void intel_fbc_init(struct drm_i915_private *i915);
> +void intel_fbc_init_pipe_state(struct drm_i915_private *i915);
> void intel_fbc_enable(struct intel_crtc *crtc,
> struct intel_crtc_state *crtc_state,
> struct intel_plane_state *plane_state);
> void intel_fbc_disable(struct intel_crtc *crtc);
> -void intel_fbc_global_disable(struct drm_i915_private *dev_priv);
> -void intel_fbc_invalidate(struct drm_i915_private *dev_priv,
> +void intel_fbc_global_disable(struct drm_i915_private *i915);
> +void intel_fbc_invalidate(struct drm_i915_private *i915,
> unsigned int frontbuffer_bits,
> enum fb_op_origin origin);
> -void intel_fbc_flush(struct drm_i915_private *dev_priv,
> +void intel_fbc_flush(struct drm_i915_private *i915,
> unsigned int frontbuffer_bits, enum fb_op_origin origin);
> -void intel_fbc_cleanup_cfb(struct drm_i915_private *dev_priv);
> -void intel_fbc_handle_fifo_underrun_irq(struct drm_i915_private *dev_priv);
> -int intel_fbc_reset_underrun(struct drm_i915_private *dev_priv);
> +void intel_fbc_cleanup_cfb(struct drm_i915_private *i915);
> +void intel_fbc_handle_fifo_underrun_irq(struct drm_i915_private *i915);
> +int intel_fbc_reset_underrun(struct drm_i915_private *i915);
>
> #endif /* __INTEL_FBC_H__ */
> diff --git a/drivers/gpu/drm/i915/intel_fbdev.c b/drivers/gpu/drm/i915/intel_fbdev.c
> index 0d3a6fa674e6..93dee9ff5a58 100644
> --- a/drivers/gpu/drm/i915/intel_fbdev.c
> +++ b/drivers/gpu/drm/i915/intel_fbdev.c
> @@ -117,7 +117,7 @@ static int intelfb_alloc(struct drm_fb_helper *helper,
> container_of(helper, struct intel_fbdev, helper);
> struct drm_framebuffer *fb;
> struct drm_device *dev = helper->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct drm_mode_fb_cmd2 mode_cmd = {};
> struct drm_i915_gem_object *obj;
> int size, ret;
> @@ -141,10 +141,10 @@ static int intelfb_alloc(struct drm_fb_helper *helper,
> * important and we should probably use that space with FBC or other
> * features. */
> obj = NULL;
> - if (size * 2 < dev_priv->stolen_usable_size)
> - obj = i915_gem_object_create_stolen(dev_priv, size);
> + if (size * 2 < i915->stolen_usable_size)
> + obj = i915_gem_object_create_stolen(i915, size);
> if (obj == NULL)
> - obj = i915_gem_object_create_shmem(dev_priv, size);
> + obj = i915_gem_object_create_shmem(i915, size);
> if (IS_ERR(obj)) {
> DRM_ERROR("failed to allocate framebuffer\n");
> ret = PTR_ERR(obj);
> @@ -174,9 +174,9 @@ static int intelfb_create(struct drm_fb_helper *helper,
> container_of(helper, struct intel_fbdev, helper);
> struct intel_framebuffer *intel_fb = ifbdev->fb;
> struct drm_device *dev = helper->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> - struct i915_ggtt *ggtt = &dev_priv->ggtt;
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct pci_dev *pdev = i915->drm.pdev;
> + struct i915_ggtt *ggtt = &i915->ggtt;
> const struct i915_ggtt_view view = {
> .type = I915_GGTT_VIEW_NORMAL,
> };
> @@ -213,7 +213,7 @@ static int intelfb_create(struct drm_fb_helper *helper,
> }
>
> mutex_lock(&dev->struct_mutex);
> - wakeref = intel_runtime_pm_get(dev_priv);
> + wakeref = intel_runtime_pm_get(i915);
>
> /* Pin the GGTT vma for our access via info->screen_base.
> * This also validates that any existing fb inherited from the
> @@ -272,7 +272,7 @@ static int intelfb_create(struct drm_fb_helper *helper,
> ifbdev->vma = vma;
> ifbdev->vma_flags = flags;
>
> - intel_runtime_pm_put(dev_priv, wakeref);
> + intel_runtime_pm_put(i915, wakeref);
> mutex_unlock(&dev->struct_mutex);
> vga_switcheroo_client_fb_set(pdev, info);
> return 0;
> @@ -280,7 +280,7 @@ static int intelfb_create(struct drm_fb_helper *helper,
> out_unpin:
> intel_unpin_fb_vma(vma, flags);
> out_unlock:
> - intel_runtime_pm_put(dev_priv, wakeref);
> + intel_runtime_pm_put(i915, wakeref);
> mutex_unlock(&dev->struct_mutex);
> return ret;
> }
> @@ -447,11 +447,11 @@ static void intel_fbdev_suspend_worker(struct work_struct *work)
>
> int intel_fbdev_init(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_fbdev *ifbdev;
> int ret;
>
> - if (WARN_ON(!HAS_DISPLAY(dev_priv)))
> + if (WARN_ON(!HAS_DISPLAY(i915)))
> return -ENODEV;
>
> ifbdev = kzalloc(sizeof(struct intel_fbdev), GFP_KERNEL);
> @@ -470,8 +470,8 @@ int intel_fbdev_init(struct drm_device *dev)
> return ret;
> }
>
> - dev_priv->fbdev = ifbdev;
> - INIT_WORK(&dev_priv->fbdev_suspend_work, intel_fbdev_suspend_worker);
> + i915->fbdev = ifbdev;
> + INIT_WORK(&i915->fbdev_suspend_work, intel_fbdev_suspend_worker);
>
> drm_fb_helper_single_add_all_connectors(&ifbdev->helper);
>
> @@ -508,23 +508,23 @@ static void intel_fbdev_sync(struct intel_fbdev *ifbdev)
> ifbdev->cookie = 0;
> }
>
> -void intel_fbdev_unregister(struct drm_i915_private *dev_priv)
> +void intel_fbdev_unregister(struct drm_i915_private *i915)
> {
> - struct intel_fbdev *ifbdev = dev_priv->fbdev;
> + struct intel_fbdev *ifbdev = i915->fbdev;
>
> if (!ifbdev)
> return;
>
> - cancel_work_sync(&dev_priv->fbdev_suspend_work);
> + cancel_work_sync(&i915->fbdev_suspend_work);
> if (!current_is_async())
> intel_fbdev_sync(ifbdev);
>
> drm_fb_helper_unregister_fbi(&ifbdev->helper);
> }
>
> -void intel_fbdev_fini(struct drm_i915_private *dev_priv)
> +void intel_fbdev_fini(struct drm_i915_private *i915)
> {
> - struct intel_fbdev *ifbdev = fetch_and_zero(&dev_priv->fbdev);
> + struct intel_fbdev *ifbdev = fetch_and_zero(&i915->fbdev);
>
> if (!ifbdev)
> return;
> @@ -554,8 +554,8 @@ static void intel_fbdev_hpd_set_suspend(struct intel_fbdev *ifbdev, int state)
>
> void intel_fbdev_set_suspend(struct drm_device *dev, int state, bool synchronous)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct intel_fbdev *ifbdev = dev_priv->fbdev;
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct intel_fbdev *ifbdev = i915->fbdev;
> struct fb_info *info;
>
> if (!ifbdev || !ifbdev->vma)
> @@ -572,7 +572,7 @@ void intel_fbdev_set_suspend(struct drm_device *dev, int state, bool synchronous
> * ourselves, so only flush outstanding work upon suspend!
> */
> if (state != FBINFO_STATE_RUNNING)
> - flush_work(&dev_priv->fbdev_suspend_work);
> + flush_work(&i915->fbdev_suspend_work);
>
> console_lock();
> } else {
> @@ -586,7 +586,7 @@ void intel_fbdev_set_suspend(struct drm_device *dev, int state, bool synchronous
> /* Don't block our own workqueue as this can
> * be run in parallel with other i915.ko tasks.
> */
> - schedule_work(&dev_priv->fbdev_suspend_work);
> + schedule_work(&i915->fbdev_suspend_work);
> return;
> }
> }
> diff --git a/drivers/gpu/drm/i915/intel_fbdev.h b/drivers/gpu/drm/i915/intel_fbdev.h
> index de7c84250eb5..4d0c8c851a5e 100644
> --- a/drivers/gpu/drm/i915/intel_fbdev.h
> +++ b/drivers/gpu/drm/i915/intel_fbdev.h
> @@ -14,8 +14,8 @@ struct drm_i915_private;
> #ifdef CONFIG_DRM_FBDEV_EMULATION
> int intel_fbdev_init(struct drm_device *dev);
> void intel_fbdev_initial_config_async(struct drm_device *dev);
> -void intel_fbdev_unregister(struct drm_i915_private *dev_priv);
> -void intel_fbdev_fini(struct drm_i915_private *dev_priv);
> +void intel_fbdev_unregister(struct drm_i915_private *i915);
> +void intel_fbdev_fini(struct drm_i915_private *i915);
> void intel_fbdev_set_suspend(struct drm_device *dev, int state, bool synchronous);
> void intel_fbdev_output_poll_changed(struct drm_device *dev);
> void intel_fbdev_restore_mode(struct drm_device *dev);
> @@ -29,11 +29,11 @@ static inline void intel_fbdev_initial_config_async(struct drm_device *dev)
> {
> }
>
> -static inline void intel_fbdev_unregister(struct drm_i915_private *dev_priv)
> +static inline void intel_fbdev_unregister(struct drm_i915_private *i915)
> {
> }
>
> -static inline void intel_fbdev_fini(struct drm_i915_private *dev_priv)
> +static inline void intel_fbdev_fini(struct drm_i915_private *i915)
> {
> }
>
> diff --git a/drivers/gpu/drm/i915/intel_fifo_underrun.c b/drivers/gpu/drm/i915/intel_fifo_underrun.c
> index 8545ad32bb50..08f092addae2 100644
> --- a/drivers/gpu/drm/i915/intel_fifo_underrun.c
> +++ b/drivers/gpu/drm/i915/intel_fifo_underrun.c
> @@ -52,14 +52,14 @@
>
> static bool ivb_can_enable_err_int(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_crtc *crtc;
> enum pipe pipe;
>
> - lockdep_assert_held(&dev_priv->irq_lock);
> + lockdep_assert_held(&i915->irq_lock);
>
> - for_each_pipe(dev_priv, pipe) {
> - crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> + for_each_pipe(i915, pipe) {
> + crtc = intel_get_crtc_for_pipe(i915, pipe);
>
> if (crtc->cpu_fifo_underrun_disabled)
> return false;
> @@ -70,14 +70,14 @@ static bool ivb_can_enable_err_int(struct drm_device *dev)
>
> static bool cpt_can_enable_serr_int(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> enum pipe pipe;
> struct intel_crtc *crtc;
>
> - lockdep_assert_held(&dev_priv->irq_lock);
> + lockdep_assert_held(&i915->irq_lock);
>
> - for_each_pipe(dev_priv, pipe) {
> - crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> + for_each_pipe(i915, pipe) {
> + crtc = intel_get_crtc_for_pipe(i915, pipe);
>
> if (crtc->pch_fifo_underrun_disabled)
> return false;
> @@ -88,20 +88,20 @@ static bool cpt_can_enable_serr_int(struct drm_device *dev)
>
> static void i9xx_check_fifo_underruns(struct intel_crtc *crtc)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> i915_reg_t reg = PIPESTAT(crtc->pipe);
> u32 enable_mask;
>
> - lockdep_assert_held(&dev_priv->irq_lock);
> + lockdep_assert_held(&i915->irq_lock);
>
> if ((I915_READ(reg) & PIPE_FIFO_UNDERRUN_STATUS) == 0)
> return;
>
> - enable_mask = i915_pipestat_enable_mask(dev_priv, crtc->pipe);
> + enable_mask = i915_pipestat_enable_mask(i915, crtc->pipe);
> I915_WRITE(reg, enable_mask | PIPE_FIFO_UNDERRUN_STATUS);
> POSTING_READ(reg);
>
> - trace_intel_cpu_fifo_underrun(dev_priv, crtc->pipe);
> + trace_intel_cpu_fifo_underrun(i915, crtc->pipe);
> DRM_ERROR("pipe %c underrun\n", pipe_name(crtc->pipe));
> }
>
> @@ -109,13 +109,13 @@ static void i9xx_set_fifo_underrun_reporting(struct drm_device *dev,
> enum pipe pipe,
> bool enable, bool old)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> i915_reg_t reg = PIPESTAT(pipe);
>
> - lockdep_assert_held(&dev_priv->irq_lock);
> + lockdep_assert_held(&i915->irq_lock);
>
> if (enable) {
> - u32 enable_mask = i915_pipestat_enable_mask(dev_priv, pipe);
> + u32 enable_mask = i915_pipestat_enable_mask(i915, pipe);
>
> I915_WRITE(reg, enable_mask | PIPE_FIFO_UNDERRUN_STATUS);
> POSTING_READ(reg);
> @@ -128,23 +128,23 @@ static void i9xx_set_fifo_underrun_reporting(struct drm_device *dev,
> static void ironlake_set_fifo_underrun_reporting(struct drm_device *dev,
> enum pipe pipe, bool enable)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> u32 bit = (pipe == PIPE_A) ?
> DE_PIPEA_FIFO_UNDERRUN : DE_PIPEB_FIFO_UNDERRUN;
>
> if (enable)
> - ilk_enable_display_irq(dev_priv, bit);
> + ilk_enable_display_irq(i915, bit);
> else
> - ilk_disable_display_irq(dev_priv, bit);
> + ilk_disable_display_irq(i915, bit);
> }
>
> static void ivybridge_check_fifo_underruns(struct intel_crtc *crtc)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
> u32 err_int = I915_READ(GEN7_ERR_INT);
>
> - lockdep_assert_held(&dev_priv->irq_lock);
> + lockdep_assert_held(&i915->irq_lock);
>
> if ((err_int & ERR_INT_FIFO_UNDERRUN(pipe)) == 0)
> return;
> @@ -152,7 +152,7 @@ static void ivybridge_check_fifo_underruns(struct intel_crtc *crtc)
> I915_WRITE(GEN7_ERR_INT, ERR_INT_FIFO_UNDERRUN(pipe));
> POSTING_READ(GEN7_ERR_INT);
>
> - trace_intel_cpu_fifo_underrun(dev_priv, pipe);
> + trace_intel_cpu_fifo_underrun(i915, pipe);
> DRM_ERROR("fifo underrun on pipe %c\n", pipe_name(pipe));
> }
>
> @@ -160,16 +160,16 @@ static void ivybridge_set_fifo_underrun_reporting(struct drm_device *dev,
> enum pipe pipe,
> bool enable, bool old)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> if (enable) {
> I915_WRITE(GEN7_ERR_INT, ERR_INT_FIFO_UNDERRUN(pipe));
>
> if (!ivb_can_enable_err_int(dev))
> return;
>
> - ilk_enable_display_irq(dev_priv, DE_ERR_INT_IVB);
> + ilk_enable_display_irq(i915, DE_ERR_INT_IVB);
> } else {
> - ilk_disable_display_irq(dev_priv, DE_ERR_INT_IVB);
> + ilk_disable_display_irq(i915, DE_ERR_INT_IVB);
>
> if (old &&
> I915_READ(GEN7_ERR_INT) & ERR_INT_FIFO_UNDERRUN(pipe)) {
> @@ -182,35 +182,35 @@ static void ivybridge_set_fifo_underrun_reporting(struct drm_device *dev,
> static void broadwell_set_fifo_underrun_reporting(struct drm_device *dev,
> enum pipe pipe, bool enable)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
>
> if (enable)
> - bdw_enable_pipe_irq(dev_priv, pipe, GEN8_PIPE_FIFO_UNDERRUN);
> + bdw_enable_pipe_irq(i915, pipe, GEN8_PIPE_FIFO_UNDERRUN);
> else
> - bdw_disable_pipe_irq(dev_priv, pipe, GEN8_PIPE_FIFO_UNDERRUN);
> + bdw_disable_pipe_irq(i915, pipe, GEN8_PIPE_FIFO_UNDERRUN);
> }
>
> static void ibx_set_fifo_underrun_reporting(struct drm_device *dev,
> enum pipe pch_transcoder,
> bool enable)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> u32 bit = (pch_transcoder == PIPE_A) ?
> SDE_TRANSA_FIFO_UNDER : SDE_TRANSB_FIFO_UNDER;
>
> if (enable)
> - ibx_enable_display_interrupt(dev_priv, bit);
> + ibx_enable_display_interrupt(i915, bit);
> else
> - ibx_disable_display_interrupt(dev_priv, bit);
> + ibx_disable_display_interrupt(i915, bit);
> }
>
> static void cpt_check_pch_fifo_underruns(struct intel_crtc *crtc)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pch_transcoder = crtc->pipe;
> u32 serr_int = I915_READ(SERR_INT);
>
> - lockdep_assert_held(&dev_priv->irq_lock);
> + lockdep_assert_held(&i915->irq_lock);
>
> if ((serr_int & SERR_INT_TRANS_FIFO_UNDERRUN(pch_transcoder)) == 0)
> return;
> @@ -218,7 +218,7 @@ static void cpt_check_pch_fifo_underruns(struct intel_crtc *crtc)
> I915_WRITE(SERR_INT, SERR_INT_TRANS_FIFO_UNDERRUN(pch_transcoder));
> POSTING_READ(SERR_INT);
>
> - trace_intel_pch_fifo_underrun(dev_priv, pch_transcoder);
> + trace_intel_pch_fifo_underrun(i915, pch_transcoder);
> DRM_ERROR("pch fifo underrun on pch transcoder %c\n",
> pipe_name(pch_transcoder));
> }
> @@ -227,7 +227,7 @@ static void cpt_set_fifo_underrun_reporting(struct drm_device *dev,
> enum pipe pch_transcoder,
> bool enable, bool old)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
>
> if (enable) {
> I915_WRITE(SERR_INT,
> @@ -236,9 +236,9 @@ static void cpt_set_fifo_underrun_reporting(struct drm_device *dev,
> if (!cpt_can_enable_serr_int(dev))
> return;
>
> - ibx_enable_display_interrupt(dev_priv, SDE_ERROR_CPT);
> + ibx_enable_display_interrupt(i915, SDE_ERROR_CPT);
> } else {
> - ibx_disable_display_interrupt(dev_priv, SDE_ERROR_CPT);
> + ibx_disable_display_interrupt(i915, SDE_ERROR_CPT);
>
> if (old && I915_READ(SERR_INT) &
> SERR_INT_TRANS_FIFO_UNDERRUN(pch_transcoder)) {
> @@ -251,22 +251,22 @@ static void cpt_set_fifo_underrun_reporting(struct drm_device *dev,
> static bool __intel_set_cpu_fifo_underrun_reporting(struct drm_device *dev,
> enum pipe pipe, bool enable)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct intel_crtc *crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct intel_crtc *crtc = intel_get_crtc_for_pipe(i915, pipe);
> bool old;
>
> - lockdep_assert_held(&dev_priv->irq_lock);
> + lockdep_assert_held(&i915->irq_lock);
>
> old = !crtc->cpu_fifo_underrun_disabled;
> crtc->cpu_fifo_underrun_disabled = !enable;
>
> - if (HAS_GMCH(dev_priv))
> + if (HAS_GMCH(i915))
> i9xx_set_fifo_underrun_reporting(dev, pipe, enable, old);
> - else if (IS_GEN_RANGE(dev_priv, 5, 6))
> + else if (IS_GEN_RANGE(i915, 5, 6))
> ironlake_set_fifo_underrun_reporting(dev, pipe, enable);
> - else if (IS_GEN(dev_priv, 7))
> + else if (IS_GEN(i915, 7))
> ivybridge_set_fifo_underrun_reporting(dev, pipe, enable, old);
> - else if (INTEL_GEN(dev_priv) >= 8)
> + else if (INTEL_GEN(i915) >= 8)
> broadwell_set_fifo_underrun_reporting(dev, pipe, enable);
>
> return old;
> @@ -274,7 +274,7 @@ static bool __intel_set_cpu_fifo_underrun_reporting(struct drm_device *dev,
>
> /**
> * intel_set_cpu_fifo_underrun_reporting - set cpu fifo underrrun reporting state
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> * @pipe: (CPU) pipe to set state for
> * @enable: whether underruns should be reported or not
> *
> @@ -288,23 +288,23 @@ static bool __intel_set_cpu_fifo_underrun_reporting(struct drm_device *dev,
> *
> * Returns the previous state of underrun reporting.
> */
> -bool intel_set_cpu_fifo_underrun_reporting(struct drm_i915_private *dev_priv,
> +bool intel_set_cpu_fifo_underrun_reporting(struct drm_i915_private *i915,
> enum pipe pipe, bool enable)
> {
> unsigned long flags;
> bool ret;
>
> - spin_lock_irqsave(&dev_priv->irq_lock, flags);
> - ret = __intel_set_cpu_fifo_underrun_reporting(&dev_priv->drm, pipe,
> + spin_lock_irqsave(&i915->irq_lock, flags);
> + ret = __intel_set_cpu_fifo_underrun_reporting(&i915->drm, pipe,
> enable);
> - spin_unlock_irqrestore(&dev_priv->irq_lock, flags);
> + spin_unlock_irqrestore(&i915->irq_lock, flags);
>
> return ret;
> }
>
> /**
> * intel_set_pch_fifo_underrun_reporting - set PCH fifo underrun reporting state
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> * @pch_transcoder: the PCH transcoder (same as pipe on IVB and older)
> * @enable: whether underruns should be reported or not
> *
> @@ -316,12 +316,12 @@ bool intel_set_cpu_fifo_underrun_reporting(struct drm_i915_private *dev_priv,
> *
> * Returns the previous state of underrun reporting.
> */
> -bool intel_set_pch_fifo_underrun_reporting(struct drm_i915_private *dev_priv,
> +bool intel_set_pch_fifo_underrun_reporting(struct drm_i915_private *i915,
> enum pipe pch_transcoder,
> bool enable)
> {
> struct intel_crtc *crtc =
> - intel_get_crtc_for_pipe(dev_priv, pch_transcoder);
> + intel_get_crtc_for_pipe(i915, pch_transcoder);
> unsigned long flags;
> bool old;
>
> @@ -334,71 +334,71 @@ bool intel_set_pch_fifo_underrun_reporting(struct drm_i915_private *dev_priv,
> * crtc on LPT won't cause issues.
> */
>
> - spin_lock_irqsave(&dev_priv->irq_lock, flags);
> + spin_lock_irqsave(&i915->irq_lock, flags);
>
> old = !crtc->pch_fifo_underrun_disabled;
> crtc->pch_fifo_underrun_disabled = !enable;
>
> - if (HAS_PCH_IBX(dev_priv))
> - ibx_set_fifo_underrun_reporting(&dev_priv->drm,
> + if (HAS_PCH_IBX(i915))
> + ibx_set_fifo_underrun_reporting(&i915->drm,
> pch_transcoder,
> enable);
> else
> - cpt_set_fifo_underrun_reporting(&dev_priv->drm,
> + cpt_set_fifo_underrun_reporting(&i915->drm,
> pch_transcoder,
> enable, old);
>
> - spin_unlock_irqrestore(&dev_priv->irq_lock, flags);
> + spin_unlock_irqrestore(&i915->irq_lock, flags);
> return old;
> }
>
> /**
> * intel_cpu_fifo_underrun_irq_handler - handle CPU fifo underrun interrupt
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> * @pipe: (CPU) pipe to set state for
> *
> * This handles a CPU fifo underrun interrupt, generating an underrun warning
> * into dmesg if underrun reporting is enabled and then disables the underrun
> * interrupt to avoid an irq storm.
> */
> -void intel_cpu_fifo_underrun_irq_handler(struct drm_i915_private *dev_priv,
> +void intel_cpu_fifo_underrun_irq_handler(struct drm_i915_private *i915,
> enum pipe pipe)
> {
> - struct intel_crtc *crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> + struct intel_crtc *crtc = intel_get_crtc_for_pipe(i915, pipe);
>
> /* We may be called too early in init, thanks BIOS! */
> if (crtc == NULL)
> return;
>
> /* GMCH can't disable fifo underruns, filter them. */
> - if (HAS_GMCH(dev_priv) &&
> + if (HAS_GMCH(i915) &&
> crtc->cpu_fifo_underrun_disabled)
> return;
>
> - if (intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, false)) {
> - trace_intel_cpu_fifo_underrun(dev_priv, pipe);
> + if (intel_set_cpu_fifo_underrun_reporting(i915, pipe, false)) {
> + trace_intel_cpu_fifo_underrun(i915, pipe);
> DRM_ERROR("CPU pipe %c FIFO underrun\n",
> pipe_name(pipe));
> }
>
> - intel_fbc_handle_fifo_underrun_irq(dev_priv);
> + intel_fbc_handle_fifo_underrun_irq(i915);
> }
>
> /**
> * intel_pch_fifo_underrun_irq_handler - handle PCH fifo underrun interrupt
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> * @pch_transcoder: the PCH transcoder (same as pipe on IVB and older)
> *
> * This handles a PCH fifo underrun interrupt, generating an underrun warning
> * into dmesg if underrun reporting is enabled and then disables the underrun
> * interrupt to avoid an irq storm.
> */
> -void intel_pch_fifo_underrun_irq_handler(struct drm_i915_private *dev_priv,
> +void intel_pch_fifo_underrun_irq_handler(struct drm_i915_private *i915,
> enum pipe pch_transcoder)
> {
> - if (intel_set_pch_fifo_underrun_reporting(dev_priv, pch_transcoder,
> + if (intel_set_pch_fifo_underrun_reporting(i915, pch_transcoder,
> false)) {
> - trace_intel_pch_fifo_underrun(dev_priv, pch_transcoder);
> + trace_intel_pch_fifo_underrun(i915, pch_transcoder);
> DRM_ERROR("PCH transcoder %c FIFO underrun\n",
> pipe_name(pch_transcoder));
> }
> @@ -406,53 +406,53 @@ void intel_pch_fifo_underrun_irq_handler(struct drm_i915_private *dev_priv,
>
> /**
> * intel_check_cpu_fifo_underruns - check for CPU fifo underruns immediately
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> *
> * Check for CPU fifo underruns immediately. Useful on IVB/HSW where the shared
> * error interrupt may have been disabled, and so CPU fifo underruns won't
> * necessarily raise an interrupt, and on GMCH platforms where underruns never
> * raise an interrupt.
> */
> -void intel_check_cpu_fifo_underruns(struct drm_i915_private *dev_priv)
> +void intel_check_cpu_fifo_underruns(struct drm_i915_private *i915)
> {
> struct intel_crtc *crtc;
>
> - spin_lock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
>
> - for_each_intel_crtc(&dev_priv->drm, crtc) {
> + for_each_intel_crtc(&i915->drm, crtc) {
> if (crtc->cpu_fifo_underrun_disabled)
> continue;
>
> - if (HAS_GMCH(dev_priv))
> + if (HAS_GMCH(i915))
> i9xx_check_fifo_underruns(crtc);
> - else if (IS_GEN(dev_priv, 7))
> + else if (IS_GEN(i915, 7))
> ivybridge_check_fifo_underruns(crtc);
> }
>
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_unlock_irq(&i915->irq_lock);
> }
>
> /**
> * intel_check_pch_fifo_underruns - check for PCH fifo underruns immediately
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> *
> * Check for PCH fifo underruns immediately. Useful on CPT/PPT where the shared
> * error interrupt may have been disabled, and so PCH fifo underruns won't
> * necessarily raise an interrupt.
> */
> -void intel_check_pch_fifo_underruns(struct drm_i915_private *dev_priv)
> +void intel_check_pch_fifo_underruns(struct drm_i915_private *i915)
> {
> struct intel_crtc *crtc;
>
> - spin_lock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
>
> - for_each_intel_crtc(&dev_priv->drm, crtc) {
> + for_each_intel_crtc(&i915->drm, crtc) {
> if (crtc->pch_fifo_underrun_disabled)
> continue;
>
> - if (HAS_PCH_CPT(dev_priv))
> + if (HAS_PCH_CPT(i915))
> cpt_check_pch_fifo_underruns(crtc);
> }
>
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_unlock_irq(&i915->irq_lock);
> }
> diff --git a/drivers/gpu/drm/i915/intel_fifo_underrun.h b/drivers/gpu/drm/i915/intel_fifo_underrun.h
> index e04f22ac1f49..290909a67de8 100644
> --- a/drivers/gpu/drm/i915/intel_fifo_underrun.h
> +++ b/drivers/gpu/drm/i915/intel_fifo_underrun.h
> @@ -12,16 +12,16 @@
>
> struct drm_i915_private;
>
> -bool intel_set_cpu_fifo_underrun_reporting(struct drm_i915_private *dev_priv,
> +bool intel_set_cpu_fifo_underrun_reporting(struct drm_i915_private *i915,
> enum pipe pipe, bool enable);
> -bool intel_set_pch_fifo_underrun_reporting(struct drm_i915_private *dev_priv,
> +bool intel_set_pch_fifo_underrun_reporting(struct drm_i915_private *i915,
> enum pipe pch_transcoder,
> bool enable);
> -void intel_cpu_fifo_underrun_irq_handler(struct drm_i915_private *dev_priv,
> +void intel_cpu_fifo_underrun_irq_handler(struct drm_i915_private *i915,
> enum pipe pipe);
> -void intel_pch_fifo_underrun_irq_handler(struct drm_i915_private *dev_priv,
> +void intel_pch_fifo_underrun_irq_handler(struct drm_i915_private *i915,
> enum pipe pch_transcoder);
> -void intel_check_cpu_fifo_underruns(struct drm_i915_private *dev_priv);
> -void intel_check_pch_fifo_underruns(struct drm_i915_private *dev_priv);
> +void intel_check_cpu_fifo_underruns(struct drm_i915_private *i915);
> +void intel_check_pch_fifo_underruns(struct drm_i915_private *i915);
>
> #endif /* __INTEL_FIFO_UNDERRUN_H__ */
> diff --git a/drivers/gpu/drm/i915/intel_frontbuffer.c b/drivers/gpu/drm/i915/intel_frontbuffer.c
> index d6036b9ad16a..161139de3f28 100644
> --- a/drivers/gpu/drm/i915/intel_frontbuffer.c
> +++ b/drivers/gpu/drm/i915/intel_frontbuffer.c
> @@ -66,24 +66,24 @@ void __intel_fb_obj_invalidate(struct drm_i915_gem_object *obj,
> enum fb_op_origin origin,
> unsigned int frontbuffer_bits)
> {
> - struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
> + struct drm_i915_private *i915 = to_i915(obj->base.dev);
>
> if (origin == ORIGIN_CS) {
> - spin_lock(&dev_priv->fb_tracking.lock);
> - dev_priv->fb_tracking.busy_bits |= frontbuffer_bits;
> - dev_priv->fb_tracking.flip_bits &= ~frontbuffer_bits;
> - spin_unlock(&dev_priv->fb_tracking.lock);
> + spin_lock(&i915->fb_tracking.lock);
> + i915->fb_tracking.busy_bits |= frontbuffer_bits;
> + i915->fb_tracking.flip_bits &= ~frontbuffer_bits;
> + spin_unlock(&i915->fb_tracking.lock);
> }
>
> might_sleep();
> - intel_psr_invalidate(dev_priv, frontbuffer_bits, origin);
> - intel_edp_drrs_invalidate(dev_priv, frontbuffer_bits);
> - intel_fbc_invalidate(dev_priv, frontbuffer_bits, origin);
> + intel_psr_invalidate(i915, frontbuffer_bits, origin);
> + intel_edp_drrs_invalidate(i915, frontbuffer_bits);
> + intel_fbc_invalidate(i915, frontbuffer_bits, origin);
> }
>
> /**
> * intel_frontbuffer_flush - flush frontbuffer
> - * @dev_priv: i915 device
> + * @i915: i915 device
> * @frontbuffer_bits: frontbuffer plane tracking bits
> * @origin: which operation caused the flush
> *
> @@ -93,45 +93,45 @@ void __intel_fb_obj_invalidate(struct drm_i915_gem_object *obj,
> *
> * Can be called without any locks held.
> */
> -static void intel_frontbuffer_flush(struct drm_i915_private *dev_priv,
> +static void intel_frontbuffer_flush(struct drm_i915_private *i915,
> unsigned frontbuffer_bits,
> enum fb_op_origin origin)
> {
> /* Delay flushing when rings are still busy.*/
> - spin_lock(&dev_priv->fb_tracking.lock);
> - frontbuffer_bits &= ~dev_priv->fb_tracking.busy_bits;
> - spin_unlock(&dev_priv->fb_tracking.lock);
> + spin_lock(&i915->fb_tracking.lock);
> + frontbuffer_bits &= ~i915->fb_tracking.busy_bits;
> + spin_unlock(&i915->fb_tracking.lock);
>
> if (!frontbuffer_bits)
> return;
>
> might_sleep();
> - intel_edp_drrs_flush(dev_priv, frontbuffer_bits);
> - intel_psr_flush(dev_priv, frontbuffer_bits, origin);
> - intel_fbc_flush(dev_priv, frontbuffer_bits, origin);
> + intel_edp_drrs_flush(i915, frontbuffer_bits);
> + intel_psr_flush(i915, frontbuffer_bits, origin);
> + intel_fbc_flush(i915, frontbuffer_bits, origin);
> }
>
> void __intel_fb_obj_flush(struct drm_i915_gem_object *obj,
> enum fb_op_origin origin,
> unsigned int frontbuffer_bits)
> {
> - struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
> + struct drm_i915_private *i915 = to_i915(obj->base.dev);
>
> if (origin == ORIGIN_CS) {
> - spin_lock(&dev_priv->fb_tracking.lock);
> + spin_lock(&i915->fb_tracking.lock);
> /* Filter out new bits since rendering started. */
> - frontbuffer_bits &= dev_priv->fb_tracking.busy_bits;
> - dev_priv->fb_tracking.busy_bits &= ~frontbuffer_bits;
> - spin_unlock(&dev_priv->fb_tracking.lock);
> + frontbuffer_bits &= i915->fb_tracking.busy_bits;
> + i915->fb_tracking.busy_bits &= ~frontbuffer_bits;
> + spin_unlock(&i915->fb_tracking.lock);
> }
>
> if (frontbuffer_bits)
> - intel_frontbuffer_flush(dev_priv, frontbuffer_bits, origin);
> + intel_frontbuffer_flush(i915, frontbuffer_bits, origin);
> }
>
> /**
> * intel_frontbuffer_flip_prepare - prepare asynchronous frontbuffer flip
> - * @dev_priv: i915 device
> + * @i915: i915 device
> * @frontbuffer_bits: frontbuffer plane tracking bits
> *
> * This function gets called after scheduling a flip on @obj. The actual
> @@ -141,19 +141,19 @@ void __intel_fb_obj_flush(struct drm_i915_gem_object *obj,
> *
> * Can be called without any locks held.
> */
> -void intel_frontbuffer_flip_prepare(struct drm_i915_private *dev_priv,
> +void intel_frontbuffer_flip_prepare(struct drm_i915_private *i915,
> unsigned frontbuffer_bits)
> {
> - spin_lock(&dev_priv->fb_tracking.lock);
> - dev_priv->fb_tracking.flip_bits |= frontbuffer_bits;
> + spin_lock(&i915->fb_tracking.lock);
> + i915->fb_tracking.flip_bits |= frontbuffer_bits;
> /* Remove stale busy bits due to the old buffer. */
> - dev_priv->fb_tracking.busy_bits &= ~frontbuffer_bits;
> - spin_unlock(&dev_priv->fb_tracking.lock);
> + i915->fb_tracking.busy_bits &= ~frontbuffer_bits;
> + spin_unlock(&i915->fb_tracking.lock);
> }
>
> /**
> * intel_frontbuffer_flip_complete - complete asynchronous frontbuffer flip
> - * @dev_priv: i915 device
> + * @i915: i915 device
> * @frontbuffer_bits: frontbuffer plane tracking bits
> *
> * This function gets called after the flip has been latched and will complete
> @@ -161,23 +161,23 @@ void intel_frontbuffer_flip_prepare(struct drm_i915_private *dev_priv,
> *
> * Can be called without any locks held.
> */
> -void intel_frontbuffer_flip_complete(struct drm_i915_private *dev_priv,
> +void intel_frontbuffer_flip_complete(struct drm_i915_private *i915,
> unsigned frontbuffer_bits)
> {
> - spin_lock(&dev_priv->fb_tracking.lock);
> + spin_lock(&i915->fb_tracking.lock);
> /* Mask any cancelled flips. */
> - frontbuffer_bits &= dev_priv->fb_tracking.flip_bits;
> - dev_priv->fb_tracking.flip_bits &= ~frontbuffer_bits;
> - spin_unlock(&dev_priv->fb_tracking.lock);
> + frontbuffer_bits &= i915->fb_tracking.flip_bits;
> + i915->fb_tracking.flip_bits &= ~frontbuffer_bits;
> + spin_unlock(&i915->fb_tracking.lock);
>
> if (frontbuffer_bits)
> - intel_frontbuffer_flush(dev_priv,
> + intel_frontbuffer_flush(i915,
> frontbuffer_bits, ORIGIN_FLIP);
> }
>
> /**
> * intel_frontbuffer_flip - synchronous frontbuffer flip
> - * @dev_priv: i915 device
> + * @i915: i915 device
> * @frontbuffer_bits: frontbuffer plane tracking bits
> *
> * This function gets called after scheduling a flip on @obj. This is for
> @@ -186,13 +186,13 @@ void intel_frontbuffer_flip_complete(struct drm_i915_private *dev_priv,
> *
> * Can be called without any locks held.
> */
> -void intel_frontbuffer_flip(struct drm_i915_private *dev_priv,
> +void intel_frontbuffer_flip(struct drm_i915_private *i915,
> unsigned frontbuffer_bits)
> {
> - spin_lock(&dev_priv->fb_tracking.lock);
> + spin_lock(&i915->fb_tracking.lock);
> /* Remove stale busy bits due to the old buffer. */
> - dev_priv->fb_tracking.busy_bits &= ~frontbuffer_bits;
> - spin_unlock(&dev_priv->fb_tracking.lock);
> + i915->fb_tracking.busy_bits &= ~frontbuffer_bits;
> + spin_unlock(&i915->fb_tracking.lock);
>
> - intel_frontbuffer_flush(dev_priv, frontbuffer_bits, ORIGIN_FLIP);
> + intel_frontbuffer_flush(i915, frontbuffer_bits, ORIGIN_FLIP);
> }
> diff --git a/drivers/gpu/drm/i915/intel_frontbuffer.h b/drivers/gpu/drm/i915/intel_frontbuffer.h
> index 5727320c8084..55c42efee7ce 100644
> --- a/drivers/gpu/drm/i915/intel_frontbuffer.h
> +++ b/drivers/gpu/drm/i915/intel_frontbuffer.h
> @@ -37,11 +37,11 @@ enum fb_op_origin {
> ORIGIN_DIRTYFB,
> };
>
> -void intel_frontbuffer_flip_prepare(struct drm_i915_private *dev_priv,
> +void intel_frontbuffer_flip_prepare(struct drm_i915_private *i915,
> unsigned frontbuffer_bits);
> -void intel_frontbuffer_flip_complete(struct drm_i915_private *dev_priv,
> +void intel_frontbuffer_flip_complete(struct drm_i915_private *i915,
> unsigned frontbuffer_bits);
> -void intel_frontbuffer_flip(struct drm_i915_private *dev_priv,
> +void intel_frontbuffer_flip(struct drm_i915_private *i915,
> unsigned frontbuffer_bits);
>
> void __intel_fb_obj_invalidate(struct drm_i915_gem_object *obj,
> diff --git a/drivers/gpu/drm/i915/intel_gmbus.c b/drivers/gpu/drm/i915/intel_gmbus.c
> index aa88e6e7cc65..f1b9844b591e 100644
> --- a/drivers/gpu/drm/i915/intel_gmbus.c
> +++ b/drivers/gpu/drm/i915/intel_gmbus.c
> @@ -89,42 +89,42 @@ static const struct gmbus_pin gmbus_pins_icp[] = {
> };
>
> /* pin is expected to be valid */
> -static const struct gmbus_pin *get_gmbus_pin(struct drm_i915_private *dev_priv,
> +static const struct gmbus_pin *get_gmbus_pin(struct drm_i915_private *i915,
> unsigned int pin)
> {
> - if (HAS_PCH_ICP(dev_priv))
> + if (HAS_PCH_ICP(i915))
> return &gmbus_pins_icp[pin];
> - else if (HAS_PCH_CNP(dev_priv))
> + else if (HAS_PCH_CNP(i915))
> return &gmbus_pins_cnp[pin];
> - else if (IS_GEN9_LP(dev_priv))
> + else if (IS_GEN9_LP(i915))
> return &gmbus_pins_bxt[pin];
> - else if (IS_GEN9_BC(dev_priv))
> + else if (IS_GEN9_BC(i915))
> return &gmbus_pins_skl[pin];
> - else if (IS_BROADWELL(dev_priv))
> + else if (IS_BROADWELL(i915))
> return &gmbus_pins_bdw[pin];
> else
> return &gmbus_pins[pin];
> }
>
> -bool intel_gmbus_is_valid_pin(struct drm_i915_private *dev_priv,
> +bool intel_gmbus_is_valid_pin(struct drm_i915_private *i915,
> unsigned int pin)
> {
> unsigned int size;
>
> - if (HAS_PCH_ICP(dev_priv))
> + if (HAS_PCH_ICP(i915))
> size = ARRAY_SIZE(gmbus_pins_icp);
> - else if (HAS_PCH_CNP(dev_priv))
> + else if (HAS_PCH_CNP(i915))
> size = ARRAY_SIZE(gmbus_pins_cnp);
> - else if (IS_GEN9_LP(dev_priv))
> + else if (IS_GEN9_LP(i915))
> size = ARRAY_SIZE(gmbus_pins_bxt);
> - else if (IS_GEN9_BC(dev_priv))
> + else if (IS_GEN9_BC(i915))
> size = ARRAY_SIZE(gmbus_pins_skl);
> - else if (IS_BROADWELL(dev_priv))
> + else if (IS_BROADWELL(i915))
> size = ARRAY_SIZE(gmbus_pins_bdw);
> else
> size = ARRAY_SIZE(gmbus_pins);
>
> - return pin < size && get_gmbus_pin(dev_priv, pin)->name;
> + return pin < size && get_gmbus_pin(i915, pin)->name;
> }
>
> /* Intel GPIO access functions */
> @@ -138,13 +138,13 @@ to_intel_gmbus(struct i2c_adapter *i2c)
> }
>
> void
> -intel_gmbus_reset(struct drm_i915_private *dev_priv)
> +intel_gmbus_reset(struct drm_i915_private *i915)
> {
> I915_WRITE(GMBUS0, 0);
> I915_WRITE(GMBUS4, 0);
> }
>
> -static void pnv_gmbus_clock_gating(struct drm_i915_private *dev_priv,
> +static void pnv_gmbus_clock_gating(struct drm_i915_private *i915,
> bool enable)
> {
> u32 val;
> @@ -158,7 +158,7 @@ static void pnv_gmbus_clock_gating(struct drm_i915_private *dev_priv,
> I915_WRITE(DSPCLK_GATE_D, val);
> }
>
> -static void pch_gmbus_clock_gating(struct drm_i915_private *dev_priv,
> +static void pch_gmbus_clock_gating(struct drm_i915_private *i915,
> bool enable)
> {
> u32 val;
> @@ -171,7 +171,7 @@ static void pch_gmbus_clock_gating(struct drm_i915_private *dev_priv,
> I915_WRITE(SOUTH_DSPCLK_GATE_D, val);
> }
>
> -static void bxt_gmbus_clock_gating(struct drm_i915_private *dev_priv,
> +static void bxt_gmbus_clock_gating(struct drm_i915_private *i915,
> bool enable)
> {
> u32 val;
> @@ -186,7 +186,7 @@ static void bxt_gmbus_clock_gating(struct drm_i915_private *dev_priv,
>
> static u32 get_reserved(struct intel_gmbus *bus)
> {
> - struct drm_i915_private *i915 = bus->dev_priv;
> + struct drm_i915_private *i915 = bus->i915;
> struct intel_uncore *uncore = &i915->uncore;
> u32 reserved = 0;
>
> @@ -202,7 +202,7 @@ static u32 get_reserved(struct intel_gmbus *bus)
> static int get_clock(void *data)
> {
> struct intel_gmbus *bus = data;
> - struct intel_uncore *uncore = &bus->dev_priv->uncore;
> + struct intel_uncore *uncore = &bus->i915->uncore;
> u32 reserved = get_reserved(bus);
>
> intel_uncore_write_notrace(uncore,
> @@ -217,7 +217,7 @@ static int get_clock(void *data)
> static int get_data(void *data)
> {
> struct intel_gmbus *bus = data;
> - struct intel_uncore *uncore = &bus->dev_priv->uncore;
> + struct intel_uncore *uncore = &bus->i915->uncore;
> u32 reserved = get_reserved(bus);
>
> intel_uncore_write_notrace(uncore,
> @@ -232,7 +232,7 @@ static int get_data(void *data)
> static void set_clock(void *data, int state_high)
> {
> struct intel_gmbus *bus = data;
> - struct intel_uncore *uncore = &bus->dev_priv->uncore;
> + struct intel_uncore *uncore = &bus->i915->uncore;
> u32 reserved = get_reserved(bus);
> u32 clock_bits;
>
> @@ -251,7 +251,7 @@ static void set_clock(void *data, int state_high)
> static void set_data(void *data, int state_high)
> {
> struct intel_gmbus *bus = data;
> - struct intel_uncore *uncore = &bus->dev_priv->uncore;
> + struct intel_uncore *uncore = &bus->i915->uncore;
> u32 reserved = get_reserved(bus);
> u32 data_bits;
>
> @@ -271,12 +271,12 @@ intel_gpio_pre_xfer(struct i2c_adapter *adapter)
> struct intel_gmbus *bus = container_of(adapter,
> struct intel_gmbus,
> adapter);
> - struct drm_i915_private *dev_priv = bus->dev_priv;
> + struct drm_i915_private *i915 = bus->i915;
>
> - intel_gmbus_reset(dev_priv);
> + intel_gmbus_reset(i915);
>
> - if (IS_PINEVIEW(dev_priv))
> - pnv_gmbus_clock_gating(dev_priv, false);
> + if (IS_PINEVIEW(i915))
> + pnv_gmbus_clock_gating(i915, false);
>
> set_data(bus, 1);
> set_clock(bus, 1);
> @@ -290,24 +290,24 @@ intel_gpio_post_xfer(struct i2c_adapter *adapter)
> struct intel_gmbus *bus = container_of(adapter,
> struct intel_gmbus,
> adapter);
> - struct drm_i915_private *dev_priv = bus->dev_priv;
> + struct drm_i915_private *i915 = bus->i915;
>
> set_data(bus, 1);
> set_clock(bus, 1);
>
> - if (IS_PINEVIEW(dev_priv))
> - pnv_gmbus_clock_gating(dev_priv, true);
> + if (IS_PINEVIEW(i915))
> + pnv_gmbus_clock_gating(i915, true);
> }
>
> static void
> intel_gpio_setup(struct intel_gmbus *bus, unsigned int pin)
> {
> - struct drm_i915_private *dev_priv = bus->dev_priv;
> + struct drm_i915_private *i915 = bus->i915;
> struct i2c_algo_bit_data *algo;
>
> algo = &bus->bit_algo;
>
> - bus->gpio_reg = GPIO(get_gmbus_pin(dev_priv, pin)->gpio);
> + bus->gpio_reg = GPIO(get_gmbus_pin(i915, pin)->gpio);
> bus->adapter.algo_data = algo;
> algo->setsda = set_data;
> algo->setscl = set_clock;
> @@ -320,7 +320,7 @@ intel_gpio_setup(struct intel_gmbus *bus, unsigned int pin)
> algo->data = bus;
> }
>
> -static int gmbus_wait(struct drm_i915_private *dev_priv, u32 status, u32 irq_en)
> +static int gmbus_wait(struct drm_i915_private *i915, u32 status, u32 irq_en)
> {
> DEFINE_WAIT(wait);
> u32 gmbus2;
> @@ -330,10 +330,10 @@ static int gmbus_wait(struct drm_i915_private *dev_priv, u32 status, u32 irq_en)
> * we also need to check for NAKs besides the hw ready/idle signal, we
> * need to wake up periodically and check that ourselves.
> */
> - if (!HAS_GMBUS_IRQ(dev_priv))
> + if (!HAS_GMBUS_IRQ(i915))
> irq_en = 0;
>
> - add_wait_queue(&dev_priv->gmbus_wait_queue, &wait);
> + add_wait_queue(&i915->gmbus_wait_queue, &wait);
> I915_WRITE_FW(GMBUS4, irq_en);
>
> status |= GMBUS_SATOER;
> @@ -342,7 +342,7 @@ static int gmbus_wait(struct drm_i915_private *dev_priv, u32 status, u32 irq_en)
> ret = wait_for((gmbus2 = I915_READ_FW(GMBUS2)) & status, 50);
>
> I915_WRITE_FW(GMBUS4, 0);
> - remove_wait_queue(&dev_priv->gmbus_wait_queue, &wait);
> + remove_wait_queue(&i915->gmbus_wait_queue, &wait);
>
> if (gmbus2 & GMBUS_SATOER)
> return -ENXIO;
> @@ -351,7 +351,7 @@ static int gmbus_wait(struct drm_i915_private *dev_priv, u32 status, u32 irq_en)
> }
>
> static int
> -gmbus_wait_idle(struct drm_i915_private *dev_priv)
> +gmbus_wait_idle(struct drm_i915_private *i915)
> {
> DEFINE_WAIT(wait);
> u32 irq_enable;
> @@ -359,36 +359,36 @@ gmbus_wait_idle(struct drm_i915_private *dev_priv)
>
> /* Important: The hw handles only the first bit, so set only one! */
> irq_enable = 0;
> - if (HAS_GMBUS_IRQ(dev_priv))
> + if (HAS_GMBUS_IRQ(i915))
> irq_enable = GMBUS_IDLE_EN;
>
> - add_wait_queue(&dev_priv->gmbus_wait_queue, &wait);
> + add_wait_queue(&i915->gmbus_wait_queue, &wait);
> I915_WRITE_FW(GMBUS4, irq_enable);
>
> - ret = intel_wait_for_register_fw(&dev_priv->uncore,
> + ret = intel_wait_for_register_fw(&i915->uncore,
> GMBUS2, GMBUS_ACTIVE, 0,
> 10);
>
> I915_WRITE_FW(GMBUS4, 0);
> - remove_wait_queue(&dev_priv->gmbus_wait_queue, &wait);
> + remove_wait_queue(&i915->gmbus_wait_queue, &wait);
>
> return ret;
> }
>
> static inline
> -unsigned int gmbus_max_xfer_size(struct drm_i915_private *dev_priv)
> +unsigned int gmbus_max_xfer_size(struct drm_i915_private *i915)
> {
> - return INTEL_GEN(dev_priv) >= 9 ? GEN9_GMBUS_BYTE_COUNT_MAX :
> + return INTEL_GEN(i915) >= 9 ? GEN9_GMBUS_BYTE_COUNT_MAX :
> GMBUS_BYTE_COUNT_MAX;
> }
>
> static int
> -gmbus_xfer_read_chunk(struct drm_i915_private *dev_priv,
> +gmbus_xfer_read_chunk(struct drm_i915_private *i915,
> unsigned short addr, u8 *buf, unsigned int len,
> u32 gmbus0_reg, u32 gmbus1_index)
> {
> unsigned int size = len;
> - bool burst_read = len > gmbus_max_xfer_size(dev_priv);
> + bool burst_read = len > gmbus_max_xfer_size(i915);
> bool extra_byte_added = false;
>
> if (burst_read) {
> @@ -414,7 +414,7 @@ gmbus_xfer_read_chunk(struct drm_i915_private *dev_priv,
> int ret;
> u32 val, loop = 0;
>
> - ret = gmbus_wait(dev_priv, GMBUS_HW_RDY, GMBUS_HW_RDY_EN);
> + ret = gmbus_wait(i915, GMBUS_HW_RDY, GMBUS_HW_RDY_EN);
> if (ret)
> return ret;
>
> @@ -446,7 +446,7 @@ gmbus_xfer_read_chunk(struct drm_i915_private *dev_priv,
> #define INTEL_GMBUS_BURST_READ_MAX_LEN 767U
>
> static int
> -gmbus_xfer_read(struct drm_i915_private *dev_priv, struct i2c_msg *msg,
> +gmbus_xfer_read(struct drm_i915_private *i915, struct i2c_msg *msg,
> u32 gmbus0_reg, u32 gmbus1_index)
> {
> u8 *buf = msg->buf;
> @@ -455,12 +455,12 @@ gmbus_xfer_read(struct drm_i915_private *dev_priv, struct i2c_msg *msg,
> int ret;
>
> do {
> - if (HAS_GMBUS_BURST_READ(dev_priv))
> + if (HAS_GMBUS_BURST_READ(i915))
> len = min(rx_size, INTEL_GMBUS_BURST_READ_MAX_LEN);
> else
> - len = min(rx_size, gmbus_max_xfer_size(dev_priv));
> + len = min(rx_size, gmbus_max_xfer_size(i915));
>
> - ret = gmbus_xfer_read_chunk(dev_priv, msg->addr, buf, len,
> + ret = gmbus_xfer_read_chunk(i915, msg->addr, buf, len,
> gmbus0_reg, gmbus1_index);
> if (ret)
> return ret;
> @@ -473,7 +473,7 @@ gmbus_xfer_read(struct drm_i915_private *dev_priv, struct i2c_msg *msg,
> }
>
> static int
> -gmbus_xfer_write_chunk(struct drm_i915_private *dev_priv,
> +gmbus_xfer_write_chunk(struct drm_i915_private *i915,
> unsigned short addr, u8 *buf, unsigned int len,
> u32 gmbus1_index)
> {
> @@ -502,7 +502,7 @@ gmbus_xfer_write_chunk(struct drm_i915_private *dev_priv,
>
> I915_WRITE_FW(GMBUS3, val);
>
> - ret = gmbus_wait(dev_priv, GMBUS_HW_RDY, GMBUS_HW_RDY_EN);
> + ret = gmbus_wait(i915, GMBUS_HW_RDY, GMBUS_HW_RDY_EN);
> if (ret)
> return ret;
> }
> @@ -511,7 +511,7 @@ gmbus_xfer_write_chunk(struct drm_i915_private *dev_priv,
> }
>
> static int
> -gmbus_xfer_write(struct drm_i915_private *dev_priv, struct i2c_msg *msg,
> +gmbus_xfer_write(struct drm_i915_private *i915, struct i2c_msg *msg,
> u32 gmbus1_index)
> {
> u8 *buf = msg->buf;
> @@ -520,9 +520,9 @@ gmbus_xfer_write(struct drm_i915_private *dev_priv, struct i2c_msg *msg,
> int ret;
>
> do {
> - len = min(tx_size, gmbus_max_xfer_size(dev_priv));
> + len = min(tx_size, gmbus_max_xfer_size(i915));
>
> - ret = gmbus_xfer_write_chunk(dev_priv, msg->addr, buf, len,
> + ret = gmbus_xfer_write_chunk(i915, msg->addr, buf, len,
> gmbus1_index);
> if (ret)
> return ret;
> @@ -549,7 +549,7 @@ gmbus_is_index_xfer(struct i2c_msg *msgs, int i, int num)
> }
>
> static int
> -gmbus_index_xfer(struct drm_i915_private *dev_priv, struct i2c_msg *msgs,
> +gmbus_index_xfer(struct drm_i915_private *i915, struct i2c_msg *msgs,
> u32 gmbus0_reg)
> {
> u32 gmbus1_index = 0;
> @@ -568,10 +568,10 @@ gmbus_index_xfer(struct drm_i915_private *dev_priv, struct i2c_msg *msgs,
> I915_WRITE_FW(GMBUS5, gmbus5);
>
> if (msgs[1].flags & I2C_M_RD)
> - ret = gmbus_xfer_read(dev_priv, &msgs[1], gmbus0_reg,
> + ret = gmbus_xfer_read(i915, &msgs[1], gmbus0_reg,
> gmbus1_index);
> else
> - ret = gmbus_xfer_write(dev_priv, &msgs[1], gmbus1_index);
> + ret = gmbus_xfer_write(i915, &msgs[1], gmbus1_index);
>
> /* Clear GMBUS5 after each index transfer */
> if (gmbus5)
> @@ -587,15 +587,15 @@ do_gmbus_xfer(struct i2c_adapter *adapter, struct i2c_msg *msgs, int num,
> struct intel_gmbus *bus = container_of(adapter,
> struct intel_gmbus,
> adapter);
> - struct drm_i915_private *dev_priv = bus->dev_priv;
> + struct drm_i915_private *i915 = bus->i915;
> int i = 0, inc, try = 0;
> int ret = 0;
>
> /* Display WA #0868: skl,bxt,kbl,cfl,glk,cnl */
> - if (IS_GEN9_LP(dev_priv))
> - bxt_gmbus_clock_gating(dev_priv, false);
> - else if (HAS_PCH_SPT(dev_priv) || HAS_PCH_CNP(dev_priv))
> - pch_gmbus_clock_gating(dev_priv, false);
> + if (IS_GEN9_LP(i915))
> + bxt_gmbus_clock_gating(i915, false);
> + else if (HAS_PCH_SPT(i915) || HAS_PCH_CNP(i915))
> + pch_gmbus_clock_gating(i915, false);
>
> retry:
> I915_WRITE_FW(GMBUS0, gmbus0_source | bus->reg0);
> @@ -603,18 +603,18 @@ do_gmbus_xfer(struct i2c_adapter *adapter, struct i2c_msg *msgs, int num,
> for (; i < num; i += inc) {
> inc = 1;
> if (gmbus_is_index_xfer(msgs, i, num)) {
> - ret = gmbus_index_xfer(dev_priv, &msgs[i],
> + ret = gmbus_index_xfer(i915, &msgs[i],
> gmbus0_source | bus->reg0);
> inc = 2; /* an index transmission is two msgs */
> } else if (msgs[i].flags & I2C_M_RD) {
> - ret = gmbus_xfer_read(dev_priv, &msgs[i],
> + ret = gmbus_xfer_read(i915, &msgs[i],
> gmbus0_source | bus->reg0, 0);
> } else {
> - ret = gmbus_xfer_write(dev_priv, &msgs[i], 0);
> + ret = gmbus_xfer_write(i915, &msgs[i], 0);
> }
>
> if (!ret)
> - ret = gmbus_wait(dev_priv,
> + ret = gmbus_wait(i915,
> GMBUS_HW_WAIT_PHASE, GMBUS_HW_WAIT_EN);
> if (ret == -ETIMEDOUT)
> goto timeout;
> @@ -632,7 +632,7 @@ do_gmbus_xfer(struct i2c_adapter *adapter, struct i2c_msg *msgs, int num,
> * We will re-enable it at the start of the next xfer,
> * till then let it sleep.
> */
> - if (gmbus_wait_idle(dev_priv)) {
> + if (gmbus_wait_idle(i915)) {
> DRM_DEBUG_KMS("GMBUS [%s] timed out waiting for idle\n",
> adapter->name);
> ret = -ETIMEDOUT;
> @@ -656,7 +656,7 @@ do_gmbus_xfer(struct i2c_adapter *adapter, struct i2c_msg *msgs, int num,
> * it's slow responding and only answers on the 2nd retry.
> */
> ret = -ENXIO;
> - if (gmbus_wait_idle(dev_priv)) {
> + if (gmbus_wait_idle(i915)) {
> DRM_DEBUG_KMS("GMBUS [%s] timed out after NAK\n",
> adapter->name);
> ret = -ETIMEDOUT;
> @@ -701,10 +701,10 @@ do_gmbus_xfer(struct i2c_adapter *adapter, struct i2c_msg *msgs, int num,
>
> out:
> /* Display WA #0868: skl,bxt,kbl,cfl,glk,cnl */
> - if (IS_GEN9_LP(dev_priv))
> - bxt_gmbus_clock_gating(dev_priv, true);
> - else if (HAS_PCH_SPT(dev_priv) || HAS_PCH_CNP(dev_priv))
> - pch_gmbus_clock_gating(dev_priv, true);
> + if (IS_GEN9_LP(i915))
> + bxt_gmbus_clock_gating(i915, true);
> + else if (HAS_PCH_SPT(i915) || HAS_PCH_CNP(i915))
> + pch_gmbus_clock_gating(i915, true);
>
> return ret;
> }
> @@ -714,11 +714,11 @@ gmbus_xfer(struct i2c_adapter *adapter, struct i2c_msg *msgs, int num)
> {
> struct intel_gmbus *bus =
> container_of(adapter, struct intel_gmbus, adapter);
> - struct drm_i915_private *dev_priv = bus->dev_priv;
> + struct drm_i915_private *i915 = bus->i915;
> intel_wakeref_t wakeref;
> int ret;
>
> - wakeref = intel_display_power_get(dev_priv, POWER_DOMAIN_GMBUS);
> + wakeref = intel_display_power_get(i915, POWER_DOMAIN_GMBUS);
>
> if (bus->force_bit) {
> ret = i2c_bit_algo.master_xfer(adapter, msgs, num);
> @@ -730,7 +730,7 @@ gmbus_xfer(struct i2c_adapter *adapter, struct i2c_msg *msgs, int num)
> bus->force_bit |= GMBUS_FORCE_BIT_RETRY;
> }
>
> - intel_display_power_put(dev_priv, POWER_DOMAIN_GMBUS, wakeref);
> + intel_display_power_put(i915, POWER_DOMAIN_GMBUS, wakeref);
>
> return ret;
> }
> @@ -739,7 +739,7 @@ int intel_gmbus_output_aksv(struct i2c_adapter *adapter)
> {
> struct intel_gmbus *bus =
> container_of(adapter, struct intel_gmbus, adapter);
> - struct drm_i915_private *dev_priv = bus->dev_priv;
> + struct drm_i915_private *i915 = bus->i915;
> u8 cmd = DRM_HDCP_DDC_AKSV;
> u8 buf[DRM_HDCP_KSV_LEN] = { 0 };
> struct i2c_msg msgs[] = {
> @@ -759,8 +759,8 @@ int intel_gmbus_output_aksv(struct i2c_adapter *adapter)
> intel_wakeref_t wakeref;
> int ret;
>
> - wakeref = intel_display_power_get(dev_priv, POWER_DOMAIN_GMBUS);
> - mutex_lock(&dev_priv->gmbus_mutex);
> + wakeref = intel_display_power_get(i915, POWER_DOMAIN_GMBUS);
> + mutex_lock(&i915->gmbus_mutex);
>
> /*
> * In order to output Aksv to the receiver, use an indexed write to
> @@ -769,8 +769,8 @@ int intel_gmbus_output_aksv(struct i2c_adapter *adapter)
> */
> ret = do_gmbus_xfer(adapter, msgs, ARRAY_SIZE(msgs), GMBUS_AKSV_SELECT);
>
> - mutex_unlock(&dev_priv->gmbus_mutex);
> - intel_display_power_put(dev_priv, POWER_DOMAIN_GMBUS, wakeref);
> + mutex_unlock(&i915->gmbus_mutex);
> + intel_display_power_put(i915, POWER_DOMAIN_GMBUS, wakeref);
>
> return ret;
> }
> @@ -793,27 +793,27 @@ static void gmbus_lock_bus(struct i2c_adapter *adapter,
> unsigned int flags)
> {
> struct intel_gmbus *bus = to_intel_gmbus(adapter);
> - struct drm_i915_private *dev_priv = bus->dev_priv;
> + struct drm_i915_private *i915 = bus->i915;
>
> - mutex_lock(&dev_priv->gmbus_mutex);
> + mutex_lock(&i915->gmbus_mutex);
> }
>
> static int gmbus_trylock_bus(struct i2c_adapter *adapter,
> unsigned int flags)
> {
> struct intel_gmbus *bus = to_intel_gmbus(adapter);
> - struct drm_i915_private *dev_priv = bus->dev_priv;
> + struct drm_i915_private *i915 = bus->i915;
>
> - return mutex_trylock(&dev_priv->gmbus_mutex);
> + return mutex_trylock(&i915->gmbus_mutex);
> }
>
> static void gmbus_unlock_bus(struct i2c_adapter *adapter,
> unsigned int flags)
> {
> struct intel_gmbus *bus = to_intel_gmbus(adapter);
> - struct drm_i915_private *dev_priv = bus->dev_priv;
> + struct drm_i915_private *i915 = bus->i915;
>
> - mutex_unlock(&dev_priv->gmbus_mutex);
> + mutex_unlock(&i915->gmbus_mutex);
> }
>
> static const struct i2c_lock_operations gmbus_lock_ops = {
> @@ -824,45 +824,45 @@ static const struct i2c_lock_operations gmbus_lock_ops = {
>
> /**
> * intel_gmbus_setup - instantiate all Intel i2c GMBuses
> - * @dev_priv: i915 device private
> + * @i915: i915 device private
> */
> -int intel_gmbus_setup(struct drm_i915_private *dev_priv)
> +int intel_gmbus_setup(struct drm_i915_private *i915)
> {
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct pci_dev *pdev = i915->drm.pdev;
> struct intel_gmbus *bus;
> unsigned int pin;
> int ret;
>
> - if (!HAS_DISPLAY(dev_priv))
> + if (!HAS_DISPLAY(i915))
> return 0;
>
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> - dev_priv->gpio_mmio_base = VLV_DISPLAY_BASE;
> - else if (!HAS_GMCH(dev_priv))
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> + i915->gpio_mmio_base = VLV_DISPLAY_BASE;
> + else if (!HAS_GMCH(i915))
> /*
> * Broxton uses the same PCH offsets for South Display Engine,
> * even though it doesn't have a PCH.
> */
> - dev_priv->gpio_mmio_base = PCH_DISPLAY_BASE;
> + i915->gpio_mmio_base = PCH_DISPLAY_BASE;
>
> - mutex_init(&dev_priv->gmbus_mutex);
> - init_waitqueue_head(&dev_priv->gmbus_wait_queue);
> + mutex_init(&i915->gmbus_mutex);
> + init_waitqueue_head(&i915->gmbus_wait_queue);
>
> - for (pin = 0; pin < ARRAY_SIZE(dev_priv->gmbus); pin++) {
> - if (!intel_gmbus_is_valid_pin(dev_priv, pin))
> + for (pin = 0; pin < ARRAY_SIZE(i915->gmbus); pin++) {
> + if (!intel_gmbus_is_valid_pin(i915, pin))
> continue;
>
> - bus = &dev_priv->gmbus[pin];
> + bus = &i915->gmbus[pin];
>
> bus->adapter.owner = THIS_MODULE;
> bus->adapter.class = I2C_CLASS_DDC;
> snprintf(bus->adapter.name,
> sizeof(bus->adapter.name),
> "i915 gmbus %s",
> - get_gmbus_pin(dev_priv, pin)->name);
> + get_gmbus_pin(i915, pin)->name);
>
> bus->adapter.dev.parent = &pdev->dev;
> - bus->dev_priv = dev_priv;
> + bus->i915 = i915;
>
> bus->adapter.algo = &gmbus_algorithm;
> bus->adapter.lock_ops = &gmbus_lock_ops;
> @@ -877,7 +877,7 @@ int intel_gmbus_setup(struct drm_i915_private *dev_priv)
> bus->reg0 = pin | GMBUS_RATE_100KHZ;
>
> /* gmbus seems to be broken on i830 */
> - if (IS_I830(dev_priv))
> + if (IS_I830(i915))
> bus->force_bit = 1;
>
> intel_gpio_setup(bus, pin);
> @@ -887,28 +887,28 @@ int intel_gmbus_setup(struct drm_i915_private *dev_priv)
> goto err;
> }
>
> - intel_gmbus_reset(dev_priv);
> + intel_gmbus_reset(i915);
>
> return 0;
>
> err:
> while (pin--) {
> - if (!intel_gmbus_is_valid_pin(dev_priv, pin))
> + if (!intel_gmbus_is_valid_pin(i915, pin))
> continue;
>
> - bus = &dev_priv->gmbus[pin];
> + bus = &i915->gmbus[pin];
> i2c_del_adapter(&bus->adapter);
> }
> return ret;
> }
>
> -struct i2c_adapter *intel_gmbus_get_adapter(struct drm_i915_private *dev_priv,
> +struct i2c_adapter *intel_gmbus_get_adapter(struct drm_i915_private *i915,
> unsigned int pin)
> {
> - if (WARN_ON(!intel_gmbus_is_valid_pin(dev_priv, pin)))
> + if (WARN_ON(!intel_gmbus_is_valid_pin(i915, pin)))
> return NULL;
>
> - return &dev_priv->gmbus[pin].adapter;
> + return &i915->gmbus[pin].adapter;
> }
>
> void intel_gmbus_set_speed(struct i2c_adapter *adapter, int speed)
> @@ -921,16 +921,16 @@ void intel_gmbus_set_speed(struct i2c_adapter *adapter, int speed)
> void intel_gmbus_force_bit(struct i2c_adapter *adapter, bool force_bit)
> {
> struct intel_gmbus *bus = to_intel_gmbus(adapter);
> - struct drm_i915_private *dev_priv = bus->dev_priv;
> + struct drm_i915_private *i915 = bus->i915;
>
> - mutex_lock(&dev_priv->gmbus_mutex);
> + mutex_lock(&i915->gmbus_mutex);
>
> bus->force_bit += force_bit ? 1 : -1;
> DRM_DEBUG_KMS("%sabling bit-banging on %s. force bit now %d\n",
> force_bit ? "en" : "dis", adapter->name,
> bus->force_bit);
>
> - mutex_unlock(&dev_priv->gmbus_mutex);
> + mutex_unlock(&i915->gmbus_mutex);
> }
>
> bool intel_gmbus_is_forced_bit(struct i2c_adapter *adapter)
> @@ -940,16 +940,16 @@ bool intel_gmbus_is_forced_bit(struct i2c_adapter *adapter)
> return bus->force_bit;
> }
>
> -void intel_gmbus_teardown(struct drm_i915_private *dev_priv)
> +void intel_gmbus_teardown(struct drm_i915_private *i915)
> {
> struct intel_gmbus *bus;
> unsigned int pin;
>
> - for (pin = 0; pin < ARRAY_SIZE(dev_priv->gmbus); pin++) {
> - if (!intel_gmbus_is_valid_pin(dev_priv, pin))
> + for (pin = 0; pin < ARRAY_SIZE(i915->gmbus); pin++) {
> + if (!intel_gmbus_is_valid_pin(i915, pin))
> continue;
>
> - bus = &dev_priv->gmbus[pin];
> + bus = &i915->gmbus[pin];
> i2c_del_adapter(&bus->adapter);
> }
> }
> diff --git a/drivers/gpu/drm/i915/intel_gmbus.h b/drivers/gpu/drm/i915/intel_gmbus.h
> index d989085b8d22..4a4196ca8343 100644
> --- a/drivers/gpu/drm/i915/intel_gmbus.h
> +++ b/drivers/gpu/drm/i915/intel_gmbus.h
> @@ -11,17 +11,17 @@
> struct drm_i915_private;
> struct i2c_adapter;
>
> -int intel_gmbus_setup(struct drm_i915_private *dev_priv);
> -void intel_gmbus_teardown(struct drm_i915_private *dev_priv);
> -bool intel_gmbus_is_valid_pin(struct drm_i915_private *dev_priv,
> +int intel_gmbus_setup(struct drm_i915_private *i915);
> +void intel_gmbus_teardown(struct drm_i915_private *i915);
> +bool intel_gmbus_is_valid_pin(struct drm_i915_private *i915,
> unsigned int pin);
> int intel_gmbus_output_aksv(struct i2c_adapter *adapter);
>
> struct i2c_adapter *
> -intel_gmbus_get_adapter(struct drm_i915_private *dev_priv, unsigned int pin);
> +intel_gmbus_get_adapter(struct drm_i915_private *i915, unsigned int pin);
> void intel_gmbus_set_speed(struct i2c_adapter *adapter, int speed);
> void intel_gmbus_force_bit(struct i2c_adapter *adapter, bool force_bit);
> bool intel_gmbus_is_forced_bit(struct i2c_adapter *adapter);
> -void intel_gmbus_reset(struct drm_i915_private *dev_priv);
> +void intel_gmbus_reset(struct drm_i915_private *i915);
>
> #endif /* __INTEL_GMBUS_H__ */
> diff --git a/drivers/gpu/drm/i915/intel_guc.c b/drivers/gpu/drm/i915/intel_guc.c
> index c40a6efdd33a..f08ac29c2a4f 100644
> --- a/drivers/gpu/drm/i915/intel_guc.c
> +++ b/drivers/gpu/drm/i915/intel_guc.c
> @@ -29,14 +29,14 @@
>
> static void gen8_guc_raise_irq(struct intel_guc *guc)
> {
> - struct drm_i915_private *dev_priv = guc_to_i915(guc);
> + struct drm_i915_private *i915 = guc_to_i915(guc);
>
> I915_WRITE(GUC_SEND_INTERRUPT, GUC_SEND_TRIGGER);
> }
>
> static void gen11_guc_raise_irq(struct intel_guc *guc)
> {
> - struct drm_i915_private *dev_priv = guc_to_i915(guc);
> + struct drm_i915_private *i915 = guc_to_i915(guc);
>
> I915_WRITE(GEN11_GUC_HOST_INTERRUPT, 0);
> }
> @@ -52,11 +52,11 @@ static inline i915_reg_t guc_send_reg(struct intel_guc *guc, u32 i)
>
> void intel_guc_init_send_regs(struct intel_guc *guc)
> {
> - struct drm_i915_private *dev_priv = guc_to_i915(guc);
> + struct drm_i915_private *i915 = guc_to_i915(guc);
> enum forcewake_domains fw_domains = 0;
> unsigned int i;
>
> - if (INTEL_GEN(dev_priv) >= 11) {
> + if (INTEL_GEN(i915) >= 11) {
> guc->send_regs.base =
> i915_mmio_reg_offset(GEN11_SOFT_SCRATCH(0));
> guc->send_regs.count = GEN11_SOFT_SCRATCH_COUNT;
> @@ -67,7 +67,7 @@ void intel_guc_init_send_regs(struct intel_guc *guc)
> }
>
> for (i = 0; i < guc->send_regs.count; i++) {
> - fw_domains |= intel_uncore_forcewake_for_reg(&dev_priv->uncore,
> + fw_domains |= intel_uncore_forcewake_for_reg(&i915->uncore,
> guc_send_reg(guc, i),
> FW_REG_READ | FW_REG_WRITE);
> }
> @@ -101,7 +101,7 @@ void intel_guc_init_early(struct intel_guc *guc)
>
> static int guc_init_wq(struct intel_guc *guc)
> {
> - struct drm_i915_private *dev_priv = guc_to_i915(guc);
> + struct drm_i915_private *i915 = guc_to_i915(guc);
>
> /*
> * GuC log buffer flush work item has to do register access to
> @@ -135,8 +135,8 @@ static int guc_init_wq(struct intel_guc *guc)
> * to make sure we're always sending a single preemption request with a
> * single workitem.
> */
> - if (HAS_LOGICAL_RING_PREEMPTION(dev_priv) &&
> - USES_GUC_SUBMISSION(dev_priv)) {
> + if (HAS_LOGICAL_RING_PREEMPTION(i915) &&
> + USES_GUC_SUBMISSION(i915)) {
> guc->preempt_wq = alloc_ordered_workqueue("i915-guc_preempt",
> WQ_HIGHPRI);
> if (!guc->preempt_wq) {
> @@ -211,7 +211,7 @@ static void guc_shared_data_destroy(struct intel_guc *guc)
>
> int intel_guc_init(struct intel_guc *guc)
> {
> - struct drm_i915_private *dev_priv = guc_to_i915(guc);
> + struct drm_i915_private *i915 = guc_to_i915(guc);
> int ret;
>
> ret = intel_uc_fw_init(&guc->fw);
> @@ -237,7 +237,7 @@ int intel_guc_init(struct intel_guc *guc)
> goto err_ads;
>
> /* We need to notify the guc whenever we change the GGTT */
> - i915_ggtt_enable_guc(dev_priv);
> + i915_ggtt_enable_guc(i915);
>
> return 0;
>
> @@ -256,9 +256,9 @@ int intel_guc_init(struct intel_guc *guc)
>
> void intel_guc_fini(struct intel_guc *guc)
> {
> - struct drm_i915_private *dev_priv = guc_to_i915(guc);
> + struct drm_i915_private *i915 = guc_to_i915(guc);
>
> - i915_ggtt_disable_guc(dev_priv);
> + i915_ggtt_disable_guc(i915);
>
> intel_guc_ct_fini(&guc->ct);
>
> @@ -366,7 +366,7 @@ static u32 guc_ctl_ads_flags(struct intel_guc *guc)
> */
> void intel_guc_init_params(struct intel_guc *guc)
> {
> - struct drm_i915_private *dev_priv = guc_to_i915(guc);
> + struct drm_i915_private *i915 = guc_to_i915(guc);
> u32 params[GUC_CTL_MAX_DWORDS];
> int i;
>
> @@ -386,14 +386,14 @@ void intel_guc_init_params(struct intel_guc *guc)
> * they are power context saved so it's ok to release forcewake
> * when we are done here and take it again at xfer time.
> */
> - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_BLITTER);
> + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_BLITTER);
>
> I915_WRITE(SOFT_SCRATCH(0), 0);
>
> for (i = 0; i < GUC_CTL_MAX_DWORDS; i++)
> I915_WRITE(SOFT_SCRATCH(1 + i), params[i]);
>
> - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_BLITTER);
> + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_BLITTER);
> }
>
> int intel_guc_send_nop(struct intel_guc *guc, const u32 *action, u32 len,
> @@ -414,8 +414,8 @@ void intel_guc_to_host_event_handler_nop(struct intel_guc *guc)
> int intel_guc_send_mmio(struct intel_guc *guc, const u32 *action, u32 len,
> u32 *response_buf, u32 response_buf_size)
> {
> - struct drm_i915_private *dev_priv = guc_to_i915(guc);
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct drm_i915_private *i915 = guc_to_i915(guc);
> + struct intel_uncore *uncore = &i915->uncore;
> u32 status;
> int i;
> int ret;
> @@ -497,12 +497,12 @@ int intel_guc_to_host_process_recv_msg(struct intel_guc *guc,
>
> int intel_guc_sample_forcewake(struct intel_guc *guc)
> {
> - struct drm_i915_private *dev_priv = guc_to_i915(guc);
> + struct drm_i915_private *i915 = guc_to_i915(guc);
> u32 action[2];
>
> action[0] = INTEL_GUC_ACTION_SAMPLE_FORCEWAKE;
> /* WaRsDisableCoarsePowerGating:skl,cnl */
> - if (!HAS_RC6(dev_priv) || NEEDS_WaRsDisableCoarsePowerGating(dev_priv))
> + if (!HAS_RC6(i915) || NEEDS_WaRsDisableCoarsePowerGating(i915))
> action[1] = 0;
> else
> /* bit 0 and 1 are for Render and Media domain separately */
> @@ -538,7 +538,7 @@ int intel_guc_auth_huc(struct intel_guc *guc, u32 rsa_offset)
> */
> int intel_guc_suspend(struct intel_guc *guc)
> {
> - struct drm_i915_private *dev_priv = guc_to_i915(guc);
> + struct drm_i915_private *i915 = guc_to_i915(guc);
> int ret;
> u32 status;
> u32 action[] = {
> @@ -562,7 +562,7 @@ int intel_guc_suspend(struct intel_guc *guc)
> if (ret)
> return ret;
>
> - ret = __intel_wait_for_register(&dev_priv->uncore, SOFT_SCRATCH(14),
> + ret = __intel_wait_for_register(&i915->uncore, SOFT_SCRATCH(14),
> INTEL_GUC_SLEEP_STATE_INVALID_MASK,
> 0, 0, 10, &status);
> if (ret)
> @@ -658,17 +658,17 @@ int intel_guc_resume(struct intel_guc *guc)
> */
> struct i915_vma *intel_guc_allocate_vma(struct intel_guc *guc, u32 size)
> {
> - struct drm_i915_private *dev_priv = guc_to_i915(guc);
> + struct drm_i915_private *i915 = guc_to_i915(guc);
> struct drm_i915_gem_object *obj;
> struct i915_vma *vma;
> u64 flags;
> int ret;
>
> - obj = i915_gem_object_create_shmem(dev_priv, size);
> + obj = i915_gem_object_create_shmem(i915, size);
> if (IS_ERR(obj))
> return ERR_CAST(obj);
>
> - vma = i915_vma_instance(obj, &dev_priv->ggtt.vm, NULL);
> + vma = i915_vma_instance(obj, &i915->ggtt.vm, NULL);
> if (IS_ERR(vma))
> goto err;
>
> diff --git a/drivers/gpu/drm/i915/intel_guc_ads.c b/drivers/gpu/drm/i915/intel_guc_ads.c
> index ecb69fc94218..4ac9ed2dd467 100644
> --- a/drivers/gpu/drm/i915/intel_guc_ads.c
> +++ b/drivers/gpu/drm/i915/intel_guc_ads.c
> @@ -85,7 +85,7 @@ struct __guc_ads_blob {
>
> static int __guc_ads_init(struct intel_guc *guc)
> {
> - struct drm_i915_private *dev_priv = guc_to_i915(guc);
> + struct drm_i915_private *i915 = guc_to_i915(guc);
> struct __guc_ads_blob *blob;
> const u32 skipped_size = LRC_PPHWSP_SZ * PAGE_SIZE + LR_HW_CONTEXT_SIZE;
> u32 base;
> @@ -115,18 +115,18 @@ static int __guc_ads_init(struct intel_guc *guc)
> */
> blob->ads.golden_context_lrca[engine_class] = 0;
> blob->ads.eng_state_size[engine_class] =
> - intel_engine_context_size(dev_priv, engine_class) -
> + intel_engine_context_size(i915, engine_class) -
> skipped_size;
> }
>
> /* System info */
> - blob->system_info.slice_enabled = hweight8(RUNTIME_INFO(dev_priv)->sseu.slice_mask);
> + blob->system_info.slice_enabled = hweight8(RUNTIME_INFO(i915)->sseu.slice_mask);
> blob->system_info.rcs_enabled = 1;
> blob->system_info.bcs_enabled = 1;
>
> - blob->system_info.vdbox_enable_mask = VDBOX_MASK(dev_priv);
> - blob->system_info.vebox_enable_mask = VEBOX_MASK(dev_priv);
> - blob->system_info.vdbox_sfc_support_mask = RUNTIME_INFO(dev_priv)->vdbox_sfc_access;
> + blob->system_info.vdbox_enable_mask = VDBOX_MASK(i915);
> + blob->system_info.vebox_enable_mask = VEBOX_MASK(i915);
> + blob->system_info.vdbox_sfc_support_mask = RUNTIME_INFO(i915)->vdbox_sfc_access;
>
> base = intel_guc_ggtt_offset(guc, guc->ads_vma);
>
> diff --git a/drivers/gpu/drm/i915/intel_guc_fw.c b/drivers/gpu/drm/i915/intel_guc_fw.c
> index 72cdafd9636a..2beb90c62a40 100644
> --- a/drivers/gpu/drm/i915/intel_guc_fw.c
> +++ b/drivers/gpu/drm/i915/intel_guc_fw.c
> @@ -125,7 +125,7 @@ void intel_guc_fw_init_early(struct intel_guc *guc)
>
> static void guc_prepare_xfer(struct intel_guc *guc)
> {
> - struct drm_i915_private *dev_priv = guc_to_i915(guc);
> + struct drm_i915_private *i915 = guc_to_i915(guc);
>
> /* Must program this register before loading the ucode with DMA */
> I915_WRITE(GUC_SHIM_CONTROL, GUC_DISABLE_SRAM_INIT_TO_ZEROES |
> @@ -135,12 +135,12 @@ static void guc_prepare_xfer(struct intel_guc *guc)
> GUC_ENABLE_READ_CACHE_FOR_WOPCM_DATA |
> GUC_ENABLE_MIA_CLOCK_GATING);
>
> - if (IS_GEN9_LP(dev_priv))
> + if (IS_GEN9_LP(i915))
> I915_WRITE(GEN9LP_GT_PM_CONFIG, GT_DOORBELL_ENABLE);
> else
> I915_WRITE(GEN9_GT_PM_CONFIG, GT_DOORBELL_ENABLE);
>
> - if (IS_GEN(dev_priv, 9)) {
> + if (IS_GEN(i915, 9)) {
> /* DOP Clock Gating Enable for GuC clocks */
> I915_WRITE(GEN7_MISCCPCTL, (GEN8_DOP_CLOCK_GATE_GUC_ENABLE |
> I915_READ(GEN7_MISCCPCTL)));
> @@ -153,7 +153,7 @@ static void guc_prepare_xfer(struct intel_guc *guc)
> /* Copy RSA signature from the fw image to HW for verification */
> static void guc_xfer_rsa(struct intel_guc *guc)
> {
> - struct drm_i915_private *dev_priv = guc_to_i915(guc);
> + struct drm_i915_private *i915 = guc_to_i915(guc);
> struct intel_uc_fw *fw = &guc->fw;
> struct sg_table *pages = fw->obj->mm.pages;
> u32 rsa[UOS_RSA_SCRATCH_COUNT];
> @@ -168,7 +168,7 @@ static void guc_xfer_rsa(struct intel_guc *guc)
>
> static bool guc_xfer_completed(struct intel_guc *guc, u32 *status)
> {
> - struct drm_i915_private *dev_priv = guc_to_i915(guc);
> + struct drm_i915_private *i915 = guc_to_i915(guc);
>
> /* Did we complete the xfer? */
> *status = I915_READ(DMA_CTRL);
> @@ -186,7 +186,7 @@ static bool guc_xfer_completed(struct intel_guc *guc, u32 *status)
> */
> static inline bool guc_ready(struct intel_guc *guc, u32 *status)
> {
> - struct drm_i915_private *dev_priv = guc_to_i915(guc);
> + struct drm_i915_private *i915 = guc_to_i915(guc);
> u32 val = I915_READ(GUC_STATUS);
> u32 uk_val = val & GS_UKERNEL_MASK;
>
> @@ -234,7 +234,7 @@ static int guc_wait_ucode(struct intel_guc *guc)
> */
> static int guc_xfer_ucode(struct intel_guc *guc)
> {
> - struct drm_i915_private *dev_priv = guc_to_i915(guc);
> + struct drm_i915_private *i915 = guc_to_i915(guc);
> struct intel_uc_fw *guc_fw = &guc->fw;
> unsigned long offset;
>
> @@ -267,12 +267,12 @@ static int guc_xfer_ucode(struct intel_guc *guc)
> static int guc_fw_xfer(struct intel_uc_fw *guc_fw)
> {
> struct intel_guc *guc = container_of(guc_fw, struct intel_guc, fw);
> - struct drm_i915_private *dev_priv = guc_to_i915(guc);
> + struct drm_i915_private *i915 = guc_to_i915(guc);
> int ret;
>
> GEM_BUG_ON(guc_fw->type != INTEL_UC_FW_TYPE_GUC);
>
> - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
>
> guc_prepare_xfer(guc);
>
> @@ -285,7 +285,7 @@ static int guc_fw_xfer(struct intel_uc_fw *guc_fw)
>
> ret = guc_xfer_ucode(guc);
>
> - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
>
> return ret;
> }
> diff --git a/drivers/gpu/drm/i915/intel_guc_log.c b/drivers/gpu/drm/i915/intel_guc_log.c
> index 67eadc82c396..300ed940f0b7 100644
> --- a/drivers/gpu/drm/i915/intel_guc_log.c
> +++ b/drivers/gpu/drm/i915/intel_guc_log.c
> @@ -390,7 +390,7 @@ void intel_guc_log_init_early(struct intel_guc_log *log)
> static int guc_log_relay_create(struct intel_guc_log *log)
> {
> struct intel_guc *guc = log_to_guc(log);
> - struct drm_i915_private *dev_priv = guc_to_i915(guc);
> + struct drm_i915_private *i915 = guc_to_i915(guc);
> struct rchan *guc_log_relay_chan;
> size_t n_subbufs, subbuf_size;
> int ret;
> @@ -409,9 +409,9 @@ static int guc_log_relay_create(struct intel_guc_log *log)
> n_subbufs = 8;
>
> guc_log_relay_chan = relay_open("guc_log",
> - dev_priv->drm.primary->debugfs_root,
> + i915->drm.primary->debugfs_root,
> subbuf_size, n_subbufs,
> - &relay_callbacks, dev_priv);
> + &relay_callbacks, i915);
> if (!guc_log_relay_chan) {
> DRM_ERROR("Couldn't create relay chan for GuC logging\n");
>
> @@ -436,7 +436,7 @@ static void guc_log_relay_destroy(struct intel_guc_log *log)
> static void guc_log_capture_logs(struct intel_guc_log *log)
> {
> struct intel_guc *guc = log_to_guc(log);
> - struct drm_i915_private *dev_priv = guc_to_i915(guc);
> + struct drm_i915_private *i915 = guc_to_i915(guc);
> intel_wakeref_t wakeref;
>
> guc_read_update_log_buffer(log);
> @@ -445,7 +445,7 @@ static void guc_log_capture_logs(struct intel_guc_log *log)
> * Generally device is expected to be active only at this
> * time, so get/put should be really quick.
> */
> - with_intel_runtime_pm(dev_priv, wakeref)
> + with_intel_runtime_pm(i915, wakeref)
> guc_action_flush_log_complete(guc);
> }
>
> @@ -505,7 +505,7 @@ void intel_guc_log_destroy(struct intel_guc_log *log)
> int intel_guc_log_set_level(struct intel_guc_log *log, u32 level)
> {
> struct intel_guc *guc = log_to_guc(log);
> - struct drm_i915_private *dev_priv = guc_to_i915(guc);
> + struct drm_i915_private *i915 = guc_to_i915(guc);
> intel_wakeref_t wakeref;
> int ret = 0;
>
> @@ -519,12 +519,12 @@ int intel_guc_log_set_level(struct intel_guc_log *log, u32 level)
> if (level < GUC_LOG_LEVEL_DISABLED || level > GUC_LOG_LEVEL_MAX)
> return -EINVAL;
>
> - mutex_lock(&dev_priv->drm.struct_mutex);
> + mutex_lock(&i915->drm.struct_mutex);
>
> if (log->level == level)
> goto out_unlock;
>
> - with_intel_runtime_pm(dev_priv, wakeref)
> + with_intel_runtime_pm(i915, wakeref)
> ret = guc_action_control_log(guc,
> GUC_LOG_LEVEL_IS_VERBOSE(level),
> GUC_LOG_LEVEL_IS_ENABLED(level),
> @@ -537,7 +537,7 @@ int intel_guc_log_set_level(struct intel_guc_log *log, u32 level)
> log->level = level;
>
> out_unlock:
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + mutex_unlock(&i915->drm.struct_mutex);
>
> return ret;
> }
> diff --git a/drivers/gpu/drm/i915/intel_guc_submission.c b/drivers/gpu/drm/i915/intel_guc_submission.c
> index 97f6970d8da8..013d85278270 100644
> --- a/drivers/gpu/drm/i915/intel_guc_submission.c
> +++ b/drivers/gpu/drm/i915/intel_guc_submission.c
> @@ -203,7 +203,7 @@ static struct guc_doorbell_info *__get_doorbell(struct intel_guc_client *client)
>
> static bool __doorbell_valid(struct intel_guc *guc, u16 db_id)
> {
> - struct drm_i915_private *dev_priv = guc_to_i915(guc);
> + struct drm_i915_private *i915 = guc_to_i915(guc);
>
> GEM_BUG_ON(db_id >= GUC_NUM_DOORBELLS);
> return I915_READ(GEN8_DRBREGL(db_id)) & GEN8_DRB_VALID;
> @@ -1014,7 +1014,7 @@ static bool guc_verify_doorbells(struct intel_guc *guc)
>
> /**
> * guc_client_alloc() - Allocate an intel_guc_client
> - * @dev_priv: driver private data structure
> + * @i915: driver private data structure
> * @engines: The set of engines to enable for this client
> * @priority: four levels priority _CRITICAL, _HIGH, _NORMAL and _LOW
> * The kernel client to replace ExecList submission is created with
> @@ -1026,13 +1026,13 @@ static bool guc_verify_doorbells(struct intel_guc *guc)
> * Return: An intel_guc_client object if success, else NULL.
> */
> static struct intel_guc_client *
> -guc_client_alloc(struct drm_i915_private *dev_priv,
> +guc_client_alloc(struct drm_i915_private *i915,
> u32 engines,
> u32 priority,
> struct i915_gem_context *ctx)
> {
> struct intel_guc_client *client;
> - struct intel_guc *guc = &dev_priv->guc;
> + struct intel_guc *guc = &i915->guc;
> struct i915_vma *vma;
> void *vaddr;
> int ret;
> @@ -1129,27 +1129,27 @@ static inline bool ctx_save_restore_disabled(struct intel_context *ce)
>
> static int guc_clients_create(struct intel_guc *guc)
> {
> - struct drm_i915_private *dev_priv = guc_to_i915(guc);
> + struct drm_i915_private *i915 = guc_to_i915(guc);
> struct intel_guc_client *client;
>
> GEM_BUG_ON(guc->execbuf_client);
> GEM_BUG_ON(guc->preempt_client);
>
> - client = guc_client_alloc(dev_priv,
> - INTEL_INFO(dev_priv)->engine_mask,
> + client = guc_client_alloc(i915,
> + INTEL_INFO(i915)->engine_mask,
> GUC_CLIENT_PRIORITY_KMD_NORMAL,
> - dev_priv->kernel_context);
> + i915->kernel_context);
> if (IS_ERR(client)) {
> DRM_ERROR("Failed to create GuC client for submission!\n");
> return PTR_ERR(client);
> }
> guc->execbuf_client = client;
>
> - if (dev_priv->preempt_context) {
> - client = guc_client_alloc(dev_priv,
> - INTEL_INFO(dev_priv)->engine_mask,
> + if (i915->preempt_context) {
> + client = guc_client_alloc(i915,
> + INTEL_INFO(i915)->engine_mask,
> GUC_CLIENT_PRIORITY_KMD_HIGH,
> - dev_priv->preempt_context);
> + i915->preempt_context);
> if (IS_ERR(client)) {
> DRM_ERROR("Failed to create GuC client for preemption!\n");
> guc_client_free(guc->execbuf_client);
> @@ -1244,7 +1244,7 @@ static void guc_clients_disable(struct intel_guc *guc)
> */
> int intel_guc_submission_init(struct intel_guc *guc)
> {
> - struct drm_i915_private *dev_priv = guc_to_i915(guc);
> + struct drm_i915_private *i915 = guc_to_i915(guc);
> struct intel_engine_cs *engine;
> enum intel_engine_id id;
> int ret;
> @@ -1266,7 +1266,7 @@ int intel_guc_submission_init(struct intel_guc *guc)
> if (ret)
> goto err_pool;
>
> - for_each_engine(engine, dev_priv, id) {
> + for_each_engine(engine, i915, id) {
> guc->preempt_work[id].engine = engine;
> INIT_WORK(&guc->preempt_work[id].work, inject_preempt_context);
> }
> @@ -1280,11 +1280,11 @@ int intel_guc_submission_init(struct intel_guc *guc)
>
> void intel_guc_submission_fini(struct intel_guc *guc)
> {
> - struct drm_i915_private *dev_priv = guc_to_i915(guc);
> + struct drm_i915_private *i915 = guc_to_i915(guc);
> struct intel_engine_cs *engine;
> enum intel_engine_id id;
>
> - for_each_engine(engine, dev_priv, id)
> + for_each_engine(engine, i915, id)
> cancel_work_sync(&guc->preempt_work[id].work);
>
> guc_clients_destroy(guc);
> @@ -1294,9 +1294,9 @@ void intel_guc_submission_fini(struct intel_guc *guc)
> guc_stage_desc_pool_destroy(guc);
> }
>
> -static void guc_interrupts_capture(struct drm_i915_private *dev_priv)
> +static void guc_interrupts_capture(struct drm_i915_private *i915)
> {
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct intel_rps *rps = &i915->gt_pm.rps;
> struct intel_engine_cs *engine;
> enum intel_engine_id id;
> int irqs;
> @@ -1305,7 +1305,7 @@ static void guc_interrupts_capture(struct drm_i915_private *dev_priv)
> * to GuC
> */
> irqs = _MASKED_BIT_ENABLE(GFX_INTERRUPT_STEERING);
> - for_each_engine(engine, dev_priv, id)
> + for_each_engine(engine, i915, id)
> ENGINE_WRITE(engine, RING_MODE_GEN7, irqs);
>
> /* route USER_INTERRUPT to Host, all others are sent to GuC. */
> @@ -1339,9 +1339,9 @@ static void guc_interrupts_capture(struct drm_i915_private *dev_priv)
> rps->pm_intrmsk_mbz &= ~GEN8_PMINTR_DISABLE_REDIRECT_TO_GUC;
> }
>
> -static void guc_interrupts_release(struct drm_i915_private *dev_priv)
> +static void guc_interrupts_release(struct drm_i915_private *i915)
> {
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct intel_rps *rps = &i915->gt_pm.rps;
> struct intel_engine_cs *engine;
> enum intel_engine_id id;
> int irqs;
> @@ -1352,7 +1352,7 @@ static void guc_interrupts_release(struct drm_i915_private *dev_priv)
> */
> irqs = _MASKED_FIELD(GFX_FORWARD_VBLANK_MASK, GFX_FORWARD_VBLANK_NEVER);
> irqs |= _MASKED_BIT_DISABLE(GFX_INTERRUPT_STEERING);
> - for_each_engine(engine, dev_priv, id)
> + for_each_engine(engine, i915, id)
> ENGINE_WRITE(engine, RING_MODE_GEN7, irqs);
>
> /* route all GT interrupts to the host */
> @@ -1408,7 +1408,7 @@ static void guc_set_default_submission(struct intel_engine_cs *engine)
>
> int intel_guc_submission_enable(struct intel_guc *guc)
> {
> - struct drm_i915_private *dev_priv = guc_to_i915(guc);
> + struct drm_i915_private *i915 = guc_to_i915(guc);
> struct intel_engine_cs *engine;
> enum intel_engine_id id;
> int err;
> @@ -1433,9 +1433,9 @@ int intel_guc_submission_enable(struct intel_guc *guc)
> return err;
>
> /* Take over from manual control of ELSP (execlists) */
> - guc_interrupts_capture(dev_priv);
> + guc_interrupts_capture(i915);
>
> - for_each_engine(engine, dev_priv, id) {
> + for_each_engine(engine, i915, id) {
> engine->set_default_submission = guc_set_default_submission;
> engine->set_default_submission(engine);
> }
> @@ -1445,11 +1445,11 @@ int intel_guc_submission_enable(struct intel_guc *guc)
>
> void intel_guc_submission_disable(struct intel_guc *guc)
> {
> - struct drm_i915_private *dev_priv = guc_to_i915(guc);
> + struct drm_i915_private *i915 = guc_to_i915(guc);
>
> - GEM_BUG_ON(dev_priv->gt.awake); /* GT should be parked first */
> + GEM_BUG_ON(i915->gt.awake); /* GT should be parked first */
>
> - guc_interrupts_release(dev_priv);
> + guc_interrupts_release(i915);
> guc_clients_disable(guc);
> }
>
> diff --git a/drivers/gpu/drm/i915/intel_gvt.c b/drivers/gpu/drm/i915/intel_gvt.c
> index 1d7d26e4cf14..b50d31a2e095 100644
> --- a/drivers/gpu/drm/i915/intel_gvt.c
> +++ b/drivers/gpu/drm/i915/intel_gvt.c
> @@ -39,17 +39,17 @@
> * doc is available on https://01.org/group/2230/documentation-list.
> */
>
> -static bool is_supported_device(struct drm_i915_private *dev_priv)
> +static bool is_supported_device(struct drm_i915_private *i915)
> {
> - if (IS_BROADWELL(dev_priv))
> + if (IS_BROADWELL(i915))
> return true;
> - if (IS_SKYLAKE(dev_priv))
> + if (IS_SKYLAKE(i915))
> return true;
> - if (IS_KABYLAKE(dev_priv))
> + if (IS_KABYLAKE(i915))
> return true;
> - if (IS_BROXTON(dev_priv))
> + if (IS_BROXTON(i915))
> return true;
> - if (IS_COFFEELAKE(dev_priv))
> + if (IS_COFFEELAKE(i915))
> return true;
>
> return false;
> @@ -57,21 +57,21 @@ static bool is_supported_device(struct drm_i915_private *dev_priv)
>
> /**
> * intel_gvt_sanitize_options - sanitize GVT related options
> - * @dev_priv: drm i915 private data
> + * @i915: drm i915 private data
> *
> * This function is called at the i915 options sanitize stage.
> */
> -void intel_gvt_sanitize_options(struct drm_i915_private *dev_priv)
> +void intel_gvt_sanitize_options(struct drm_i915_private *i915)
> {
> if (!i915_modparams.enable_gvt)
> return;
>
> - if (intel_vgpu_active(dev_priv)) {
> + if (intel_vgpu_active(i915)) {
> DRM_INFO("GVT-g is disabled for guest\n");
> goto bail;
> }
>
> - if (!is_supported_device(dev_priv)) {
> + if (!is_supported_device(i915)) {
> DRM_INFO("Unsupported device. GVT-g is disabled\n");
> goto bail;
> }
> @@ -83,7 +83,7 @@ void intel_gvt_sanitize_options(struct drm_i915_private *dev_priv)
>
> /**
> * intel_gvt_init - initialize GVT components
> - * @dev_priv: drm i915 private data
> + * @i915: drm i915 private data
> *
> * This function is called at the initialization stage to create a GVT device.
> *
> @@ -91,7 +91,7 @@ void intel_gvt_sanitize_options(struct drm_i915_private *dev_priv)
> * Zero on success, negative error code if failed.
> *
> */
> -int intel_gvt_init(struct drm_i915_private *dev_priv)
> +int intel_gvt_init(struct drm_i915_private *i915)
> {
> int ret;
>
> @@ -103,12 +103,12 @@ int intel_gvt_init(struct drm_i915_private *dev_priv)
> return 0;
> }
>
> - if (USES_GUC_SUBMISSION(dev_priv)) {
> + if (USES_GUC_SUBMISSION(i915)) {
> DRM_ERROR("i915 GVT-g loading failed due to Graphics virtualization is not yet supported with GuC submission\n");
> return -EIO;
> }
>
> - ret = intel_gvt_init_device(dev_priv);
> + ret = intel_gvt_init_device(i915);
> if (ret) {
> DRM_DEBUG_DRIVER("Fail to init GVT device\n");
> goto bail;
> @@ -123,15 +123,15 @@ int intel_gvt_init(struct drm_i915_private *dev_priv)
>
> /**
> * intel_gvt_cleanup - cleanup GVT components when i915 driver is unloading
> - * @dev_priv: drm i915 private *
> + * @i915: drm i915 private *
> *
> * This function is called at the i915 driver unloading stage, to shutdown
> * GVT components and release the related resources.
> */
> -void intel_gvt_cleanup(struct drm_i915_private *dev_priv)
> +void intel_gvt_cleanup(struct drm_i915_private *i915)
> {
> - if (!intel_gvt_active(dev_priv))
> + if (!intel_gvt_active(i915))
> return;
>
> - intel_gvt_clean_device(dev_priv);
> + intel_gvt_clean_device(i915);
> }
> diff --git a/drivers/gpu/drm/i915/intel_gvt.h b/drivers/gpu/drm/i915/intel_gvt.h
> index 61b246470282..6f81e26cb9ad 100644
> --- a/drivers/gpu/drm/i915/intel_gvt.h
> +++ b/drivers/gpu/drm/i915/intel_gvt.h
> @@ -27,22 +27,22 @@
> struct intel_gvt;
>
> #ifdef CONFIG_DRM_I915_GVT
> -int intel_gvt_init(struct drm_i915_private *dev_priv);
> -void intel_gvt_cleanup(struct drm_i915_private *dev_priv);
> -int intel_gvt_init_device(struct drm_i915_private *dev_priv);
> -void intel_gvt_clean_device(struct drm_i915_private *dev_priv);
> +int intel_gvt_init(struct drm_i915_private *i915);
> +void intel_gvt_cleanup(struct drm_i915_private *i915);
> +int intel_gvt_init_device(struct drm_i915_private *i915);
> +void intel_gvt_clean_device(struct drm_i915_private *i915);
> int intel_gvt_init_host(void);
> -void intel_gvt_sanitize_options(struct drm_i915_private *dev_priv);
> +void intel_gvt_sanitize_options(struct drm_i915_private *i915);
> #else
> -static inline int intel_gvt_init(struct drm_i915_private *dev_priv)
> +static inline int intel_gvt_init(struct drm_i915_private *i915)
> {
> return 0;
> }
> -static inline void intel_gvt_cleanup(struct drm_i915_private *dev_priv)
> +static inline void intel_gvt_cleanup(struct drm_i915_private *i915)
> {
> }
>
> -static inline void intel_gvt_sanitize_options(struct drm_i915_private *dev_priv)
> +static inline void intel_gvt_sanitize_options(struct drm_i915_private *i915)
> {
> }
> #endif
> diff --git a/drivers/gpu/drm/i915/intel_hdcp.c b/drivers/gpu/drm/i915/intel_hdcp.c
> index bc3a94d491c4..7df5f19b52d5 100644
> --- a/drivers/gpu/drm/i915/intel_hdcp.c
> +++ b/drivers/gpu/drm/i915/intel_hdcp.c
> @@ -81,7 +81,7 @@ bool intel_hdcp_capable(struct intel_connector *connector)
> /* Is HDCP2.2 capable on Platform and Sink */
> bool intel_hdcp2_capable(struct intel_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_digital_port *intel_dig_port = conn_to_dig_port(connector);
> struct intel_hdcp *hdcp = &connector->hdcp;
> bool capable = false;
> @@ -91,12 +91,12 @@ bool intel_hdcp2_capable(struct intel_connector *connector)
> return false;
>
> /* MEI interface is solid */
> - mutex_lock(&dev_priv->hdcp_comp_mutex);
> - if (!dev_priv->hdcp_comp_added || !dev_priv->hdcp_master) {
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + mutex_lock(&i915->hdcp_comp_mutex);
> + if (!i915->hdcp_comp_added || !i915->hdcp_master) {
> + mutex_unlock(&i915->hdcp_comp_mutex);
> return false;
> }
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + mutex_unlock(&i915->hdcp_comp_mutex);
>
> /* Sink's capability for HDCP2.2 */
> hdcp->shim->hdcp_2_2_capable(intel_dig_port, &capable);
> @@ -106,7 +106,7 @@ bool intel_hdcp2_capable(struct intel_connector *connector)
>
> static inline bool intel_hdcp_in_use(struct intel_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> enum port port = connector->encoder->port;
> u32 reg;
>
> @@ -116,7 +116,7 @@ static inline bool intel_hdcp_in_use(struct intel_connector *connector)
>
> static inline bool intel_hdcp2_in_use(struct intel_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> enum port port = connector->encoder->port;
> u32 reg;
>
> @@ -145,9 +145,9 @@ static int intel_hdcp_poll_ksv_fifo(struct intel_digital_port *intel_dig_port,
> return 0;
> }
>
> -static bool hdcp_key_loadable(struct drm_i915_private *dev_priv)
> +static bool hdcp_key_loadable(struct drm_i915_private *i915)
> {
> - struct i915_power_domains *power_domains = &dev_priv->power_domains;
> + struct i915_power_domains *power_domains = &i915->power_domains;
> struct i915_power_well *power_well;
> enum i915_power_well_id id;
> bool enabled = false;
> @@ -156,7 +156,7 @@ static bool hdcp_key_loadable(struct drm_i915_private *dev_priv)
> * On HSW and BDW, Display HW loads the Key as soon as Display resumes.
> * On all BXT+, SW can load the keys only when the PW#1 is turned on.
> */
> - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> + if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> id = HSW_DISP_PW_GLOBAL;
> else
> id = SKL_DISP_PW_1;
> @@ -164,9 +164,9 @@ static bool hdcp_key_loadable(struct drm_i915_private *dev_priv)
> mutex_lock(&power_domains->lock);
>
> /* PG1 (power well #1) needs to be enabled */
> - for_each_power_well(dev_priv, power_well) {
> + for_each_power_well(i915, power_well) {
> if (power_well->desc->id == id) {
> - enabled = power_well->desc->ops->is_enabled(dev_priv,
> + enabled = power_well->desc->ops->is_enabled(i915,
> power_well);
> break;
> }
> @@ -182,14 +182,14 @@ static bool hdcp_key_loadable(struct drm_i915_private *dev_priv)
> return enabled;
> }
>
> -static void intel_hdcp_clear_keys(struct drm_i915_private *dev_priv)
> +static void intel_hdcp_clear_keys(struct drm_i915_private *i915)
> {
> I915_WRITE(HDCP_KEY_CONF, HDCP_CLEAR_KEYS_TRIGGER);
> I915_WRITE(HDCP_KEY_STATUS, HDCP_KEY_LOAD_DONE | HDCP_KEY_LOAD_STATUS |
> HDCP_FUSE_IN_PROGRESS | HDCP_FUSE_ERROR | HDCP_FUSE_DONE);
> }
>
> -static int intel_hdcp_load_keys(struct drm_i915_private *dev_priv)
> +static int intel_hdcp_load_keys(struct drm_i915_private *i915)
> {
> int ret;
> u32 val;
> @@ -202,7 +202,7 @@ static int intel_hdcp_load_keys(struct drm_i915_private *dev_priv)
> * On HSW and BDW HW loads the HDCP1.4 Key when Display comes
> * out of reset. So if Key is not already loaded, its an error state.
> */
> - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> + if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> if (!(I915_READ(HDCP_KEY_STATUS) & HDCP_KEY_LOAD_DONE))
> return -ENXIO;
>
> @@ -213,8 +213,8 @@ static int intel_hdcp_load_keys(struct drm_i915_private *dev_priv)
> * platforms except BXT and GLK, differ in the key load trigger process
> * from other platforms. So GEN9_BC uses the GT Driver Mailbox i/f.
> */
> - if (IS_GEN9_BC(dev_priv)) {
> - ret = sandybridge_pcode_write(dev_priv,
> + if (IS_GEN9_BC(i915)) {
> + ret = sandybridge_pcode_write(i915,
> SKL_PCODE_LOAD_HDCP_KEYS, 1);
> if (ret) {
> DRM_ERROR("Failed to initiate HDCP key load (%d)\n",
> @@ -226,7 +226,7 @@ static int intel_hdcp_load_keys(struct drm_i915_private *dev_priv)
> }
>
> /* Wait for the keys to load (500us) */
> - ret = __intel_wait_for_register(&dev_priv->uncore, HDCP_KEY_STATUS,
> + ret = __intel_wait_for_register(&i915->uncore, HDCP_KEY_STATUS,
> HDCP_KEY_LOAD_DONE, HDCP_KEY_LOAD_DONE,
> 10, 1, &val);
> if (ret)
> @@ -241,10 +241,10 @@ static int intel_hdcp_load_keys(struct drm_i915_private *dev_priv)
> }
>
> /* Returns updated SHA-1 index */
> -static int intel_write_sha_text(struct drm_i915_private *dev_priv, u32 sha_text)
> +static int intel_write_sha_text(struct drm_i915_private *i915, u32 sha_text)
> {
> I915_WRITE(HDCP_SHA_TEXT, sha_text);
> - if (intel_wait_for_register(&dev_priv->uncore, HDCP_REP_CTL,
> + if (intel_wait_for_register(&i915->uncore, HDCP_REP_CTL,
> HDCP_SHA1_READY, HDCP_SHA1_READY, 1)) {
> DRM_ERROR("Timed out waiting for SHA1 ready\n");
> return -ETIMEDOUT;
> @@ -279,11 +279,11 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port,
> const struct intel_hdcp_shim *shim,
> u8 *ksv_fifo, u8 num_downstream, u8 *bstatus)
> {
> - struct drm_i915_private *dev_priv;
> + struct drm_i915_private *i915;
> u32 vprime, sha_text, sha_leftovers, rep_ctl;
> int ret, i, j, sha_idx;
>
> - dev_priv = intel_dig_port->base.base.dev->dev_private;
> + i915 = intel_dig_port->base.base.dev->dev_private;
>
> /* Process V' values from the receiver */
> for (i = 0; i < DRM_HDCP_V_PRIME_NUM_PARTS; i++) {
> @@ -317,7 +317,7 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port,
> for (j = 0; j < sha_empty; j++)
> sha_text |= ksv[j] << ((sizeof(sha_text) - j - 1) * 8);
>
> - ret = intel_write_sha_text(dev_priv, sha_text);
> + ret = intel_write_sha_text(i915, sha_text);
> if (ret < 0)
> return ret;
>
> @@ -340,7 +340,7 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port,
> if (sizeof(sha_text) > sha_leftovers)
> continue;
>
> - ret = intel_write_sha_text(dev_priv, sha_text);
> + ret = intel_write_sha_text(i915, sha_text);
> if (ret < 0)
> return ret;
> sha_leftovers = 0;
> @@ -357,7 +357,7 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port,
> if (sha_leftovers == 0) {
> /* Write 16 bits of text, 16 bits of M0 */
> I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_16);
> - ret = intel_write_sha_text(dev_priv,
> + ret = intel_write_sha_text(i915,
> bstatus[0] << 8 | bstatus[1]);
> if (ret < 0)
> return ret;
> @@ -365,14 +365,14 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port,
>
> /* Write 32 bits of M0 */
> I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_0);
> - ret = intel_write_sha_text(dev_priv, 0);
> + ret = intel_write_sha_text(i915, 0);
> if (ret < 0)
> return ret;
> sha_idx += sizeof(sha_text);
>
> /* Write 16 bits of M0 */
> I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_16);
> - ret = intel_write_sha_text(dev_priv, 0);
> + ret = intel_write_sha_text(i915, 0);
> if (ret < 0)
> return ret;
> sha_idx += sizeof(sha_text);
> @@ -383,21 +383,21 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port,
> sha_text |= bstatus[0] << 16 | bstatus[1] << 8;
> /* Only 24-bits of data, must be in the LSB */
> sha_text = (sha_text & 0xffffff00) >> 8;
> - ret = intel_write_sha_text(dev_priv, sha_text);
> + ret = intel_write_sha_text(i915, sha_text);
> if (ret < 0)
> return ret;
> sha_idx += sizeof(sha_text);
>
> /* Write 32 bits of M0 */
> I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_0);
> - ret = intel_write_sha_text(dev_priv, 0);
> + ret = intel_write_sha_text(i915, 0);
> if (ret < 0)
> return ret;
> sha_idx += sizeof(sha_text);
>
> /* Write 24 bits of M0 */
> I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_8);
> - ret = intel_write_sha_text(dev_priv, 0);
> + ret = intel_write_sha_text(i915, 0);
> if (ret < 0)
> return ret;
> sha_idx += sizeof(sha_text);
> @@ -406,7 +406,7 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port,
> /* Write 32 bits of text */
> I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_32);
> sha_text |= bstatus[0] << 24 | bstatus[1] << 16;
> - ret = intel_write_sha_text(dev_priv, sha_text);
> + ret = intel_write_sha_text(i915, sha_text);
> if (ret < 0)
> return ret;
> sha_idx += sizeof(sha_text);
> @@ -414,7 +414,7 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port,
> /* Write 64 bits of M0 */
> I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_0);
> for (i = 0; i < 2; i++) {
> - ret = intel_write_sha_text(dev_priv, 0);
> + ret = intel_write_sha_text(i915, 0);
> if (ret < 0)
> return ret;
> sha_idx += sizeof(sha_text);
> @@ -423,28 +423,28 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port,
> /* Write 32 bits of text */
> I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_32);
> sha_text |= bstatus[0] << 24;
> - ret = intel_write_sha_text(dev_priv, sha_text);
> + ret = intel_write_sha_text(i915, sha_text);
> if (ret < 0)
> return ret;
> sha_idx += sizeof(sha_text);
>
> /* Write 8 bits of text, 24 bits of M0 */
> I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_8);
> - ret = intel_write_sha_text(dev_priv, bstatus[1]);
> + ret = intel_write_sha_text(i915, bstatus[1]);
> if (ret < 0)
> return ret;
> sha_idx += sizeof(sha_text);
>
> /* Write 32 bits of M0 */
> I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_0);
> - ret = intel_write_sha_text(dev_priv, 0);
> + ret = intel_write_sha_text(i915, 0);
> if (ret < 0)
> return ret;
> sha_idx += sizeof(sha_text);
>
> /* Write 8 bits of M0 */
> I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_24);
> - ret = intel_write_sha_text(dev_priv, 0);
> + ret = intel_write_sha_text(i915, 0);
> if (ret < 0)
> return ret;
> sha_idx += sizeof(sha_text);
> @@ -457,7 +457,7 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port,
> I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_32);
> /* Fill up to 64-4 bytes with zeros (leave the last write for length) */
> while ((sha_idx % 64) < (64 - sizeof(sha_text))) {
> - ret = intel_write_sha_text(dev_priv, 0);
> + ret = intel_write_sha_text(i915, 0);
> if (ret < 0)
> return ret;
> sha_idx += sizeof(sha_text);
> @@ -469,13 +469,13 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port,
> * - 10 bytes for BINFO/BSTATUS(2), M0(8)
> */
> sha_text = (num_downstream * 5 + 10) * 8;
> - ret = intel_write_sha_text(dev_priv, sha_text);
> + ret = intel_write_sha_text(i915, sha_text);
> if (ret < 0)
> return ret;
>
> /* Tell the HW we're done with the hash and wait for it to ACK */
> I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_COMPLETE_HASH);
> - if (intel_wait_for_register(&dev_priv->uncore, HDCP_REP_CTL,
> + if (intel_wait_for_register(&i915->uncore, HDCP_REP_CTL,
> HDCP_SHA1_COMPLETE,
> HDCP_SHA1_COMPLETE, 1)) {
> DRM_ERROR("Timed out waiting for SHA1 complete\n");
> @@ -571,7 +571,7 @@ static int intel_hdcp_auth(struct intel_connector *connector)
> struct intel_hdcp *hdcp = &connector->hdcp;
> struct drm_device *dev = connector->base.dev;
> const struct intel_hdcp_shim *shim = hdcp->shim;
> - struct drm_i915_private *dev_priv;
> + struct drm_i915_private *i915;
> enum port port;
> unsigned long r0_prime_gen_start;
> int ret, i, tries = 2;
> @@ -589,7 +589,7 @@ static int intel_hdcp_auth(struct intel_connector *connector)
> } ri;
> bool repeater_present, hdcp_capable;
>
> - dev_priv = intel_dig_port->base.base.dev->dev_private;
> + i915 = intel_dig_port->base.base.dev->dev_private;
>
> port = intel_dig_port->base.port;
>
> @@ -615,7 +615,7 @@ static int intel_hdcp_auth(struct intel_connector *connector)
> I915_WRITE(PORT_HDCP_CONF(port), HDCP_CONF_CAPTURE_AN);
>
> /* Wait for An to be acquired */
> - if (intel_wait_for_register(&dev_priv->uncore, PORT_HDCP_STATUS(port),
> + if (intel_wait_for_register(&i915->uncore, PORT_HDCP_STATUS(port),
> HDCP_STATUS_AN_READY,
> HDCP_STATUS_AN_READY, 1)) {
> DRM_ERROR("Timed out waiting for An\n");
> @@ -701,7 +701,7 @@ static int intel_hdcp_auth(struct intel_connector *connector)
> }
>
> /* Wait for encryption confirmation */
> - if (intel_wait_for_register(&dev_priv->uncore, PORT_HDCP_STATUS(port),
> + if (intel_wait_for_register(&i915->uncore, PORT_HDCP_STATUS(port),
> HDCP_STATUS_ENC, HDCP_STATUS_ENC,
> ENCRYPT_STATUS_CHANGE_TIMEOUT_MS)) {
> DRM_ERROR("Timed out waiting for encryption\n");
> @@ -723,7 +723,7 @@ static int intel_hdcp_auth(struct intel_connector *connector)
> static int _intel_hdcp_disable(struct intel_connector *connector)
> {
> struct intel_hdcp *hdcp = &connector->hdcp;
> - struct drm_i915_private *dev_priv = connector->base.dev->dev_private;
> + struct drm_i915_private *i915 = connector->base.dev->dev_private;
> struct intel_digital_port *intel_dig_port = conn_to_dig_port(connector);
> enum port port = intel_dig_port->base.port;
> int ret;
> @@ -733,7 +733,7 @@ static int _intel_hdcp_disable(struct intel_connector *connector)
>
> hdcp->hdcp_encrypted = false;
> I915_WRITE(PORT_HDCP_CONF(port), 0);
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> PORT_HDCP_STATUS(port), ~0, 0,
> ENCRYPT_STATUS_CHANGE_TIMEOUT_MS)) {
> DRM_ERROR("Failed to disable HDCP, timeout clearing status\n");
> @@ -753,22 +753,22 @@ static int _intel_hdcp_disable(struct intel_connector *connector)
> static int _intel_hdcp_enable(struct intel_connector *connector)
> {
> struct intel_hdcp *hdcp = &connector->hdcp;
> - struct drm_i915_private *dev_priv = connector->base.dev->dev_private;
> + struct drm_i915_private *i915 = connector->base.dev->dev_private;
> int i, ret, tries = 3;
>
> DRM_DEBUG_KMS("[%s:%d] HDCP is being enabled...\n",
> connector->base.name, connector->base.base.id);
>
> - if (!hdcp_key_loadable(dev_priv)) {
> + if (!hdcp_key_loadable(i915)) {
> DRM_ERROR("HDCP key Load is not possible\n");
> return -ENXIO;
> }
>
> for (i = 0; i < KEY_LOAD_TRIES; i++) {
> - ret = intel_hdcp_load_keys(dev_priv);
> + ret = intel_hdcp_load_keys(i915);
> if (!ret)
> break;
> - intel_hdcp_clear_keys(dev_priv);
> + intel_hdcp_clear_keys(i915);
> }
> if (ret) {
> DRM_ERROR("Could not load HDCP keys, (%d)\n", ret);
> @@ -803,7 +803,7 @@ struct intel_connector *intel_hdcp_to_connector(struct intel_hdcp *hdcp)
> static int intel_hdcp_check_link(struct intel_connector *connector)
> {
> struct intel_hdcp *hdcp = &connector->hdcp;
> - struct drm_i915_private *dev_priv = connector->base.dev->dev_private;
> + struct drm_i915_private *i915 = connector->base.dev->dev_private;
> struct intel_digital_port *intel_dig_port = conn_to_dig_port(connector);
> enum port port = intel_dig_port->base.port;
> int ret = 0;
> @@ -884,10 +884,10 @@ static void intel_hdcp_prop_work(struct work_struct *work)
> drm_modeset_unlock(&dev->mode_config.connection_mutex);
> }
>
> -bool is_hdcp_supported(struct drm_i915_private *dev_priv, enum port port)
> +bool is_hdcp_supported(struct drm_i915_private *i915, enum port port)
> {
> /* PORT E doesn't have HDCP, and PORT F is disabled */
> - return INTEL_GEN(dev_priv) >= 9 && port < PORT_E;
> + return INTEL_GEN(i915) >= 9 && port < PORT_E;
> }
>
> static int
> @@ -895,22 +895,22 @@ hdcp2_prepare_ake_init(struct intel_connector *connector,
> struct hdcp2_ake_init *ake_data)
> {
> struct hdcp_port_data *data = &connector->hdcp.port_data;
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct i915_hdcp_comp_master *comp;
> int ret;
>
> - mutex_lock(&dev_priv->hdcp_comp_mutex);
> - comp = dev_priv->hdcp_master;
> + mutex_lock(&i915->hdcp_comp_mutex);
> + comp = i915->hdcp_master;
>
> if (!comp || !comp->ops) {
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + mutex_unlock(&i915->hdcp_comp_mutex);
> return -EINVAL;
> }
>
> ret = comp->ops->initiate_hdcp2_session(comp->mei_dev, data, ake_data);
> if (ret)
> DRM_DEBUG_KMS("Prepare_ake_init failed. %d\n", ret);
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + mutex_unlock(&i915->hdcp_comp_mutex);
>
> return ret;
> }
> @@ -923,15 +923,15 @@ hdcp2_verify_rx_cert_prepare_km(struct intel_connector *connector,
> size_t *msg_sz)
> {
> struct hdcp_port_data *data = &connector->hdcp.port_data;
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct i915_hdcp_comp_master *comp;
> int ret;
>
> - mutex_lock(&dev_priv->hdcp_comp_mutex);
> - comp = dev_priv->hdcp_master;
> + mutex_lock(&i915->hdcp_comp_mutex);
> + comp = i915->hdcp_master;
>
> if (!comp || !comp->ops) {
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + mutex_unlock(&i915->hdcp_comp_mutex);
> return -EINVAL;
> }
>
> @@ -940,7 +940,7 @@ hdcp2_verify_rx_cert_prepare_km(struct intel_connector *connector,
> ek_pub_km, msg_sz);
> if (ret < 0)
> DRM_DEBUG_KMS("Verify rx_cert failed. %d\n", ret);
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + mutex_unlock(&i915->hdcp_comp_mutex);
>
> return ret;
> }
> @@ -949,22 +949,22 @@ static int hdcp2_verify_hprime(struct intel_connector *connector,
> struct hdcp2_ake_send_hprime *rx_hprime)
> {
> struct hdcp_port_data *data = &connector->hdcp.port_data;
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct i915_hdcp_comp_master *comp;
> int ret;
>
> - mutex_lock(&dev_priv->hdcp_comp_mutex);
> - comp = dev_priv->hdcp_master;
> + mutex_lock(&i915->hdcp_comp_mutex);
> + comp = i915->hdcp_master;
>
> if (!comp || !comp->ops) {
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + mutex_unlock(&i915->hdcp_comp_mutex);
> return -EINVAL;
> }
>
> ret = comp->ops->verify_hprime(comp->mei_dev, data, rx_hprime);
> if (ret < 0)
> DRM_DEBUG_KMS("Verify hprime failed. %d\n", ret);
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + mutex_unlock(&i915->hdcp_comp_mutex);
>
> return ret;
> }
> @@ -974,22 +974,22 @@ hdcp2_store_pairing_info(struct intel_connector *connector,
> struct hdcp2_ake_send_pairing_info *pairing_info)
> {
> struct hdcp_port_data *data = &connector->hdcp.port_data;
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct i915_hdcp_comp_master *comp;
> int ret;
>
> - mutex_lock(&dev_priv->hdcp_comp_mutex);
> - comp = dev_priv->hdcp_master;
> + mutex_lock(&i915->hdcp_comp_mutex);
> + comp = i915->hdcp_master;
>
> if (!comp || !comp->ops) {
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + mutex_unlock(&i915->hdcp_comp_mutex);
> return -EINVAL;
> }
>
> ret = comp->ops->store_pairing_info(comp->mei_dev, data, pairing_info);
> if (ret < 0)
> DRM_DEBUG_KMS("Store pairing info failed. %d\n", ret);
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + mutex_unlock(&i915->hdcp_comp_mutex);
>
> return ret;
> }
> @@ -999,22 +999,22 @@ hdcp2_prepare_lc_init(struct intel_connector *connector,
> struct hdcp2_lc_init *lc_init)
> {
> struct hdcp_port_data *data = &connector->hdcp.port_data;
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct i915_hdcp_comp_master *comp;
> int ret;
>
> - mutex_lock(&dev_priv->hdcp_comp_mutex);
> - comp = dev_priv->hdcp_master;
> + mutex_lock(&i915->hdcp_comp_mutex);
> + comp = i915->hdcp_master;
>
> if (!comp || !comp->ops) {
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + mutex_unlock(&i915->hdcp_comp_mutex);
> return -EINVAL;
> }
>
> ret = comp->ops->initiate_locality_check(comp->mei_dev, data, lc_init);
> if (ret < 0)
> DRM_DEBUG_KMS("Prepare lc_init failed. %d\n", ret);
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + mutex_unlock(&i915->hdcp_comp_mutex);
>
> return ret;
> }
> @@ -1024,22 +1024,22 @@ hdcp2_verify_lprime(struct intel_connector *connector,
> struct hdcp2_lc_send_lprime *rx_lprime)
> {
> struct hdcp_port_data *data = &connector->hdcp.port_data;
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct i915_hdcp_comp_master *comp;
> int ret;
>
> - mutex_lock(&dev_priv->hdcp_comp_mutex);
> - comp = dev_priv->hdcp_master;
> + mutex_lock(&i915->hdcp_comp_mutex);
> + comp = i915->hdcp_master;
>
> if (!comp || !comp->ops) {
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + mutex_unlock(&i915->hdcp_comp_mutex);
> return -EINVAL;
> }
>
> ret = comp->ops->verify_lprime(comp->mei_dev, data, rx_lprime);
> if (ret < 0)
> DRM_DEBUG_KMS("Verify L_Prime failed. %d\n", ret);
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + mutex_unlock(&i915->hdcp_comp_mutex);
>
> return ret;
> }
> @@ -1048,22 +1048,22 @@ static int hdcp2_prepare_skey(struct intel_connector *connector,
> struct hdcp2_ske_send_eks *ske_data)
> {
> struct hdcp_port_data *data = &connector->hdcp.port_data;
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct i915_hdcp_comp_master *comp;
> int ret;
>
> - mutex_lock(&dev_priv->hdcp_comp_mutex);
> - comp = dev_priv->hdcp_master;
> + mutex_lock(&i915->hdcp_comp_mutex);
> + comp = i915->hdcp_master;
>
> if (!comp || !comp->ops) {
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + mutex_unlock(&i915->hdcp_comp_mutex);
> return -EINVAL;
> }
>
> ret = comp->ops->get_session_key(comp->mei_dev, data, ske_data);
> if (ret < 0)
> DRM_DEBUG_KMS("Get session key failed. %d\n", ret);
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + mutex_unlock(&i915->hdcp_comp_mutex);
>
> return ret;
> }
> @@ -1075,15 +1075,15 @@ hdcp2_verify_rep_topology_prepare_ack(struct intel_connector *connector,
> struct hdcp2_rep_send_ack *rep_send_ack)
> {
> struct hdcp_port_data *data = &connector->hdcp.port_data;
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct i915_hdcp_comp_master *comp;
> int ret;
>
> - mutex_lock(&dev_priv->hdcp_comp_mutex);
> - comp = dev_priv->hdcp_master;
> + mutex_lock(&i915->hdcp_comp_mutex);
> + comp = i915->hdcp_master;
>
> if (!comp || !comp->ops) {
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + mutex_unlock(&i915->hdcp_comp_mutex);
> return -EINVAL;
> }
>
> @@ -1092,7 +1092,7 @@ hdcp2_verify_rep_topology_prepare_ack(struct intel_connector *connector,
> rep_send_ack);
> if (ret < 0)
> DRM_DEBUG_KMS("Verify rep topology failed. %d\n", ret);
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + mutex_unlock(&i915->hdcp_comp_mutex);
>
> return ret;
> }
> @@ -1102,22 +1102,22 @@ hdcp2_verify_mprime(struct intel_connector *connector,
> struct hdcp2_rep_stream_ready *stream_ready)
> {
> struct hdcp_port_data *data = &connector->hdcp.port_data;
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct i915_hdcp_comp_master *comp;
> int ret;
>
> - mutex_lock(&dev_priv->hdcp_comp_mutex);
> - comp = dev_priv->hdcp_master;
> + mutex_lock(&i915->hdcp_comp_mutex);
> + comp = i915->hdcp_master;
>
> if (!comp || !comp->ops) {
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + mutex_unlock(&i915->hdcp_comp_mutex);
> return -EINVAL;
> }
>
> ret = comp->ops->verify_mprime(comp->mei_dev, data, stream_ready);
> if (ret < 0)
> DRM_DEBUG_KMS("Verify mprime failed. %d\n", ret);
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + mutex_unlock(&i915->hdcp_comp_mutex);
>
> return ret;
> }
> @@ -1125,43 +1125,43 @@ hdcp2_verify_mprime(struct intel_connector *connector,
> static int hdcp2_authenticate_port(struct intel_connector *connector)
> {
> struct hdcp_port_data *data = &connector->hdcp.port_data;
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct i915_hdcp_comp_master *comp;
> int ret;
>
> - mutex_lock(&dev_priv->hdcp_comp_mutex);
> - comp = dev_priv->hdcp_master;
> + mutex_lock(&i915->hdcp_comp_mutex);
> + comp = i915->hdcp_master;
>
> if (!comp || !comp->ops) {
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + mutex_unlock(&i915->hdcp_comp_mutex);
> return -EINVAL;
> }
>
> ret = comp->ops->enable_hdcp_authentication(comp->mei_dev, data);
> if (ret < 0)
> DRM_DEBUG_KMS("Enable hdcp auth failed. %d\n", ret);
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + mutex_unlock(&i915->hdcp_comp_mutex);
>
> return ret;
> }
>
> static int hdcp2_close_mei_session(struct intel_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct i915_hdcp_comp_master *comp;
> int ret;
>
> - mutex_lock(&dev_priv->hdcp_comp_mutex);
> - comp = dev_priv->hdcp_master;
> + mutex_lock(&i915->hdcp_comp_mutex);
> + comp = i915->hdcp_master;
>
> if (!comp || !comp->ops) {
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + mutex_unlock(&i915->hdcp_comp_mutex);
> return -EINVAL;
> }
>
> ret = comp->ops->close_hdcp_session(comp->mei_dev,
> &connector->hdcp.port_data);
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + mutex_unlock(&i915->hdcp_comp_mutex);
>
> return ret;
> }
> @@ -1488,7 +1488,7 @@ static int hdcp2_authenticate_sink(struct intel_connector *connector)
> static int hdcp2_enable_encryption(struct intel_connector *connector)
> {
> struct intel_digital_port *intel_dig_port = conn_to_dig_port(connector);
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_hdcp *hdcp = &connector->hdcp;
> enum port port = connector->encoder->port;
> int ret;
> @@ -1511,7 +1511,7 @@ static int hdcp2_enable_encryption(struct intel_connector *connector)
> CTL_LINK_ENCRYPTION_REQ);
> }
>
> - ret = intel_wait_for_register(&dev_priv->uncore, HDCP2_STATUS_DDI(port),
> + ret = intel_wait_for_register(&i915->uncore, HDCP2_STATUS_DDI(port),
> LINK_ENCRYPTION_STATUS,
> LINK_ENCRYPTION_STATUS,
> ENCRYPT_STATUS_CHANGE_TIMEOUT_MS);
> @@ -1522,7 +1522,7 @@ static int hdcp2_enable_encryption(struct intel_connector *connector)
> static int hdcp2_disable_encryption(struct intel_connector *connector)
> {
> struct intel_digital_port *intel_dig_port = conn_to_dig_port(connector);
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_hdcp *hdcp = &connector->hdcp;
> enum port port = connector->encoder->port;
> int ret;
> @@ -1532,7 +1532,7 @@ static int hdcp2_disable_encryption(struct intel_connector *connector)
> I915_WRITE(HDCP2_CTL_DDI(port),
> I915_READ(HDCP2_CTL_DDI(port)) & ~CTL_LINK_ENCRYPTION_REQ);
>
> - ret = intel_wait_for_register(&dev_priv->uncore, HDCP2_STATUS_DDI(port),
> + ret = intel_wait_for_register(&i915->uncore, HDCP2_STATUS_DDI(port),
> LINK_ENCRYPTION_STATUS, 0x0,
> ENCRYPT_STATUS_CHANGE_TIMEOUT_MS);
> if (ret == -ETIMEDOUT)
> @@ -1628,7 +1628,7 @@ static int _intel_hdcp2_disable(struct intel_connector *connector)
> static int intel_hdcp2_check_link(struct intel_connector *connector)
> {
> struct intel_digital_port *intel_dig_port = conn_to_dig_port(connector);
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_hdcp *hdcp = &connector->hdcp;
> enum port port = connector->encoder->port;
> int ret = 0;
> @@ -1721,13 +1721,13 @@ static void intel_hdcp_check_work(struct work_struct *work)
> static int i915_hdcp_component_bind(struct device *i915_kdev,
> struct device *mei_kdev, void *data)
> {
> - struct drm_i915_private *dev_priv = kdev_to_i915(i915_kdev);
> + struct drm_i915_private *i915 = kdev_to_i915(i915_kdev);
>
> DRM_DEBUG("I915 HDCP comp bind\n");
> - mutex_lock(&dev_priv->hdcp_comp_mutex);
> - dev_priv->hdcp_master = (struct i915_hdcp_comp_master *)data;
> - dev_priv->hdcp_master->mei_dev = mei_kdev;
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + mutex_lock(&i915->hdcp_comp_mutex);
> + i915->hdcp_master = (struct i915_hdcp_comp_master *)data;
> + i915->hdcp_master->mei_dev = mei_kdev;
> + mutex_unlock(&i915->hdcp_comp_mutex);
>
> return 0;
> }
> @@ -1735,12 +1735,12 @@ static int i915_hdcp_component_bind(struct device *i915_kdev,
> static void i915_hdcp_component_unbind(struct device *i915_kdev,
> struct device *mei_kdev, void *data)
> {
> - struct drm_i915_private *dev_priv = kdev_to_i915(i915_kdev);
> + struct drm_i915_private *i915 = kdev_to_i915(i915_kdev);
>
> DRM_DEBUG("I915 HDCP comp unbind\n");
> - mutex_lock(&dev_priv->hdcp_comp_mutex);
> - dev_priv->hdcp_master = NULL;
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + mutex_lock(&i915->hdcp_comp_mutex);
> + i915->hdcp_master = NULL;
> + mutex_unlock(&i915->hdcp_comp_mutex);
> }
>
> static const struct component_ops i915_hdcp_component_ops = {
> @@ -1773,34 +1773,34 @@ static inline int initialize_hdcp_port_data(struct intel_connector *connector)
> return 0;
> }
>
> -static bool is_hdcp2_supported(struct drm_i915_private *dev_priv)
> +static bool is_hdcp2_supported(struct drm_i915_private *i915)
> {
> if (!IS_ENABLED(CONFIG_INTEL_MEI_HDCP))
> return false;
>
> - return (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv) ||
> - IS_KABYLAKE(dev_priv));
> + return (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915) ||
> + IS_KABYLAKE(i915));
> }
>
> -void intel_hdcp_component_init(struct drm_i915_private *dev_priv)
> +void intel_hdcp_component_init(struct drm_i915_private *i915)
> {
> int ret;
>
> - if (!is_hdcp2_supported(dev_priv))
> + if (!is_hdcp2_supported(i915))
> return;
>
> - mutex_lock(&dev_priv->hdcp_comp_mutex);
> - WARN_ON(dev_priv->hdcp_comp_added);
> + mutex_lock(&i915->hdcp_comp_mutex);
> + WARN_ON(i915->hdcp_comp_added);
>
> - dev_priv->hdcp_comp_added = true;
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> - ret = component_add_typed(dev_priv->drm.dev, &i915_hdcp_component_ops,
> + i915->hdcp_comp_added = true;
> + mutex_unlock(&i915->hdcp_comp_mutex);
> + ret = component_add_typed(i915->drm.dev, &i915_hdcp_component_ops,
> I915_COMPONENT_HDCP);
> if (ret < 0) {
> DRM_DEBUG_KMS("Failed at component add(%d)\n", ret);
> - mutex_lock(&dev_priv->hdcp_comp_mutex);
> - dev_priv->hdcp_comp_added = false;
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + mutex_lock(&i915->hdcp_comp_mutex);
> + i915->hdcp_comp_added = false;
> + mutex_unlock(&i915->hdcp_comp_mutex);
> return;
> }
> }
> @@ -1822,7 +1822,7 @@ static void intel_hdcp2_init(struct intel_connector *connector)
> int intel_hdcp_init(struct intel_connector *connector,
> const struct intel_hdcp_shim *shim)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_hdcp *hdcp = &connector->hdcp;
> int ret;
>
> @@ -1838,7 +1838,7 @@ int intel_hdcp_init(struct intel_connector *connector,
> INIT_DELAYED_WORK(&hdcp->check_work, intel_hdcp_check_work);
> INIT_WORK(&hdcp->prop_work, intel_hdcp_prop_work);
>
> - if (is_hdcp2_supported(dev_priv))
> + if (is_hdcp2_supported(i915))
> intel_hdcp2_init(connector);
> init_waitqueue_head(&hdcp->cp_irq_queue);
>
> @@ -1905,18 +1905,18 @@ int intel_hdcp_disable(struct intel_connector *connector)
> return ret;
> }
>
> -void intel_hdcp_component_fini(struct drm_i915_private *dev_priv)
> +void intel_hdcp_component_fini(struct drm_i915_private *i915)
> {
> - mutex_lock(&dev_priv->hdcp_comp_mutex);
> - if (!dev_priv->hdcp_comp_added) {
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + mutex_lock(&i915->hdcp_comp_mutex);
> + if (!i915->hdcp_comp_added) {
> + mutex_unlock(&i915->hdcp_comp_mutex);
> return;
> }
>
> - dev_priv->hdcp_comp_added = false;
> - mutex_unlock(&dev_priv->hdcp_comp_mutex);
> + i915->hdcp_comp_added = false;
> + mutex_unlock(&i915->hdcp_comp_mutex);
>
> - component_del(dev_priv->drm.dev, &i915_hdcp_component_ops);
> + component_del(i915->drm.dev, &i915_hdcp_component_ops);
> }
>
> void intel_hdcp_cleanup(struct intel_connector *connector)
> diff --git a/drivers/gpu/drm/i915/intel_hdcp.h b/drivers/gpu/drm/i915/intel_hdcp.h
> index be8da85c866a..24079dda1ac4 100644
> --- a/drivers/gpu/drm/i915/intel_hdcp.h
> +++ b/drivers/gpu/drm/i915/intel_hdcp.h
> @@ -23,11 +23,11 @@ int intel_hdcp_init(struct intel_connector *connector,
> const struct intel_hdcp_shim *hdcp_shim);
> int intel_hdcp_enable(struct intel_connector *connector);
> int intel_hdcp_disable(struct intel_connector *connector);
> -bool is_hdcp_supported(struct drm_i915_private *dev_priv, enum port port);
> +bool is_hdcp_supported(struct drm_i915_private *i915, enum port port);
> bool intel_hdcp_capable(struct intel_connector *connector);
> bool intel_hdcp2_capable(struct intel_connector *connector);
> -void intel_hdcp_component_init(struct drm_i915_private *dev_priv);
> -void intel_hdcp_component_fini(struct drm_i915_private *dev_priv);
> +void intel_hdcp_component_init(struct drm_i915_private *i915);
> +void intel_hdcp_component_fini(struct drm_i915_private *i915);
> void intel_hdcp_cleanup(struct intel_connector *connector);
> void intel_hdcp_handle_cp_irq(struct intel_connector *connector);
>
> diff --git a/drivers/gpu/drm/i915/intel_hdmi.c b/drivers/gpu/drm/i915/intel_hdmi.c
> index 187a2b828b97..c5be41ba36da 100644
> --- a/drivers/gpu/drm/i915/intel_hdmi.c
> +++ b/drivers/gpu/drm/i915/intel_hdmi.c
> @@ -67,17 +67,17 @@ static void
> assert_hdmi_port_disabled(struct intel_hdmi *intel_hdmi)
> {
> struct drm_device *dev = intel_hdmi_to_dev(intel_hdmi);
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> u32 enabled_bits;
>
> - enabled_bits = HAS_DDI(dev_priv) ? DDI_BUF_CTL_ENABLE : SDVO_ENABLE;
> + enabled_bits = HAS_DDI(i915) ? DDI_BUF_CTL_ENABLE : SDVO_ENABLE;
>
> WARN(I915_READ(intel_hdmi->hdmi_reg) & enabled_bits,
> "HDMI port enabled, expecting disabled\n");
> }
>
> static void
> -assert_hdmi_transcoder_func_disabled(struct drm_i915_private *dev_priv,
> +assert_hdmi_transcoder_func_disabled(struct drm_i915_private *i915,
> enum transcoder cpu_transcoder)
> {
> WARN(I915_READ(TRANS_DDI_FUNC_CTL(cpu_transcoder)) &
> @@ -163,7 +163,7 @@ static u32 hsw_infoframe_enable(unsigned int type)
> }
>
> static i915_reg_t
> -hsw_dip_data_reg(struct drm_i915_private *dev_priv,
> +hsw_dip_data_reg(struct drm_i915_private *i915,
> enum transcoder cpu_transcoder,
> unsigned int type,
> int i)
> @@ -207,7 +207,7 @@ static void g4x_write_infoframe(struct intel_encoder *encoder,
> const void *frame, ssize_t len)
> {
> const u32 *data = frame;
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> u32 val = I915_READ(VIDEO_DIP_CTL);
> int i;
>
> @@ -241,7 +241,7 @@ static void g4x_read_infoframe(struct intel_encoder *encoder,
> unsigned int type,
> void *frame, ssize_t len)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> u32 val, *data = frame;
> int i;
>
> @@ -259,7 +259,7 @@ static void g4x_read_infoframe(struct intel_encoder *encoder,
> static u32 g4x_infoframes_enabled(struct intel_encoder *encoder,
> const struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> u32 val = I915_READ(VIDEO_DIP_CTL);
>
> if ((val & VIDEO_DIP_ENABLE) == 0)
> @@ -278,7 +278,7 @@ static void ibx_write_infoframe(struct intel_encoder *encoder,
> const void *frame, ssize_t len)
> {
> const u32 *data = frame;
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc);
> i915_reg_t reg = TVIDEO_DIP_CTL(intel_crtc->pipe);
> u32 val = I915_READ(reg);
> @@ -314,7 +314,7 @@ static void ibx_read_infoframe(struct intel_encoder *encoder,
> unsigned int type,
> void *frame, ssize_t len)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> u32 val, *data = frame;
> int i;
> @@ -333,7 +333,7 @@ static void ibx_read_infoframe(struct intel_encoder *encoder,
> static u32 ibx_infoframes_enabled(struct intel_encoder *encoder,
> const struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> enum pipe pipe = to_intel_crtc(pipe_config->base.crtc)->pipe;
> i915_reg_t reg = TVIDEO_DIP_CTL(pipe);
> u32 val = I915_READ(reg);
> @@ -355,7 +355,7 @@ static void cpt_write_infoframe(struct intel_encoder *encoder,
> const void *frame, ssize_t len)
> {
> const u32 *data = frame;
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc);
> i915_reg_t reg = TVIDEO_DIP_CTL(intel_crtc->pipe);
> u32 val = I915_READ(reg);
> @@ -394,7 +394,7 @@ static void cpt_read_infoframe(struct intel_encoder *encoder,
> unsigned int type,
> void *frame, ssize_t len)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> u32 val, *data = frame;
> int i;
> @@ -413,7 +413,7 @@ static void cpt_read_infoframe(struct intel_encoder *encoder,
> static u32 cpt_infoframes_enabled(struct intel_encoder *encoder,
> const struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> enum pipe pipe = to_intel_crtc(pipe_config->base.crtc)->pipe;
> u32 val = I915_READ(TVIDEO_DIP_CTL(pipe));
>
> @@ -431,7 +431,7 @@ static void vlv_write_infoframe(struct intel_encoder *encoder,
> const void *frame, ssize_t len)
> {
> const u32 *data = frame;
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc);
> i915_reg_t reg = VLV_TVIDEO_DIP_CTL(intel_crtc->pipe);
> u32 val = I915_READ(reg);
> @@ -467,7 +467,7 @@ static void vlv_read_infoframe(struct intel_encoder *encoder,
> unsigned int type,
> void *frame, ssize_t len)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> u32 val, *data = frame;
> int i;
> @@ -486,7 +486,7 @@ static void vlv_read_infoframe(struct intel_encoder *encoder,
> static u32 vlv_infoframes_enabled(struct intel_encoder *encoder,
> const struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> enum pipe pipe = to_intel_crtc(pipe_config->base.crtc)->pipe;
> u32 val = I915_READ(VLV_TVIDEO_DIP_CTL(pipe));
>
> @@ -507,7 +507,7 @@ static void hsw_write_infoframe(struct intel_encoder *encoder,
> const void *frame, ssize_t len)
> {
> const u32 *data = frame;
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> i915_reg_t ctl_reg = HSW_TVIDEO_DIP_CTL(cpu_transcoder);
> int data_size;
> @@ -520,13 +520,13 @@ static void hsw_write_infoframe(struct intel_encoder *encoder,
> I915_WRITE(ctl_reg, val);
>
> for (i = 0; i < len; i += 4) {
> - I915_WRITE(hsw_dip_data_reg(dev_priv, cpu_transcoder,
> + I915_WRITE(hsw_dip_data_reg(i915, cpu_transcoder,
> type, i >> 2), *data);
> data++;
> }
> /* Write every possible data byte to force correct ECC calculation. */
> for (; i < data_size; i += 4)
> - I915_WRITE(hsw_dip_data_reg(dev_priv, cpu_transcoder,
> + I915_WRITE(hsw_dip_data_reg(i915, cpu_transcoder,
> type, i >> 2), 0);
>
> val |= hsw_infoframe_enable(type);
> @@ -539,7 +539,7 @@ static void hsw_read_infoframe(struct intel_encoder *encoder,
> unsigned int type,
> void *frame, ssize_t len)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> u32 val, *data = frame;
> int i;
> @@ -547,14 +547,14 @@ static void hsw_read_infoframe(struct intel_encoder *encoder,
> val = I915_READ(HSW_TVIDEO_DIP_CTL(cpu_transcoder));
>
> for (i = 0; i < len; i += 4)
> - *data++ = I915_READ(hsw_dip_data_reg(dev_priv, cpu_transcoder,
> + *data++ = I915_READ(hsw_dip_data_reg(i915, cpu_transcoder,
> type, i >> 2));
> }
>
> static u32 hsw_infoframes_enabled(struct intel_encoder *encoder,
> const struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> u32 val = I915_READ(HSW_TVIDEO_DIP_CTL(pipe_config->cpu_transcoder));
> u32 mask;
>
> @@ -562,7 +562,7 @@ static u32 hsw_infoframes_enabled(struct intel_encoder *encoder,
> VIDEO_DIP_ENABLE_GCP_HSW | VIDEO_DIP_ENABLE_VS_HSW |
> VIDEO_DIP_ENABLE_GMP_HSW | VIDEO_DIP_ENABLE_SPD_HSW);
>
> - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> mask |= VIDEO_DIP_ENABLE_DRM_GLK;
>
> return val & mask;
> @@ -593,7 +593,7 @@ u32 intel_hdmi_infoframe_enable(unsigned int type)
> u32 intel_hdmi_infoframes_enabled(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base);
> u32 val, ret = 0;
> int i;
> @@ -604,7 +604,7 @@ u32 intel_hdmi_infoframes_enabled(struct intel_encoder *encoder,
> for (i = 0; i < ARRAY_SIZE(infoframe_type_to_idx); i++) {
> unsigned int type = infoframe_type_to_idx[i];
>
> - if (HAS_DDI(dev_priv)) {
> + if (HAS_DDI(i915)) {
> if (val & hsw_infoframe_enable(type))
> ret |= BIT(i);
> } else {
> @@ -804,10 +804,10 @@ intel_hdmi_compute_drm_infoframe(struct intel_encoder *encoder,
> struct drm_connector_state *conn_state)
> {
> struct hdmi_drm_infoframe *frame = &crtc_state->infoframes.drm.drm;
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> int ret;
>
> - if (!(INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)))
> + if (!(INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)))
> return true;
>
> if (!crtc_state->has_infoframe)
> @@ -837,7 +837,7 @@ static void g4x_set_infoframes(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state,
> const struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_digital_port *intel_dig_port = enc_to_dig_port(&encoder->base);
> struct intel_hdmi *intel_hdmi = &intel_dig_port->hdmi;
> i915_reg_t reg = VIDEO_DIP_CTL;
> @@ -947,7 +947,7 @@ static bool intel_hdmi_set_gcp_infoframe(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state,
> const struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> i915_reg_t reg;
>
> @@ -955,11 +955,11 @@ static bool intel_hdmi_set_gcp_infoframe(struct intel_encoder *encoder,
> intel_hdmi_infoframe_enable(HDMI_PACKET_TYPE_GENERAL_CONTROL)) == 0)
> return false;
>
> - if (HAS_DDI(dev_priv))
> + if (HAS_DDI(i915))
> reg = HSW_TVIDEO_DIP_GCP(crtc_state->cpu_transcoder);
> - else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> + else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> reg = VLV_TVIDEO_DIP_GCP(crtc->pipe);
> - else if (HAS_PCH_SPLIT(dev_priv))
> + else if (HAS_PCH_SPLIT(i915))
> reg = TVIDEO_DIP_GCP(crtc->pipe);
> else
> return false;
> @@ -972,7 +972,7 @@ static bool intel_hdmi_set_gcp_infoframe(struct intel_encoder *encoder,
> void intel_hdmi_read_gcp_infoframe(struct intel_encoder *encoder,
> struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> i915_reg_t reg;
>
> @@ -980,11 +980,11 @@ void intel_hdmi_read_gcp_infoframe(struct intel_encoder *encoder,
> intel_hdmi_infoframe_enable(HDMI_PACKET_TYPE_GENERAL_CONTROL)) == 0)
> return;
>
> - if (HAS_DDI(dev_priv))
> + if (HAS_DDI(i915))
> reg = HSW_TVIDEO_DIP_GCP(crtc_state->cpu_transcoder);
> - else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> + else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> reg = VLV_TVIDEO_DIP_GCP(crtc->pipe);
> - else if (HAS_PCH_SPLIT(dev_priv))
> + else if (HAS_PCH_SPLIT(i915))
> reg = TVIDEO_DIP_GCP(crtc->pipe);
> else
> return;
> @@ -996,9 +996,9 @@ static void intel_hdmi_compute_gcp_infoframe(struct intel_encoder *encoder,
> struct intel_crtc_state *crtc_state,
> struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
>
> - if (IS_G4X(dev_priv) || !crtc_state->has_infoframe)
> + if (IS_G4X(i915) || !crtc_state->has_infoframe)
> return;
>
> crtc_state->infoframes.enable |=
> @@ -1019,7 +1019,7 @@ static void ibx_set_infoframes(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state,
> const struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc);
> struct intel_digital_port *intel_dig_port = enc_to_dig_port(&encoder->base);
> struct intel_hdmi *intel_hdmi = &intel_dig_port->hdmi;
> @@ -1078,7 +1078,7 @@ static void cpt_set_infoframes(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state,
> const struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc);
> struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base);
> i915_reg_t reg = TVIDEO_DIP_CTL(intel_crtc->pipe);
> @@ -1127,7 +1127,7 @@ static void vlv_set_infoframes(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state,
> const struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc);
> struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base);
> i915_reg_t reg = VLV_TVIDEO_DIP_CTL(intel_crtc->pipe);
> @@ -1185,11 +1185,11 @@ static void hsw_set_infoframes(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state,
> const struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> i915_reg_t reg = HSW_TVIDEO_DIP_CTL(crtc_state->cpu_transcoder);
> u32 val = I915_READ(reg);
>
> - assert_hdmi_transcoder_func_disabled(dev_priv,
> + assert_hdmi_transcoder_func_disabled(i915,
> crtc_state->cpu_transcoder);
>
> val &= ~(VIDEO_DIP_ENABLE_VSC_HSW | VIDEO_DIP_ENABLE_AVI_HSW |
> @@ -1225,9 +1225,9 @@ static void hsw_set_infoframes(struct intel_encoder *encoder,
>
> void intel_dp_dual_mode_set_tmds_output(struct intel_hdmi *hdmi, bool enable)
> {
> - struct drm_i915_private *dev_priv = to_i915(intel_hdmi_to_dev(hdmi));
> + struct drm_i915_private *i915 = to_i915(intel_hdmi_to_dev(hdmi));
> struct i2c_adapter *adapter =
> - intel_gmbus_get_adapter(dev_priv, hdmi->ddc_bus);
> + intel_gmbus_get_adapter(i915, hdmi->ddc_bus);
>
> if (hdmi->dp_dual_mode.type < DRM_DP_DUAL_MODE_TYPE2_DVI)
> return;
> @@ -1243,9 +1243,9 @@ static int intel_hdmi_hdcp_read(struct intel_digital_port *intel_dig_port,
> unsigned int offset, void *buffer, size_t size)
> {
> struct intel_hdmi *hdmi = &intel_dig_port->hdmi;
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> intel_dig_port->base.base.dev->dev_private;
> - struct i2c_adapter *adapter = intel_gmbus_get_adapter(dev_priv,
> + struct i2c_adapter *adapter = intel_gmbus_get_adapter(i915,
> hdmi->ddc_bus);
> int ret;
> u8 start = offset & 0xff;
> @@ -1273,9 +1273,9 @@ static int intel_hdmi_hdcp_write(struct intel_digital_port *intel_dig_port,
> unsigned int offset, void *buffer, size_t size)
> {
> struct intel_hdmi *hdmi = &intel_dig_port->hdmi;
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> intel_dig_port->base.base.dev->dev_private;
> - struct i2c_adapter *adapter = intel_gmbus_get_adapter(dev_priv,
> + struct i2c_adapter *adapter = intel_gmbus_get_adapter(i915,
> hdmi->ddc_bus);
> int ret;
> u8 *write_buf;
> @@ -1308,9 +1308,9 @@ int intel_hdmi_hdcp_write_an_aksv(struct intel_digital_port *intel_dig_port,
> u8 *an)
> {
> struct intel_hdmi *hdmi = &intel_dig_port->hdmi;
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> intel_dig_port->base.base.dev->dev_private;
> - struct i2c_adapter *adapter = intel_gmbus_get_adapter(dev_priv,
> + struct i2c_adapter *adapter = intel_gmbus_get_adapter(i915,
> hdmi->ddc_bus);
> int ret;
>
> @@ -1428,7 +1428,7 @@ int intel_hdmi_hdcp_read_v_prime_part(struct intel_digital_port *intel_dig_port,
>
> static int kbl_repositioning_enc_en_signal(struct intel_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_digital_port *intel_dig_port = conn_to_dig_port(connector);
> struct drm_crtc *crtc = connector->base.state->crtc;
> struct intel_crtc *intel_crtc = container_of(crtc,
> @@ -1463,7 +1463,7 @@ int intel_hdmi_hdcp_toggle_signalling(struct intel_digital_port *intel_dig_port,
> {
> struct intel_hdmi *hdmi = &intel_dig_port->hdmi;
> struct intel_connector *connector = hdmi->attached_connector;
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> int ret;
>
> if (!enable)
> @@ -1480,7 +1480,7 @@ int intel_hdmi_hdcp_toggle_signalling(struct intel_digital_port *intel_dig_port,
> * WA: To fix incorrect positioning of the window of
> * opportunity and enc_en signalling in KABYLAKE.
> */
> - if (IS_KABYLAKE(dev_priv) && enable)
> + if (IS_KABYLAKE(i915) && enable)
> return kbl_repositioning_enc_en_signal(connector);
>
> return 0;
> @@ -1489,7 +1489,7 @@ int intel_hdmi_hdcp_toggle_signalling(struct intel_digital_port *intel_dig_port,
> static
> bool intel_hdmi_hdcp_check_link(struct intel_digital_port *intel_dig_port)
> {
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> intel_dig_port->base.base.dev->dev_private;
> enum port port = intel_dig_port->base.port;
> int ret;
> @@ -1720,7 +1720,7 @@ static void intel_hdmi_prepare(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state)
> {
> struct drm_device *dev = encoder->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base);
> const struct drm_display_mode *adjusted_mode = &crtc_state->base.adjusted_mode;
> @@ -1729,7 +1729,7 @@ static void intel_hdmi_prepare(struct intel_encoder *encoder,
> intel_dp_dual_mode_set_tmds_output(intel_hdmi, true);
>
> hdmi_val = SDVO_ENCODING_HDMI;
> - if (!HAS_PCH_SPLIT(dev_priv) && crtc_state->limited_color_range)
> + if (!HAS_PCH_SPLIT(i915) && crtc_state->limited_color_range)
> hdmi_val |= HDMI_COLOR_RANGE_16_235;
> if (adjusted_mode->flags & DRM_MODE_FLAG_PVSYNC)
> hdmi_val |= SDVO_VSYNC_ACTIVE_HIGH;
> @@ -1744,9 +1744,9 @@ static void intel_hdmi_prepare(struct intel_encoder *encoder,
> if (crtc_state->has_hdmi_sink)
> hdmi_val |= HDMI_MODE_SELECT_HDMI;
>
> - if (HAS_PCH_CPT(dev_priv))
> + if (HAS_PCH_CPT(i915))
> hdmi_val |= SDVO_PIPE_SEL_CPT(crtc->pipe);
> - else if (IS_CHERRYVIEW(dev_priv))
> + else if (IS_CHERRYVIEW(i915))
> hdmi_val |= SDVO_PIPE_SEL_CHV(crtc->pipe);
> else
> hdmi_val |= SDVO_PIPE_SEL(crtc->pipe);
> @@ -1758,19 +1758,19 @@ static void intel_hdmi_prepare(struct intel_encoder *encoder,
> static bool intel_hdmi_get_hw_state(struct intel_encoder *encoder,
> enum pipe *pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base);
> intel_wakeref_t wakeref;
> bool ret;
>
> - wakeref = intel_display_power_get_if_enabled(dev_priv,
> + wakeref = intel_display_power_get_if_enabled(i915,
> encoder->power_domain);
> if (!wakeref)
> return false;
>
> - ret = intel_sdvo_port_enabled(dev_priv, intel_hdmi->hdmi_reg, pipe);
> + ret = intel_sdvo_port_enabled(i915, intel_hdmi->hdmi_reg, pipe);
>
> - intel_display_power_put(dev_priv, encoder->power_domain, wakeref);
> + intel_display_power_put(i915, encoder->power_domain, wakeref);
>
> return ret;
> }
> @@ -1780,7 +1780,7 @@ static void intel_hdmi_get_config(struct intel_encoder *encoder,
> {
> struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base);
> struct drm_device *dev = encoder->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> u32 tmp, flags = 0;
> int dotclock;
>
> @@ -1810,7 +1810,7 @@ static void intel_hdmi_get_config(struct intel_encoder *encoder,
> if (tmp & HDMI_AUDIO_ENABLE)
> pipe_config->has_audio = true;
>
> - if (!HAS_PCH_SPLIT(dev_priv) &&
> + if (!HAS_PCH_SPLIT(i915) &&
> tmp & HDMI_COLOR_RANGE_16_235)
> pipe_config->limited_color_range = true;
>
> @@ -1858,7 +1858,7 @@ static void g4x_enable_hdmi(struct intel_encoder *encoder,
> const struct drm_connector_state *conn_state)
> {
> struct drm_device *dev = encoder->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base);
> u32 temp;
>
> @@ -1880,7 +1880,7 @@ static void ibx_enable_hdmi(struct intel_encoder *encoder,
> const struct drm_connector_state *conn_state)
> {
> struct drm_device *dev = encoder->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base);
> u32 temp;
>
> @@ -1930,7 +1930,7 @@ static void cpt_enable_hdmi(struct intel_encoder *encoder,
> const struct drm_connector_state *conn_state)
> {
> struct drm_device *dev = encoder->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_crtc *crtc = to_intel_crtc(pipe_config->base.crtc);
> struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base);
> enum pipe pipe = crtc->pipe;
> @@ -1991,7 +1991,7 @@ static void intel_disable_hdmi(struct intel_encoder *encoder,
> const struct drm_connector_state *old_conn_state)
> {
> struct drm_device *dev = encoder->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base);
> struct intel_digital_port *intel_dig_port =
> hdmi_to_dig_port(intel_hdmi);
> @@ -2009,13 +2009,13 @@ static void intel_disable_hdmi(struct intel_encoder *encoder,
> * to transcoder A after disabling it to allow the
> * matching DP port to be enabled on transcoder A.
> */
> - if (HAS_PCH_IBX(dev_priv) && crtc->pipe == PIPE_B) {
> + if (HAS_PCH_IBX(i915) && crtc->pipe == PIPE_B) {
> /*
> * We get CPU/PCH FIFO underruns on the other pipe when
> * doing the workaround. Sweep them under the rug.
> */
> - intel_set_cpu_fifo_underrun_reporting(dev_priv, PIPE_A, false);
> - intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, false);
> + intel_set_cpu_fifo_underrun_reporting(i915, PIPE_A, false);
> + intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, false);
>
> temp &= ~SDVO_PIPE_SEL_MASK;
> temp |= SDVO_ENABLE | SDVO_PIPE_SEL(PIPE_A);
> @@ -2032,9 +2032,9 @@ static void intel_disable_hdmi(struct intel_encoder *encoder,
> I915_WRITE(intel_hdmi->hdmi_reg, temp);
> POSTING_READ(intel_hdmi->hdmi_reg);
>
> - intel_wait_for_vblank_if_active(dev_priv, PIPE_A);
> - intel_set_cpu_fifo_underrun_reporting(dev_priv, PIPE_A, true);
> - intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, true);
> + intel_wait_for_vblank_if_active(i915, PIPE_A);
> + intel_set_cpu_fifo_underrun_reporting(i915, PIPE_A, true);
> + intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, true);
> }
>
> intel_dig_port->set_infoframes(encoder,
> @@ -2073,16 +2073,16 @@ static void pch_post_disable_hdmi(struct intel_encoder *encoder,
>
> static int intel_hdmi_source_max_tmds_clock(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> const struct ddi_vbt_port_info *info =
> - &dev_priv->vbt.ddi_port_info[encoder->port];
> + &i915->vbt.ddi_port_info[encoder->port];
> int max_tmds_clock;
>
> - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> max_tmds_clock = 594000;
> - else if (INTEL_GEN(dev_priv) >= 8 || IS_HASWELL(dev_priv))
> + else if (INTEL_GEN(i915) >= 8 || IS_HASWELL(i915))
> max_tmds_clock = 300000;
> - else if (INTEL_GEN(dev_priv) >= 5)
> + else if (INTEL_GEN(i915) >= 5)
> max_tmds_clock = 225000;
> else
> max_tmds_clock = 165000;
> @@ -2123,7 +2123,7 @@ hdmi_port_clock_valid(struct intel_hdmi *hdmi,
> int clock, bool respect_downstream_limits,
> bool force_dvi)
> {
> - struct drm_i915_private *dev_priv = to_i915(intel_hdmi_to_dev(hdmi));
> + struct drm_i915_private *i915 = to_i915(intel_hdmi_to_dev(hdmi));
>
> if (clock < 25000)
> return MODE_CLOCK_LOW;
> @@ -2131,11 +2131,11 @@ hdmi_port_clock_valid(struct intel_hdmi *hdmi,
> return MODE_CLOCK_HIGH;
>
> /* BXT DPLL can't generate 223-240 MHz */
> - if (IS_GEN9_LP(dev_priv) && clock > 223333 && clock < 240000)
> + if (IS_GEN9_LP(i915) && clock > 223333 && clock < 240000)
> return MODE_CLOCK_RANGE;
>
> /* CHV DPLL can't generate 216-240 MHz */
> - if (IS_CHERRYVIEW(dev_priv) && clock > 216000 && clock < 240000)
> + if (IS_CHERRYVIEW(i915) && clock > 216000 && clock < 240000)
> return MODE_CLOCK_RANGE;
>
> return MODE_OK;
> @@ -2147,7 +2147,7 @@ intel_hdmi_mode_valid(struct drm_connector *connector,
> {
> struct intel_hdmi *hdmi = intel_attached_hdmi(connector);
> struct drm_device *dev = intel_hdmi_to_dev(hdmi);
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> enum drm_mode_status status;
> int clock;
> int max_dotclk = to_i915(connector->dev)->max_dotclk_freq;
> @@ -2176,12 +2176,12 @@ intel_hdmi_mode_valid(struct drm_connector *connector,
>
> if (hdmi->has_hdmi_sink && !force_dvi) {
> /* if we can't do 8bpc we may still be able to do 12bpc */
> - if (status != MODE_OK && !HAS_GMCH(dev_priv))
> + if (status != MODE_OK && !HAS_GMCH(i915))
> status = hdmi_port_clock_valid(hdmi, clock * 3 / 2,
> true, force_dvi);
>
> /* if we can't do 8,12bpc we may still be able to do 10bpc */
> - if (status != MODE_OK && INTEL_GEN(dev_priv) >= 11)
> + if (status != MODE_OK && INTEL_GEN(i915) >= 11)
> status = hdmi_port_clock_valid(hdmi, clock * 5 / 4,
> true, force_dvi);
> }
> @@ -2192,7 +2192,7 @@ intel_hdmi_mode_valid(struct drm_connector *connector,
> static bool hdmi_deep_color_possible(const struct intel_crtc_state *crtc_state,
> int bpc)
> {
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> to_i915(crtc_state->base.crtc->dev);
> struct drm_atomic_state *state = crtc_state->base.state;
> struct drm_connector_state *connector_state;
> @@ -2201,10 +2201,10 @@ static bool hdmi_deep_color_possible(const struct intel_crtc_state *crtc_state,
> &crtc_state->base.adjusted_mode;
> int i;
>
> - if (HAS_GMCH(dev_priv))
> + if (HAS_GMCH(i915))
> return false;
>
> - if (bpc == 10 && INTEL_GEN(dev_priv) < 11)
> + if (bpc == 10 && INTEL_GEN(i915) < 11)
> return false;
>
> if (crtc_state->pipe_bpp < bpc * 3)
> @@ -2246,13 +2246,13 @@ static bool hdmi_deep_color_possible(const struct intel_crtc_state *crtc_state,
> }
>
> /* Display WA #1139: glk */
> - if (bpc == 12 && IS_GLK_REVID(dev_priv, 0, GLK_REVID_A1) &&
> + if (bpc == 12 && IS_GLK_REVID(i915, 0, GLK_REVID_A1) &&
> adjusted_mode->htotal > 5460)
> return false;
>
> /* Display Wa_1405510057:icl */
> if (crtc_state->output_format == INTEL_OUTPUT_FORMAT_YCBCR420 &&
> - bpc == 10 && INTEL_GEN(dev_priv) >= 11 &&
> + bpc == 10 && INTEL_GEN(i915) >= 11 &&
> (adjusted_mode->crtc_hblank_end -
> adjusted_mode->crtc_hblank_start) % 8 == 2)
> return false;
> @@ -2297,7 +2297,7 @@ int intel_hdmi_compute_config(struct intel_encoder *encoder,
> struct drm_connector_state *conn_state)
> {
> struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base);
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct drm_display_mode *adjusted_mode = &pipe_config->base.adjusted_mode;
> struct drm_connector *connector = conn_state->connector;
> struct drm_scdc *scdc = &connector->display_info.hdmi.scdc;
> @@ -2345,7 +2345,7 @@ int intel_hdmi_compute_config(struct intel_encoder *encoder,
> }
> }
>
> - if (HAS_PCH_SPLIT(dev_priv) && !HAS_DDI(dev_priv))
> + if (HAS_PCH_SPLIT(i915) && !HAS_DDI(i915))
> pipe_config->has_pch_encoder = true;
>
> if (pipe_config->has_hdmi_sink) {
> @@ -2399,8 +2399,8 @@ int intel_hdmi_compute_config(struct intel_encoder *encoder,
>
> pipe_config->lane_count = 4;
>
> - if (scdc->scrambling.supported && (INTEL_GEN(dev_priv) >= 10 ||
> - IS_GEMINILAKE(dev_priv))) {
> + if (scdc->scrambling.supported && (INTEL_GEN(i915) >= 10 ||
> + IS_GEMINILAKE(i915))) {
> if (scdc->scrambling.low_rates)
> pipe_config->hdmi_scrambling = true;
>
> @@ -2453,11 +2453,11 @@ intel_hdmi_unset_edid(struct drm_connector *connector)
> static void
> intel_hdmi_dp_dual_mode_detect(struct drm_connector *connector, bool has_edid)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->dev);
> + struct drm_i915_private *i915 = to_i915(connector->dev);
> struct intel_hdmi *hdmi = intel_attached_hdmi(connector);
> enum port port = hdmi_to_dig_port(hdmi)->base.port;
> struct i2c_adapter *adapter =
> - intel_gmbus_get_adapter(dev_priv, hdmi->ddc_bus);
> + intel_gmbus_get_adapter(i915, hdmi->ddc_bus);
> enum drm_dp_dual_mode_type type = drm_dp_dual_mode_detect(adapter);
>
> /*
> @@ -2477,7 +2477,7 @@ intel_hdmi_dp_dual_mode_detect(struct drm_connector *connector, bool has_edid)
> * Make sure not to set limits for that port.
> */
> if (has_edid && !connector->override_edid &&
> - intel_bios_is_port_dp_dual_mode(dev_priv, port)) {
> + intel_bios_is_port_dp_dual_mode(i915, port)) {
> DRM_DEBUG_KMS("Assuming DP dual mode adaptor presence based on VBT\n");
> type = DRM_DP_DUAL_MODE_TYPE1_DVI;
> } else {
> @@ -2500,16 +2500,16 @@ intel_hdmi_dp_dual_mode_detect(struct drm_connector *connector, bool has_edid)
> static bool
> intel_hdmi_set_edid(struct drm_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->dev);
> + struct drm_i915_private *i915 = to_i915(connector->dev);
> struct intel_hdmi *intel_hdmi = intel_attached_hdmi(connector);
> intel_wakeref_t wakeref;
> struct edid *edid;
> bool connected = false;
> struct i2c_adapter *i2c;
>
> - wakeref = intel_display_power_get(dev_priv, POWER_DOMAIN_GMBUS);
> + wakeref = intel_display_power_get(i915, POWER_DOMAIN_GMBUS);
>
> - i2c = intel_gmbus_get_adapter(dev_priv, intel_hdmi->ddc_bus);
> + i2c = intel_gmbus_get_adapter(i915, intel_hdmi->ddc_bus);
>
> edid = drm_get_edid(connector, i2c);
>
> @@ -2522,7 +2522,7 @@ intel_hdmi_set_edid(struct drm_connector *connector)
>
> intel_hdmi_dp_dual_mode_detect(connector, edid != NULL);
>
> - intel_display_power_put(dev_priv, POWER_DOMAIN_GMBUS, wakeref);
> + intel_display_power_put(i915, POWER_DOMAIN_GMBUS, wakeref);
>
> to_intel_connector(connector)->detect_edid = edid;
> if (edid && edid->input & DRM_EDID_INPUT_DIGITAL) {
> @@ -2541,7 +2541,7 @@ static enum drm_connector_status
> intel_hdmi_detect(struct drm_connector *connector, bool force)
> {
> enum drm_connector_status status = connector_status_disconnected;
> - struct drm_i915_private *dev_priv = to_i915(connector->dev);
> + struct drm_i915_private *i915 = to_i915(connector->dev);
> struct intel_hdmi *intel_hdmi = intel_attached_hdmi(connector);
> struct intel_encoder *encoder = &hdmi_to_dig_port(intel_hdmi)->base;
> intel_wakeref_t wakeref;
> @@ -2549,9 +2549,9 @@ intel_hdmi_detect(struct drm_connector *connector, bool force)
> DRM_DEBUG_KMS("[CONNECTOR:%d:%s]\n",
> connector->base.id, connector->name);
>
> - wakeref = intel_display_power_get(dev_priv, POWER_DOMAIN_GMBUS);
> + wakeref = intel_display_power_get(i915, POWER_DOMAIN_GMBUS);
>
> - if (INTEL_GEN(dev_priv) >= 11 &&
> + if (INTEL_GEN(i915) >= 11 &&
> !intel_digital_port_connected(encoder))
> goto out;
>
> @@ -2561,7 +2561,7 @@ intel_hdmi_detect(struct drm_connector *connector, bool force)
> status = connector_status_connected;
>
> out:
> - intel_display_power_put(dev_priv, POWER_DOMAIN_GMBUS, wakeref);
> + intel_display_power_put(i915, POWER_DOMAIN_GMBUS, wakeref);
>
> if (status != connector_status_connected)
> cec_notifier_phys_addr_invalidate(intel_hdmi->cec_notifier);
> @@ -2613,7 +2613,7 @@ static void vlv_hdmi_pre_enable(struct intel_encoder *encoder,
> const struct drm_connector_state *conn_state)
> {
> struct intel_digital_port *dport = enc_to_dig_port(&encoder->base);
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
>
> vlv_phy_pre_encoder_enable(encoder, pipe_config);
>
> @@ -2627,7 +2627,7 @@ static void vlv_hdmi_pre_enable(struct intel_encoder *encoder,
>
> g4x_enable_hdmi(encoder, pipe_config, conn_state);
>
> - vlv_wait_port_ready(dev_priv, dport, 0x0);
> + vlv_wait_port_ready(i915, dport, 0x0);
> }
>
> static void vlv_hdmi_pre_pll_enable(struct intel_encoder *encoder,
> @@ -2668,14 +2668,14 @@ static void chv_hdmi_post_disable(struct intel_encoder *encoder,
> const struct drm_connector_state *old_conn_state)
> {
> struct drm_device *dev = encoder->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
>
> - vlv_dpio_get(dev_priv);
> + vlv_dpio_get(i915);
>
> /* Assert data lane reset */
> chv_data_lane_soft_reset(encoder, old_crtc_state, true);
>
> - vlv_dpio_put(dev_priv);
> + vlv_dpio_put(i915);
> }
>
> static void chv_hdmi_pre_enable(struct intel_encoder *encoder,
> @@ -2684,7 +2684,7 @@ static void chv_hdmi_pre_enable(struct intel_encoder *encoder,
> {
> struct intel_digital_port *dport = enc_to_dig_port(&encoder->base);
> struct drm_device *dev = encoder->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
>
> chv_phy_pre_encoder_enable(encoder, pipe_config);
>
> @@ -2698,7 +2698,7 @@ static void chv_hdmi_pre_enable(struct intel_encoder *encoder,
>
> g4x_enable_hdmi(encoder, pipe_config, conn_state);
>
> - vlv_wait_port_ready(dev_priv, dport, 0x0);
> + vlv_wait_port_ready(i915, dport, 0x0);
>
> /* Second common lane will stay alive on its own now */
> chv_phy_release_cl2_override(encoder);
> @@ -2707,10 +2707,10 @@ static void chv_hdmi_pre_enable(struct intel_encoder *encoder,
> static struct i2c_adapter *
> intel_hdmi_get_i2c_adapter(struct drm_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->dev);
> + struct drm_i915_private *i915 = to_i915(connector->dev);
> struct intel_hdmi *intel_hdmi = intel_attached_hdmi(connector);
>
> - return intel_gmbus_get_adapter(dev_priv, intel_hdmi->ddc_bus);
> + return intel_gmbus_get_adapter(i915, intel_hdmi->ddc_bus);
> }
>
> static void intel_hdmi_create_i2c_symlink(struct drm_connector *connector)
> @@ -2791,7 +2791,7 @@ static const struct drm_encoder_funcs intel_hdmi_enc_funcs = {
> static void
> intel_hdmi_add_properties(struct intel_hdmi *intel_hdmi, struct drm_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->dev);
> + struct drm_i915_private *i915 = to_i915(connector->dev);
> struct intel_digital_port *intel_dig_port =
> hdmi_to_dig_port(intel_hdmi);
>
> @@ -2810,11 +2810,11 @@ intel_hdmi_add_properties(struct intel_hdmi *intel_hdmi, struct drm_connector *c
> drm_connector_attach_content_type_property(connector);
> connector->state->picture_aspect_ratio = HDMI_PICTURE_ASPECT_NONE;
>
> - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> drm_object_attach_property(&connector->base,
> connector->dev->mode_config.hdr_output_metadata_property, 0);
>
> - if (!HAS_GMCH(dev_priv))
> + if (!HAS_GMCH(i915))
> drm_connector_attach_max_bpc_property(connector, 8, 12);
> }
>
> @@ -2841,12 +2841,12 @@ bool intel_hdmi_handle_sink_scrambling(struct intel_encoder *encoder,
> bool high_tmds_clock_ratio,
> bool scrambling)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base);
> struct drm_scrambling *sink_scrambling =
> &connector->display_info.hdmi.scdc.scrambling;
> struct i2c_adapter *adapter =
> - intel_gmbus_get_adapter(dev_priv, intel_hdmi->ddc_bus);
> + intel_gmbus_get_adapter(i915, intel_hdmi->ddc_bus);
>
> if (!sink_scrambling->supported)
> return true;
> @@ -2861,7 +2861,7 @@ bool intel_hdmi_handle_sink_scrambling(struct intel_encoder *encoder,
> drm_scdc_set_scrambling(adapter, scrambling);
> }
>
> -static u8 chv_port_to_ddc_pin(struct drm_i915_private *dev_priv, enum port port)
> +static u8 chv_port_to_ddc_pin(struct drm_i915_private *i915, enum port port)
> {
> u8 ddc_pin;
>
> @@ -2883,7 +2883,7 @@ static u8 chv_port_to_ddc_pin(struct drm_i915_private *dev_priv, enum port port)
> return ddc_pin;
> }
>
> -static u8 bxt_port_to_ddc_pin(struct drm_i915_private *dev_priv, enum port port)
> +static u8 bxt_port_to_ddc_pin(struct drm_i915_private *i915, enum port port)
> {
> u8 ddc_pin;
>
> @@ -2902,7 +2902,7 @@ static u8 bxt_port_to_ddc_pin(struct drm_i915_private *dev_priv, enum port port)
> return ddc_pin;
> }
>
> -static u8 cnp_port_to_ddc_pin(struct drm_i915_private *dev_priv,
> +static u8 cnp_port_to_ddc_pin(struct drm_i915_private *i915,
> enum port port)
> {
> u8 ddc_pin;
> @@ -2928,7 +2928,7 @@ static u8 cnp_port_to_ddc_pin(struct drm_i915_private *dev_priv,
> return ddc_pin;
> }
>
> -static u8 icl_port_to_ddc_pin(struct drm_i915_private *dev_priv, enum port port)
> +static u8 icl_port_to_ddc_pin(struct drm_i915_private *i915, enum port port)
> {
> u8 ddc_pin;
>
> @@ -2959,7 +2959,7 @@ static u8 icl_port_to_ddc_pin(struct drm_i915_private *dev_priv, enum port port)
> return ddc_pin;
> }
>
> -static u8 g4x_port_to_ddc_pin(struct drm_i915_private *dev_priv,
> +static u8 g4x_port_to_ddc_pin(struct drm_i915_private *i915,
> enum port port)
> {
> u8 ddc_pin;
> @@ -2982,11 +2982,11 @@ static u8 g4x_port_to_ddc_pin(struct drm_i915_private *dev_priv,
> return ddc_pin;
> }
>
> -static u8 intel_hdmi_ddc_pin(struct drm_i915_private *dev_priv,
> +static u8 intel_hdmi_ddc_pin(struct drm_i915_private *i915,
> enum port port)
> {
> const struct ddi_vbt_port_info *info =
> - &dev_priv->vbt.ddi_port_info[port];
> + &i915->vbt.ddi_port_info[port];
> u8 ddc_pin;
>
> if (info->alternate_ddc_pin) {
> @@ -2995,16 +2995,16 @@ static u8 intel_hdmi_ddc_pin(struct drm_i915_private *dev_priv,
> return info->alternate_ddc_pin;
> }
>
> - if (HAS_PCH_ICP(dev_priv))
> - ddc_pin = icl_port_to_ddc_pin(dev_priv, port);
> - else if (HAS_PCH_CNP(dev_priv))
> - ddc_pin = cnp_port_to_ddc_pin(dev_priv, port);
> - else if (IS_GEN9_LP(dev_priv))
> - ddc_pin = bxt_port_to_ddc_pin(dev_priv, port);
> - else if (IS_CHERRYVIEW(dev_priv))
> - ddc_pin = chv_port_to_ddc_pin(dev_priv, port);
> + if (HAS_PCH_ICP(i915))
> + ddc_pin = icl_port_to_ddc_pin(i915, port);
> + else if (HAS_PCH_CNP(i915))
> + ddc_pin = cnp_port_to_ddc_pin(i915, port);
> + else if (IS_GEN9_LP(i915))
> + ddc_pin = bxt_port_to_ddc_pin(i915, port);
> + else if (IS_CHERRYVIEW(i915))
> + ddc_pin = chv_port_to_ddc_pin(i915, port);
> else
> - ddc_pin = g4x_port_to_ddc_pin(dev_priv, port);
> + ddc_pin = g4x_port_to_ddc_pin(i915, port);
>
> DRM_DEBUG_KMS("Using DDC pin 0x%x for port %c (platform default)\n",
> ddc_pin, port_name(port));
> @@ -3014,20 +3014,20 @@ static u8 intel_hdmi_ddc_pin(struct drm_i915_private *dev_priv,
>
> void intel_infoframe_init(struct intel_digital_port *intel_dig_port)
> {
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> to_i915(intel_dig_port->base.base.dev);
>
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> intel_dig_port->write_infoframe = vlv_write_infoframe;
> intel_dig_port->read_infoframe = vlv_read_infoframe;
> intel_dig_port->set_infoframes = vlv_set_infoframes;
> intel_dig_port->infoframes_enabled = vlv_infoframes_enabled;
> - } else if (IS_G4X(dev_priv)) {
> + } else if (IS_G4X(i915)) {
> intel_dig_port->write_infoframe = g4x_write_infoframe;
> intel_dig_port->read_infoframe = g4x_read_infoframe;
> intel_dig_port->set_infoframes = g4x_set_infoframes;
> intel_dig_port->infoframes_enabled = g4x_infoframes_enabled;
> - } else if (HAS_DDI(dev_priv)) {
> + } else if (HAS_DDI(i915)) {
> if (intel_dig_port->lspcon.active) {
> intel_dig_port->write_infoframe = lspcon_write_infoframe;
> intel_dig_port->read_infoframe = lspcon_read_infoframe;
> @@ -3039,7 +3039,7 @@ void intel_infoframe_init(struct intel_digital_port *intel_dig_port)
> intel_dig_port->set_infoframes = hsw_set_infoframes;
> intel_dig_port->infoframes_enabled = hsw_infoframes_enabled;
> }
> - } else if (HAS_PCH_IBX(dev_priv)) {
> + } else if (HAS_PCH_IBX(i915)) {
> intel_dig_port->write_infoframe = ibx_write_infoframe;
> intel_dig_port->read_infoframe = ibx_read_infoframe;
> intel_dig_port->set_infoframes = ibx_set_infoframes;
> @@ -3059,7 +3059,7 @@ void intel_hdmi_init_connector(struct intel_digital_port *intel_dig_port,
> struct intel_hdmi *intel_hdmi = &intel_dig_port->hdmi;
> struct intel_encoder *intel_encoder = &intel_dig_port->base;
> struct drm_device *dev = intel_encoder->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> enum port port = intel_encoder->port;
>
> DRM_DEBUG_KMS("Adding HDMI connector on port %c\n",
> @@ -3078,16 +3078,16 @@ void intel_hdmi_init_connector(struct intel_digital_port *intel_dig_port,
> connector->doublescan_allowed = 0;
> connector->stereo_allowed = 1;
>
> - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> connector->ycbcr_420_allowed = true;
>
> - intel_hdmi->ddc_bus = intel_hdmi_ddc_pin(dev_priv, port);
> + intel_hdmi->ddc_bus = intel_hdmi_ddc_pin(i915, port);
>
> if (WARN_ON(port == PORT_A))
> return;
> - intel_encoder->hpd_pin = intel_hpd_pin_default(dev_priv, port);
> + intel_encoder->hpd_pin = intel_hpd_pin_default(i915, port);
>
> - if (HAS_DDI(dev_priv))
> + if (HAS_DDI(i915))
> intel_connector->get_hw_state = intel_ddi_connector_get_hw_state;
> else
> intel_connector->get_hw_state = intel_connector_get_hw_state;
> @@ -3097,7 +3097,7 @@ void intel_hdmi_init_connector(struct intel_digital_port *intel_dig_port,
> intel_connector_attach_encoder(intel_connector, intel_encoder);
> intel_hdmi->attached_connector = intel_connector;
>
> - if (is_hdcp_supported(dev_priv, port)) {
> + if (is_hdcp_supported(i915, port)) {
> int ret = intel_hdcp_init(intel_connector,
> &intel_hdmi_hdcp_shim);
> if (ret)
> @@ -3108,7 +3108,7 @@ void intel_hdmi_init_connector(struct intel_digital_port *intel_dig_port,
> * 0xd. Failure to do so will result in spurious interrupts being
> * generated on the port when a cable is not attached.
> */
> - if (IS_G45(dev_priv)) {
> + if (IS_G45(i915)) {
> u32 temp = I915_READ(PEG_BAND_GAP_DATA);
> I915_WRITE(PEG_BAND_GAP_DATA, (temp & ~0xf) | 0xd);
> }
> @@ -3119,7 +3119,7 @@ void intel_hdmi_init_connector(struct intel_digital_port *intel_dig_port,
> DRM_DEBUG_KMS("CEC notifier get failed\n");
> }
>
> -void intel_hdmi_init(struct drm_i915_private *dev_priv,
> +void intel_hdmi_init(struct drm_i915_private *i915,
> i915_reg_t hdmi_reg, enum port port)
> {
> struct intel_digital_port *intel_dig_port;
> @@ -3138,13 +3138,13 @@ void intel_hdmi_init(struct drm_i915_private *dev_priv,
>
> intel_encoder = &intel_dig_port->base;
>
> - drm_encoder_init(&dev_priv->drm, &intel_encoder->base,
> + drm_encoder_init(&i915->drm, &intel_encoder->base,
> &intel_hdmi_enc_funcs, DRM_MODE_ENCODER_TMDS,
> "HDMI %c", port_name(port));
>
> intel_encoder->hotplug = intel_encoder_hotplug;
> intel_encoder->compute_config = intel_hdmi_compute_config;
> - if (HAS_PCH_SPLIT(dev_priv)) {
> + if (HAS_PCH_SPLIT(i915)) {
> intel_encoder->disable = pch_disable_hdmi;
> intel_encoder->post_disable = pch_post_disable_hdmi;
> } else {
> @@ -3152,22 +3152,22 @@ void intel_hdmi_init(struct drm_i915_private *dev_priv,
> }
> intel_encoder->get_hw_state = intel_hdmi_get_hw_state;
> intel_encoder->get_config = intel_hdmi_get_config;
> - if (IS_CHERRYVIEW(dev_priv)) {
> + if (IS_CHERRYVIEW(i915)) {
> intel_encoder->pre_pll_enable = chv_hdmi_pre_pll_enable;
> intel_encoder->pre_enable = chv_hdmi_pre_enable;
> intel_encoder->enable = vlv_enable_hdmi;
> intel_encoder->post_disable = chv_hdmi_post_disable;
> intel_encoder->post_pll_disable = chv_hdmi_post_pll_disable;
> - } else if (IS_VALLEYVIEW(dev_priv)) {
> + } else if (IS_VALLEYVIEW(i915)) {
> intel_encoder->pre_pll_enable = vlv_hdmi_pre_pll_enable;
> intel_encoder->pre_enable = vlv_hdmi_pre_enable;
> intel_encoder->enable = vlv_enable_hdmi;
> intel_encoder->post_disable = vlv_hdmi_post_disable;
> } else {
> intel_encoder->pre_enable = intel_hdmi_pre_enable;
> - if (HAS_PCH_CPT(dev_priv))
> + if (HAS_PCH_CPT(i915))
> intel_encoder->enable = cpt_enable_hdmi;
> - else if (HAS_PCH_IBX(dev_priv))
> + else if (HAS_PCH_IBX(i915))
> intel_encoder->enable = ibx_enable_hdmi;
> else
> intel_encoder->enable = g4x_enable_hdmi;
> @@ -3176,7 +3176,7 @@ void intel_hdmi_init(struct drm_i915_private *dev_priv,
> intel_encoder->type = INTEL_OUTPUT_HDMI;
> intel_encoder->power_domain = intel_port_to_power_domain(port);
> intel_encoder->port = port;
> - if (IS_CHERRYVIEW(dev_priv)) {
> + if (IS_CHERRYVIEW(i915)) {
> if (port == PORT_D)
> intel_encoder->crtc_mask = 1 << 2;
> else
> @@ -3190,7 +3190,7 @@ void intel_hdmi_init(struct drm_i915_private *dev_priv,
> * to work on real hardware. And since g4x can send infoframes to
> * only one port anyway, nothing is lost by allowing it.
> */
> - if (IS_G4X(dev_priv))
> + if (IS_G4X(i915))
> intel_encoder->cloneable |= 1 << INTEL_OUTPUT_HDMI;
>
> intel_dig_port->hdmi.hdmi_reg = hdmi_reg;
> @@ -3199,6 +3199,6 @@ void intel_hdmi_init(struct drm_i915_private *dev_priv,
>
> intel_infoframe_init(intel_dig_port);
>
> - intel_dig_port->aux_ch = intel_bios_port_aux_ch(dev_priv, port);
> + intel_dig_port->aux_ch = intel_bios_port_aux_ch(i915, port);
> intel_hdmi_init_connector(intel_dig_port, intel_connector);
> }
> diff --git a/drivers/gpu/drm/i915/intel_hdmi.h b/drivers/gpu/drm/i915/intel_hdmi.h
> index 106c2e0bc3c9..46f03356b54b 100644
> --- a/drivers/gpu/drm/i915/intel_hdmi.h
> +++ b/drivers/gpu/drm/i915/intel_hdmi.h
> @@ -24,7 +24,7 @@ struct intel_hdmi;
> struct drm_connector_state;
> union hdmi_infoframe;
>
> -void intel_hdmi_init(struct drm_i915_private *dev_priv, i915_reg_t hdmi_reg,
> +void intel_hdmi_init(struct drm_i915_private *i915, i915_reg_t hdmi_reg,
> enum port port);
> void intel_hdmi_init_connector(struct intel_digital_port *intel_dig_port,
> struct intel_connector *intel_connector);
> diff --git a/drivers/gpu/drm/i915/intel_hotplug.c b/drivers/gpu/drm/i915/intel_hotplug.c
> index ff9eb3c855d3..a551b9ab2e17 100644
> --- a/drivers/gpu/drm/i915/intel_hotplug.c
> +++ b/drivers/gpu/drm/i915/intel_hotplug.c
> @@ -78,7 +78,7 @@
>
> /**
> * intel_hpd_pin_default - return default pin associated with certain port.
> - * @dev_priv: private driver data pointer
> + * @i915: private driver data pointer
> * @port: the hpd port to get associated pin
> *
> * It is only valid and used by digital port encoder.
> @@ -86,7 +86,7 @@
> * Return pin that is associatade with @port and HDP_NONE if no pin is
> * hard associated with that @port.
> */
> -enum hpd_pin intel_hpd_pin_default(struct drm_i915_private *dev_priv,
> +enum hpd_pin intel_hpd_pin_default(struct drm_i915_private *i915,
> enum port port)
> {
> switch (port) {
> @@ -101,7 +101,7 @@ enum hpd_pin intel_hpd_pin_default(struct drm_i915_private *dev_priv,
> case PORT_E:
> return HPD_PORT_E;
> case PORT_F:
> - if (IS_CNL_WITH_PORT_F(dev_priv))
> + if (IS_CNL_WITH_PORT_F(i915))
> return HPD_PORT_E;
> return HPD_PORT_F;
> default:
> @@ -115,7 +115,7 @@ enum hpd_pin intel_hpd_pin_default(struct drm_i915_private *dev_priv,
>
> /**
> * intel_hpd_irq_storm_detect - gather stats and detect HPD IRQ storm on a pin
> - * @dev_priv: private driver data pointer
> + * @i915: private driver data pointer
> * @pin: the pin to gather stats on
> * @long_hpd: whether the HPD IRQ was long or short
> *
> @@ -124,13 +124,13 @@ enum hpd_pin intel_hpd_pin_default(struct drm_i915_private *dev_priv,
> * responsible for further action.
> *
> * The number of IRQs that are allowed within @HPD_STORM_DETECT_PERIOD is
> - * stored in @dev_priv->hotplug.hpd_storm_threshold which defaults to
> + * stored in @i915->hotplug.hpd_storm_threshold which defaults to
> * @HPD_STORM_DEFAULT_THRESHOLD. Long IRQs count as +10 to this threshold, and
> * short IRQs count as +1. If this threshold is exceeded, it's considered an
> * IRQ storm and the IRQ state is set to @HPD_MARK_DISABLED.
> *
> * By default, most systems will only count long IRQs towards
> - * &dev_priv->hotplug.hpd_storm_threshold. However, some older systems also
> + * &i915->hotplug.hpd_storm_threshold. However, some older systems also
> * suffer from short IRQ storms and must also track these. Because short IRQ
> * storms are naturally caused by sideband interactions with DP MST devices,
> * short IRQ detection is only enabled for systems without DP MST support.
> @@ -142,10 +142,10 @@ enum hpd_pin intel_hpd_pin_default(struct drm_i915_private *dev_priv,
> *
> * Return true if an IRQ storm was detected on @pin.
> */
> -static bool intel_hpd_irq_storm_detect(struct drm_i915_private *dev_priv,
> +static bool intel_hpd_irq_storm_detect(struct drm_i915_private *i915,
> enum hpd_pin pin, bool long_hpd)
> {
> - struct i915_hotplug *hpd = &dev_priv->hotplug;
> + struct i915_hotplug *hpd = &i915->hotplug;
> unsigned long start = hpd->stats[pin].last_jiffies;
> unsigned long end = start + msecs_to_jiffies(HPD_STORM_DETECT_PERIOD);
> const int increment = long_hpd ? 10 : 1;
> @@ -153,7 +153,7 @@ static bool intel_hpd_irq_storm_detect(struct drm_i915_private *dev_priv,
> bool storm = false;
>
> if (!threshold ||
> - (!long_hpd && !dev_priv->hotplug.hpd_short_storm_enabled))
> + (!long_hpd && !i915->hotplug.hpd_short_storm_enabled))
> return false;
>
> if (!time_in_range(jiffies, start, end)) {
> @@ -175,9 +175,9 @@ static bool intel_hpd_irq_storm_detect(struct drm_i915_private *dev_priv,
> }
>
> static void
> -intel_hpd_irq_storm_switch_to_polling(struct drm_i915_private *dev_priv)
> +intel_hpd_irq_storm_switch_to_polling(struct drm_i915_private *i915)
> {
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_device *dev = &i915->drm;
> struct intel_connector *intel_connector;
> struct intel_encoder *intel_encoder;
> struct drm_connector *connector;
> @@ -185,7 +185,7 @@ intel_hpd_irq_storm_switch_to_polling(struct drm_i915_private *dev_priv)
> enum hpd_pin pin;
> bool hpd_disabled = false;
>
> - lockdep_assert_held(&dev_priv->irq_lock);
> + lockdep_assert_held(&i915->irq_lock);
>
> drm_connector_list_iter_begin(dev, &conn_iter);
> drm_for_each_connector_iter(connector, &conn_iter) {
> @@ -199,14 +199,14 @@ intel_hpd_irq_storm_switch_to_polling(struct drm_i915_private *dev_priv)
>
> pin = intel_encoder->hpd_pin;
> if (pin == HPD_NONE ||
> - dev_priv->hotplug.stats[pin].state != HPD_MARK_DISABLED)
> + i915->hotplug.stats[pin].state != HPD_MARK_DISABLED)
> continue;
>
> DRM_INFO("HPD interrupt storm detected on connector %s: "
> "switching from hotplug detection to polling\n",
> connector->name);
>
> - dev_priv->hotplug.stats[pin].state = HPD_DISABLED;
> + i915->hotplug.stats[pin].state = HPD_DISABLED;
> connector->polled = DRM_CONNECTOR_POLL_CONNECT
> | DRM_CONNECTOR_POLL_DISCONNECT;
> hpd_disabled = true;
> @@ -216,31 +216,31 @@ intel_hpd_irq_storm_switch_to_polling(struct drm_i915_private *dev_priv)
> /* Enable polling and queue hotplug re-enabling. */
> if (hpd_disabled) {
> drm_kms_helper_poll_enable(dev);
> - mod_delayed_work(system_wq, &dev_priv->hotplug.reenable_work,
> + mod_delayed_work(system_wq, &i915->hotplug.reenable_work,
> msecs_to_jiffies(HPD_STORM_REENABLE_DELAY));
> }
> }
>
> static void intel_hpd_irq_storm_reenable_work(struct work_struct *work)
> {
> - struct drm_i915_private *dev_priv =
> - container_of(work, typeof(*dev_priv),
> + struct drm_i915_private *i915 =
> + container_of(work, typeof(*i915),
> hotplug.reenable_work.work);
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_device *dev = &i915->drm;
> intel_wakeref_t wakeref;
> enum hpd_pin pin;
>
> - wakeref = intel_runtime_pm_get(dev_priv);
> + wakeref = intel_runtime_pm_get(i915);
>
> - spin_lock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
> for_each_hpd_pin(pin) {
> struct drm_connector *connector;
> struct drm_connector_list_iter conn_iter;
>
> - if (dev_priv->hotplug.stats[pin].state != HPD_DISABLED)
> + if (i915->hotplug.stats[pin].state != HPD_DISABLED)
> continue;
>
> - dev_priv->hotplug.stats[pin].state = HPD_ENABLED;
> + i915->hotplug.stats[pin].state = HPD_ENABLED;
>
> drm_connector_list_iter_begin(dev, &conn_iter);
> drm_for_each_connector_iter(connector, &conn_iter) {
> @@ -259,11 +259,11 @@ static void intel_hpd_irq_storm_reenable_work(struct work_struct *work)
> }
> drm_connector_list_iter_end(&conn_iter);
> }
> - if (dev_priv->display_irqs_enabled && dev_priv->display.hpd_irq_setup)
> - dev_priv->display.hpd_irq_setup(dev_priv);
> - spin_unlock_irq(&dev_priv->irq_lock);
> + if (i915->display_irqs_enabled && i915->display.hpd_irq_setup)
> + i915->display.hpd_irq_setup(i915);
> + spin_unlock_irq(&i915->irq_lock);
>
> - intel_runtime_pm_put(dev_priv, wakeref);
> + intel_runtime_pm_put(i915, wakeref);
> }
>
> bool intel_encoder_hotplug(struct intel_encoder *encoder,
> @@ -298,20 +298,20 @@ static bool intel_encoder_has_hpd_pulse(struct intel_encoder *encoder)
>
> static void i915_digport_work_func(struct work_struct *work)
> {
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> container_of(work, struct drm_i915_private, hotplug.dig_port_work);
> u32 long_port_mask, short_port_mask;
> struct intel_encoder *encoder;
> u32 old_bits = 0;
>
> - spin_lock_irq(&dev_priv->irq_lock);
> - long_port_mask = dev_priv->hotplug.long_port_mask;
> - dev_priv->hotplug.long_port_mask = 0;
> - short_port_mask = dev_priv->hotplug.short_port_mask;
> - dev_priv->hotplug.short_port_mask = 0;
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
> + long_port_mask = i915->hotplug.long_port_mask;
> + i915->hotplug.long_port_mask = 0;
> + short_port_mask = i915->hotplug.short_port_mask;
> + i915->hotplug.short_port_mask = 0;
> + spin_unlock_irq(&i915->irq_lock);
>
> - for_each_intel_encoder(&dev_priv->drm, encoder) {
> + for_each_intel_encoder(&i915->drm, encoder) {
> struct intel_digital_port *dig_port;
> enum port port = encoder->port;
> bool long_hpd, short_hpd;
> @@ -336,10 +336,10 @@ static void i915_digport_work_func(struct work_struct *work)
> }
>
> if (old_bits) {
> - spin_lock_irq(&dev_priv->irq_lock);
> - dev_priv->hotplug.event_bits |= old_bits;
> - spin_unlock_irq(&dev_priv->irq_lock);
> - schedule_work(&dev_priv->hotplug.hotplug_work);
> + spin_lock_irq(&i915->irq_lock);
> + i915->hotplug.event_bits |= old_bits;
> + spin_unlock_irq(&i915->irq_lock);
> + schedule_work(&i915->hotplug.hotplug_work);
> }
> }
>
> @@ -348,9 +348,9 @@ static void i915_digport_work_func(struct work_struct *work)
> */
> static void i915_hotplug_work_func(struct work_struct *work)
> {
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> container_of(work, struct drm_i915_private, hotplug.hotplug_work);
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_device *dev = &i915->drm;
> struct intel_connector *intel_connector;
> struct intel_encoder *intel_encoder;
> struct drm_connector *connector;
> @@ -361,15 +361,15 @@ static void i915_hotplug_work_func(struct work_struct *work)
> mutex_lock(&dev->mode_config.mutex);
> DRM_DEBUG_KMS("running encoder hotplug functions\n");
>
> - spin_lock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
>
> - hpd_event_bits = dev_priv->hotplug.event_bits;
> - dev_priv->hotplug.event_bits = 0;
> + hpd_event_bits = i915->hotplug.event_bits;
> + i915->hotplug.event_bits = 0;
>
> /* Enable polling for connectors which had HPD IRQ storms */
> - intel_hpd_irq_storm_switch_to_polling(dev_priv);
> + intel_hpd_irq_storm_switch_to_polling(i915);
>
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_unlock_irq(&i915->irq_lock);
>
> drm_connector_list_iter_begin(dev, &conn_iter);
> drm_for_each_connector_iter(connector, &conn_iter) {
> @@ -395,7 +395,7 @@ static void i915_hotplug_work_func(struct work_struct *work)
>
> /**
> * intel_hpd_irq_handler - main hotplug irq handler
> - * @dev_priv: drm_i915_private
> + * @i915: drm_i915_private
> * @pin_mask: a mask of hpd pins that have triggered the irq
> * @long_mask: a mask of hpd pins that may be long hpd pulses
> *
> @@ -409,7 +409,7 @@ static void i915_hotplug_work_func(struct work_struct *work)
> * Here, we do hotplug irq storm detection and mitigation, and pass further
> * processing to appropriate bottom halves.
> */
> -void intel_hpd_irq_handler(struct drm_i915_private *dev_priv,
> +void intel_hpd_irq_handler(struct drm_i915_private *i915,
> u32 pin_mask, u32 long_mask)
> {
> struct intel_encoder *encoder;
> @@ -422,7 +422,7 @@ void intel_hpd_irq_handler(struct drm_i915_private *dev_priv,
> if (!pin_mask)
> return;
>
> - spin_lock(&dev_priv->irq_lock);
> + spin_lock(&i915->irq_lock);
>
> /*
> * Determine whether ->hpd_pulse() exists for each pin, and
> @@ -430,7 +430,7 @@ void intel_hpd_irq_handler(struct drm_i915_private *dev_priv,
> * as each pin may have up to two encoders (HDMI and DP) and
> * only the one of them (DP) will have ->hpd_pulse().
> */
> - for_each_intel_encoder(&dev_priv->drm, encoder) {
> + for_each_intel_encoder(&i915->drm, encoder) {
> bool has_hpd_pulse = intel_encoder_has_hpd_pulse(encoder);
> enum port port = encoder->port;
> bool long_hpd;
> @@ -450,10 +450,10 @@ void intel_hpd_irq_handler(struct drm_i915_private *dev_priv,
>
> if (long_hpd) {
> long_hpd_pulse_mask |= BIT(pin);
> - dev_priv->hotplug.long_port_mask |= BIT(port);
> + i915->hotplug.long_port_mask |= BIT(port);
> } else {
> short_hpd_pulse_mask |= BIT(pin);
> - dev_priv->hotplug.short_port_mask |= BIT(port);
> + i915->hotplug.short_port_mask |= BIT(port);
> }
> }
>
> @@ -464,19 +464,19 @@ void intel_hpd_irq_handler(struct drm_i915_private *dev_priv,
> if (!(BIT(pin) & pin_mask))
> continue;
>
> - if (dev_priv->hotplug.stats[pin].state == HPD_DISABLED) {
> + if (i915->hotplug.stats[pin].state == HPD_DISABLED) {
> /*
> * On GMCH platforms the interrupt mask bits only
> * prevent irq generation, not the setting of the
> * hotplug bits itself. So only WARN about unexpected
> * interrupts on saner platforms.
> */
> - WARN_ONCE(!HAS_GMCH(dev_priv),
> + WARN_ONCE(!HAS_GMCH(i915),
> "Received HPD interrupt on pin %d although disabled\n", pin);
> continue;
> }
>
> - if (dev_priv->hotplug.stats[pin].state != HPD_ENABLED)
> + if (i915->hotplug.stats[pin].state != HPD_ENABLED)
> continue;
>
> /*
> @@ -487,13 +487,13 @@ void intel_hpd_irq_handler(struct drm_i915_private *dev_priv,
> if (((short_hpd_pulse_mask | long_hpd_pulse_mask) & BIT(pin))) {
> long_hpd = long_hpd_pulse_mask & BIT(pin);
> } else {
> - dev_priv->hotplug.event_bits |= BIT(pin);
> + i915->hotplug.event_bits |= BIT(pin);
> long_hpd = true;
> queue_hp = true;
> }
>
> - if (intel_hpd_irq_storm_detect(dev_priv, pin, long_hpd)) {
> - dev_priv->hotplug.event_bits &= ~BIT(pin);
> + if (intel_hpd_irq_storm_detect(i915, pin, long_hpd)) {
> + i915->hotplug.event_bits &= ~BIT(pin);
> storm_detected = true;
> queue_hp = true;
> }
> @@ -503,9 +503,9 @@ void intel_hpd_irq_handler(struct drm_i915_private *dev_priv,
> * Disable any IRQs that storms were detected on. Polling enablement
> * happens later in our hotplug work.
> */
> - if (storm_detected && dev_priv->display_irqs_enabled)
> - dev_priv->display.hpd_irq_setup(dev_priv);
> - spin_unlock(&dev_priv->irq_lock);
> + if (storm_detected && i915->display_irqs_enabled)
> + i915->display.hpd_irq_setup(i915);
> + spin_unlock(&i915->irq_lock);
>
> /*
> * Our hotplug handler can grab modeset locks (by calling down into the
> @@ -514,14 +514,14 @@ void intel_hpd_irq_handler(struct drm_i915_private *dev_priv,
> * deadlock.
> */
> if (queue_dig)
> - queue_work(dev_priv->hotplug.dp_wq, &dev_priv->hotplug.dig_port_work);
> + queue_work(i915->hotplug.dp_wq, &i915->hotplug.dig_port_work);
> if (queue_hp)
> - schedule_work(&dev_priv->hotplug.hotplug_work);
> + schedule_work(&i915->hotplug.hotplug_work);
> }
>
> /**
> * intel_hpd_init - initializes and enables hpd support
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> *
> * This function enables the hotplug support. It requires that interrupts have
> * already been enabled with intel_irq_init_hw(). From this point on hotplug and
> @@ -533,43 +533,43 @@ void intel_hpd_irq_handler(struct drm_i915_private *dev_priv,
> *
> * Also see: intel_hpd_poll_init(), which enables connector polling
> */
> -void intel_hpd_init(struct drm_i915_private *dev_priv)
> +void intel_hpd_init(struct drm_i915_private *i915)
> {
> int i;
>
> for_each_hpd_pin(i) {
> - dev_priv->hotplug.stats[i].count = 0;
> - dev_priv->hotplug.stats[i].state = HPD_ENABLED;
> + i915->hotplug.stats[i].count = 0;
> + i915->hotplug.stats[i].state = HPD_ENABLED;
> }
>
> - WRITE_ONCE(dev_priv->hotplug.poll_enabled, false);
> - schedule_work(&dev_priv->hotplug.poll_init_work);
> + WRITE_ONCE(i915->hotplug.poll_enabled, false);
> + schedule_work(&i915->hotplug.poll_init_work);
>
> /*
> * Interrupt setup is already guaranteed to be single-threaded, this is
> * just to make the assert_spin_locked checks happy.
> */
> - if (dev_priv->display_irqs_enabled && dev_priv->display.hpd_irq_setup) {
> - spin_lock_irq(&dev_priv->irq_lock);
> - if (dev_priv->display_irqs_enabled)
> - dev_priv->display.hpd_irq_setup(dev_priv);
> - spin_unlock_irq(&dev_priv->irq_lock);
> + if (i915->display_irqs_enabled && i915->display.hpd_irq_setup) {
> + spin_lock_irq(&i915->irq_lock);
> + if (i915->display_irqs_enabled)
> + i915->display.hpd_irq_setup(i915);
> + spin_unlock_irq(&i915->irq_lock);
> }
> }
>
> static void i915_hpd_poll_init_work(struct work_struct *work)
> {
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> container_of(work, struct drm_i915_private,
> hotplug.poll_init_work);
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_device *dev = &i915->drm;
> struct drm_connector *connector;
> struct drm_connector_list_iter conn_iter;
> bool enabled;
>
> mutex_lock(&dev->mode_config.mutex);
>
> - enabled = READ_ONCE(dev_priv->hotplug.poll_enabled);
> + enabled = READ_ONCE(i915->hotplug.poll_enabled);
>
> drm_connector_list_iter_begin(dev, &conn_iter);
> drm_for_each_connector_iter(connector, &conn_iter) {
> @@ -582,7 +582,7 @@ static void i915_hpd_poll_init_work(struct work_struct *work)
> if (intel_connector->mst_port)
> continue;
>
> - if (!connector->polled && I915_HAS_HOTPLUG(dev_priv) &&
> + if (!connector->polled && I915_HAS_HOTPLUG(i915) &&
> intel_connector->encoder->hpd_pin > HPD_NONE) {
> connector->polled = enabled ?
> DRM_CONNECTOR_POLL_CONNECT |
> @@ -607,7 +607,7 @@ static void i915_hpd_poll_init_work(struct work_struct *work)
>
> /**
> * intel_hpd_poll_init - enables/disables polling for connectors with hpd
> - * @dev_priv: i915 device instance
> + * @i915: i915 device instance
> *
> * This function enables polling for all connectors, regardless of whether or
> * not they support hotplug detection. Under certain conditions HPD may not be
> @@ -621,9 +621,9 @@ static void i915_hpd_poll_init_work(struct work_struct *work)
> *
> * Also see: intel_hpd_init(), which restores hpd handling.
> */
> -void intel_hpd_poll_init(struct drm_i915_private *dev_priv)
> +void intel_hpd_poll_init(struct drm_i915_private *i915)
> {
> - WRITE_ONCE(dev_priv->hotplug.poll_enabled, true);
> + WRITE_ONCE(i915->hotplug.poll_enabled, true);
>
> /*
> * We might already be holding dev->mode_config.mutex, so do this in a
> @@ -631,57 +631,57 @@ void intel_hpd_poll_init(struct drm_i915_private *dev_priv)
> * As well, there's no issue if we race here since we always reschedule
> * this worker anyway
> */
> - schedule_work(&dev_priv->hotplug.poll_init_work);
> + schedule_work(&i915->hotplug.poll_init_work);
> }
>
> -void intel_hpd_init_work(struct drm_i915_private *dev_priv)
> +void intel_hpd_init_work(struct drm_i915_private *i915)
> {
> - INIT_WORK(&dev_priv->hotplug.hotplug_work, i915_hotplug_work_func);
> - INIT_WORK(&dev_priv->hotplug.dig_port_work, i915_digport_work_func);
> - INIT_WORK(&dev_priv->hotplug.poll_init_work, i915_hpd_poll_init_work);
> - INIT_DELAYED_WORK(&dev_priv->hotplug.reenable_work,
> + INIT_WORK(&i915->hotplug.hotplug_work, i915_hotplug_work_func);
> + INIT_WORK(&i915->hotplug.dig_port_work, i915_digport_work_func);
> + INIT_WORK(&i915->hotplug.poll_init_work, i915_hpd_poll_init_work);
> + INIT_DELAYED_WORK(&i915->hotplug.reenable_work,
> intel_hpd_irq_storm_reenable_work);
> }
>
> -void intel_hpd_cancel_work(struct drm_i915_private *dev_priv)
> +void intel_hpd_cancel_work(struct drm_i915_private *i915)
> {
> - spin_lock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
>
> - dev_priv->hotplug.long_port_mask = 0;
> - dev_priv->hotplug.short_port_mask = 0;
> - dev_priv->hotplug.event_bits = 0;
> + i915->hotplug.long_port_mask = 0;
> + i915->hotplug.short_port_mask = 0;
> + i915->hotplug.event_bits = 0;
>
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_unlock_irq(&i915->irq_lock);
>
> - cancel_work_sync(&dev_priv->hotplug.dig_port_work);
> - cancel_work_sync(&dev_priv->hotplug.hotplug_work);
> - cancel_work_sync(&dev_priv->hotplug.poll_init_work);
> - cancel_delayed_work_sync(&dev_priv->hotplug.reenable_work);
> + cancel_work_sync(&i915->hotplug.dig_port_work);
> + cancel_work_sync(&i915->hotplug.hotplug_work);
> + cancel_work_sync(&i915->hotplug.poll_init_work);
> + cancel_delayed_work_sync(&i915->hotplug.reenable_work);
> }
>
> -bool intel_hpd_disable(struct drm_i915_private *dev_priv, enum hpd_pin pin)
> +bool intel_hpd_disable(struct drm_i915_private *i915, enum hpd_pin pin)
> {
> bool ret = false;
>
> if (pin == HPD_NONE)
> return false;
>
> - spin_lock_irq(&dev_priv->irq_lock);
> - if (dev_priv->hotplug.stats[pin].state == HPD_ENABLED) {
> - dev_priv->hotplug.stats[pin].state = HPD_DISABLED;
> + spin_lock_irq(&i915->irq_lock);
> + if (i915->hotplug.stats[pin].state == HPD_ENABLED) {
> + i915->hotplug.stats[pin].state = HPD_DISABLED;
> ret = true;
> }
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_unlock_irq(&i915->irq_lock);
>
> return ret;
> }
>
> -void intel_hpd_enable(struct drm_i915_private *dev_priv, enum hpd_pin pin)
> +void intel_hpd_enable(struct drm_i915_private *i915, enum hpd_pin pin)
> {
> if (pin == HPD_NONE)
> return;
>
> - spin_lock_irq(&dev_priv->irq_lock);
> - dev_priv->hotplug.stats[pin].state = HPD_ENABLED;
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_lock_irq(&i915->irq_lock);
> + i915->hotplug.stats[pin].state = HPD_ENABLED;
> + spin_unlock_irq(&i915->irq_lock);
> }
> diff --git a/drivers/gpu/drm/i915/intel_hotplug.h b/drivers/gpu/drm/i915/intel_hotplug.h
> index 805f897dbb7a..2fcc3148b369 100644
> --- a/drivers/gpu/drm/i915/intel_hotplug.h
> +++ b/drivers/gpu/drm/i915/intel_hotplug.h
> @@ -14,17 +14,17 @@ struct drm_i915_private;
> struct intel_connector;
> struct intel_encoder;
>
> -void intel_hpd_poll_init(struct drm_i915_private *dev_priv);
> +void intel_hpd_poll_init(struct drm_i915_private *i915);
> bool intel_encoder_hotplug(struct intel_encoder *encoder,
> struct intel_connector *connector);
> -void intel_hpd_irq_handler(struct drm_i915_private *dev_priv,
> +void intel_hpd_irq_handler(struct drm_i915_private *i915,
> u32 pin_mask, u32 long_mask);
> -void intel_hpd_init(struct drm_i915_private *dev_priv);
> -void intel_hpd_init_work(struct drm_i915_private *dev_priv);
> -void intel_hpd_cancel_work(struct drm_i915_private *dev_priv);
> -enum hpd_pin intel_hpd_pin_default(struct drm_i915_private *dev_priv,
> +void intel_hpd_init(struct drm_i915_private *i915);
> +void intel_hpd_init_work(struct drm_i915_private *i915);
> +void intel_hpd_cancel_work(struct drm_i915_private *i915);
> +enum hpd_pin intel_hpd_pin_default(struct drm_i915_private *i915,
> enum port port);
> -bool intel_hpd_disable(struct drm_i915_private *dev_priv, enum hpd_pin pin);
> -void intel_hpd_enable(struct drm_i915_private *dev_priv, enum hpd_pin pin);
> +bool intel_hpd_disable(struct drm_i915_private *i915, enum hpd_pin pin);
> +void intel_hpd_enable(struct drm_i915_private *i915, enum hpd_pin pin);
>
> #endif /* __INTEL_HOTPLUG_H__ */
> diff --git a/drivers/gpu/drm/i915/intel_huc.c b/drivers/gpu/drm/i915/intel_huc.c
> index 8572a0588efc..2b23662698a9 100644
> --- a/drivers/gpu/drm/i915/intel_huc.c
> +++ b/drivers/gpu/drm/i915/intel_huc.c
> @@ -167,14 +167,14 @@ int intel_huc_auth(struct intel_huc *huc)
> */
> int intel_huc_check_status(struct intel_huc *huc)
> {
> - struct drm_i915_private *dev_priv = huc_to_i915(huc);
> + struct drm_i915_private *i915 = huc_to_i915(huc);
> intel_wakeref_t wakeref;
> bool status = false;
>
> - if (!HAS_HUC(dev_priv))
> + if (!HAS_HUC(i915))
> return -ENODEV;
>
> - with_intel_runtime_pm(dev_priv, wakeref)
> + with_intel_runtime_pm(i915, wakeref)
> status = (I915_READ(huc->status.reg) & huc->status.mask) ==
> huc->status.value;
>
> diff --git a/drivers/gpu/drm/i915/intel_huc_fw.c b/drivers/gpu/drm/i915/intel_huc_fw.c
> index 05cbf8338f53..03aed0664726 100644
> --- a/drivers/gpu/drm/i915/intel_huc_fw.c
> +++ b/drivers/gpu/drm/i915/intel_huc_fw.c
> @@ -69,34 +69,34 @@ MODULE_FIRMWARE(I915_ICL_HUC_UCODE);
> static void huc_fw_select(struct intel_uc_fw *huc_fw)
> {
> struct intel_huc *huc = container_of(huc_fw, struct intel_huc, fw);
> - struct drm_i915_private *dev_priv = huc_to_i915(huc);
> + struct drm_i915_private *i915 = huc_to_i915(huc);
>
> GEM_BUG_ON(huc_fw->type != INTEL_UC_FW_TYPE_HUC);
>
> - if (!HAS_HUC(dev_priv))
> + if (!HAS_HUC(i915))
> return;
>
> if (i915_modparams.huc_firmware_path) {
> huc_fw->path = i915_modparams.huc_firmware_path;
> huc_fw->major_ver_wanted = 0;
> huc_fw->minor_ver_wanted = 0;
> - } else if (IS_SKYLAKE(dev_priv)) {
> + } else if (IS_SKYLAKE(i915)) {
> huc_fw->path = I915_SKL_HUC_UCODE;
> huc_fw->major_ver_wanted = SKL_HUC_FW_MAJOR;
> huc_fw->minor_ver_wanted = SKL_HUC_FW_MINOR;
> - } else if (IS_BROXTON(dev_priv)) {
> + } else if (IS_BROXTON(i915)) {
> huc_fw->path = I915_BXT_HUC_UCODE;
> huc_fw->major_ver_wanted = BXT_HUC_FW_MAJOR;
> huc_fw->minor_ver_wanted = BXT_HUC_FW_MINOR;
> - } else if (IS_KABYLAKE(dev_priv) || IS_COFFEELAKE(dev_priv)) {
> + } else if (IS_KABYLAKE(i915) || IS_COFFEELAKE(i915)) {
> huc_fw->path = I915_KBL_HUC_UCODE;
> huc_fw->major_ver_wanted = KBL_HUC_FW_MAJOR;
> huc_fw->minor_ver_wanted = KBL_HUC_FW_MINOR;
> - } else if (IS_GEMINILAKE(dev_priv)) {
> + } else if (IS_GEMINILAKE(i915)) {
> huc_fw->path = I915_GLK_HUC_UCODE;
> huc_fw->major_ver_wanted = GLK_HUC_FW_MAJOR;
> huc_fw->minor_ver_wanted = GLK_HUC_FW_MINOR;
> - } else if (IS_ICELAKE(dev_priv)) {
> + } else if (IS_ICELAKE(i915)) {
> huc_fw->path = I915_ICL_HUC_UCODE;
> huc_fw->major_ver_wanted = ICL_HUC_FW_MAJOR;
> huc_fw->minor_ver_wanted = ICL_HUC_FW_MINOR;
> @@ -135,8 +135,8 @@ static void huc_xfer_rsa(struct intel_huc *huc)
> static int huc_xfer_ucode(struct intel_huc *huc)
> {
> struct intel_uc_fw *huc_fw = &huc->fw;
> - struct drm_i915_private *dev_priv = huc_to_i915(huc);
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct drm_i915_private *i915 = huc_to_i915(huc);
> + struct intel_uncore *uncore = &i915->uncore;
> unsigned long offset = 0;
> u32 size;
> int ret;
> diff --git a/drivers/gpu/drm/i915/intel_lpe_audio.c b/drivers/gpu/drm/i915/intel_lpe_audio.c
> index b19800b58442..d95ae05ccebc 100644
> --- a/drivers/gpu/drm/i915/intel_lpe_audio.c
> +++ b/drivers/gpu/drm/i915/intel_lpe_audio.c
> @@ -73,12 +73,12 @@
> #include "i915_drv.h"
> #include "intel_lpe_audio.h"
>
> -#define HAS_LPE_AUDIO(dev_priv) ((dev_priv)->lpe_audio.platdev != NULL)
> +#define HAS_LPE_AUDIO(i915) ((i915)->lpe_audio.platdev != NULL)
>
> static struct platform_device *
> -lpe_audio_platdev_create(struct drm_i915_private *dev_priv)
> +lpe_audio_platdev_create(struct drm_i915_private *i915)
> {
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_device *dev = &i915->drm;
> struct platform_device_info pinfo = {};
> struct resource *rsc;
> struct platform_device *platdev;
> @@ -94,7 +94,7 @@ lpe_audio_platdev_create(struct drm_i915_private *dev_priv)
> return ERR_PTR(-ENOMEM);
> }
>
> - rsc[0].start = rsc[0].end = dev_priv->lpe_audio.irq;
> + rsc[0].start = rsc[0].end = i915->lpe_audio.irq;
> rsc[0].flags = IORESOURCE_IRQ;
> rsc[0].name = "hdmi-lpe-audio-irq";
>
> @@ -114,8 +114,8 @@ lpe_audio_platdev_create(struct drm_i915_private *dev_priv)
> pinfo.size_data = sizeof(*pdata);
> pinfo.dma_mask = DMA_BIT_MASK(32);
>
> - pdata->num_pipes = INTEL_INFO(dev_priv)->num_pipes;
> - pdata->num_ports = IS_CHERRYVIEW(dev_priv) ? 3 : 2; /* B,C,D or B,C */
> + pdata->num_pipes = INTEL_INFO(i915)->num_pipes;
> + pdata->num_ports = IS_CHERRYVIEW(i915) ? 3 : 2; /* B,C,D or B,C */
> pdata->port[0].pipe = -1;
> pdata->port[1].pipe = -1;
> pdata->port[2].pipe = -1;
> @@ -135,7 +135,7 @@ lpe_audio_platdev_create(struct drm_i915_private *dev_priv)
> return platdev;
> }
>
> -static void lpe_audio_platdev_destroy(struct drm_i915_private *dev_priv)
> +static void lpe_audio_platdev_destroy(struct drm_i915_private *i915)
> {
> /* XXX Note that platform_device_register_full() allocates a dma_mask
> * and never frees it. We can't free it here as we cannot guarantee
> @@ -145,7 +145,7 @@ static void lpe_audio_platdev_destroy(struct drm_i915_private *dev_priv)
> * than us fiddle with its internals.
> */
>
> - platform_device_unregister(dev_priv->lpe_audio.platdev);
> + platform_device_unregister(i915->lpe_audio.platdev);
> }
>
> static void lpe_audio_irq_unmask(struct irq_data *d)
> @@ -162,24 +162,24 @@ static struct irq_chip lpe_audio_irqchip = {
> .irq_unmask = lpe_audio_irq_unmask,
> };
>
> -static int lpe_audio_irq_init(struct drm_i915_private *dev_priv)
> +static int lpe_audio_irq_init(struct drm_i915_private *i915)
> {
> - int irq = dev_priv->lpe_audio.irq;
> + int irq = i915->lpe_audio.irq;
>
> - WARN_ON(!intel_irqs_enabled(dev_priv));
> + WARN_ON(!intel_irqs_enabled(i915));
> irq_set_chip_and_handler_name(irq,
> &lpe_audio_irqchip,
> handle_simple_irq,
> "hdmi_lpe_audio_irq_handler");
>
> - return irq_set_chip_data(irq, dev_priv);
> + return irq_set_chip_data(irq, i915);
> }
>
> -static bool lpe_audio_detect(struct drm_i915_private *dev_priv)
> +static bool lpe_audio_detect(struct drm_i915_private *i915)
> {
> int lpe_present = false;
>
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> static const struct pci_device_id atom_hdaudio_ids[] = {
> /* Baytrail */
> {PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x0f04)},
> @@ -196,21 +196,21 @@ static bool lpe_audio_detect(struct drm_i915_private *dev_priv)
> return lpe_present;
> }
>
> -static int lpe_audio_setup(struct drm_i915_private *dev_priv)
> +static int lpe_audio_setup(struct drm_i915_private *i915)
> {
> int ret;
>
> - dev_priv->lpe_audio.irq = irq_alloc_desc(0);
> - if (dev_priv->lpe_audio.irq < 0) {
> + i915->lpe_audio.irq = irq_alloc_desc(0);
> + if (i915->lpe_audio.irq < 0) {
> DRM_ERROR("Failed to allocate IRQ desc: %d\n",
> - dev_priv->lpe_audio.irq);
> - ret = dev_priv->lpe_audio.irq;
> + i915->lpe_audio.irq);
> + ret = i915->lpe_audio.irq;
> goto err;
> }
>
> - DRM_DEBUG("irq = %d\n", dev_priv->lpe_audio.irq);
> + DRM_DEBUG("irq = %d\n", i915->lpe_audio.irq);
>
> - ret = lpe_audio_irq_init(dev_priv);
> + ret = lpe_audio_irq_init(i915);
>
> if (ret) {
> DRM_ERROR("Failed to initialize irqchip for lpe audio: %d\n",
> @@ -218,10 +218,10 @@ static int lpe_audio_setup(struct drm_i915_private *dev_priv)
> goto err_free_irq;
> }
>
> - dev_priv->lpe_audio.platdev = lpe_audio_platdev_create(dev_priv);
> + i915->lpe_audio.platdev = lpe_audio_platdev_create(i915);
>
> - if (IS_ERR(dev_priv->lpe_audio.platdev)) {
> - ret = PTR_ERR(dev_priv->lpe_audio.platdev);
> + if (IS_ERR(i915->lpe_audio.platdev)) {
> + ret = PTR_ERR(i915->lpe_audio.platdev);
> DRM_ERROR("Failed to create lpe audio platform device: %d\n",
> ret);
> goto err_free_irq;
> @@ -234,28 +234,28 @@ static int lpe_audio_setup(struct drm_i915_private *dev_priv)
>
> return 0;
> err_free_irq:
> - irq_free_desc(dev_priv->lpe_audio.irq);
> + irq_free_desc(i915->lpe_audio.irq);
> err:
> - dev_priv->lpe_audio.irq = -1;
> - dev_priv->lpe_audio.platdev = NULL;
> + i915->lpe_audio.irq = -1;
> + i915->lpe_audio.platdev = NULL;
> return ret;
> }
>
> /**
> * intel_lpe_audio_irq_handler() - forwards the LPE audio irq
> - * @dev_priv: the i915 drm device private data
> + * @i915: the i915 drm device private data
> *
> * the LPE Audio irq is forwarded to the irq handler registered by LPE audio
> * driver.
> */
> -void intel_lpe_audio_irq_handler(struct drm_i915_private *dev_priv)
> +void intel_lpe_audio_irq_handler(struct drm_i915_private *i915)
> {
> int ret;
>
> - if (!HAS_LPE_AUDIO(dev_priv))
> + if (!HAS_LPE_AUDIO(i915))
> return;
>
> - ret = generic_handle_irq(dev_priv->lpe_audio.irq);
> + ret = generic_handle_irq(i915->lpe_audio.irq);
> if (ret)
> DRM_ERROR_RATELIMITED("error handling LPE audio irq: %d\n",
> ret);
> @@ -264,17 +264,17 @@ void intel_lpe_audio_irq_handler(struct drm_i915_private *dev_priv)
> /**
> * intel_lpe_audio_init() - detect and setup the bridge between HDMI LPE Audio
> * driver and i915
> - * @dev_priv: the i915 drm device private data
> + * @i915: the i915 drm device private data
> *
> * Return: 0 if successful. non-zero if detection or
> * llocation/initialization fails
> */
> -int intel_lpe_audio_init(struct drm_i915_private *dev_priv)
> +int intel_lpe_audio_init(struct drm_i915_private *i915)
> {
> int ret = -ENODEV;
>
> - if (lpe_audio_detect(dev_priv)) {
> - ret = lpe_audio_setup(dev_priv);
> + if (lpe_audio_detect(i915)) {
> + ret = lpe_audio_setup(i915);
> if (ret < 0)
> DRM_ERROR("failed to setup LPE Audio bridge\n");
> }
> @@ -284,31 +284,31 @@ int intel_lpe_audio_init(struct drm_i915_private *dev_priv)
> /**
> * intel_lpe_audio_teardown() - destroy the bridge between HDMI LPE
> * audio driver and i915
> - * @dev_priv: the i915 drm device private data
> + * @i915: the i915 drm device private data
> *
> * release all the resources for LPE audio <-> i915 bridge.
> */
> -void intel_lpe_audio_teardown(struct drm_i915_private *dev_priv)
> +void intel_lpe_audio_teardown(struct drm_i915_private *i915)
> {
> struct irq_desc *desc;
>
> - if (!HAS_LPE_AUDIO(dev_priv))
> + if (!HAS_LPE_AUDIO(i915))
> return;
>
> - desc = irq_to_desc(dev_priv->lpe_audio.irq);
> + desc = irq_to_desc(i915->lpe_audio.irq);
>
> - lpe_audio_platdev_destroy(dev_priv);
> + lpe_audio_platdev_destroy(i915);
>
> - irq_free_desc(dev_priv->lpe_audio.irq);
> + irq_free_desc(i915->lpe_audio.irq);
>
> - dev_priv->lpe_audio.irq = -1;
> - dev_priv->lpe_audio.platdev = NULL;
> + i915->lpe_audio.irq = -1;
> + i915->lpe_audio.platdev = NULL;
> }
>
> /**
> * intel_lpe_audio_notify() - notify lpe audio event
> * audio driver and i915
> - * @dev_priv: the i915 drm device private data
> + * @i915: the i915 drm device private data
> * @pipe: pipe
> * @port: port
> * @eld : ELD data
> @@ -317,7 +317,7 @@ void intel_lpe_audio_teardown(struct drm_i915_private *dev_priv)
> *
> * Notify lpe audio driver of eld change.
> */
> -void intel_lpe_audio_notify(struct drm_i915_private *dev_priv,
> +void intel_lpe_audio_notify(struct drm_i915_private *i915,
> enum pipe pipe, enum port port,
> const void *eld, int ls_clock, bool dp_output)
> {
> @@ -326,10 +326,10 @@ void intel_lpe_audio_notify(struct drm_i915_private *dev_priv,
> struct intel_hdmi_lpe_audio_port_pdata *ppdata;
> u32 audio_enable;
>
> - if (!HAS_LPE_AUDIO(dev_priv))
> + if (!HAS_LPE_AUDIO(i915))
> return;
>
> - pdata = dev_get_platdata(&dev_priv->lpe_audio.platdev->dev);
> + pdata = dev_get_platdata(&i915->lpe_audio.platdev->dev);
> ppdata = &pdata->port[port - PORT_B];
>
> spin_lock_irqsave(&pdata->lpe_audio_slock, irqflags);
> @@ -357,7 +357,7 @@ void intel_lpe_audio_notify(struct drm_i915_private *dev_priv,
> }
>
> if (pdata->notify_audio_lpe)
> - pdata->notify_audio_lpe(dev_priv->lpe_audio.platdev, port - PORT_B);
> + pdata->notify_audio_lpe(i915->lpe_audio.platdev, port - PORT_B);
>
> spin_unlock_irqrestore(&pdata->lpe_audio_slock, irqflags);
> }
> diff --git a/drivers/gpu/drm/i915/intel_lpe_audio.h b/drivers/gpu/drm/i915/intel_lpe_audio.h
> index f848c5038714..df43abce3c95 100644
> --- a/drivers/gpu/drm/i915/intel_lpe_audio.h
> +++ b/drivers/gpu/drm/i915/intel_lpe_audio.h
> @@ -12,10 +12,10 @@ enum pipe;
> enum port;
> struct drm_i915_private;
>
> -int intel_lpe_audio_init(struct drm_i915_private *dev_priv);
> -void intel_lpe_audio_teardown(struct drm_i915_private *dev_priv);
> -void intel_lpe_audio_irq_handler(struct drm_i915_private *dev_priv);
> -void intel_lpe_audio_notify(struct drm_i915_private *dev_priv,
> +int intel_lpe_audio_init(struct drm_i915_private *i915);
> +void intel_lpe_audio_teardown(struct drm_i915_private *i915);
> +void intel_lpe_audio_irq_handler(struct drm_i915_private *i915);
> +void intel_lpe_audio_notify(struct drm_i915_private *i915,
> enum pipe pipe, enum port port,
> const void *eld, int ls_clock, bool dp_output);
>
> diff --git a/drivers/gpu/drm/i915/intel_lspcon.c b/drivers/gpu/drm/i915/intel_lspcon.c
> index 7028d0cf3bb1..695465f97195 100644
> --- a/drivers/gpu/drm/i915/intel_lspcon.c
> +++ b/drivers/gpu/drm/i915/intel_lspcon.c
> @@ -555,10 +555,10 @@ bool lspcon_init(struct intel_digital_port *intel_dig_port)
> struct intel_dp *dp = &intel_dig_port->dp;
> struct intel_lspcon *lspcon = &intel_dig_port->lspcon;
> struct drm_device *dev = intel_dig_port->base.base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct drm_connector *connector = &dp->attached_connector->base;
>
> - if (!HAS_LSPCON(dev_priv)) {
> + if (!HAS_LSPCON(i915)) {
> DRM_ERROR("LSPCON is not supported on this platform\n");
> return false;
> }
> diff --git a/drivers/gpu/drm/i915/intel_lvds.c b/drivers/gpu/drm/i915/intel_lvds.c
> index efefed62a7f8..0405f9834827 100644
> --- a/drivers/gpu/drm/i915/intel_lvds.c
> +++ b/drivers/gpu/drm/i915/intel_lvds.c
> @@ -80,7 +80,7 @@ static struct intel_lvds_encoder *to_lvds_encoder(struct drm_encoder *encoder)
> return container_of(encoder, struct intel_lvds_encoder, base.base);
> }
>
> -bool intel_lvds_port_enabled(struct drm_i915_private *dev_priv,
> +bool intel_lvds_port_enabled(struct drm_i915_private *i915,
> i915_reg_t lvds_reg, enum pipe *pipe)
> {
> u32 val;
> @@ -88,7 +88,7 @@ bool intel_lvds_port_enabled(struct drm_i915_private *dev_priv,
> val = I915_READ(lvds_reg);
>
> /* asserts want to know the pipe even if the port is disabled */
> - if (HAS_PCH_CPT(dev_priv))
> + if (HAS_PCH_CPT(i915))
> *pipe = (val & LVDS_PIPE_SEL_MASK_CPT) >> LVDS_PIPE_SEL_SHIFT_CPT;
> else
> *pipe = (val & LVDS_PIPE_SEL_MASK) >> LVDS_PIPE_SEL_SHIFT;
> @@ -99,19 +99,19 @@ bool intel_lvds_port_enabled(struct drm_i915_private *dev_priv,
> static bool intel_lvds_get_hw_state(struct intel_encoder *encoder,
> enum pipe *pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_lvds_encoder *lvds_encoder = to_lvds_encoder(&encoder->base);
> intel_wakeref_t wakeref;
> bool ret;
>
> - wakeref = intel_display_power_get_if_enabled(dev_priv,
> + wakeref = intel_display_power_get_if_enabled(i915,
> encoder->power_domain);
> if (!wakeref)
> return false;
>
> - ret = intel_lvds_port_enabled(dev_priv, lvds_encoder->reg, pipe);
> + ret = intel_lvds_port_enabled(i915, lvds_encoder->reg, pipe);
>
> - intel_display_power_put(dev_priv, encoder->power_domain, wakeref);
> + intel_display_power_put(i915, encoder->power_domain, wakeref);
>
> return ret;
> }
> @@ -119,7 +119,7 @@ static bool intel_lvds_get_hw_state(struct intel_encoder *encoder,
> static void intel_lvds_get_config(struct intel_encoder *encoder,
> struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_lvds_encoder *lvds_encoder = to_lvds_encoder(&encoder->base);
> u32 tmp, flags = 0;
>
> @@ -137,12 +137,12 @@ static void intel_lvds_get_config(struct intel_encoder *encoder,
>
> pipe_config->base.adjusted_mode.flags |= flags;
>
> - if (INTEL_GEN(dev_priv) < 5)
> + if (INTEL_GEN(i915) < 5)
> pipe_config->gmch_pfit.lvds_border_bits =
> tmp & LVDS_BORDER_ENABLE;
>
> /* gen2/3 store dither state in pfit control, needs to match */
> - if (INTEL_GEN(dev_priv) < 4) {
> + if (INTEL_GEN(i915) < 4) {
> tmp = I915_READ(PFIT_CONTROL);
>
> pipe_config->gmch_pfit.control |= tmp & PANEL_8TO6_DITHER_ENABLE;
> @@ -151,7 +151,7 @@ static void intel_lvds_get_config(struct intel_encoder *encoder,
> pipe_config->base.adjusted_mode.crtc_clock = pipe_config->port_clock;
> }
>
> -static void intel_lvds_pps_get_hw_state(struct drm_i915_private *dev_priv,
> +static void intel_lvds_pps_get_hw_state(struct drm_i915_private *i915,
> struct intel_lvds_pps *pps)
> {
> u32 val;
> @@ -180,7 +180,7 @@ static void intel_lvds_pps_get_hw_state(struct drm_i915_private *dev_priv,
> /* Convert from 100ms to 100us units */
> pps->t4 = val * 1000;
>
> - if (INTEL_GEN(dev_priv) <= 4 &&
> + if (INTEL_GEN(i915) <= 4 &&
> pps->t1_t2 == 0 && pps->t5 == 0 && pps->t3 == 0 && pps->tx == 0) {
> DRM_DEBUG_KMS("Panel power timings uninitialized, "
> "setting defaults\n");
> @@ -198,7 +198,7 @@ static void intel_lvds_pps_get_hw_state(struct drm_i915_private *dev_priv,
> pps->divider, pps->port, pps->powerdown_on_reset);
> }
>
> -static void intel_lvds_pps_init_hw(struct drm_i915_private *dev_priv,
> +static void intel_lvds_pps_init_hw(struct drm_i915_private *i915,
> struct intel_lvds_pps *pps)
> {
> u32 val;
> @@ -229,26 +229,26 @@ static void intel_pre_enable_lvds(struct intel_encoder *encoder,
> const struct drm_connector_state *conn_state)
> {
> struct intel_lvds_encoder *lvds_encoder = to_lvds_encoder(&encoder->base);
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crtc *crtc = to_intel_crtc(pipe_config->base.crtc);
> const struct drm_display_mode *adjusted_mode = &pipe_config->base.adjusted_mode;
> int pipe = crtc->pipe;
> u32 temp;
>
> - if (HAS_PCH_SPLIT(dev_priv)) {
> - assert_fdi_rx_pll_disabled(dev_priv, pipe);
> - assert_shared_dpll_disabled(dev_priv,
> + if (HAS_PCH_SPLIT(i915)) {
> + assert_fdi_rx_pll_disabled(i915, pipe);
> + assert_shared_dpll_disabled(i915,
> pipe_config->shared_dpll);
> } else {
> - assert_pll_disabled(dev_priv, pipe);
> + assert_pll_disabled(i915, pipe);
> }
>
> - intel_lvds_pps_init_hw(dev_priv, &lvds_encoder->init_pps);
> + intel_lvds_pps_init_hw(i915, &lvds_encoder->init_pps);
>
> temp = lvds_encoder->init_lvds_val;
> temp |= LVDS_PORT_EN | LVDS_A0A2_CLKA_POWER_UP;
>
> - if (HAS_PCH_CPT(dev_priv)) {
> + if (HAS_PCH_CPT(i915)) {
> temp &= ~LVDS_PIPE_SEL_MASK_CPT;
> temp |= LVDS_PIPE_SEL_CPT(pipe);
> } else {
> @@ -283,7 +283,7 @@ static void intel_pre_enable_lvds(struct intel_encoder *encoder,
> * special lvds dither control bit on pch-split platforms, dithering is
> * only controlled through the PIPECONF reg.
> */
> - if (IS_GEN(dev_priv, 4)) {
> + if (IS_GEN(i915, 4)) {
> /*
> * Bspec wording suggests that LVDS port dithering only exists
> * for 18bpp panels.
> @@ -311,14 +311,14 @@ static void intel_enable_lvds(struct intel_encoder *encoder,
> {
> struct drm_device *dev = encoder->base.dev;
> struct intel_lvds_encoder *lvds_encoder = to_lvds_encoder(&encoder->base);
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
>
> I915_WRITE(lvds_encoder->reg, I915_READ(lvds_encoder->reg) | LVDS_PORT_EN);
>
> I915_WRITE(PP_CONTROL(0), I915_READ(PP_CONTROL(0)) | PANEL_POWER_ON);
> POSTING_READ(lvds_encoder->reg);
>
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> PP_STATUS(0), PP_ON, PP_ON, 5000))
> DRM_ERROR("timed out waiting for panel to power on\n");
>
> @@ -330,10 +330,10 @@ static void intel_disable_lvds(struct intel_encoder *encoder,
> const struct drm_connector_state *old_conn_state)
> {
> struct intel_lvds_encoder *lvds_encoder = to_lvds_encoder(&encoder->base);
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
>
> I915_WRITE(PP_CONTROL(0), I915_READ(PP_CONTROL(0)) & ~PANEL_POWER_ON);
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> PP_STATUS(0), PP_ON, 0, 1000))
> DRM_ERROR("timed out waiting for panel to power off\n");
>
> @@ -389,7 +389,7 @@ static int intel_lvds_compute_config(struct intel_encoder *intel_encoder,
> struct intel_crtc_state *pipe_config,
> struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(intel_encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(intel_encoder->base.dev);
> struct intel_lvds_encoder *lvds_encoder =
> to_lvds_encoder(&intel_encoder->base);
> struct intel_connector *intel_connector =
> @@ -399,7 +399,7 @@ static int intel_lvds_compute_config(struct intel_encoder *intel_encoder,
> unsigned int lvds_bpp;
>
> /* Should never happen!! */
> - if (INTEL_GEN(dev_priv) < 4 && intel_crtc->pipe == 0) {
> + if (INTEL_GEN(i915) < 4 && intel_crtc->pipe == 0) {
> DRM_ERROR("Can't support LVDS on pipe A\n");
> return -EINVAL;
> }
> @@ -429,7 +429,7 @@ static int intel_lvds_compute_config(struct intel_encoder *intel_encoder,
> if (adjusted_mode->flags & DRM_MODE_FLAG_DBLSCAN)
> return -EINVAL;
>
> - if (HAS_PCH_SPLIT(dev_priv)) {
> + if (HAS_PCH_SPLIT(i915)) {
> pipe_config->has_pch_encoder = true;
>
> intel_pch_panel_fitting(intel_crtc, pipe_config,
> @@ -751,11 +751,11 @@ static const struct dmi_system_id intel_dual_link_lvds[] = {
> { } /* terminating entry */
> };
>
> -struct intel_encoder *intel_get_lvds_encoder(struct drm_i915_private *dev_priv)
> +struct intel_encoder *intel_get_lvds_encoder(struct drm_i915_private *i915)
> {
> struct intel_encoder *encoder;
>
> - for_each_intel_encoder(&dev_priv->drm, encoder) {
> + for_each_intel_encoder(&i915->drm, encoder) {
> if (encoder->type == INTEL_OUTPUT_LVDS)
> return encoder;
> }
> @@ -763,9 +763,9 @@ struct intel_encoder *intel_get_lvds_encoder(struct drm_i915_private *dev_priv)
> return NULL;
> }
>
> -bool intel_is_dual_link_lvds(struct drm_i915_private *dev_priv)
> +bool intel_is_dual_link_lvds(struct drm_i915_private *i915)
> {
> - struct intel_encoder *encoder = intel_get_lvds_encoder(dev_priv);
> + struct intel_encoder *encoder = intel_get_lvds_encoder(i915);
>
> return encoder && to_lvds_encoder(&encoder->base)->is_dual_link;
> }
> @@ -774,7 +774,7 @@ static bool compute_is_dual_link_lvds(struct intel_lvds_encoder *lvds_encoder)
> {
> struct drm_device *dev = lvds_encoder->base.base.dev;
> unsigned int val;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
>
> /* use the module option value if specified */
> if (i915_modparams.lvds_channel_mode > 0)
> @@ -794,26 +794,26 @@ static bool compute_is_dual_link_lvds(struct intel_lvds_encoder *lvds_encoder)
> * register is uninitialized.
> */
> val = I915_READ(lvds_encoder->reg);
> - if (HAS_PCH_CPT(dev_priv))
> + if (HAS_PCH_CPT(i915))
> val &= ~(LVDS_DETECTED | LVDS_PIPE_SEL_MASK_CPT);
> else
> val &= ~(LVDS_DETECTED | LVDS_PIPE_SEL_MASK);
> if (val == 0)
> - val = dev_priv->vbt.bios_lvds_val;
> + val = i915->vbt.bios_lvds_val;
>
> return (val & LVDS_CLKB_POWER_MASK) == LVDS_CLKB_POWER_UP;
> }
>
> /**
> * intel_lvds_init - setup LVDS connectors on this device
> - * @dev_priv: i915 device
> + * @i915: i915 device
> *
> * Create the connector, register the LVDS DDC bus, and try to figure out what
> * modes we can display on the LVDS panel (if present).
> */
> -void intel_lvds_init(struct drm_i915_private *dev_priv)
> +void intel_lvds_init(struct drm_i915_private *i915)
> {
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_device *dev = &i915->drm;
> struct intel_lvds_encoder *lvds_encoder;
> struct intel_encoder *intel_encoder;
> struct intel_connector *intel_connector;
> @@ -829,30 +829,30 @@ void intel_lvds_init(struct drm_i915_private *dev_priv)
>
> /* Skip init on machines we know falsely report LVDS */
> if (dmi_check_system(intel_no_lvds)) {
> - WARN(!dev_priv->vbt.int_lvds_support,
> + WARN(!i915->vbt.int_lvds_support,
> "Useless DMI match. Internal LVDS support disabled by VBT\n");
> return;
> }
>
> - if (!dev_priv->vbt.int_lvds_support) {
> + if (!i915->vbt.int_lvds_support) {
> DRM_DEBUG_KMS("Internal LVDS support disabled by VBT\n");
> return;
> }
>
> - if (HAS_PCH_SPLIT(dev_priv))
> + if (HAS_PCH_SPLIT(i915))
> lvds_reg = PCH_LVDS;
> else
> lvds_reg = LVDS;
>
> lvds = I915_READ(lvds_reg);
>
> - if (HAS_PCH_SPLIT(dev_priv)) {
> + if (HAS_PCH_SPLIT(i915)) {
> if ((lvds & LVDS_DETECTED) == 0)
> return;
> }
>
> pin = GMBUS_PIN_PANEL;
> - if (!intel_bios_is_lvds_present(dev_priv, &pin)) {
> + if (!intel_bios_is_lvds_present(i915, &pin)) {
> if ((lvds & LVDS_PORT_EN) == 0) {
> DRM_DEBUG_KMS("LVDS is not present in VBT\n");
> return;
> @@ -884,7 +884,7 @@ void intel_lvds_init(struct drm_i915_private *dev_priv)
> intel_encoder->enable = intel_enable_lvds;
> intel_encoder->pre_enable = intel_pre_enable_lvds;
> intel_encoder->compute_config = intel_lvds_compute_config;
> - if (HAS_PCH_SPLIT(dev_priv)) {
> + if (HAS_PCH_SPLIT(i915)) {
> intel_encoder->disable = pch_disable_lvds;
> intel_encoder->post_disable = pch_post_disable_lvds;
> } else {
> @@ -901,9 +901,9 @@ void intel_lvds_init(struct drm_i915_private *dev_priv)
> intel_encoder->power_domain = POWER_DOMAIN_PORT_OTHER;
> intel_encoder->port = PORT_NONE;
> intel_encoder->cloneable = 0;
> - if (HAS_PCH_SPLIT(dev_priv))
> + if (HAS_PCH_SPLIT(i915))
> intel_encoder->crtc_mask = (1 << 0) | (1 << 1) | (1 << 2);
> - else if (IS_GEN(dev_priv, 4))
> + else if (IS_GEN(i915, 4))
> intel_encoder->crtc_mask = (1 << 0) | (1 << 1);
> else
> intel_encoder->crtc_mask = (1 << 1);
> @@ -922,7 +922,7 @@ void intel_lvds_init(struct drm_i915_private *dev_priv)
> drm_connector_attach_scaling_mode_property(connector, allowed_scalers);
> connector->state->scaling_mode = DRM_MODE_SCALE_ASPECT;
>
> - intel_lvds_pps_get_hw_state(dev_priv, &lvds_encoder->init_pps);
> + intel_lvds_pps_get_hw_state(i915, &lvds_encoder->init_pps);
> lvds_encoder->init_lvds_val = lvds;
>
> /*
> @@ -940,10 +940,10 @@ void intel_lvds_init(struct drm_i915_private *dev_priv)
> mutex_lock(&dev->mode_config.mutex);
> if (vga_switcheroo_handler_flags() & VGA_SWITCHEROO_CAN_SWITCH_DDC)
> edid = drm_get_edid_switcheroo(connector,
> - intel_gmbus_get_adapter(dev_priv, pin));
> + intel_gmbus_get_adapter(i915, pin));
> else
> edid = drm_get_edid(connector,
> - intel_gmbus_get_adapter(dev_priv, pin));
> + intel_gmbus_get_adapter(i915, pin));
> if (edid) {
> if (drm_add_edid_modes(connector, edid)) {
> drm_connector_update_edid_property(connector,
> diff --git a/drivers/gpu/drm/i915/intel_lvds.h b/drivers/gpu/drm/i915/intel_lvds.h
> index bc9c8b84ba2f..4afdcb4d6352 100644
> --- a/drivers/gpu/drm/i915/intel_lvds.h
> +++ b/drivers/gpu/drm/i915/intel_lvds.h
> @@ -13,10 +13,10 @@
> enum pipe;
> struct drm_i915_private;
>
> -bool intel_lvds_port_enabled(struct drm_i915_private *dev_priv,
> +bool intel_lvds_port_enabled(struct drm_i915_private *i915,
> i915_reg_t lvds_reg, enum pipe *pipe);
> -void intel_lvds_init(struct drm_i915_private *dev_priv);
> -struct intel_encoder *intel_get_lvds_encoder(struct drm_i915_private *dev_priv);
> -bool intel_is_dual_link_lvds(struct drm_i915_private *dev_priv);
> +void intel_lvds_init(struct drm_i915_private *i915);
> +struct intel_encoder *intel_get_lvds_encoder(struct drm_i915_private *i915);
> +bool intel_is_dual_link_lvds(struct drm_i915_private *i915);
>
> #endif /* __INTEL_LVDS_H__ */
> diff --git a/drivers/gpu/drm/i915/intel_opregion.c b/drivers/gpu/drm/i915/intel_opregion.c
> index 8fa1159d097f..6cbe775d7537 100644
> --- a/drivers/gpu/drm/i915/intel_opregion.c
> +++ b/drivers/gpu/drm/i915/intel_opregion.c
> @@ -266,11 +266,11 @@ struct opregion_asle_ext {
>
> #define MAX_DSLP 1500
>
> -static int swsci(struct drm_i915_private *dev_priv,
> +static int swsci(struct drm_i915_private *i915,
> u32 function, u32 parm, u32 *parm_out)
> {
> - struct opregion_swsci *swsci = dev_priv->opregion.swsci;
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct opregion_swsci *swsci = i915->opregion.swsci;
> + struct pci_dev *pdev = i915->drm.pdev;
> u32 main_function, sub_function, scic;
> u16 swsci_val;
> u32 dslp;
> @@ -285,11 +285,11 @@ static int swsci(struct drm_i915_private *dev_priv,
>
> /* Check if we can call the function. See swsci_setup for details. */
> if (main_function == SWSCI_SBCB) {
> - if ((dev_priv->opregion.swsci_sbcb_sub_functions &
> + if ((i915->opregion.swsci_sbcb_sub_functions &
> (1 << sub_function)) == 0)
> return -EINVAL;
> } else if (main_function == SWSCI_GBDA) {
> - if ((dev_priv->opregion.swsci_gbda_sub_functions &
> + if ((i915->opregion.swsci_gbda_sub_functions &
> (1 << sub_function)) == 0)
> return -EINVAL;
> }
> @@ -363,13 +363,13 @@ static int swsci(struct drm_i915_private *dev_priv,
> int intel_opregion_notify_encoder(struct intel_encoder *intel_encoder,
> bool enable)
> {
> - struct drm_i915_private *dev_priv = to_i915(intel_encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(intel_encoder->base.dev);
> u32 parm = 0;
> u32 type = 0;
> u32 port;
>
> /* don't care about old stuff for now */
> - if (!HAS_DDI(dev_priv))
> + if (!HAS_DDI(i915))
> return 0;
>
> if (intel_encoder->type == INTEL_OUTPUT_DSI)
> @@ -409,7 +409,7 @@ int intel_opregion_notify_encoder(struct intel_encoder *intel_encoder,
>
> parm |= type << (16 + port * 3);
>
> - return swsci(dev_priv, SWSCI_SBCB_DISPLAY_POWER_STATE, parm, NULL);
> + return swsci(i915, SWSCI_SBCB_DISPLAY_POWER_STATE, parm, NULL);
> }
>
> static const struct {
> @@ -423,29 +423,29 @@ static const struct {
> { PCI_D3cold, 0x04 },
> };
>
> -int intel_opregion_notify_adapter(struct drm_i915_private *dev_priv,
> +int intel_opregion_notify_adapter(struct drm_i915_private *i915,
> pci_power_t state)
> {
> int i;
>
> - if (!HAS_DDI(dev_priv))
> + if (!HAS_DDI(i915))
> return 0;
>
> for (i = 0; i < ARRAY_SIZE(power_state_map); i++) {
> if (state == power_state_map[i].pci_power_state)
> - return swsci(dev_priv, SWSCI_SBCB_ADAPTER_POWER_STATE,
> + return swsci(i915, SWSCI_SBCB_ADAPTER_POWER_STATE,
> power_state_map[i].parm, NULL);
> }
>
> return -EINVAL;
> }
>
> -static u32 asle_set_backlight(struct drm_i915_private *dev_priv, u32 bclp)
> +static u32 asle_set_backlight(struct drm_i915_private *i915, u32 bclp)
> {
> struct intel_connector *connector;
> struct drm_connector_list_iter conn_iter;
> - struct opregion_asle *asle = dev_priv->opregion.asle;
> - struct drm_device *dev = &dev_priv->drm;
> + struct opregion_asle *asle = i915->opregion.asle;
> + struct drm_device *dev = &i915->drm;
>
> DRM_DEBUG_DRIVER("bclp = 0x%08x\n", bclp);
>
> @@ -480,7 +480,7 @@ static u32 asle_set_backlight(struct drm_i915_private *dev_priv, u32 bclp)
> return 0;
> }
>
> -static u32 asle_set_als_illum(struct drm_i915_private *dev_priv, u32 alsi)
> +static u32 asle_set_als_illum(struct drm_i915_private *i915, u32 alsi)
> {
> /* alsi is the current ALS reading in lux. 0 indicates below sensor
> range, 0xffff indicates above sensor range. 1-0xfffe are valid */
> @@ -488,13 +488,13 @@ static u32 asle_set_als_illum(struct drm_i915_private *dev_priv, u32 alsi)
> return ASLC_ALS_ILLUM_FAILED;
> }
>
> -static u32 asle_set_pwm_freq(struct drm_i915_private *dev_priv, u32 pfmb)
> +static u32 asle_set_pwm_freq(struct drm_i915_private *i915, u32 pfmb)
> {
> DRM_DEBUG_DRIVER("PWM freq is not supported\n");
> return ASLC_PWM_FREQ_FAILED;
> }
>
> -static u32 asle_set_pfit(struct drm_i915_private *dev_priv, u32 pfit)
> +static u32 asle_set_pfit(struct drm_i915_private *i915, u32 pfit)
> {
> /* Panel fitting is currently controlled by the X code, so this is a
> noop until modesetting support works fully */
> @@ -502,13 +502,13 @@ static u32 asle_set_pfit(struct drm_i915_private *dev_priv, u32 pfit)
> return ASLC_PFIT_FAILED;
> }
>
> -static u32 asle_set_supported_rotation_angles(struct drm_i915_private *dev_priv, u32 srot)
> +static u32 asle_set_supported_rotation_angles(struct drm_i915_private *i915, u32 srot)
> {
> DRM_DEBUG_DRIVER("SROT is not supported\n");
> return ASLC_ROTATION_ANGLES_FAILED;
> }
>
> -static u32 asle_set_button_array(struct drm_i915_private *dev_priv, u32 iuer)
> +static u32 asle_set_button_array(struct drm_i915_private *i915, u32 iuer)
> {
> if (!iuer)
> DRM_DEBUG_DRIVER("Button array event is not supported (nothing)\n");
> @@ -526,7 +526,7 @@ static u32 asle_set_button_array(struct drm_i915_private *dev_priv, u32 iuer)
> return ASLC_BUTTON_ARRAY_FAILED;
> }
>
> -static u32 asle_set_convertible(struct drm_i915_private *dev_priv, u32 iuer)
> +static u32 asle_set_convertible(struct drm_i915_private *i915, u32 iuer)
> {
> if (iuer & ASLE_IUER_CONVERTIBLE)
> DRM_DEBUG_DRIVER("Convertible is not supported (clamshell)\n");
> @@ -536,7 +536,7 @@ static u32 asle_set_convertible(struct drm_i915_private *dev_priv, u32 iuer)
> return ASLC_CONVERTIBLE_FAILED;
> }
>
> -static u32 asle_set_docking(struct drm_i915_private *dev_priv, u32 iuer)
> +static u32 asle_set_docking(struct drm_i915_private *i915, u32 iuer)
> {
> if (iuer & ASLE_IUER_DOCKING)
> DRM_DEBUG_DRIVER("Docking is not supported (docked)\n");
> @@ -546,7 +546,7 @@ static u32 asle_set_docking(struct drm_i915_private *dev_priv, u32 iuer)
> return ASLC_DOCKING_FAILED;
> }
>
> -static u32 asle_isct_state(struct drm_i915_private *dev_priv)
> +static u32 asle_isct_state(struct drm_i915_private *i915)
> {
> DRM_DEBUG_DRIVER("ISCT is not supported\n");
> return ASLC_ISCT_STATE_FAILED;
> @@ -556,9 +556,9 @@ static void asle_work(struct work_struct *work)
> {
> struct intel_opregion *opregion =
> container_of(work, struct intel_opregion, asle_work);
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> container_of(opregion, struct drm_i915_private, opregion);
> - struct opregion_asle *asle = dev_priv->opregion.asle;
> + struct opregion_asle *asle = i915->opregion.asle;
> u32 aslc_stat = 0;
> u32 aslc_req;
>
> @@ -574,40 +574,40 @@ static void asle_work(struct work_struct *work)
> }
>
> if (aslc_req & ASLC_SET_ALS_ILLUM)
> - aslc_stat |= asle_set_als_illum(dev_priv, asle->alsi);
> + aslc_stat |= asle_set_als_illum(i915, asle->alsi);
>
> if (aslc_req & ASLC_SET_BACKLIGHT)
> - aslc_stat |= asle_set_backlight(dev_priv, asle->bclp);
> + aslc_stat |= asle_set_backlight(i915, asle->bclp);
>
> if (aslc_req & ASLC_SET_PFIT)
> - aslc_stat |= asle_set_pfit(dev_priv, asle->pfit);
> + aslc_stat |= asle_set_pfit(i915, asle->pfit);
>
> if (aslc_req & ASLC_SET_PWM_FREQ)
> - aslc_stat |= asle_set_pwm_freq(dev_priv, asle->pfmb);
> + aslc_stat |= asle_set_pwm_freq(i915, asle->pfmb);
>
> if (aslc_req & ASLC_SUPPORTED_ROTATION_ANGLES)
> - aslc_stat |= asle_set_supported_rotation_angles(dev_priv,
> + aslc_stat |= asle_set_supported_rotation_angles(i915,
> asle->srot);
>
> if (aslc_req & ASLC_BUTTON_ARRAY)
> - aslc_stat |= asle_set_button_array(dev_priv, asle->iuer);
> + aslc_stat |= asle_set_button_array(i915, asle->iuer);
>
> if (aslc_req & ASLC_CONVERTIBLE_INDICATOR)
> - aslc_stat |= asle_set_convertible(dev_priv, asle->iuer);
> + aslc_stat |= asle_set_convertible(i915, asle->iuer);
>
> if (aslc_req & ASLC_DOCKING_INDICATOR)
> - aslc_stat |= asle_set_docking(dev_priv, asle->iuer);
> + aslc_stat |= asle_set_docking(i915, asle->iuer);
>
> if (aslc_req & ASLC_ISCT_STATE_CHANGE)
> - aslc_stat |= asle_isct_state(dev_priv);
> + aslc_stat |= asle_isct_state(i915);
>
> asle->aslc = aslc_stat;
> }
>
> -void intel_opregion_asle_intr(struct drm_i915_private *dev_priv)
> +void intel_opregion_asle_intr(struct drm_i915_private *i915)
> {
> - if (dev_priv->opregion.asle)
> - schedule_work(&dev_priv->opregion.asle_work);
> + if (i915->opregion.asle)
> + schedule_work(&i915->opregion.asle_work);
> }
>
> #define ACPI_EV_DISPLAY_SWITCH (1<<0)
> @@ -702,9 +702,9 @@ static u32 acpi_display_type(struct intel_connector *connector)
> return display_type;
> }
>
> -static void intel_didl_outputs(struct drm_i915_private *dev_priv)
> +static void intel_didl_outputs(struct drm_i915_private *i915)
> {
> - struct intel_opregion *opregion = &dev_priv->opregion;
> + struct intel_opregion *opregion = &i915->opregion;
> struct intel_connector *connector;
> struct drm_connector_list_iter conn_iter;
> int i = 0, max_outputs;
> @@ -720,7 +720,7 @@ static void intel_didl_outputs(struct drm_i915_private *dev_priv)
> max_outputs = ARRAY_SIZE(opregion->acpi->didl) +
> ARRAY_SIZE(opregion->acpi->did2);
>
> - drm_connector_list_iter_begin(&dev_priv->drm, &conn_iter);
> + drm_connector_list_iter_begin(&i915->drm, &conn_iter);
> for_each_intel_connector_iter(connector, &conn_iter) {
> u32 device_id, type;
>
> @@ -749,9 +749,9 @@ static void intel_didl_outputs(struct drm_i915_private *dev_priv)
> set_did(opregion, i, 0);
> }
>
> -static void intel_setup_cadls(struct drm_i915_private *dev_priv)
> +static void intel_setup_cadls(struct drm_i915_private *i915)
> {
> - struct intel_opregion *opregion = &dev_priv->opregion;
> + struct intel_opregion *opregion = &i915->opregion;
> struct intel_connector *connector;
> struct drm_connector_list_iter conn_iter;
> int i = 0;
> @@ -766,7 +766,7 @@ static void intel_setup_cadls(struct drm_i915_private *dev_priv)
> * Note that internal panels should be at the front of the connector
> * list already, ensuring they're not left out.
> */
> - drm_connector_list_iter_begin(&dev_priv->drm, &conn_iter);
> + drm_connector_list_iter_begin(&i915->drm, &conn_iter);
> for_each_intel_connector_iter(connector, &conn_iter) {
> if (i >= ARRAY_SIZE(opregion->acpi->cadl))
> break;
> @@ -779,9 +779,9 @@ static void intel_setup_cadls(struct drm_i915_private *dev_priv)
> opregion->acpi->cadl[i] = 0;
> }
>
> -static void swsci_setup(struct drm_i915_private *dev_priv)
> +static void swsci_setup(struct drm_i915_private *i915)
> {
> - struct intel_opregion *opregion = &dev_priv->opregion;
> + struct intel_opregion *opregion = &i915->opregion;
> bool requested_callbacks = false;
> u32 tmp;
>
> @@ -790,7 +790,7 @@ static void swsci_setup(struct drm_i915_private *dev_priv)
> opregion->swsci_sbcb_sub_functions = 1;
>
> /* We use GBDA to ask for supported GBDA calls. */
> - if (swsci(dev_priv, SWSCI_GBDA_SUPPORTED_CALLS, 0, &tmp) == 0) {
> + if (swsci(i915, SWSCI_GBDA_SUPPORTED_CALLS, 0, &tmp) == 0) {
> /* make the bits match the sub-function codes */
> tmp <<= 1;
> opregion->swsci_gbda_sub_functions |= tmp;
> @@ -801,7 +801,7 @@ static void swsci_setup(struct drm_i915_private *dev_priv)
> * must not call interfaces that are not specifically requested by the
> * bios.
> */
> - if (swsci(dev_priv, SWSCI_GBDA_REQUESTED_CALLBACKS, 0, &tmp) == 0) {
> + if (swsci(i915, SWSCI_GBDA_REQUESTED_CALLBACKS, 0, &tmp) == 0) {
> /* here, the bits already match sub-function codes */
> opregion->swsci_sbcb_sub_functions |= tmp;
> requested_callbacks = true;
> @@ -812,7 +812,7 @@ static void swsci_setup(struct drm_i915_private *dev_priv)
> * the callback is _requested_. But we still can't call interfaces that
> * are not requested.
> */
> - if (swsci(dev_priv, SWSCI_SBCB_SUPPORTED_CALLBACKS, 0, &tmp) == 0) {
> + if (swsci(i915, SWSCI_SBCB_SUPPORTED_CALLBACKS, 0, &tmp) == 0) {
> /* make the bits match the sub-function codes */
> u32 low = tmp & 0x7ff;
> u32 high = tmp & ~0xfff; /* bit 11 is reserved */
> @@ -854,9 +854,9 @@ static const struct dmi_system_id intel_no_opregion_vbt[] = {
> { }
> };
>
> -static int intel_load_vbt_firmware(struct drm_i915_private *dev_priv)
> +static int intel_load_vbt_firmware(struct drm_i915_private *i915)
> {
> - struct intel_opregion *opregion = &dev_priv->opregion;
> + struct intel_opregion *opregion = &i915->opregion;
> const struct firmware *fw = NULL;
> const char *name = i915_modparams.vbt_firmware;
> int ret;
> @@ -864,7 +864,7 @@ static int intel_load_vbt_firmware(struct drm_i915_private *dev_priv)
> if (!name || !*name)
> return -ENOENT;
>
> - ret = request_firmware(&fw, name, &dev_priv->drm.pdev->dev);
> + ret = request_firmware(&fw, name, &i915->drm.pdev->dev);
> if (ret) {
> DRM_ERROR("Requesting VBT firmware \"%s\" failed (%d)\n",
> name, ret);
> @@ -891,10 +891,10 @@ static int intel_load_vbt_firmware(struct drm_i915_private *dev_priv)
> return ret;
> }
>
> -int intel_opregion_setup(struct drm_i915_private *dev_priv)
> +int intel_opregion_setup(struct drm_i915_private *i915)
> {
> - struct intel_opregion *opregion = &dev_priv->opregion;
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct intel_opregion *opregion = &i915->opregion;
> + struct pci_dev *pdev = i915->drm.pdev;
> u32 asls, mboxes;
> char buf[sizeof(OPREGION_SIGNATURE)];
> int err = 0;
> @@ -945,7 +945,7 @@ int intel_opregion_setup(struct drm_i915_private *dev_priv)
> if (mboxes & MBOX_SWSCI) {
> DRM_DEBUG_DRIVER("SWSCI supported\n");
> opregion->swsci = base + OPREGION_SWSCI_OFFSET;
> - swsci_setup(dev_priv);
> + swsci_setup(i915);
> }
>
> if (mboxes & MBOX_ASLE) {
> @@ -958,7 +958,7 @@ int intel_opregion_setup(struct drm_i915_private *dev_priv)
> if (mboxes & MBOX_ASLE_EXT)
> DRM_DEBUG_DRIVER("ASLE extension supported\n");
>
> - if (intel_load_vbt_firmware(dev_priv) == 0)
> + if (intel_load_vbt_firmware(i915) == 0)
> goto out;
>
> if (dmi_check_system(intel_no_opregion_vbt))
> @@ -1043,12 +1043,12 @@ static const struct dmi_system_id intel_use_opregion_panel_type[] = {
> };
>
> int
> -intel_opregion_get_panel_type(struct drm_i915_private *dev_priv)
> +intel_opregion_get_panel_type(struct drm_i915_private *i915)
> {
> u32 panel_details;
> int ret;
>
> - ret = swsci(dev_priv, SWSCI_GBDA_PANEL_DETAILS, 0x0, &panel_details);
> + ret = swsci(i915, SWSCI_GBDA_PANEL_DETAILS, 0x0, &panel_details);
> if (ret) {
> DRM_DEBUG_KMS("Failed to get panel details from OpRegion (%d)\n",
> ret);
> diff --git a/drivers/gpu/drm/i915/intel_opregion.h b/drivers/gpu/drm/i915/intel_opregion.h
> index 4aa68ffbd30e..8e0b7bdecbe4 100644
> --- a/drivers/gpu/drm/i915/intel_opregion.h
> +++ b/drivers/gpu/drm/i915/intel_opregion.h
> @@ -56,47 +56,47 @@ struct intel_opregion {
>
> #ifdef CONFIG_ACPI
>
> -int intel_opregion_setup(struct drm_i915_private *dev_priv);
> +int intel_opregion_setup(struct drm_i915_private *i915);
>
> -void intel_opregion_register(struct drm_i915_private *dev_priv);
> -void intel_opregion_unregister(struct drm_i915_private *dev_priv);
> +void intel_opregion_register(struct drm_i915_private *i915);
> +void intel_opregion_unregister(struct drm_i915_private *i915);
>
> -void intel_opregion_resume(struct drm_i915_private *dev_priv);
> -void intel_opregion_suspend(struct drm_i915_private *dev_priv,
> +void intel_opregion_resume(struct drm_i915_private *i915);
> +void intel_opregion_suspend(struct drm_i915_private *i915,
> pci_power_t state);
>
> -void intel_opregion_asle_intr(struct drm_i915_private *dev_priv);
> +void intel_opregion_asle_intr(struct drm_i915_private *i915);
> int intel_opregion_notify_encoder(struct intel_encoder *intel_encoder,
> bool enable);
> -int intel_opregion_notify_adapter(struct drm_i915_private *dev_priv,
> +int intel_opregion_notify_adapter(struct drm_i915_private *i915,
> pci_power_t state);
> -int intel_opregion_get_panel_type(struct drm_i915_private *dev_priv);
> +int intel_opregion_get_panel_type(struct drm_i915_private *i915);
>
> #else /* CONFIG_ACPI*/
>
> -static inline int intel_opregion_setup(struct drm_i915_private *dev_priv)
> +static inline int intel_opregion_setup(struct drm_i915_private *i915)
> {
> return 0;
> }
>
> -static inline void intel_opregion_register(struct drm_i915_private *dev_priv)
> +static inline void intel_opregion_register(struct drm_i915_private *i915)
> {
> }
>
> -static inline void intel_opregion_unregister(struct drm_i915_private *dev_priv)
> +static inline void intel_opregion_unregister(struct drm_i915_private *i915)
> {
> }
>
> -static inline void intel_opregion_resume(struct drm_i915_private *dev_priv)
> +static inline void intel_opregion_resume(struct drm_i915_private *i915)
> {
> }
>
> -static inline void intel_opregion_suspend(struct drm_i915_private *dev_priv,
> +static inline void intel_opregion_suspend(struct drm_i915_private *i915,
> pci_power_t state)
> {
> }
>
> -static inline void intel_opregion_asle_intr(struct drm_i915_private *dev_priv)
> +static inline void intel_opregion_asle_intr(struct drm_i915_private *i915)
> {
> }
>
> diff --git a/drivers/gpu/drm/i915/intel_overlay.c b/drivers/gpu/drm/i915/intel_overlay.c
> index a2ac06a08715..802bc4c72ecd 100644
> --- a/drivers/gpu/drm/i915/intel_overlay.c
> +++ b/drivers/gpu/drm/i915/intel_overlay.c
> @@ -193,10 +193,10 @@ struct intel_overlay {
> struct i915_active_request last_flip;
> };
>
> -static void i830_overlay_clock_gating(struct drm_i915_private *dev_priv,
> +static void i830_overlay_clock_gating(struct drm_i915_private *i915,
> bool enable)
> {
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct pci_dev *pdev = i915->drm.pdev;
> u8 val;
>
> /* WA_OVERLAY_CLKGATE:alm */
> @@ -247,7 +247,7 @@ static struct i915_request *alloc_request(struct intel_overlay *overlay)
> /* overlay needs to be disable in OCMD reg */
> static int intel_overlay_on(struct intel_overlay *overlay)
> {
> - struct drm_i915_private *dev_priv = overlay->i915;
> + struct drm_i915_private *i915 = overlay->i915;
> struct i915_request *rq;
> u32 *cs;
>
> @@ -265,8 +265,8 @@ static int intel_overlay_on(struct intel_overlay *overlay)
>
> overlay->active = true;
>
> - if (IS_I830(dev_priv))
> - i830_overlay_clock_gating(dev_priv, false);
> + if (IS_I830(i915))
> + i830_overlay_clock_gating(i915, false);
>
> *cs++ = MI_OVERLAY_FLIP | MI_OVERLAY_ON;
> *cs++ = overlay->flip_addr | OFC_UPDATE;
> @@ -303,7 +303,7 @@ static int intel_overlay_continue(struct intel_overlay *overlay,
> struct i915_vma *vma,
> bool load_polyphase_filter)
> {
> - struct drm_i915_private *dev_priv = overlay->i915;
> + struct drm_i915_private *i915 = overlay->i915;
> struct i915_request *rq;
> u32 flip_addr = overlay->flip_addr;
> u32 tmp, *cs;
> @@ -369,7 +369,7 @@ static void intel_overlay_off_tail(struct i915_active_request *active,
> {
> struct intel_overlay *overlay =
> container_of(active, typeof(*overlay), last_flip);
> - struct drm_i915_private *dev_priv = overlay->i915;
> + struct drm_i915_private *i915 = overlay->i915;
>
> intel_overlay_release_old_vma(overlay);
>
> @@ -377,8 +377,8 @@ static void intel_overlay_off_tail(struct i915_active_request *active,
> overlay->crtc = NULL;
> overlay->active = false;
>
> - if (IS_I830(dev_priv))
> - i830_overlay_clock_gating(dev_priv, true);
> + if (IS_I830(i915))
> + i830_overlay_clock_gating(i915, true);
> }
>
> /* overlay needs to be disabled in OCMD reg */
> @@ -437,11 +437,11 @@ static int intel_overlay_recover_from_interrupt(struct intel_overlay *overlay)
> */
> static int intel_overlay_release_old_vid(struct intel_overlay *overlay)
> {
> - struct drm_i915_private *dev_priv = overlay->i915;
> + struct drm_i915_private *i915 = overlay->i915;
> u32 *cs;
> int ret;
>
> - lockdep_assert_held(&dev_priv->drm.struct_mutex);
> + lockdep_assert_held(&i915->drm.struct_mutex);
>
> /* Only wait if there is actually an old frame to release to
> * guarantee forward progress.
> @@ -477,9 +477,9 @@ static int intel_overlay_release_old_vid(struct intel_overlay *overlay)
> return 0;
> }
>
> -void intel_overlay_reset(struct drm_i915_private *dev_priv)
> +void intel_overlay_reset(struct drm_i915_private *i915)
> {
> - struct intel_overlay *overlay = dev_priv->overlay;
> + struct intel_overlay *overlay = i915->overlay;
>
> if (!overlay)
> return;
> @@ -540,11 +540,11 @@ static int uv_vsubsampling(u32 format)
> }
> }
>
> -static u32 calc_swidthsw(struct drm_i915_private *dev_priv, u32 offset, u32 width)
> +static u32 calc_swidthsw(struct drm_i915_private *i915, u32 offset, u32 width)
> {
> u32 sw;
>
> - if (IS_GEN(dev_priv, 2))
> + if (IS_GEN(i915, 2))
> sw = ALIGN((offset & 31) + width, 32);
> else
> sw = ALIGN((offset & 63) + width, 64);
> @@ -749,21 +749,21 @@ static int intel_overlay_do_put_image(struct intel_overlay *overlay,
> struct drm_intel_overlay_put_image *params)
> {
> struct overlay_registers __iomem *regs = overlay->regs;
> - struct drm_i915_private *dev_priv = overlay->i915;
> + struct drm_i915_private *i915 = overlay->i915;
> u32 swidth, swidthsw, sheight, ostride;
> enum pipe pipe = overlay->crtc->pipe;
> bool scale_changed = false;
> struct i915_vma *vma;
> int ret, tmp_width;
>
> - lockdep_assert_held(&dev_priv->drm.struct_mutex);
> - WARN_ON(!drm_modeset_is_locked(&dev_priv->drm.mode_config.connection_mutex));
> + lockdep_assert_held(&i915->drm.struct_mutex);
> + WARN_ON(!drm_modeset_is_locked(&i915->drm.mode_config.connection_mutex));
>
> ret = intel_overlay_release_old_vid(overlay);
> if (ret != 0)
> return ret;
>
> - atomic_inc(&dev_priv->gpu_error.pending_fb_pin);
> + atomic_inc(&i915->gpu_error.pending_fb_pin);
>
> i915_gem_object_lock(new_bo);
> vma = i915_gem_object_pin_to_display_plane(new_bo,
> @@ -783,7 +783,7 @@ static int intel_overlay_do_put_image(struct intel_overlay *overlay,
> u32 oconfig;
>
> oconfig = OCONF_CC_OUT_8BIT;
> - if (IS_GEN(dev_priv, 4))
> + if (IS_GEN(i915, 4))
> oconfig |= OCONF_CSC_MODE_BT709;
> oconfig |= pipe == 0 ?
> OCONF_PIPE_A : OCONF_PIPE_B;
> @@ -804,7 +804,7 @@ static int intel_overlay_do_put_image(struct intel_overlay *overlay,
> tmp_width = params->src_width;
>
> swidth = params->src_width;
> - swidthsw = calc_swidthsw(dev_priv, params->offset_Y, tmp_width);
> + swidthsw = calc_swidthsw(i915, params->offset_Y, tmp_width);
> sheight = params->src_height;
> iowrite32(i915_ggtt_offset(vma) + params->offset_Y, ®s->OBUF_0Y);
> ostride = params->stride_Y;
> @@ -817,9 +817,9 @@ static int intel_overlay_do_put_image(struct intel_overlay *overlay,
> swidth |= (params->src_width / uv_hscale) << 16;
> sheight |= (params->src_height / uv_vscale) << 16;
>
> - tmp_U = calc_swidthsw(dev_priv, params->offset_U,
> + tmp_U = calc_swidthsw(i915, params->offset_U,
> params->src_width / uv_hscale);
> - tmp_V = calc_swidthsw(dev_priv, params->offset_V,
> + tmp_V = calc_swidthsw(i915, params->offset_V,
> params->src_width / uv_hscale);
> swidthsw |= max(tmp_U, tmp_V) << 16;
>
> @@ -851,18 +851,18 @@ static int intel_overlay_do_put_image(struct intel_overlay *overlay,
> out_unpin:
> i915_gem_object_unpin_from_display_plane(vma);
> out_pin_section:
> - atomic_dec(&dev_priv->gpu_error.pending_fb_pin);
> + atomic_dec(&i915->gpu_error.pending_fb_pin);
>
> return ret;
> }
>
> int intel_overlay_switch_off(struct intel_overlay *overlay)
> {
> - struct drm_i915_private *dev_priv = overlay->i915;
> + struct drm_i915_private *i915 = overlay->i915;
> int ret;
>
> - lockdep_assert_held(&dev_priv->drm.struct_mutex);
> - WARN_ON(!drm_modeset_is_locked(&dev_priv->drm.mode_config.connection_mutex));
> + lockdep_assert_held(&i915->drm.struct_mutex);
> + WARN_ON(!drm_modeset_is_locked(&i915->drm.mode_config.connection_mutex));
>
> ret = intel_overlay_recover_from_interrupt(overlay);
> if (ret != 0)
> @@ -895,14 +895,14 @@ static int check_overlay_possible_on_crtc(struct intel_overlay *overlay,
>
> static void update_pfit_vscale_ratio(struct intel_overlay *overlay)
> {
> - struct drm_i915_private *dev_priv = overlay->i915;
> + struct drm_i915_private *i915 = overlay->i915;
> u32 pfit_control = I915_READ(PFIT_CONTROL);
> u32 ratio;
>
> /* XXX: This is not the same logic as in the xorg driver, but more in
> * line with the intel documentation for the i965
> */
> - if (INTEL_GEN(dev_priv) >= 4) {
> + if (INTEL_GEN(i915) >= 4) {
> /* on i965 use the PGM reg to read out the autoscaler values */
> ratio = I915_READ(PFIT_PGM_RATIOS) >> PFIT_VERT_SCALE_SHIFT_965;
> } else {
> @@ -947,7 +947,7 @@ static int check_overlay_scaling(struct drm_intel_overlay_put_image *rec)
> return 0;
> }
>
> -static int check_overlay_src(struct drm_i915_private *dev_priv,
> +static int check_overlay_src(struct drm_i915_private *i915,
> struct drm_intel_overlay_put_image *rec,
> struct drm_i915_gem_object *new_bo)
> {
> @@ -958,7 +958,7 @@ static int check_overlay_src(struct drm_i915_private *dev_priv,
> u32 tmp;
>
> /* check src dimensions */
> - if (IS_I845G(dev_priv) || IS_I830(dev_priv)) {
> + if (IS_I845G(i915) || IS_I830(i915)) {
> if (rec->src_height > IMAGE_MAX_HEIGHT_LEGACY ||
> rec->src_width > IMAGE_MAX_WIDTH_LEGACY)
> return -EINVAL;
> @@ -1010,14 +1010,14 @@ static int check_overlay_src(struct drm_i915_private *dev_priv,
> return -EINVAL;
>
> /* stride checking */
> - if (IS_I830(dev_priv) || IS_I845G(dev_priv))
> + if (IS_I830(i915) || IS_I845G(i915))
> stride_mask = 255;
> else
> stride_mask = 63;
>
> if (rec->stride_Y & stride_mask || rec->stride_UV & stride_mask)
> return -EINVAL;
> - if (IS_GEN(dev_priv, 4) && rec->stride_Y < 512)
> + if (IS_GEN(i915, 4) && rec->stride_Y < 512)
> return -EINVAL;
>
> tmp = (rec->flags & I915_OVERLAY_TYPE_MASK) == I915_OVERLAY_YUV_PLANAR ?
> @@ -1062,14 +1062,14 @@ int intel_overlay_put_image_ioctl(struct drm_device *dev, void *data,
> struct drm_file *file_priv)
> {
> struct drm_intel_overlay_put_image *params = data;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_overlay *overlay;
> struct drm_crtc *drmmode_crtc;
> struct intel_crtc *crtc;
> struct drm_i915_gem_object *new_bo;
> int ret;
>
> - overlay = dev_priv->overlay;
> + overlay = i915->overlay;
> if (!overlay) {
> DRM_DEBUG("userspace bug: no overlay\n");
> return -ENODEV;
> @@ -1148,7 +1148,7 @@ int intel_overlay_put_image_ioctl(struct drm_device *dev, void *data,
> goto out_unlock;
> }
>
> - ret = check_overlay_src(dev_priv, params, new_bo);
> + ret = check_overlay_src(i915, params, new_bo);
> if (ret != 0)
> goto out_unlock;
>
> @@ -1231,11 +1231,11 @@ int intel_overlay_attrs_ioctl(struct drm_device *dev, void *data,
> struct drm_file *file_priv)
> {
> struct drm_intel_overlay_attrs *attrs = data;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_overlay *overlay;
> int ret;
>
> - overlay = dev_priv->overlay;
> + overlay = i915->overlay;
> if (!overlay) {
> DRM_DEBUG("userspace bug: no overlay\n");
> return -ENODEV;
> @@ -1251,7 +1251,7 @@ int intel_overlay_attrs_ioctl(struct drm_device *dev, void *data,
> attrs->contrast = overlay->contrast;
> attrs->saturation = overlay->saturation;
>
> - if (!IS_GEN(dev_priv, 2)) {
> + if (!IS_GEN(i915, 2)) {
> attrs->gamma0 = I915_READ(OGAMC0);
> attrs->gamma1 = I915_READ(OGAMC1);
> attrs->gamma2 = I915_READ(OGAMC2);
> @@ -1275,7 +1275,7 @@ int intel_overlay_attrs_ioctl(struct drm_device *dev, void *data,
> update_reg_attrs(overlay, overlay->regs);
>
> if (attrs->flags & I915_OVERLAY_UPDATE_GAMMA) {
> - if (IS_GEN(dev_priv, 2))
> + if (IS_GEN(i915, 2))
> goto out_unlock;
>
> if (overlay->active) {
> @@ -1351,19 +1351,19 @@ static int get_registers(struct intel_overlay *overlay, bool use_phys)
> return err;
> }
>
> -void intel_overlay_setup(struct drm_i915_private *dev_priv)
> +void intel_overlay_setup(struct drm_i915_private *i915)
> {
> struct intel_overlay *overlay;
> int ret;
>
> - if (!HAS_OVERLAY(dev_priv))
> + if (!HAS_OVERLAY(i915))
> return;
>
> overlay = kzalloc(sizeof(*overlay), GFP_KERNEL);
> if (!overlay)
> return;
>
> - overlay->i915 = dev_priv;
> + overlay->i915 = i915;
>
> overlay->color_key = 0x0101fe;
> overlay->color_key_enabled = true;
> @@ -1373,7 +1373,7 @@ void intel_overlay_setup(struct drm_i915_private *dev_priv)
>
> INIT_ACTIVE_REQUEST(&overlay->last_flip);
>
> - ret = get_registers(overlay, OVERLAY_NEEDS_PHYSICAL(dev_priv));
> + ret = get_registers(overlay, OVERLAY_NEEDS_PHYSICAL(i915));
> if (ret)
> goto out_free;
>
> @@ -1387,7 +1387,7 @@ void intel_overlay_setup(struct drm_i915_private *dev_priv)
> update_polyphase_filter(overlay->regs);
> update_reg_attrs(overlay, overlay->regs);
>
> - dev_priv->overlay = overlay;
> + i915->overlay = overlay;
> DRM_INFO("Initialized overlay support.\n");
> return;
>
> @@ -1397,11 +1397,11 @@ void intel_overlay_setup(struct drm_i915_private *dev_priv)
> kfree(overlay);
> }
>
> -void intel_overlay_cleanup(struct drm_i915_private *dev_priv)
> +void intel_overlay_cleanup(struct drm_i915_private *i915)
> {
> struct intel_overlay *overlay;
>
> - overlay = fetch_and_zero(&dev_priv->overlay);
> + overlay = fetch_and_zero(&i915->overlay);
> if (!overlay)
> return;
>
> @@ -1427,9 +1427,9 @@ struct intel_overlay_error_state {
> };
>
> struct intel_overlay_error_state *
> -intel_overlay_capture_error_state(struct drm_i915_private *dev_priv)
> +intel_overlay_capture_error_state(struct drm_i915_private *i915)
> {
> - struct intel_overlay *overlay = dev_priv->overlay;
> + struct intel_overlay *overlay = i915->overlay;
> struct intel_overlay_error_state *error;
>
> if (!overlay || !overlay->active)
> diff --git a/drivers/gpu/drm/i915/intel_overlay.h b/drivers/gpu/drm/i915/intel_overlay.h
> index a167c28acd27..2e8771cffee9 100644
> --- a/drivers/gpu/drm/i915/intel_overlay.h
> +++ b/drivers/gpu/drm/i915/intel_overlay.h
> @@ -13,16 +13,16 @@ struct drm_i915_private;
> struct intel_overlay;
> struct intel_overlay_error_state;
>
> -void intel_overlay_setup(struct drm_i915_private *dev_priv);
> -void intel_overlay_cleanup(struct drm_i915_private *dev_priv);
> +void intel_overlay_setup(struct drm_i915_private *i915);
> +void intel_overlay_cleanup(struct drm_i915_private *i915);
> int intel_overlay_switch_off(struct intel_overlay *overlay);
> int intel_overlay_put_image_ioctl(struct drm_device *dev, void *data,
> struct drm_file *file_priv);
> int intel_overlay_attrs_ioctl(struct drm_device *dev, void *data,
> struct drm_file *file_priv);
> -void intel_overlay_reset(struct drm_i915_private *dev_priv);
> +void intel_overlay_reset(struct drm_i915_private *i915);
> struct intel_overlay_error_state *
> -intel_overlay_capture_error_state(struct drm_i915_private *dev_priv);
> +intel_overlay_capture_error_state(struct drm_i915_private *i915);
> void intel_overlay_print_error_state(struct drm_i915_error_state_buf *e,
> struct intel_overlay_error_state *error);
>
> diff --git a/drivers/gpu/drm/i915/intel_panel.c b/drivers/gpu/drm/i915/intel_panel.c
> index 9cd4e37e3934..d80115089445 100644
> --- a/drivers/gpu/drm/i915/intel_panel.c
> +++ b/drivers/gpu/drm/i915/intel_panel.c
> @@ -65,7 +65,7 @@ struct drm_display_mode *
> intel_panel_edid_downclock_mode(struct intel_connector *connector,
> const struct drm_display_mode *fixed_mode)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> const struct drm_display_mode *scan, *best_mode = NULL;
> struct drm_display_mode *downclock_mode;
> int best_clock = fixed_mode->clock;
> @@ -92,7 +92,7 @@ intel_panel_edid_downclock_mode(struct intel_connector *connector,
> if (!best_mode)
> return NULL;
>
> - downclock_mode = drm_mode_duplicate(&dev_priv->drm, best_mode);
> + downclock_mode = drm_mode_duplicate(&i915->drm, best_mode);
> if (!downclock_mode)
> return NULL;
>
> @@ -106,7 +106,7 @@ intel_panel_edid_downclock_mode(struct intel_connector *connector,
> struct drm_display_mode *
> intel_panel_edid_fixed_mode(struct intel_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> const struct drm_display_mode *scan;
> struct drm_display_mode *fixed_mode;
>
> @@ -118,7 +118,7 @@ intel_panel_edid_fixed_mode(struct intel_connector *connector)
> if ((scan->type & DRM_MODE_TYPE_PREFERRED) == 0)
> continue;
>
> - fixed_mode = drm_mode_duplicate(&dev_priv->drm, scan);
> + fixed_mode = drm_mode_duplicate(&i915->drm, scan);
> if (!fixed_mode)
> return NULL;
>
> @@ -132,7 +132,7 @@ intel_panel_edid_fixed_mode(struct intel_connector *connector)
> scan = list_first_entry(&connector->base.probed_modes,
> typeof(*scan), head);
>
> - fixed_mode = drm_mode_duplicate(&dev_priv->drm, scan);
> + fixed_mode = drm_mode_duplicate(&i915->drm, scan);
> if (!fixed_mode)
> return NULL;
>
> @@ -148,15 +148,15 @@ intel_panel_edid_fixed_mode(struct intel_connector *connector)
> struct drm_display_mode *
> intel_panel_vbt_fixed_mode(struct intel_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct drm_display_info *info = &connector->base.display_info;
> struct drm_display_mode *fixed_mode;
>
> - if (!dev_priv->vbt.lfp_lvds_vbt_mode)
> + if (!i915->vbt.lfp_lvds_vbt_mode)
> return NULL;
>
> - fixed_mode = drm_mode_duplicate(&dev_priv->drm,
> - dev_priv->vbt.lfp_lvds_vbt_mode);
> + fixed_mode = drm_mode_duplicate(&i915->drm,
> + i915->vbt.lfp_lvds_vbt_mode);
> if (!fixed_mode)
> return NULL;
>
> @@ -378,7 +378,7 @@ void intel_gmch_panel_fitting(struct intel_crtc *intel_crtc,
> struct intel_crtc_state *pipe_config,
> int fitting_mode)
> {
> - struct drm_i915_private *dev_priv = to_i915(intel_crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(intel_crtc->base.dev);
> u32 pfit_control = 0, pfit_pgm_ratios = 0, border = 0;
> struct drm_display_mode *adjusted_mode = &pipe_config->base.adjusted_mode;
>
> @@ -399,7 +399,7 @@ void intel_gmch_panel_fitting(struct intel_crtc *intel_crtc,
> break;
> case DRM_MODE_SCALE_ASPECT:
> /* Scale but preserve the aspect ratio */
> - if (INTEL_GEN(dev_priv) >= 4)
> + if (INTEL_GEN(i915) >= 4)
> i965_scale_aspect(pipe_config, &pfit_control);
> else
> i9xx_scale_aspect(pipe_config, &pfit_control,
> @@ -413,7 +413,7 @@ void intel_gmch_panel_fitting(struct intel_crtc *intel_crtc,
> if (pipe_config->pipe_src_h != adjusted_mode->crtc_vdisplay ||
> pipe_config->pipe_src_w != adjusted_mode->crtc_hdisplay) {
> pfit_control |= PFIT_ENABLE;
> - if (INTEL_GEN(dev_priv) >= 4)
> + if (INTEL_GEN(i915) >= 4)
> pfit_control |= PFIT_SCALING_AUTO;
> else
> pfit_control |= (VERT_AUTO_SCALE |
> @@ -429,7 +429,7 @@ void intel_gmch_panel_fitting(struct intel_crtc *intel_crtc,
>
> /* 965+ wants fuzzy fitting */
> /* FIXME: handle multiple panels by failing gracefully */
> - if (INTEL_GEN(dev_priv) >= 4)
> + if (INTEL_GEN(i915) >= 4)
> pfit_control |= ((intel_crtc->pipe << PFIT_PIPE_SHIFT) |
> PFIT_FILTER_FUZZY);
>
> @@ -440,7 +440,7 @@ void intel_gmch_panel_fitting(struct intel_crtc *intel_crtc,
> }
>
> /* Make sure pre-965 set dither correctly for 18bpp panels. */
> - if (INTEL_GEN(dev_priv) < 4 && pipe_config->pipe_bpp == 18)
> + if (INTEL_GEN(i915) < 4 && pipe_config->pipe_bpp == 18)
> pfit_control |= PANEL_8TO6_DITHER_ENABLE;
>
> pipe_config->gmch_pfit.control = pfit_control;
> @@ -517,7 +517,7 @@ static inline u32 scale_hw_to_user(struct intel_connector *connector,
> static u32 intel_panel_compute_brightness(struct intel_connector *connector,
> u32 val)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
>
> WARN_ON(panel->backlight.max == 0);
> @@ -526,7 +526,7 @@ static u32 intel_panel_compute_brightness(struct intel_connector *connector,
> return val;
>
> if (i915_modparams.invert_brightness > 0 ||
> - dev_priv->quirks & QUIRK_INVERT_BRIGHTNESS) {
> + i915->quirks & QUIRK_INVERT_BRIGHTNESS) {
> return panel->backlight.max - val + panel->backlight.min;
> }
>
> @@ -535,39 +535,39 @@ static u32 intel_panel_compute_brightness(struct intel_connector *connector,
>
> static u32 lpt_get_backlight(struct intel_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
>
> return I915_READ(BLC_PWM_PCH_CTL2) & BACKLIGHT_DUTY_CYCLE_MASK;
> }
>
> static u32 pch_get_backlight(struct intel_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
>
> return I915_READ(BLC_PWM_CPU_CTL) & BACKLIGHT_DUTY_CYCLE_MASK;
> }
>
> static u32 i9xx_get_backlight(struct intel_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
> u32 val;
>
> val = I915_READ(BLC_PWM_CTL) & BACKLIGHT_DUTY_CYCLE_MASK;
> - if (INTEL_GEN(dev_priv) < 4)
> + if (INTEL_GEN(i915) < 4)
> val >>= 1;
>
> if (panel->backlight.combination_mode) {
> u8 lbpc;
>
> - pci_read_config_byte(dev_priv->drm.pdev, LBPC, &lbpc);
> + pci_read_config_byte(i915->drm.pdev, LBPC, &lbpc);
> val *= lbpc;
> }
>
> return val;
> }
>
> -static u32 _vlv_get_backlight(struct drm_i915_private *dev_priv, enum pipe pipe)
> +static u32 _vlv_get_backlight(struct drm_i915_private *i915, enum pipe pipe)
> {
> if (WARN_ON(pipe != PIPE_A && pipe != PIPE_B))
> return 0;
> @@ -577,15 +577,15 @@ static u32 _vlv_get_backlight(struct drm_i915_private *dev_priv, enum pipe pipe)
>
> static u32 vlv_get_backlight(struct intel_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> enum pipe pipe = intel_connector_get_pipe(connector);
>
> - return _vlv_get_backlight(dev_priv, pipe);
> + return _vlv_get_backlight(i915, pipe);
> }
>
> static u32 bxt_get_backlight(struct intel_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
>
> return I915_READ(BXT_BLC_PWM_DUTY(panel->backlight.controller));
> @@ -603,7 +603,7 @@ static u32 pwm_get_backlight(struct intel_connector *connector)
> static void lpt_set_backlight(const struct drm_connector_state *conn_state, u32 level)
> {
> struct intel_connector *connector = to_intel_connector(conn_state->connector);
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
>
> u32 val = I915_READ(BLC_PWM_PCH_CTL2) & ~BACKLIGHT_DUTY_CYCLE_MASK;
> I915_WRITE(BLC_PWM_PCH_CTL2, val | level);
> @@ -612,7 +612,7 @@ static void lpt_set_backlight(const struct drm_connector_state *conn_state, u32
> static void pch_set_backlight(const struct drm_connector_state *conn_state, u32 level)
> {
> struct intel_connector *connector = to_intel_connector(conn_state->connector);
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> u32 tmp;
>
> tmp = I915_READ(BLC_PWM_CPU_CTL) & ~BACKLIGHT_DUTY_CYCLE_MASK;
> @@ -622,7 +622,7 @@ static void pch_set_backlight(const struct drm_connector_state *conn_state, u32
> static void i9xx_set_backlight(const struct drm_connector_state *conn_state, u32 level)
> {
> struct intel_connector *connector = to_intel_connector(conn_state->connector);
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
> u32 tmp, mask;
>
> @@ -633,10 +633,10 @@ static void i9xx_set_backlight(const struct drm_connector_state *conn_state, u32
>
> lbpc = level * 0xfe / panel->backlight.max + 1;
> level /= lbpc;
> - pci_write_config_byte(dev_priv->drm.pdev, LBPC, lbpc);
> + pci_write_config_byte(i915->drm.pdev, LBPC, lbpc);
> }
>
> - if (IS_GEN(dev_priv, 4)) {
> + if (IS_GEN(i915, 4)) {
> mask = BACKLIGHT_DUTY_CYCLE_MASK;
> } else {
> level <<= 1;
> @@ -650,7 +650,7 @@ static void i9xx_set_backlight(const struct drm_connector_state *conn_state, u32
> static void vlv_set_backlight(const struct drm_connector_state *conn_state, u32 level)
> {
> struct intel_connector *connector = to_intel_connector(conn_state->connector);
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> enum pipe pipe = to_intel_crtc(conn_state->crtc)->pipe;
> u32 tmp;
>
> @@ -661,7 +661,7 @@ static void vlv_set_backlight(const struct drm_connector_state *conn_state, u32
> static void bxt_set_backlight(const struct drm_connector_state *conn_state, u32 level)
> {
> struct intel_connector *connector = to_intel_connector(conn_state->connector);
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
>
> I915_WRITE(BXT_BLC_PWM_DUTY(panel->backlight.controller), level);
> @@ -694,7 +694,7 @@ void intel_panel_set_backlight_acpi(const struct drm_connector_state *conn_state
> u32 user_level, u32 user_max)
> {
> struct intel_connector *connector = to_intel_connector(conn_state->connector);
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
> u32 hw_level;
>
> @@ -707,7 +707,7 @@ void intel_panel_set_backlight_acpi(const struct drm_connector_state *conn_state
> if (!panel->backlight.present || !conn_state->crtc)
> return;
>
> - mutex_lock(&dev_priv->backlight_lock);
> + mutex_lock(&i915->backlight_lock);
>
> WARN_ON(panel->backlight.max == 0);
>
> @@ -723,13 +723,13 @@ void intel_panel_set_backlight_acpi(const struct drm_connector_state *conn_state
> if (panel->backlight.enabled)
> intel_panel_actually_set_backlight(conn_state, hw_level);
>
> - mutex_unlock(&dev_priv->backlight_lock);
> + mutex_unlock(&i915->backlight_lock);
> }
>
> static void lpt_disable_backlight(const struct drm_connector_state *old_conn_state)
> {
> struct intel_connector *connector = to_intel_connector(old_conn_state->connector);
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> u32 tmp;
>
> intel_panel_actually_set_backlight(old_conn_state, 0);
> @@ -755,7 +755,7 @@ static void lpt_disable_backlight(const struct drm_connector_state *old_conn_sta
> static void pch_disable_backlight(const struct drm_connector_state *old_conn_state)
> {
> struct intel_connector *connector = to_intel_connector(old_conn_state->connector);
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> u32 tmp;
>
> intel_panel_actually_set_backlight(old_conn_state, 0);
> @@ -774,7 +774,7 @@ static void i9xx_disable_backlight(const struct drm_connector_state *old_conn_st
>
> static void i965_disable_backlight(const struct drm_connector_state *old_conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(old_conn_state->connector->dev);
> + struct drm_i915_private *i915 = to_i915(old_conn_state->connector->dev);
> u32 tmp;
>
> intel_panel_actually_set_backlight(old_conn_state, 0);
> @@ -786,7 +786,7 @@ static void i965_disable_backlight(const struct drm_connector_state *old_conn_st
> static void vlv_disable_backlight(const struct drm_connector_state *old_conn_state)
> {
> struct intel_connector *connector = to_intel_connector(old_conn_state->connector);
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> enum pipe pipe = to_intel_crtc(old_conn_state->crtc)->pipe;
> u32 tmp;
>
> @@ -799,7 +799,7 @@ static void vlv_disable_backlight(const struct drm_connector_state *old_conn_sta
> static void bxt_disable_backlight(const struct drm_connector_state *old_conn_state)
> {
> struct intel_connector *connector = to_intel_connector(old_conn_state->connector);
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
> u32 tmp, val;
>
> @@ -819,7 +819,7 @@ static void bxt_disable_backlight(const struct drm_connector_state *old_conn_sta
> static void cnp_disable_backlight(const struct drm_connector_state *old_conn_state)
> {
> struct intel_connector *connector = to_intel_connector(old_conn_state->connector);
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
> u32 tmp;
>
> @@ -844,7 +844,7 @@ static void pwm_disable_backlight(const struct drm_connector_state *old_conn_sta
> void intel_panel_disable_backlight(const struct drm_connector_state *old_conn_state)
> {
> struct intel_connector *connector = to_intel_connector(old_conn_state->connector);
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
>
> if (!panel->backlight.present)
> @@ -856,26 +856,26 @@ void intel_panel_disable_backlight(const struct drm_connector_state *old_conn_st
> * backlight. This will leave the backlight on unnecessarily when
> * another client is not activated.
> */
> - if (dev_priv->drm.switch_power_state == DRM_SWITCH_POWER_CHANGING) {
> + if (i915->drm.switch_power_state == DRM_SWITCH_POWER_CHANGING) {
> DRM_DEBUG_DRIVER("Skipping backlight disable on vga switch\n");
> return;
> }
>
> - mutex_lock(&dev_priv->backlight_lock);
> + mutex_lock(&i915->backlight_lock);
>
> if (panel->backlight.device)
> panel->backlight.device->props.power = FB_BLANK_POWERDOWN;
> panel->backlight.enabled = false;
> panel->backlight.disable(old_conn_state);
>
> - mutex_unlock(&dev_priv->backlight_lock);
> + mutex_unlock(&i915->backlight_lock);
> }
>
> static void lpt_enable_backlight(const struct intel_crtc_state *crtc_state,
> const struct drm_connector_state *conn_state)
> {
> struct intel_connector *connector = to_intel_connector(conn_state->connector);
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
> u32 pch_ctl1, pch_ctl2, schicken;
>
> @@ -886,7 +886,7 @@ static void lpt_enable_backlight(const struct intel_crtc_state *crtc_state,
> I915_WRITE(BLC_PWM_PCH_CTL1, pch_ctl1);
> }
>
> - if (HAS_PCH_LPT(dev_priv)) {
> + if (HAS_PCH_LPT(i915)) {
> schicken = I915_READ(SOUTH_CHICKEN2);
> if (panel->backlight.alternate_pwm_increment)
> schicken |= LPT_PWM_GRANULARITY;
> @@ -910,7 +910,7 @@ static void lpt_enable_backlight(const struct intel_crtc_state *crtc_state,
> pch_ctl1 |= BLM_PCH_POLARITY;
>
> /* After LPT, override is the default. */
> - if (HAS_PCH_LPT(dev_priv))
> + if (HAS_PCH_LPT(i915))
> pch_ctl1 |= BLM_PCH_OVERRIDE_ENABLE;
>
> I915_WRITE(BLC_PWM_PCH_CTL1, pch_ctl1);
> @@ -925,7 +925,7 @@ static void pch_enable_backlight(const struct intel_crtc_state *crtc_state,
> const struct drm_connector_state *conn_state)
> {
> struct intel_connector *connector = to_intel_connector(conn_state->connector);
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
> enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> u32 cpu_ctl2, pch_ctl1, pch_ctl2;
> @@ -971,7 +971,7 @@ static void i9xx_enable_backlight(const struct intel_crtc_state *crtc_state,
> const struct drm_connector_state *conn_state)
> {
> struct intel_connector *connector = to_intel_connector(conn_state->connector);
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
> u32 ctl, freq;
>
> @@ -988,7 +988,7 @@ static void i9xx_enable_backlight(const struct intel_crtc_state *crtc_state,
> ctl = freq << 17;
> if (panel->backlight.combination_mode)
> ctl |= BLM_LEGACY_MODE;
> - if (IS_PINEVIEW(dev_priv) && panel->backlight.active_low_pwm)
> + if (IS_PINEVIEW(i915) && panel->backlight.active_low_pwm)
> ctl |= BLM_POLARITY_PNV;
>
> I915_WRITE(BLC_PWM_CTL, ctl);
> @@ -1002,7 +1002,7 @@ static void i9xx_enable_backlight(const struct intel_crtc_state *crtc_state,
> * 855gm only, but checking for gen2 is safe, as 855gm is the only gen2
> * that has backlight.
> */
> - if (IS_GEN(dev_priv, 2))
> + if (IS_GEN(i915, 2))
> I915_WRITE(BLC_HIST_CTL, BLM_HISTOGRAM_ENABLE);
> }
>
> @@ -1010,7 +1010,7 @@ static void i965_enable_backlight(const struct intel_crtc_state *crtc_state,
> const struct drm_connector_state *conn_state)
> {
> struct intel_connector *connector = to_intel_connector(conn_state->connector);
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
> enum pipe pipe = to_intel_crtc(conn_state->crtc)->pipe;
> u32 ctl, ctl2, freq;
> @@ -1045,7 +1045,7 @@ static void vlv_enable_backlight(const struct intel_crtc_state *crtc_state,
> const struct drm_connector_state *conn_state)
> {
> struct intel_connector *connector = to_intel_connector(conn_state->connector);
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
> enum pipe pipe = to_intel_crtc(crtc_state->base.crtc)->pipe;
> u32 ctl, ctl2;
> @@ -1075,7 +1075,7 @@ static void bxt_enable_backlight(const struct intel_crtc_state *crtc_state,
> const struct drm_connector_state *conn_state)
> {
> struct intel_connector *connector = to_intel_connector(conn_state->connector);
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
> enum pipe pipe = to_intel_crtc(crtc_state->base.crtc)->pipe;
> u32 pwm_ctl, val;
> @@ -1123,7 +1123,7 @@ static void cnp_enable_backlight(const struct intel_crtc_state *crtc_state,
> const struct drm_connector_state *conn_state)
> {
> struct intel_connector *connector = to_intel_connector(conn_state->connector);
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
> u32 pwm_ctl;
>
> @@ -1187,7 +1187,7 @@ void intel_panel_enable_backlight(const struct intel_crtc_state *crtc_state,
> const struct drm_connector_state *conn_state)
> {
> struct intel_connector *connector = to_intel_connector(conn_state->connector);
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
> enum pipe pipe = to_intel_crtc(crtc_state->base.crtc)->pipe;
>
> @@ -1196,28 +1196,28 @@ void intel_panel_enable_backlight(const struct intel_crtc_state *crtc_state,
>
> DRM_DEBUG_KMS("pipe %c\n", pipe_name(pipe));
>
> - mutex_lock(&dev_priv->backlight_lock);
> + mutex_lock(&i915->backlight_lock);
>
> __intel_panel_enable_backlight(crtc_state, conn_state);
>
> - mutex_unlock(&dev_priv->backlight_lock);
> + mutex_unlock(&i915->backlight_lock);
> }
>
> #if IS_ENABLED(CONFIG_BACKLIGHT_CLASS_DEVICE)
> static u32 intel_panel_get_backlight(struct intel_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
> u32 val = 0;
>
> - mutex_lock(&dev_priv->backlight_lock);
> + mutex_lock(&i915->backlight_lock);
>
> if (panel->backlight.enabled) {
> val = panel->backlight.get(connector);
> val = intel_panel_compute_brightness(connector, val);
> }
>
> - mutex_unlock(&dev_priv->backlight_lock);
> + mutex_unlock(&i915->backlight_lock);
>
> DRM_DEBUG_DRIVER("get backlight PWM = %d\n", val);
> return val;
> @@ -1228,14 +1228,14 @@ static void intel_panel_set_backlight(const struct drm_connector_state *conn_sta
> u32 user_level, u32 user_max)
> {
> struct intel_connector *connector = to_intel_connector(conn_state->connector);
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
> u32 hw_level;
>
> if (!panel->backlight.present)
> return;
>
> - mutex_lock(&dev_priv->backlight_lock);
> + mutex_lock(&i915->backlight_lock);
>
> WARN_ON(panel->backlight.max == 0);
>
> @@ -1245,7 +1245,7 @@ static void intel_panel_set_backlight(const struct drm_connector_state *conn_sta
> if (panel->backlight.enabled)
> intel_panel_actually_set_backlight(conn_state, hw_level);
>
> - mutex_unlock(&dev_priv->backlight_lock);
> + mutex_unlock(&i915->backlight_lock);
> }
>
> static int intel_backlight_device_update_status(struct backlight_device *bd)
> @@ -1284,11 +1284,11 @@ static int intel_backlight_device_get_brightness(struct backlight_device *bd)
> {
> struct intel_connector *connector = bl_get_data(bd);
> struct drm_device *dev = connector->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> intel_wakeref_t wakeref;
> int ret = 0;
>
> - with_intel_runtime_pm(dev_priv, wakeref) {
> + with_intel_runtime_pm(i915, wakeref) {
> u32 hw_level;
>
> drm_modeset_lock(&dev->mode_config.connection_mutex, NULL);
> @@ -1378,9 +1378,9 @@ void intel_backlight_device_unregister(struct intel_connector *connector)
> */
> static u32 cnp_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
>
> - return DIV_ROUND_CLOSEST(KHz(dev_priv->rawclk_freq), pwm_freq_hz);
> + return DIV_ROUND_CLOSEST(KHz(i915->rawclk_freq), pwm_freq_hz);
> }
>
> /*
> @@ -1416,7 +1416,7 @@ static u32 spt_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz)
> */
> static u32 lpt_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
> u32 mul, clock;
>
> @@ -1425,7 +1425,7 @@ static u32 lpt_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz)
> else
> mul = 128;
>
> - if (HAS_PCH_LPT_H(dev_priv))
> + if (HAS_PCH_LPT_H(i915))
> clock = MHz(135); /* LPT:H */
> else
> clock = MHz(24); /* LPT:LP */
> @@ -1439,9 +1439,9 @@ static u32 lpt_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz)
> */
> static u32 pch_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
>
> - return DIV_ROUND_CLOSEST(KHz(dev_priv->rawclk_freq), pwm_freq_hz * 128);
> + return DIV_ROUND_CLOSEST(KHz(i915->rawclk_freq), pwm_freq_hz * 128);
> }
>
> /*
> @@ -1454,13 +1454,13 @@ static u32 pch_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz)
> */
> static u32 i9xx_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> int clock;
>
> - if (IS_PINEVIEW(dev_priv))
> - clock = KHz(dev_priv->rawclk_freq);
> + if (IS_PINEVIEW(i915))
> + clock = KHz(i915->rawclk_freq);
> else
> - clock = KHz(dev_priv->cdclk.hw.cdclk);
> + clock = KHz(i915->cdclk.hw.cdclk);
>
> return DIV_ROUND_CLOSEST(clock, pwm_freq_hz * 32);
> }
> @@ -1472,13 +1472,13 @@ static u32 i9xx_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz)
> */
> static u32 i965_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> int clock;
>
> - if (IS_G4X(dev_priv))
> - clock = KHz(dev_priv->rawclk_freq);
> + if (IS_G4X(i915))
> + clock = KHz(i915->rawclk_freq);
> else
> - clock = KHz(dev_priv->cdclk.hw.cdclk);
> + clock = KHz(i915->cdclk.hw.cdclk);
>
> return DIV_ROUND_CLOSEST(clock, pwm_freq_hz * 128);
> }
> @@ -1490,17 +1490,17 @@ static u32 i965_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz)
> */
> static u32 vlv_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> int mul, clock;
>
> if ((I915_READ(CBR1_VLV) & CBR_PWM_CLOCK_MUX_SELECT) == 0) {
> - if (IS_CHERRYVIEW(dev_priv))
> + if (IS_CHERRYVIEW(i915))
> clock = KHz(19200);
> else
> clock = MHz(25);
> mul = 16;
> } else {
> - clock = KHz(dev_priv->rawclk_freq);
> + clock = KHz(i915->rawclk_freq);
> mul = 128;
> }
>
> @@ -1509,9 +1509,9 @@ static u32 vlv_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz)
>
> static u32 get_backlight_max_vbt(struct intel_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
> - u16 pwm_freq_hz = dev_priv->vbt.backlight.pwm_freq_hz;
> + u16 pwm_freq_hz = i915->vbt.backlight.pwm_freq_hz;
> u32 pwm;
>
> if (!panel->backlight.hz_to_pwm) {
> @@ -1542,7 +1542,7 @@ static u32 get_backlight_max_vbt(struct intel_connector *connector)
> */
> static u32 get_backlight_min_vbt(struct intel_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
> int min;
>
> @@ -1555,10 +1555,10 @@ static u32 get_backlight_min_vbt(struct intel_connector *connector)
> * against this by letting the minimum be at most (arbitrarily chosen)
> * 25% of the max.
> */
> - min = clamp_t(int, dev_priv->vbt.backlight.min_brightness, 0, 64);
> - if (min != dev_priv->vbt.backlight.min_brightness) {
> + min = clamp_t(int, i915->vbt.backlight.min_brightness, 0, 64);
> + if (min != i915->vbt.backlight.min_brightness) {
> DRM_DEBUG_KMS("clamping VBT min backlight %d/255 to %d/255\n",
> - dev_priv->vbt.backlight.min_brightness, min);
> + i915->vbt.backlight.min_brightness, min);
> }
>
> /* vbt value is a coefficient in range [0..255] */
> @@ -1567,12 +1567,12 @@ static u32 get_backlight_min_vbt(struct intel_connector *connector)
>
> static int lpt_setup_backlight(struct intel_connector *connector, enum pipe unused)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
> u32 cpu_ctl2, pch_ctl1, pch_ctl2, val;
> bool alt, cpu_mode;
>
> - if (HAS_PCH_LPT(dev_priv))
> + if (HAS_PCH_LPT(i915))
> alt = I915_READ(SOUTH_CHICKEN2) & LPT_PWM_GRANULARITY;
> else
> alt = I915_READ(SOUTH_CHICKEN1) & SPT_PWM_GRANULARITY;
> @@ -1596,7 +1596,7 @@ static int lpt_setup_backlight(struct intel_connector *connector, enum pipe unus
>
> panel->backlight.enabled = pch_ctl1 & BLM_PCH_PWM_ENABLE;
>
> - cpu_mode = panel->backlight.enabled && HAS_PCH_LPT(dev_priv) &&
> + cpu_mode = panel->backlight.enabled && HAS_PCH_LPT(i915) &&
> !(pch_ctl1 & BLM_PCH_OVERRIDE_ENABLE) &&
> (cpu_ctl2 & BLM_PWM_ENABLE);
> if (cpu_mode)
> @@ -1622,7 +1622,7 @@ static int lpt_setup_backlight(struct intel_connector *connector, enum pipe unus
>
> static int pch_setup_backlight(struct intel_connector *connector, enum pipe unused)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
> u32 cpu_ctl2, pch_ctl1, pch_ctl2, val;
>
> @@ -1654,16 +1654,16 @@ static int pch_setup_backlight(struct intel_connector *connector, enum pipe unus
>
> static int i9xx_setup_backlight(struct intel_connector *connector, enum pipe unused)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
> u32 ctl, val;
>
> ctl = I915_READ(BLC_PWM_CTL);
>
> - if (IS_GEN(dev_priv, 2) || IS_I915GM(dev_priv) || IS_I945GM(dev_priv))
> + if (IS_GEN(i915, 2) || IS_I915GM(i915) || IS_I945GM(i915))
> panel->backlight.combination_mode = ctl & BLM_LEGACY_MODE;
>
> - if (IS_PINEVIEW(dev_priv))
> + if (IS_PINEVIEW(i915))
> panel->backlight.active_low_pwm = ctl & BLM_POLARITY_PNV;
>
> panel->backlight.max = ctl >> 17;
> @@ -1693,7 +1693,7 @@ static int i9xx_setup_backlight(struct intel_connector *connector, enum pipe unu
>
> static int i965_setup_backlight(struct intel_connector *connector, enum pipe unused)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
> u32 ctl, ctl2, val;
>
> @@ -1727,7 +1727,7 @@ static int i965_setup_backlight(struct intel_connector *connector, enum pipe unu
>
> static int vlv_setup_backlight(struct intel_connector *connector, enum pipe pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
> u32 ctl, ctl2, val;
>
> @@ -1748,7 +1748,7 @@ static int vlv_setup_backlight(struct intel_connector *connector, enum pipe pipe
>
> panel->backlight.min = get_backlight_min_vbt(connector);
>
> - val = _vlv_get_backlight(dev_priv, pipe);
> + val = _vlv_get_backlight(i915, pipe);
> val = intel_panel_compute_brightness(connector, val);
> panel->backlight.level = clamp(val, panel->backlight.min,
> panel->backlight.max);
> @@ -1761,11 +1761,11 @@ static int vlv_setup_backlight(struct intel_connector *connector, enum pipe pipe
> static int
> bxt_setup_backlight(struct intel_connector *connector, enum pipe unused)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
> u32 pwm_ctl, val;
>
> - panel->backlight.controller = dev_priv->vbt.backlight.controller;
> + panel->backlight.controller = i915->vbt.backlight.controller;
>
> pwm_ctl = I915_READ(BXT_BLC_PWM_CTL(panel->backlight.controller));
>
> @@ -1801,7 +1801,7 @@ bxt_setup_backlight(struct intel_connector *connector, enum pipe unused)
> static int
> cnp_setup_backlight(struct intel_connector *connector, enum pipe unused)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
> u32 pwm_ctl, val;
>
> @@ -1881,28 +1881,28 @@ void intel_panel_update_backlight(struct intel_encoder *encoder,
> const struct drm_connector_state *conn_state)
> {
> struct intel_connector *connector = to_intel_connector(conn_state->connector);
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_panel *panel = &connector->panel;
>
> if (!panel->backlight.present)
> return;
>
> - mutex_lock(&dev_priv->backlight_lock);
> + mutex_lock(&i915->backlight_lock);
> if (!panel->backlight.enabled)
> __intel_panel_enable_backlight(crtc_state, conn_state);
>
> - mutex_unlock(&dev_priv->backlight_lock);
> + mutex_unlock(&i915->backlight_lock);
> }
>
> int intel_panel_setup_backlight(struct drm_connector *connector, enum pipe pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->dev);
> + struct drm_i915_private *i915 = to_i915(connector->dev);
> struct intel_connector *intel_connector = to_intel_connector(connector);
> struct intel_panel *panel = &intel_connector->panel;
> int ret;
>
> - if (!dev_priv->vbt.backlight.present) {
> - if (dev_priv->quirks & QUIRK_BACKLIGHT_PRESENT) {
> + if (!i915->vbt.backlight.present) {
> + if (i915->quirks & QUIRK_BACKLIGHT_PRESENT) {
> DRM_DEBUG_KMS("no backlight present per VBT, but present per quirk\n");
> } else {
> DRM_DEBUG_KMS("no backlight present per VBT\n");
> @@ -1915,9 +1915,9 @@ int intel_panel_setup_backlight(struct drm_connector *connector, enum pipe pipe)
> return -ENODEV;
>
> /* set level and max in panel struct */
> - mutex_lock(&dev_priv->backlight_lock);
> + mutex_lock(&i915->backlight_lock);
> ret = panel->backlight.setup(intel_connector, pipe);
> - mutex_unlock(&dev_priv->backlight_lock);
> + mutex_unlock(&i915->backlight_lock);
>
> if (ret) {
> DRM_DEBUG_KMS("failed to setup backlight for connector %s\n",
> @@ -1950,7 +1950,7 @@ intel_panel_init_backlight_funcs(struct intel_panel *panel)
> {
> struct intel_connector *connector =
> container_of(panel, struct intel_connector, panel);
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
>
> if (connector->base.connector_type == DRM_MODE_CONNECTOR_eDP &&
> intel_dp_aux_init_backlight_funcs(connector) == 0)
> @@ -1960,38 +1960,38 @@ intel_panel_init_backlight_funcs(struct intel_panel *panel)
> intel_dsi_dcs_init_backlight_funcs(connector) == 0)
> return;
>
> - if (IS_GEN9_LP(dev_priv)) {
> + if (IS_GEN9_LP(i915)) {
> panel->backlight.setup = bxt_setup_backlight;
> panel->backlight.enable = bxt_enable_backlight;
> panel->backlight.disable = bxt_disable_backlight;
> panel->backlight.set = bxt_set_backlight;
> panel->backlight.get = bxt_get_backlight;
> panel->backlight.hz_to_pwm = bxt_hz_to_pwm;
> - } else if (INTEL_PCH_TYPE(dev_priv) >= PCH_CNP) {
> + } else if (INTEL_PCH_TYPE(i915) >= PCH_CNP) {
> panel->backlight.setup = cnp_setup_backlight;
> panel->backlight.enable = cnp_enable_backlight;
> panel->backlight.disable = cnp_disable_backlight;
> panel->backlight.set = bxt_set_backlight;
> panel->backlight.get = bxt_get_backlight;
> panel->backlight.hz_to_pwm = cnp_hz_to_pwm;
> - } else if (INTEL_PCH_TYPE(dev_priv) >= PCH_LPT) {
> + } else if (INTEL_PCH_TYPE(i915) >= PCH_LPT) {
> panel->backlight.setup = lpt_setup_backlight;
> panel->backlight.enable = lpt_enable_backlight;
> panel->backlight.disable = lpt_disable_backlight;
> panel->backlight.set = lpt_set_backlight;
> panel->backlight.get = lpt_get_backlight;
> - if (HAS_PCH_LPT(dev_priv))
> + if (HAS_PCH_LPT(i915))
> panel->backlight.hz_to_pwm = lpt_hz_to_pwm;
> else
> panel->backlight.hz_to_pwm = spt_hz_to_pwm;
> - } else if (HAS_PCH_SPLIT(dev_priv)) {
> + } else if (HAS_PCH_SPLIT(i915)) {
> panel->backlight.setup = pch_setup_backlight;
> panel->backlight.enable = pch_enable_backlight;
> panel->backlight.disable = pch_disable_backlight;
> panel->backlight.set = pch_set_backlight;
> panel->backlight.get = pch_get_backlight;
> panel->backlight.hz_to_pwm = pch_hz_to_pwm;
> - } else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> + } else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> if (connector->base.connector_type == DRM_MODE_CONNECTOR_DSI) {
> panel->backlight.setup = pwm_setup_backlight;
> panel->backlight.enable = pwm_enable_backlight;
> @@ -2006,7 +2006,7 @@ intel_panel_init_backlight_funcs(struct intel_panel *panel)
> panel->backlight.get = vlv_get_backlight;
> panel->backlight.hz_to_pwm = vlv_hz_to_pwm;
> }
> - } else if (IS_GEN(dev_priv, 4)) {
> + } else if (IS_GEN(i915, 4)) {
> panel->backlight.setup = i965_setup_backlight;
> panel->backlight.enable = i965_enable_backlight;
> panel->backlight.disable = i965_disable_backlight;
> diff --git a/drivers/gpu/drm/i915/intel_pipe_crc.c b/drivers/gpu/drm/i915/intel_pipe_crc.c
> index 1e2c4307d05a..4a4409f4076f 100644
> --- a/drivers/gpu/drm/i915/intel_pipe_crc.c
> +++ b/drivers/gpu/drm/i915/intel_pipe_crc.c
> @@ -70,11 +70,11 @@ static int i8xx_pipe_crc_ctl_reg(enum intel_pipe_crc_source *source,
> return 0;
> }
>
> -static int i9xx_pipe_crc_auto_source(struct drm_i915_private *dev_priv,
> +static int i9xx_pipe_crc_auto_source(struct drm_i915_private *i915,
> enum pipe pipe,
> enum intel_pipe_crc_source *source)
> {
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_device *dev = &i915->drm;
> struct intel_encoder *encoder;
> struct intel_crtc *crtc;
> struct intel_digital_port *dig_port;
> @@ -124,7 +124,7 @@ static int i9xx_pipe_crc_auto_source(struct drm_i915_private *dev_priv,
> return ret;
> }
>
> -static int vlv_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv,
> +static int vlv_pipe_crc_ctl_reg(struct drm_i915_private *i915,
> enum pipe pipe,
> enum intel_pipe_crc_source *source,
> u32 *val)
> @@ -132,7 +132,7 @@ static int vlv_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv,
> bool need_stable_symbols = false;
>
> if (*source == INTEL_PIPE_CRC_SOURCE_AUTO) {
> - int ret = i9xx_pipe_crc_auto_source(dev_priv, pipe, source);
> + int ret = i9xx_pipe_crc_auto_source(i915, pipe, source);
> if (ret)
> return ret;
> }
> @@ -150,7 +150,7 @@ static int vlv_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv,
> need_stable_symbols = true;
> break;
> case INTEL_PIPE_CRC_SOURCE_DP_D:
> - if (!IS_CHERRYVIEW(dev_priv))
> + if (!IS_CHERRYVIEW(i915))
> return -EINVAL;
> *val = PIPE_CRC_ENABLE | PIPE_CRC_SOURCE_DP_D_VLV;
> need_stable_symbols = true;
> @@ -194,13 +194,13 @@ static int vlv_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv,
> return 0;
> }
>
> -static int i9xx_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv,
> +static int i9xx_pipe_crc_ctl_reg(struct drm_i915_private *i915,
> enum pipe pipe,
> enum intel_pipe_crc_source *source,
> u32 *val)
> {
> if (*source == INTEL_PIPE_CRC_SOURCE_AUTO) {
> - int ret = i9xx_pipe_crc_auto_source(dev_priv, pipe, source);
> + int ret = i9xx_pipe_crc_auto_source(i915, pipe, source);
> if (ret)
> return ret;
> }
> @@ -210,7 +210,7 @@ static int i9xx_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv,
> *val = PIPE_CRC_ENABLE | PIPE_CRC_SOURCE_PIPE_I9XX;
> break;
> case INTEL_PIPE_CRC_SOURCE_TV:
> - if (!SUPPORTS_TV(dev_priv))
> + if (!SUPPORTS_TV(i915))
> return -EINVAL;
> *val = PIPE_CRC_ENABLE | PIPE_CRC_SOURCE_TV_PRE;
> break;
> @@ -234,7 +234,7 @@ static int i9xx_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv,
> return 0;
> }
>
> -static void vlv_undo_pipe_scramble_reset(struct drm_i915_private *dev_priv,
> +static void vlv_undo_pipe_scramble_reset(struct drm_i915_private *i915,
> enum pipe pipe)
> {
> u32 tmp = I915_READ(PORT_DFT2_G4X);
> @@ -286,7 +286,7 @@ static int ilk_pipe_crc_ctl_reg(enum intel_pipe_crc_source *source,
> static void
> intel_crtc_crc_setup_workarounds(struct intel_crtc *crtc, bool enable)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> struct intel_crtc_state *pipe_config;
> struct drm_atomic_state *state;
> struct drm_modeset_acquire_ctx ctx;
> @@ -294,7 +294,7 @@ intel_crtc_crc_setup_workarounds(struct intel_crtc *crtc, bool enable)
>
> drm_modeset_acquire_init(&ctx, 0);
>
> - state = drm_atomic_state_alloc(&dev_priv->drm);
> + state = drm_atomic_state_alloc(&i915->drm);
> if (!state) {
> ret = -ENOMEM;
> goto unlock;
> @@ -312,7 +312,7 @@ intel_crtc_crc_setup_workarounds(struct intel_crtc *crtc, bool enable)
> pipe_config->base.mode_changed = pipe_config->has_psr;
> pipe_config->crc_enabled = enable;
>
> - if (IS_HASWELL(dev_priv) &&
> + if (IS_HASWELL(i915) &&
> pipe_config->base.active && crtc->pipe == PIPE_A &&
> pipe_config->cpu_transcoder == TRANSCODER_EDP)
> pipe_config->base.mode_changed = true;
> @@ -333,7 +333,7 @@ intel_crtc_crc_setup_workarounds(struct intel_crtc *crtc, bool enable)
> drm_modeset_acquire_fini(&ctx);
> }
>
> -static int ivb_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv,
> +static int ivb_pipe_crc_ctl_reg(struct drm_i915_private *i915,
> enum pipe pipe,
> enum intel_pipe_crc_source *source,
> u32 *val)
> @@ -361,7 +361,7 @@ static int ivb_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv,
> return 0;
> }
>
> -static int skl_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv,
> +static int skl_pipe_crc_ctl_reg(struct drm_i915_private *i915,
> enum pipe pipe,
> enum intel_pipe_crc_source *source,
> u32 *val)
> @@ -404,22 +404,22 @@ static int skl_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv,
> return 0;
> }
>
> -static int get_new_crc_ctl_reg(struct drm_i915_private *dev_priv,
> +static int get_new_crc_ctl_reg(struct drm_i915_private *i915,
> enum pipe pipe,
> enum intel_pipe_crc_source *source, u32 *val)
> {
> - if (IS_GEN(dev_priv, 2))
> + if (IS_GEN(i915, 2))
> return i8xx_pipe_crc_ctl_reg(source, val);
> - else if (INTEL_GEN(dev_priv) < 5)
> - return i9xx_pipe_crc_ctl_reg(dev_priv, pipe, source, val);
> - else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> - return vlv_pipe_crc_ctl_reg(dev_priv, pipe, source, val);
> - else if (IS_GEN_RANGE(dev_priv, 5, 6))
> + else if (INTEL_GEN(i915) < 5)
> + return i9xx_pipe_crc_ctl_reg(i915, pipe, source, val);
> + else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> + return vlv_pipe_crc_ctl_reg(i915, pipe, source, val);
> + else if (IS_GEN_RANGE(i915, 5, 6))
> return ilk_pipe_crc_ctl_reg(source, val);
> - else if (INTEL_GEN(dev_priv) < 9)
> - return ivb_pipe_crc_ctl_reg(dev_priv, pipe, source, val);
> + else if (INTEL_GEN(i915) < 9)
> + return ivb_pipe_crc_ctl_reg(i915, pipe, source, val);
> else
> - return skl_pipe_crc_ctl_reg(dev_priv, pipe, source, val);
> + return skl_pipe_crc_ctl_reg(i915, pipe, source, val);
> }
>
> static int
> @@ -440,18 +440,18 @@ display_crc_ctl_parse_source(const char *buf, enum intel_pipe_crc_source *s)
> return 0;
> }
>
> -void intel_display_crc_init(struct drm_i915_private *dev_priv)
> +void intel_display_crc_init(struct drm_i915_private *i915)
> {
> enum pipe pipe;
>
> - for_each_pipe(dev_priv, pipe) {
> - struct intel_pipe_crc *pipe_crc = &dev_priv->pipe_crc[pipe];
> + for_each_pipe(i915, pipe) {
> + struct intel_pipe_crc *pipe_crc = &i915->pipe_crc[pipe];
>
> spin_lock_init(&pipe_crc->lock);
> }
> }
>
> -static int i8xx_crc_source_valid(struct drm_i915_private *dev_priv,
> +static int i8xx_crc_source_valid(struct drm_i915_private *i915,
> const enum intel_pipe_crc_source source)
> {
> switch (source) {
> @@ -463,7 +463,7 @@ static int i8xx_crc_source_valid(struct drm_i915_private *dev_priv,
> }
> }
>
> -static int i9xx_crc_source_valid(struct drm_i915_private *dev_priv,
> +static int i9xx_crc_source_valid(struct drm_i915_private *i915,
> const enum intel_pipe_crc_source source)
> {
> switch (source) {
> @@ -476,7 +476,7 @@ static int i9xx_crc_source_valid(struct drm_i915_private *dev_priv,
> }
> }
>
> -static int vlv_crc_source_valid(struct drm_i915_private *dev_priv,
> +static int vlv_crc_source_valid(struct drm_i915_private *i915,
> const enum intel_pipe_crc_source source)
> {
> switch (source) {
> @@ -491,7 +491,7 @@ static int vlv_crc_source_valid(struct drm_i915_private *dev_priv,
> }
> }
>
> -static int ilk_crc_source_valid(struct drm_i915_private *dev_priv,
> +static int ilk_crc_source_valid(struct drm_i915_private *i915,
> const enum intel_pipe_crc_source source)
> {
> switch (source) {
> @@ -505,7 +505,7 @@ static int ilk_crc_source_valid(struct drm_i915_private *dev_priv,
> }
> }
>
> -static int ivb_crc_source_valid(struct drm_i915_private *dev_priv,
> +static int ivb_crc_source_valid(struct drm_i915_private *i915,
> const enum intel_pipe_crc_source source)
> {
> switch (source) {
> @@ -519,7 +519,7 @@ static int ivb_crc_source_valid(struct drm_i915_private *dev_priv,
> }
> }
>
> -static int skl_crc_source_valid(struct drm_i915_private *dev_priv,
> +static int skl_crc_source_valid(struct drm_i915_private *i915,
> const enum intel_pipe_crc_source source)
> {
> switch (source) {
> @@ -539,21 +539,21 @@ static int skl_crc_source_valid(struct drm_i915_private *dev_priv,
> }
>
> static int
> -intel_is_valid_crc_source(struct drm_i915_private *dev_priv,
> +intel_is_valid_crc_source(struct drm_i915_private *i915,
> const enum intel_pipe_crc_source source)
> {
> - if (IS_GEN(dev_priv, 2))
> - return i8xx_crc_source_valid(dev_priv, source);
> - else if (INTEL_GEN(dev_priv) < 5)
> - return i9xx_crc_source_valid(dev_priv, source);
> - else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> - return vlv_crc_source_valid(dev_priv, source);
> - else if (IS_GEN_RANGE(dev_priv, 5, 6))
> - return ilk_crc_source_valid(dev_priv, source);
> - else if (INTEL_GEN(dev_priv) < 9)
> - return ivb_crc_source_valid(dev_priv, source);
> + if (IS_GEN(i915, 2))
> + return i8xx_crc_source_valid(i915, source);
> + else if (INTEL_GEN(i915) < 5)
> + return i9xx_crc_source_valid(i915, source);
> + else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> + return vlv_crc_source_valid(i915, source);
> + else if (IS_GEN_RANGE(i915, 5, 6))
> + return ilk_crc_source_valid(i915, source);
> + else if (INTEL_GEN(i915) < 9)
> + return ivb_crc_source_valid(i915, source);
> else
> - return skl_crc_source_valid(dev_priv, source);
> + return skl_crc_source_valid(i915, source);
> }
>
> const char *const *intel_crtc_get_crc_sources(struct drm_crtc *crtc,
> @@ -566,7 +566,7 @@ const char *const *intel_crtc_get_crc_sources(struct drm_crtc *crtc,
> int intel_crtc_verify_crc_source(struct drm_crtc *crtc, const char *source_name,
> size_t *values_cnt)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc->dev);
> enum intel_pipe_crc_source source;
>
> if (display_crc_ctl_parse_source(source_name, &source) < 0) {
> @@ -575,7 +575,7 @@ int intel_crtc_verify_crc_source(struct drm_crtc *crtc, const char *source_name,
> }
>
> if (source == INTEL_PIPE_CRC_SOURCE_AUTO ||
> - intel_is_valid_crc_source(dev_priv, source) == 0) {
> + intel_is_valid_crc_source(i915, source) == 0) {
> *values_cnt = 5;
> return 0;
> }
> @@ -585,8 +585,8 @@ int intel_crtc_verify_crc_source(struct drm_crtc *crtc, const char *source_name,
>
> int intel_crtc_set_crc_source(struct drm_crtc *crtc, const char *source_name)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->dev);
> - struct intel_pipe_crc *pipe_crc = &dev_priv->pipe_crc[crtc->index];
> + struct drm_i915_private *i915 = to_i915(crtc->dev);
> + struct intel_pipe_crc *pipe_crc = &i915->pipe_crc[crtc->index];
> enum intel_display_power_domain power_domain;
> enum intel_pipe_crc_source source;
> intel_wakeref_t wakeref;
> @@ -600,7 +600,7 @@ int intel_crtc_set_crc_source(struct drm_crtc *crtc, const char *source_name)
> }
>
> power_domain = POWER_DOMAIN_PIPE(crtc->index);
> - wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain);
> + wakeref = intel_display_power_get_if_enabled(i915, power_domain);
> if (!wakeref) {
> DRM_DEBUG_KMS("Trying to capture CRC while pipe is off\n");
> return -EIO;
> @@ -610,7 +610,7 @@ int intel_crtc_set_crc_source(struct drm_crtc *crtc, const char *source_name)
> if (enable)
> intel_crtc_crc_setup_workarounds(to_intel_crtc(crtc), true);
>
> - ret = get_new_crc_ctl_reg(dev_priv, crtc->index, &source, &val);
> + ret = get_new_crc_ctl_reg(i915, crtc->index, &source, &val);
> if (ret != 0)
> goto out;
>
> @@ -619,8 +619,8 @@ int intel_crtc_set_crc_source(struct drm_crtc *crtc, const char *source_name)
> POSTING_READ(PIPE_CRC_CTL(crtc->index));
>
> if (!source) {
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> - vlv_undo_pipe_scramble_reset(dev_priv, crtc->index);
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> + vlv_undo_pipe_scramble_reset(i915, crtc->index);
> }
>
> pipe_crc->skipped = 0;
> @@ -629,7 +629,7 @@ int intel_crtc_set_crc_source(struct drm_crtc *crtc, const char *source_name)
> if (!enable)
> intel_crtc_crc_setup_workarounds(to_intel_crtc(crtc), false);
>
> - intel_display_power_put(dev_priv, power_domain, wakeref);
> + intel_display_power_put(i915, power_domain, wakeref);
>
> return ret;
> }
> @@ -637,14 +637,14 @@ int intel_crtc_set_crc_source(struct drm_crtc *crtc, const char *source_name)
> void intel_crtc_enable_pipe_crc(struct intel_crtc *intel_crtc)
> {
> struct drm_crtc *crtc = &intel_crtc->base;
> - struct drm_i915_private *dev_priv = to_i915(crtc->dev);
> - struct intel_pipe_crc *pipe_crc = &dev_priv->pipe_crc[crtc->index];
> + struct drm_i915_private *i915 = to_i915(crtc->dev);
> + struct intel_pipe_crc *pipe_crc = &i915->pipe_crc[crtc->index];
> u32 val = 0;
>
> if (!crtc->crc.opened)
> return;
>
> - if (get_new_crc_ctl_reg(dev_priv, crtc->index, &pipe_crc->source, &val) < 0)
> + if (get_new_crc_ctl_reg(i915, crtc->index, &pipe_crc->source, &val) < 0)
> return;
>
> /* Don't need pipe_crc->lock here, IRQs are not generated. */
> @@ -657,8 +657,8 @@ void intel_crtc_enable_pipe_crc(struct intel_crtc *intel_crtc)
> void intel_crtc_disable_pipe_crc(struct intel_crtc *intel_crtc)
> {
> struct drm_crtc *crtc = &intel_crtc->base;
> - struct drm_i915_private *dev_priv = to_i915(crtc->dev);
> - struct intel_pipe_crc *pipe_crc = &dev_priv->pipe_crc[crtc->index];
> + struct drm_i915_private *i915 = to_i915(crtc->dev);
> + struct intel_pipe_crc *pipe_crc = &i915->pipe_crc[crtc->index];
>
> /* Swallow crc's until we stop generating them. */
> spin_lock_irq(&pipe_crc->lock);
> @@ -667,5 +667,5 @@ void intel_crtc_disable_pipe_crc(struct intel_crtc *intel_crtc)
>
> I915_WRITE(PIPE_CRC_CTL(crtc->index), 0);
> POSTING_READ(PIPE_CRC_CTL(crtc->index));
> - synchronize_irq(dev_priv->drm.irq);
> + synchronize_irq(i915->drm.irq);
> }
> diff --git a/drivers/gpu/drm/i915/intel_pipe_crc.h b/drivers/gpu/drm/i915/intel_pipe_crc.h
> index db258a756fc6..42048ab1391a 100644
> --- a/drivers/gpu/drm/i915/intel_pipe_crc.h
> +++ b/drivers/gpu/drm/i915/intel_pipe_crc.h
> @@ -13,7 +13,7 @@ struct drm_i915_private;
> struct intel_crtc;
>
> #ifdef CONFIG_DEBUG_FS
> -void intel_display_crc_init(struct drm_i915_private *dev_priv);
> +void intel_display_crc_init(struct drm_i915_private *i915);
> int intel_crtc_set_crc_source(struct drm_crtc *crtc, const char *source_name);
> int intel_crtc_verify_crc_source(struct drm_crtc *crtc,
> const char *source_name, size_t *values_cnt);
> @@ -22,7 +22,7 @@ const char *const *intel_crtc_get_crc_sources(struct drm_crtc *crtc,
> void intel_crtc_disable_pipe_crc(struct intel_crtc *crtc);
> void intel_crtc_enable_pipe_crc(struct intel_crtc *crtc);
> #else
> -static inline void intel_display_crc_init(struct drm_i915_private *dev_priv) {}
> +static inline void intel_display_crc_init(struct drm_i915_private *i915) {}
> #define intel_crtc_set_crc_source NULL
> #define intel_crtc_verify_crc_source NULL
> #define intel_crtc_get_crc_sources NULL
> diff --git a/drivers/gpu/drm/i915/intel_pm.c b/drivers/gpu/drm/i915/intel_pm.c
> index 2c7f3ebc0117..4b939160cd95 100644
> --- a/drivers/gpu/drm/i915/intel_pm.c
> +++ b/drivers/gpu/drm/i915/intel_pm.c
> @@ -63,9 +63,9 @@
> * require higher latency to switch to and wake up.
> */
>
> -static void gen9_init_clock_gating(struct drm_i915_private *dev_priv)
> +static void gen9_init_clock_gating(struct drm_i915_private *i915)
> {
> - if (HAS_LLC(dev_priv)) {
> + if (HAS_LLC(i915)) {
> /*
> * WaCompressedResourceDisplayNewHashMode:skl,kbl
> * Display WA #0390: skl,kbl
> @@ -96,16 +96,16 @@ static void gen9_init_clock_gating(struct drm_i915_private *dev_priv)
> I915_WRITE(ILK_DPFC_CHICKEN, I915_READ(ILK_DPFC_CHICKEN) |
> ILK_DPFC_DISABLE_DUMMY0);
>
> - if (IS_SKYLAKE(dev_priv)) {
> + if (IS_SKYLAKE(i915)) {
> /* WaDisableDopClockGating */
> I915_WRITE(GEN7_MISCCPCTL, I915_READ(GEN7_MISCCPCTL)
> & ~GEN7_DOP_CLOCK_GATE_ENABLE);
> }
> }
>
> -static void bxt_init_clock_gating(struct drm_i915_private *dev_priv)
> +static void bxt_init_clock_gating(struct drm_i915_private *i915)
> {
> - gen9_init_clock_gating(dev_priv);
> + gen9_init_clock_gating(i915);
>
> /* WaDisableSDEUnitClockGating:bxt */
> I915_WRITE(GEN8_UCGCTL6, I915_READ(GEN8_UCGCTL6) |
> @@ -126,9 +126,9 @@ static void bxt_init_clock_gating(struct drm_i915_private *dev_priv)
> PWM1_GATING_DIS | PWM2_GATING_DIS);
> }
>
> -static void glk_init_clock_gating(struct drm_i915_private *dev_priv)
> +static void glk_init_clock_gating(struct drm_i915_private *i915)
> {
> - gen9_init_clock_gating(dev_priv);
> + gen9_init_clock_gating(i915);
>
> /*
> * WaDisablePWMClockGating:glk
> @@ -139,7 +139,7 @@ static void glk_init_clock_gating(struct drm_i915_private *dev_priv)
> PWM1_GATING_DIS | PWM2_GATING_DIS);
>
> /* WaDDIIOTimeout:glk */
> - if (IS_GLK_REVID(dev_priv, 0, GLK_REVID_A1)) {
> + if (IS_GLK_REVID(i915, 0, GLK_REVID_A1)) {
> u32 val = I915_READ(CHICKEN_MISC_2);
> val &= ~(GLK_CL0_PWR_DOWN |
> GLK_CL1_PWR_DOWN |
> @@ -149,7 +149,7 @@ static void glk_init_clock_gating(struct drm_i915_private *dev_priv)
>
> }
>
> -static void i915_pineview_get_mem_freq(struct drm_i915_private *dev_priv)
> +static void i915_pineview_get_mem_freq(struct drm_i915_private *i915)
> {
> u32 tmp;
>
> @@ -157,100 +157,100 @@ static void i915_pineview_get_mem_freq(struct drm_i915_private *dev_priv)
>
> switch (tmp & CLKCFG_FSB_MASK) {
> case CLKCFG_FSB_533:
> - dev_priv->fsb_freq = 533; /* 133*4 */
> + i915->fsb_freq = 533; /* 133*4 */
> break;
> case CLKCFG_FSB_800:
> - dev_priv->fsb_freq = 800; /* 200*4 */
> + i915->fsb_freq = 800; /* 200*4 */
> break;
> case CLKCFG_FSB_667:
> - dev_priv->fsb_freq = 667; /* 167*4 */
> + i915->fsb_freq = 667; /* 167*4 */
> break;
> case CLKCFG_FSB_400:
> - dev_priv->fsb_freq = 400; /* 100*4 */
> + i915->fsb_freq = 400; /* 100*4 */
> break;
> }
>
> switch (tmp & CLKCFG_MEM_MASK) {
> case CLKCFG_MEM_533:
> - dev_priv->mem_freq = 533;
> + i915->mem_freq = 533;
> break;
> case CLKCFG_MEM_667:
> - dev_priv->mem_freq = 667;
> + i915->mem_freq = 667;
> break;
> case CLKCFG_MEM_800:
> - dev_priv->mem_freq = 800;
> + i915->mem_freq = 800;
> break;
> }
>
> /* detect pineview DDR3 setting */
> tmp = I915_READ(CSHRDDR3CTL);
> - dev_priv->is_ddr3 = (tmp & CSHRDDR3CTL_DDR3) ? 1 : 0;
> + i915->is_ddr3 = (tmp & CSHRDDR3CTL_DDR3) ? 1 : 0;
> }
>
> -static void i915_ironlake_get_mem_freq(struct drm_i915_private *dev_priv)
> +static void i915_ironlake_get_mem_freq(struct drm_i915_private *i915)
> {
> u16 ddrpll, csipll;
>
> - ddrpll = intel_uncore_read16(&dev_priv->uncore, DDRMPLL1);
> - csipll = intel_uncore_read16(&dev_priv->uncore, CSIPLL0);
> + ddrpll = intel_uncore_read16(&i915->uncore, DDRMPLL1);
> + csipll = intel_uncore_read16(&i915->uncore, CSIPLL0);
>
> switch (ddrpll & 0xff) {
> case 0xc:
> - dev_priv->mem_freq = 800;
> + i915->mem_freq = 800;
> break;
> case 0x10:
> - dev_priv->mem_freq = 1066;
> + i915->mem_freq = 1066;
> break;
> case 0x14:
> - dev_priv->mem_freq = 1333;
> + i915->mem_freq = 1333;
> break;
> case 0x18:
> - dev_priv->mem_freq = 1600;
> + i915->mem_freq = 1600;
> break;
> default:
> DRM_DEBUG_DRIVER("unknown memory frequency 0x%02x\n",
> ddrpll & 0xff);
> - dev_priv->mem_freq = 0;
> + i915->mem_freq = 0;
> break;
> }
>
> - dev_priv->ips.r_t = dev_priv->mem_freq;
> + i915->ips.r_t = i915->mem_freq;
>
> switch (csipll & 0x3ff) {
> case 0x00c:
> - dev_priv->fsb_freq = 3200;
> + i915->fsb_freq = 3200;
> break;
> case 0x00e:
> - dev_priv->fsb_freq = 3733;
> + i915->fsb_freq = 3733;
> break;
> case 0x010:
> - dev_priv->fsb_freq = 4266;
> + i915->fsb_freq = 4266;
> break;
> case 0x012:
> - dev_priv->fsb_freq = 4800;
> + i915->fsb_freq = 4800;
> break;
> case 0x014:
> - dev_priv->fsb_freq = 5333;
> + i915->fsb_freq = 5333;
> break;
> case 0x016:
> - dev_priv->fsb_freq = 5866;
> + i915->fsb_freq = 5866;
> break;
> case 0x018:
> - dev_priv->fsb_freq = 6400;
> + i915->fsb_freq = 6400;
> break;
> default:
> DRM_DEBUG_DRIVER("unknown fsb frequency 0x%04x\n",
> csipll & 0x3ff);
> - dev_priv->fsb_freq = 0;
> + i915->fsb_freq = 0;
> break;
> }
>
> - if (dev_priv->fsb_freq == 3200) {
> - dev_priv->ips.c_m = 0;
> - } else if (dev_priv->fsb_freq > 3200 && dev_priv->fsb_freq <= 4800) {
> - dev_priv->ips.c_m = 1;
> + if (i915->fsb_freq == 3200) {
> + i915->ips.c_m = 0;
> + } else if (i915->fsb_freq > 3200 && i915->fsb_freq <= 4800) {
> + i915->ips.c_m = 1;
> } else {
> - dev_priv->ips.c_m = 2;
> + i915->ips.c_m = 2;
> }
> }
>
> @@ -316,61 +316,61 @@ static const struct cxsr_latency *intel_get_cxsr_latency(bool is_desktop,
> return NULL;
> }
>
> -static void chv_set_memory_dvfs(struct drm_i915_private *dev_priv, bool enable)
> +static void chv_set_memory_dvfs(struct drm_i915_private *i915, bool enable)
> {
> u32 val;
>
> - vlv_punit_get(dev_priv);
> + vlv_punit_get(i915);
>
> - val = vlv_punit_read(dev_priv, PUNIT_REG_DDR_SETUP2);
> + val = vlv_punit_read(i915, PUNIT_REG_DDR_SETUP2);
> if (enable)
> val &= ~FORCE_DDR_HIGH_FREQ;
> else
> val |= FORCE_DDR_HIGH_FREQ;
> val &= ~FORCE_DDR_LOW_FREQ;
> val |= FORCE_DDR_FREQ_REQ_ACK;
> - vlv_punit_write(dev_priv, PUNIT_REG_DDR_SETUP2, val);
> + vlv_punit_write(i915, PUNIT_REG_DDR_SETUP2, val);
>
> - if (wait_for((vlv_punit_read(dev_priv, PUNIT_REG_DDR_SETUP2) &
> + if (wait_for((vlv_punit_read(i915, PUNIT_REG_DDR_SETUP2) &
> FORCE_DDR_FREQ_REQ_ACK) == 0, 3))
> DRM_ERROR("timed out waiting for Punit DDR DVFS request\n");
>
> - vlv_punit_put(dev_priv);
> + vlv_punit_put(i915);
> }
>
> -static void chv_set_memory_pm5(struct drm_i915_private *dev_priv, bool enable)
> +static void chv_set_memory_pm5(struct drm_i915_private *i915, bool enable)
> {
> u32 val;
>
> - vlv_punit_get(dev_priv);
> + vlv_punit_get(i915);
>
> - val = vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM);
> + val = vlv_punit_read(i915, PUNIT_REG_DSPSSPM);
> if (enable)
> val |= DSP_MAXFIFO_PM5_ENABLE;
> else
> val &= ~DSP_MAXFIFO_PM5_ENABLE;
> - vlv_punit_write(dev_priv, PUNIT_REG_DSPSSPM, val);
> + vlv_punit_write(i915, PUNIT_REG_DSPSSPM, val);
>
> - vlv_punit_put(dev_priv);
> + vlv_punit_put(i915);
> }
>
> #define FW_WM(value, plane) \
> (((value) << DSPFW_ ## plane ## _SHIFT) & DSPFW_ ## plane ## _MASK)
>
> -static bool _intel_set_memory_cxsr(struct drm_i915_private *dev_priv, bool enable)
> +static bool _intel_set_memory_cxsr(struct drm_i915_private *i915, bool enable)
> {
> bool was_enabled;
> u32 val;
>
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> was_enabled = I915_READ(FW_BLC_SELF_VLV) & FW_CSPWRDWNEN;
> I915_WRITE(FW_BLC_SELF_VLV, enable ? FW_CSPWRDWNEN : 0);
> POSTING_READ(FW_BLC_SELF_VLV);
> - } else if (IS_G4X(dev_priv) || IS_I965GM(dev_priv)) {
> + } else if (IS_G4X(i915) || IS_I965GM(i915)) {
> was_enabled = I915_READ(FW_BLC_SELF) & FW_BLC_SELF_EN;
> I915_WRITE(FW_BLC_SELF, enable ? FW_BLC_SELF_EN : 0);
> POSTING_READ(FW_BLC_SELF);
> - } else if (IS_PINEVIEW(dev_priv)) {
> + } else if (IS_PINEVIEW(i915)) {
> val = I915_READ(DSPFW3);
> was_enabled = val & PINEVIEW_SELF_REFRESH_EN;
> if (enable)
> @@ -379,13 +379,13 @@ static bool _intel_set_memory_cxsr(struct drm_i915_private *dev_priv, bool enabl
> val &= ~PINEVIEW_SELF_REFRESH_EN;
> I915_WRITE(DSPFW3, val);
> POSTING_READ(DSPFW3);
> - } else if (IS_I945G(dev_priv) || IS_I945GM(dev_priv)) {
> + } else if (IS_I945G(i915) || IS_I945GM(i915)) {
> was_enabled = I915_READ(FW_BLC_SELF) & FW_BLC_SELF_EN;
> val = enable ? _MASKED_BIT_ENABLE(FW_BLC_SELF_EN) :
> _MASKED_BIT_DISABLE(FW_BLC_SELF_EN);
> I915_WRITE(FW_BLC_SELF, val);
> POSTING_READ(FW_BLC_SELF);
> - } else if (IS_I915GM(dev_priv)) {
> + } else if (IS_I915GM(i915)) {
> /*
> * FIXME can't find a bit like this for 915G, and
> * and yet it does have the related watermark in
> @@ -400,7 +400,7 @@ static bool _intel_set_memory_cxsr(struct drm_i915_private *dev_priv, bool enabl
> return false;
> }
>
> - trace_intel_memory_cxsr(dev_priv, was_enabled, enable);
> + trace_intel_memory_cxsr(i915, was_enabled, enable);
>
> DRM_DEBUG_KMS("memory self-refresh is %s (was %s)\n",
> enableddisabled(enable),
> @@ -411,7 +411,7 @@ static bool _intel_set_memory_cxsr(struct drm_i915_private *dev_priv, bool enabl
>
> /**
> * intel_set_memory_cxsr - Configure CxSR state
> - * @dev_priv: i915 device
> + * @i915: i915 device
> * @enable: Allow vs. disallow CxSR
> *
> * Allow or disallow the system to enter a special CxSR
> @@ -446,17 +446,17 @@ static bool _intel_set_memory_cxsr(struct drm_i915_private *dev_priv, bool enabl
> * the hardware w.r.t. HPLL SR when writing to plane registers.
> * Disallowing just CxSR is sufficient.
> */
> -bool intel_set_memory_cxsr(struct drm_i915_private *dev_priv, bool enable)
> +bool intel_set_memory_cxsr(struct drm_i915_private *i915, bool enable)
> {
> bool ret;
>
> - mutex_lock(&dev_priv->wm.wm_mutex);
> - ret = _intel_set_memory_cxsr(dev_priv, enable);
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> - dev_priv->wm.vlv.cxsr = enable;
> - else if (IS_G4X(dev_priv))
> - dev_priv->wm.g4x.cxsr = enable;
> - mutex_unlock(&dev_priv->wm.wm_mutex);
> + mutex_lock(&i915->wm.wm_mutex);
> + ret = _intel_set_memory_cxsr(i915, enable);
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> + i915->wm.vlv.cxsr = enable;
> + else if (IS_G4X(i915))
> + i915->wm.g4x.cxsr = enable;
> + mutex_unlock(&i915->wm.wm_mutex);
>
> return ret;
> }
> @@ -483,7 +483,7 @@ static const int pessimal_latency_ns = 5000;
> static void vlv_get_fifo_size(struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> struct vlv_fifo_state *fifo_state = &crtc_state->wm.vlv.fifo_state;
> enum pipe pipe = crtc->pipe;
> int sprite0_start, sprite1_start;
> @@ -519,7 +519,7 @@ static void vlv_get_fifo_size(struct intel_crtc_state *crtc_state)
> fifo_state->plane[PLANE_CURSOR] = 63;
> }
>
> -static int i9xx_get_fifo_size(struct drm_i915_private *dev_priv,
> +static int i9xx_get_fifo_size(struct drm_i915_private *i915,
> enum i9xx_plane_id i9xx_plane)
> {
> u32 dsparb = I915_READ(DSPARB);
> @@ -535,7 +535,7 @@ static int i9xx_get_fifo_size(struct drm_i915_private *dev_priv,
> return size;
> }
>
> -static int i830_get_fifo_size(struct drm_i915_private *dev_priv,
> +static int i830_get_fifo_size(struct drm_i915_private *i915,
> enum i9xx_plane_id i9xx_plane)
> {
> u32 dsparb = I915_READ(DSPARB);
> @@ -552,7 +552,7 @@ static int i830_get_fifo_size(struct drm_i915_private *dev_priv,
> return size;
> }
>
> -static int i845_get_fifo_size(struct drm_i915_private *dev_priv,
> +static int i845_get_fifo_size(struct drm_i915_private *i915,
> enum i9xx_plane_id i9xx_plane)
> {
> u32 dsparb = I915_READ(DSPARB);
> @@ -805,9 +805,9 @@ static bool is_enabling(int old, int new, int threshold)
> return old < threshold && new >= threshold;
> }
>
> -static int intel_wm_num_levels(struct drm_i915_private *dev_priv)
> +static int intel_wm_num_levels(struct drm_i915_private *i915)
> {
> - return dev_priv->wm.max_level + 1;
> + return i915->wm.max_level + 1;
> }
>
> static bool intel_wm_plane_visible(const struct intel_crtc_state *crtc_state,
> @@ -833,11 +833,11 @@ static bool intel_wm_plane_visible(const struct intel_crtc_state *crtc_state,
> return plane_state->base.visible;
> }
>
> -static struct intel_crtc *single_enabled_crtc(struct drm_i915_private *dev_priv)
> +static struct intel_crtc *single_enabled_crtc(struct drm_i915_private *i915)
> {
> struct intel_crtc *crtc, *enabled = NULL;
>
> - for_each_intel_crtc(&dev_priv->drm, crtc) {
> + for_each_intel_crtc(&i915->drm, crtc) {
> if (intel_crtc_active(crtc)) {
> if (enabled)
> return NULL;
> @@ -850,23 +850,23 @@ static struct intel_crtc *single_enabled_crtc(struct drm_i915_private *dev_priv)
>
> static void pineview_update_wm(struct intel_crtc *unused_crtc)
> {
> - struct drm_i915_private *dev_priv = to_i915(unused_crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(unused_crtc->base.dev);
> struct intel_crtc *crtc;
> const struct cxsr_latency *latency;
> u32 reg;
> unsigned int wm;
>
> - latency = intel_get_cxsr_latency(!IS_MOBILE(dev_priv),
> - dev_priv->is_ddr3,
> - dev_priv->fsb_freq,
> - dev_priv->mem_freq);
> + latency = intel_get_cxsr_latency(!IS_MOBILE(i915),
> + i915->is_ddr3,
> + i915->fsb_freq,
> + i915->mem_freq);
> if (!latency) {
> DRM_DEBUG_KMS("Unknown FSB/MEM found, disable CxSR\n");
> - intel_set_memory_cxsr(dev_priv, false);
> + intel_set_memory_cxsr(i915, false);
> return;
> }
>
> - crtc = single_enabled_crtc(dev_priv);
> + crtc = single_enabled_crtc(i915);
> if (crtc) {
> const struct drm_display_mode *adjusted_mode =
> &crtc->config->base.adjusted_mode;
> @@ -913,9 +913,9 @@ static void pineview_update_wm(struct intel_crtc *unused_crtc)
> I915_WRITE(DSPFW3, reg);
> DRM_DEBUG_KMS("DSPFW3 register is %x\n", reg);
>
> - intel_set_memory_cxsr(dev_priv, true);
> + intel_set_memory_cxsr(i915, true);
> } else {
> - intel_set_memory_cxsr(dev_priv, false);
> + intel_set_memory_cxsr(i915, false);
> }
> }
>
> @@ -936,13 +936,13 @@ static unsigned int g4x_tlb_miss_wa(int fifo_size, int width, int cpp)
> return max(0, tlb_miss);
> }
>
> -static void g4x_write_wm_values(struct drm_i915_private *dev_priv,
> +static void g4x_write_wm_values(struct drm_i915_private *i915,
> const struct g4x_wm_values *wm)
> {
> enum pipe pipe;
>
> - for_each_pipe(dev_priv, pipe)
> - trace_g4x_wm(intel_get_crtc_for_pipe(dev_priv, pipe), wm);
> + for_each_pipe(i915, pipe)
> + trace_g4x_wm(intel_get_crtc_for_pipe(i915, pipe), wm);
>
> I915_WRITE(DSPFW1,
> FW_WM(wm->sr.plane, SR) |
> @@ -968,13 +968,13 @@ static void g4x_write_wm_values(struct drm_i915_private *dev_priv,
> #define FW_WM_VLV(value, plane) \
> (((value) << DSPFW_ ## plane ## _SHIFT) & DSPFW_ ## plane ## _MASK_VLV)
>
> -static void vlv_write_wm_values(struct drm_i915_private *dev_priv,
> +static void vlv_write_wm_values(struct drm_i915_private *i915,
> const struct vlv_wm_values *wm)
> {
> enum pipe pipe;
>
> - for_each_pipe(dev_priv, pipe) {
> - trace_vlv_wm(intel_get_crtc_for_pipe(dev_priv, pipe), wm);
> + for_each_pipe(i915, pipe) {
> + trace_vlv_wm(intel_get_crtc_for_pipe(i915, pipe), wm);
>
> I915_WRITE(VLV_DDL(pipe),
> (wm->ddl[pipe].plane[PLANE_CURSOR] << DDL_CURSOR_SHIFT) |
> @@ -1006,7 +1006,7 @@ static void vlv_write_wm_values(struct drm_i915_private *dev_priv,
> I915_WRITE(DSPFW3,
> FW_WM(wm->sr.cursor, CURSOR_SR));
>
> - if (IS_CHERRYVIEW(dev_priv)) {
> + if (IS_CHERRYVIEW(i915)) {
> I915_WRITE(DSPFW7_CHV,
> FW_WM_VLV(wm->pipe[PIPE_B].plane[PLANE_SPRITE1], SPRITED) |
> FW_WM_VLV(wm->pipe[PIPE_B].plane[PLANE_SPRITE0], SPRITEC));
> @@ -1046,14 +1046,14 @@ static void vlv_write_wm_values(struct drm_i915_private *dev_priv,
>
> #undef FW_WM_VLV
>
> -static void g4x_setup_wm_latency(struct drm_i915_private *dev_priv)
> +static void g4x_setup_wm_latency(struct drm_i915_private *i915)
> {
> /* all latencies in usec */
> - dev_priv->wm.pri_latency[G4X_WM_LEVEL_NORMAL] = 5;
> - dev_priv->wm.pri_latency[G4X_WM_LEVEL_SR] = 12;
> - dev_priv->wm.pri_latency[G4X_WM_LEVEL_HPLL] = 35;
> + i915->wm.pri_latency[G4X_WM_LEVEL_NORMAL] = 5;
> + i915->wm.pri_latency[G4X_WM_LEVEL_SR] = 12;
> + i915->wm.pri_latency[G4X_WM_LEVEL_HPLL] = 35;
>
> - dev_priv->wm.max_level = G4X_WM_LEVEL_HPLL;
> + i915->wm.max_level = G4X_WM_LEVEL_HPLL;
> }
>
> static int g4x_plane_fifo_size(enum plane_id plane_id, int level)
> @@ -1103,10 +1103,10 @@ static u16 g4x_compute_wm(const struct intel_crtc_state *crtc_state,
> int level)
> {
> struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> const struct drm_display_mode *adjusted_mode =
> &crtc_state->base.adjusted_mode;
> - unsigned int latency = dev_priv->wm.pri_latency[level] * 10;
> + unsigned int latency = i915->wm.pri_latency[level] * 10;
> unsigned int clock, htotal, cpp, width, wm;
>
> if (latency == 0)
> @@ -1126,7 +1126,7 @@ static u16 g4x_compute_wm(const struct intel_crtc_state *crtc_state,
> * the HPLL watermark, which seems a little strange.
> * Let's use 32bpp for the HPLL watermark as well.
> */
> - if (IS_GM45(dev_priv) && plane->id == PLANE_PRIMARY &&
> + if (IS_GM45(i915) && plane->id == PLANE_PRIMARY &&
> level != G4X_WM_LEVEL_NORMAL)
> cpp = 4;
> else
> @@ -1165,10 +1165,10 @@ static u16 g4x_compute_wm(const struct intel_crtc_state *crtc_state,
> static bool g4x_raw_plane_wm_set(struct intel_crtc_state *crtc_state,
> int level, enum plane_id plane_id, u16 value)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> bool dirty = false;
>
> - for (; level < intel_wm_num_levels(dev_priv); level++) {
> + for (; level < intel_wm_num_levels(i915); level++) {
> struct g4x_pipe_wm *raw = &crtc_state->wm.g4x.raw[level];
>
> dirty |= raw->plane[plane_id] != value;
> @@ -1181,13 +1181,13 @@ static bool g4x_raw_plane_wm_set(struct intel_crtc_state *crtc_state,
> static bool g4x_raw_fbc_wm_set(struct intel_crtc_state *crtc_state,
> int level, u16 value)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> bool dirty = false;
>
> /* NORMAL level doesn't have an FBC watermark */
> level = max(level, G4X_WM_LEVEL_SR);
>
> - for (; level < intel_wm_num_levels(dev_priv); level++) {
> + for (; level < intel_wm_num_levels(i915); level++) {
> struct g4x_pipe_wm *raw = &crtc_state->wm.g4x.raw[level];
>
> dirty |= raw->fbc != value;
> @@ -1283,9 +1283,9 @@ static bool g4x_raw_plane_wm_is_valid(const struct intel_crtc_state *crtc_state,
> static bool g4x_raw_crtc_wm_is_valid(const struct intel_crtc_state *crtc_state,
> int level)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
>
> - if (level > dev_priv->wm.max_level)
> + if (level > i915->wm.max_level)
> return false;
>
> return g4x_raw_plane_wm_is_valid(crtc_state, PLANE_PRIMARY, level) &&
> @@ -1484,7 +1484,7 @@ static int g4x_compute_intermediate_wm(struct intel_crtc_state *new_crtc_state)
> return 0;
> }
>
> -static void g4x_merge_wm(struct drm_i915_private *dev_priv,
> +static void g4x_merge_wm(struct drm_i915_private *i915,
> struct g4x_wm_values *wm)
> {
> struct intel_crtc *crtc;
> @@ -1494,7 +1494,7 @@ static void g4x_merge_wm(struct drm_i915_private *dev_priv,
> wm->hpll_en = true;
> wm->fbc_en = true;
>
> - for_each_intel_crtc(&dev_priv->drm, crtc) {
> + for_each_intel_crtc(&i915->drm, crtc) {
> const struct g4x_wm_state *wm_state = &crtc->wm.active.g4x;
>
> if (!crtc->active)
> @@ -1516,7 +1516,7 @@ static void g4x_merge_wm(struct drm_i915_private *dev_priv,
> wm->fbc_en = false;
> }
>
> - for_each_intel_crtc(&dev_priv->drm, crtc) {
> + for_each_intel_crtc(&i915->drm, crtc) {
> const struct g4x_wm_state *wm_state = &crtc->wm.active.g4x;
> enum pipe pipe = crtc->pipe;
>
> @@ -1528,23 +1528,23 @@ static void g4x_merge_wm(struct drm_i915_private *dev_priv,
> }
> }
>
> -static void g4x_program_watermarks(struct drm_i915_private *dev_priv)
> +static void g4x_program_watermarks(struct drm_i915_private *i915)
> {
> - struct g4x_wm_values *old_wm = &dev_priv->wm.g4x;
> + struct g4x_wm_values *old_wm = &i915->wm.g4x;
> struct g4x_wm_values new_wm = {};
>
> - g4x_merge_wm(dev_priv, &new_wm);
> + g4x_merge_wm(i915, &new_wm);
>
> if (memcmp(old_wm, &new_wm, sizeof(new_wm)) == 0)
> return;
>
> if (is_disabling(old_wm->cxsr, new_wm.cxsr, true))
> - _intel_set_memory_cxsr(dev_priv, false);
> + _intel_set_memory_cxsr(i915, false);
>
> - g4x_write_wm_values(dev_priv, &new_wm);
> + g4x_write_wm_values(i915, &new_wm);
>
> if (is_enabling(old_wm->cxsr, new_wm.cxsr, true))
> - _intel_set_memory_cxsr(dev_priv, true);
> + _intel_set_memory_cxsr(i915, true);
>
> *old_wm = new_wm;
> }
> @@ -1552,28 +1552,28 @@ static void g4x_program_watermarks(struct drm_i915_private *dev_priv)
> static void g4x_initial_watermarks(struct intel_atomic_state *state,
> struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
>
> - mutex_lock(&dev_priv->wm.wm_mutex);
> + mutex_lock(&i915->wm.wm_mutex);
> crtc->wm.active.g4x = crtc_state->wm.g4x.intermediate;
> - g4x_program_watermarks(dev_priv);
> - mutex_unlock(&dev_priv->wm.wm_mutex);
> + g4x_program_watermarks(i915);
> + mutex_unlock(&i915->wm.wm_mutex);
> }
>
> static void g4x_optimize_watermarks(struct intel_atomic_state *state,
> struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc);
>
> if (!crtc_state->wm.need_postvbl_update)
> return;
>
> - mutex_lock(&dev_priv->wm.wm_mutex);
> + mutex_lock(&i915->wm.wm_mutex);
> intel_crtc->wm.active.g4x = crtc_state->wm.g4x.optimal;
> - g4x_program_watermarks(dev_priv);
> - mutex_unlock(&dev_priv->wm.wm_mutex);
> + g4x_program_watermarks(i915);
> + mutex_unlock(&i915->wm.wm_mutex);
> }
>
> /* latency must be in 0.1us units. */
> @@ -1592,18 +1592,18 @@ static unsigned int vlv_wm_method2(unsigned int pixel_rate,
> return ret;
> }
>
> -static void vlv_setup_wm_latency(struct drm_i915_private *dev_priv)
> +static void vlv_setup_wm_latency(struct drm_i915_private *i915)
> {
> /* all latencies in usec */
> - dev_priv->wm.pri_latency[VLV_WM_LEVEL_PM2] = 3;
> + i915->wm.pri_latency[VLV_WM_LEVEL_PM2] = 3;
>
> - dev_priv->wm.max_level = VLV_WM_LEVEL_PM2;
> + i915->wm.max_level = VLV_WM_LEVEL_PM2;
>
> - if (IS_CHERRYVIEW(dev_priv)) {
> - dev_priv->wm.pri_latency[VLV_WM_LEVEL_PM5] = 12;
> - dev_priv->wm.pri_latency[VLV_WM_LEVEL_DDR_DVFS] = 33;
> + if (IS_CHERRYVIEW(i915)) {
> + i915->wm.pri_latency[VLV_WM_LEVEL_PM5] = 12;
> + i915->wm.pri_latency[VLV_WM_LEVEL_DDR_DVFS] = 33;
>
> - dev_priv->wm.max_level = VLV_WM_LEVEL_DDR_DVFS;
> + i915->wm.max_level = VLV_WM_LEVEL_DDR_DVFS;
> }
> }
>
> @@ -1612,12 +1612,12 @@ static u16 vlv_compute_wm_level(const struct intel_crtc_state *crtc_state,
> int level)
> {
> struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> const struct drm_display_mode *adjusted_mode =
> &crtc_state->base.adjusted_mode;
> unsigned int clock, htotal, cpp, width, wm;
>
> - if (dev_priv->wm.pri_latency[level] == 0)
> + if (i915->wm.pri_latency[level] == 0)
> return USHRT_MAX;
>
> if (!intel_wm_plane_visible(crtc_state, plane_state))
> @@ -1638,7 +1638,7 @@ static u16 vlv_compute_wm_level(const struct intel_crtc_state *crtc_state,
> wm = 63;
> } else {
> wm = vlv_wm_method2(clock, htotal, width, cpp,
> - dev_priv->wm.pri_latency[level] * 10);
> + i915->wm.pri_latency[level] * 10);
> }
>
> return min_t(unsigned int, wm, USHRT_MAX);
> @@ -1736,9 +1736,9 @@ static int vlv_compute_fifo(struct intel_crtc_state *crtc_state)
> static void vlv_invalidate_wms(struct intel_crtc *crtc,
> struct vlv_wm_state *wm_state, int level)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
>
> - for (; level < intel_wm_num_levels(dev_priv); level++) {
> + for (; level < intel_wm_num_levels(i915); level++) {
> enum plane_id plane_id;
>
> for_each_plane_id_on_crtc(crtc, plane_id)
> @@ -1764,8 +1764,8 @@ static u16 vlv_invert_wm_value(u16 wm, u16 fifo_size)
> static bool vlv_raw_plane_wm_set(struct intel_crtc_state *crtc_state,
> int level, enum plane_id plane_id, u16 value)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> - int num_levels = intel_wm_num_levels(dev_priv);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> + int num_levels = intel_wm_num_levels(i915);
> bool dirty = false;
>
> for (; level < num_levels; level++) {
> @@ -1840,7 +1840,7 @@ static bool vlv_raw_crtc_wm_is_valid(const struct intel_crtc_state *crtc_state,
> static int vlv_compute_pipe_wm(struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> struct intel_atomic_state *state =
> to_intel_atomic_state(crtc_state->base.state);
> struct vlv_wm_state *wm_state = &crtc_state->wm.vlv.optimal;
> @@ -1897,7 +1897,7 @@ static int vlv_compute_pipe_wm(struct intel_crtc_state *crtc_state)
> }
>
> /* initially allow all levels */
> - wm_state->num_levels = intel_wm_num_levels(dev_priv);
> + wm_state->num_levels = intel_wm_num_levels(i915);
> /*
> * Note that enabling cxsr with no primary/sprite planes
> * enabled can wedge the pipe. Hence we only allow cxsr
> @@ -1907,7 +1907,7 @@ static int vlv_compute_pipe_wm(struct intel_crtc_state *crtc_state)
>
> for (level = 0; level < wm_state->num_levels; level++) {
> const struct g4x_pipe_wm *raw = &crtc_state->wm.vlv.raw[level];
> - const int sr_fifo_size = INTEL_INFO(dev_priv)->num_pipes * 512 - 1;
> + const int sr_fifo_size = INTEL_INFO(i915)->num_pipes * 512 - 1;
>
> if (!vlv_raw_crtc_wm_is_valid(crtc_state, level))
> break;
> @@ -1948,8 +1948,8 @@ static void vlv_atomic_update_fifo(struct intel_atomic_state *state,
> struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> + struct intel_uncore *uncore = &i915->uncore;
> const struct vlv_fifo_state *fifo_state =
> &crtc_state->wm.vlv.fifo_state;
> int sprite0_start, sprite1_start, fifo_size;
> @@ -2092,16 +2092,16 @@ static int vlv_compute_intermediate_wm(struct intel_crtc_state *new_crtc_state)
> return 0;
> }
>
> -static void vlv_merge_wm(struct drm_i915_private *dev_priv,
> +static void vlv_merge_wm(struct drm_i915_private *i915,
> struct vlv_wm_values *wm)
> {
> struct intel_crtc *crtc;
> int num_active_crtcs = 0;
>
> - wm->level = dev_priv->wm.max_level;
> + wm->level = i915->wm.max_level;
> wm->cxsr = true;
>
> - for_each_intel_crtc(&dev_priv->drm, crtc) {
> + for_each_intel_crtc(&i915->drm, crtc) {
> const struct vlv_wm_state *wm_state = &crtc->wm.active.vlv;
>
> if (!crtc->active)
> @@ -2120,7 +2120,7 @@ static void vlv_merge_wm(struct drm_i915_private *dev_priv,
> if (num_active_crtcs > 1)
> wm->level = VLV_WM_LEVEL_PM2;
>
> - for_each_intel_crtc(&dev_priv->drm, crtc) {
> + for_each_intel_crtc(&i915->drm, crtc) {
> const struct vlv_wm_state *wm_state = &crtc->wm.active.vlv;
> enum pipe pipe = crtc->pipe;
>
> @@ -2135,35 +2135,35 @@ static void vlv_merge_wm(struct drm_i915_private *dev_priv,
> }
> }
>
> -static void vlv_program_watermarks(struct drm_i915_private *dev_priv)
> +static void vlv_program_watermarks(struct drm_i915_private *i915)
> {
> - struct vlv_wm_values *old_wm = &dev_priv->wm.vlv;
> + struct vlv_wm_values *old_wm = &i915->wm.vlv;
> struct vlv_wm_values new_wm = {};
>
> - vlv_merge_wm(dev_priv, &new_wm);
> + vlv_merge_wm(i915, &new_wm);
>
> if (memcmp(old_wm, &new_wm, sizeof(new_wm)) == 0)
> return;
>
> if (is_disabling(old_wm->level, new_wm.level, VLV_WM_LEVEL_DDR_DVFS))
> - chv_set_memory_dvfs(dev_priv, false);
> + chv_set_memory_dvfs(i915, false);
>
> if (is_disabling(old_wm->level, new_wm.level, VLV_WM_LEVEL_PM5))
> - chv_set_memory_pm5(dev_priv, false);
> + chv_set_memory_pm5(i915, false);
>
> if (is_disabling(old_wm->cxsr, new_wm.cxsr, true))
> - _intel_set_memory_cxsr(dev_priv, false);
> + _intel_set_memory_cxsr(i915, false);
>
> - vlv_write_wm_values(dev_priv, &new_wm);
> + vlv_write_wm_values(i915, &new_wm);
>
> if (is_enabling(old_wm->cxsr, new_wm.cxsr, true))
> - _intel_set_memory_cxsr(dev_priv, true);
> + _intel_set_memory_cxsr(i915, true);
>
> if (is_enabling(old_wm->level, new_wm.level, VLV_WM_LEVEL_PM5))
> - chv_set_memory_pm5(dev_priv, true);
> + chv_set_memory_pm5(i915, true);
>
> if (is_enabling(old_wm->level, new_wm.level, VLV_WM_LEVEL_DDR_DVFS))
> - chv_set_memory_dvfs(dev_priv, true);
> + chv_set_memory_dvfs(i915, true);
>
> *old_wm = new_wm;
> }
> @@ -2171,40 +2171,40 @@ static void vlv_program_watermarks(struct drm_i915_private *dev_priv)
> static void vlv_initial_watermarks(struct intel_atomic_state *state,
> struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
>
> - mutex_lock(&dev_priv->wm.wm_mutex);
> + mutex_lock(&i915->wm.wm_mutex);
> crtc->wm.active.vlv = crtc_state->wm.vlv.intermediate;
> - vlv_program_watermarks(dev_priv);
> - mutex_unlock(&dev_priv->wm.wm_mutex);
> + vlv_program_watermarks(i915);
> + mutex_unlock(&i915->wm.wm_mutex);
> }
>
> static void vlv_optimize_watermarks(struct intel_atomic_state *state,
> struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc);
>
> if (!crtc_state->wm.need_postvbl_update)
> return;
>
> - mutex_lock(&dev_priv->wm.wm_mutex);
> + mutex_lock(&i915->wm.wm_mutex);
> intel_crtc->wm.active.vlv = crtc_state->wm.vlv.optimal;
> - vlv_program_watermarks(dev_priv);
> - mutex_unlock(&dev_priv->wm.wm_mutex);
> + vlv_program_watermarks(i915);
> + mutex_unlock(&i915->wm.wm_mutex);
> }
>
> static void i965_update_wm(struct intel_crtc *unused_crtc)
> {
> - struct drm_i915_private *dev_priv = to_i915(unused_crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(unused_crtc->base.dev);
> struct intel_crtc *crtc;
> int srwm = 1;
> int cursor_sr = 16;
> bool cxsr_enabled;
>
> /* Calc sr entries for one plane configs */
> - crtc = single_enabled_crtc(dev_priv);
> + crtc = single_enabled_crtc(i915);
> if (crtc) {
> /* self-refresh has much higher latency */
> static const int sr_latency_ns = 12000;
> @@ -2246,7 +2246,7 @@ static void i965_update_wm(struct intel_crtc *unused_crtc)
> } else {
> cxsr_enabled = false;
> /* Turn off self refresh if both pipes are enabled */
> - intel_set_memory_cxsr(dev_priv, false);
> + intel_set_memory_cxsr(i915, false);
> }
>
> DRM_DEBUG_KMS("Setting FIFO watermarks - A: 8, B: 8, C: 8, SR %d\n",
> @@ -2263,14 +2263,14 @@ static void i965_update_wm(struct intel_crtc *unused_crtc)
> I915_WRITE(DSPFW3, FW_WM(cursor_sr, CURSOR_SR));
>
> if (cxsr_enabled)
> - intel_set_memory_cxsr(dev_priv, true);
> + intel_set_memory_cxsr(i915, true);
> }
>
> #undef FW_WM
>
> static void i9xx_update_wm(struct intel_crtc *unused_crtc)
> {
> - struct drm_i915_private *dev_priv = to_i915(unused_crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(unused_crtc->base.dev);
> const struct intel_watermark_params *wm_info;
> u32 fwater_lo;
> u32 fwater_hi;
> @@ -2279,15 +2279,15 @@ static void i9xx_update_wm(struct intel_crtc *unused_crtc)
> int planea_wm, planeb_wm;
> struct intel_crtc *crtc, *enabled = NULL;
>
> - if (IS_I945GM(dev_priv))
> + if (IS_I945GM(i915))
> wm_info = &i945_wm_info;
> - else if (!IS_GEN(dev_priv, 2))
> + else if (!IS_GEN(i915, 2))
> wm_info = &i915_wm_info;
> else
> wm_info = &i830_a_wm_info;
>
> - fifo_size = dev_priv->display.get_fifo_size(dev_priv, PLANE_A);
> - crtc = intel_get_crtc_for_plane(dev_priv, PLANE_A);
> + fifo_size = i915->display.get_fifo_size(i915, PLANE_A);
> + crtc = intel_get_crtc_for_plane(i915, PLANE_A);
> if (intel_crtc_active(crtc)) {
> const struct drm_display_mode *adjusted_mode =
> &crtc->config->base.adjusted_mode;
> @@ -2295,7 +2295,7 @@ static void i9xx_update_wm(struct intel_crtc *unused_crtc)
> crtc->base.primary->state->fb;
> int cpp;
>
> - if (IS_GEN(dev_priv, 2))
> + if (IS_GEN(i915, 2))
> cpp = 4;
> else
> cpp = fb->format->cpp[0];
> @@ -2310,11 +2310,11 @@ static void i9xx_update_wm(struct intel_crtc *unused_crtc)
> planea_wm = wm_info->max_wm;
> }
>
> - if (IS_GEN(dev_priv, 2))
> + if (IS_GEN(i915, 2))
> wm_info = &i830_bc_wm_info;
>
> - fifo_size = dev_priv->display.get_fifo_size(dev_priv, PLANE_B);
> - crtc = intel_get_crtc_for_plane(dev_priv, PLANE_B);
> + fifo_size = i915->display.get_fifo_size(i915, PLANE_B);
> + crtc = intel_get_crtc_for_plane(i915, PLANE_B);
> if (intel_crtc_active(crtc)) {
> const struct drm_display_mode *adjusted_mode =
> &crtc->config->base.adjusted_mode;
> @@ -2322,7 +2322,7 @@ static void i9xx_update_wm(struct intel_crtc *unused_crtc)
> crtc->base.primary->state->fb;
> int cpp;
>
> - if (IS_GEN(dev_priv, 2))
> + if (IS_GEN(i915, 2))
> cpp = 4;
> else
> cpp = fb->format->cpp[0];
> @@ -2342,7 +2342,7 @@ static void i9xx_update_wm(struct intel_crtc *unused_crtc)
>
> DRM_DEBUG_KMS("FIFO watermarks - A: %d, B: %d\n", planea_wm, planeb_wm);
>
> - if (IS_I915GM(dev_priv) && enabled) {
> + if (IS_I915GM(i915) && enabled) {
> struct drm_i915_gem_object *obj;
>
> obj = intel_fb_obj(enabled->base.primary->state->fb);
> @@ -2358,10 +2358,10 @@ static void i9xx_update_wm(struct intel_crtc *unused_crtc)
> cwm = 2;
>
> /* Play safe and disable self-refresh before adjusting watermarks. */
> - intel_set_memory_cxsr(dev_priv, false);
> + intel_set_memory_cxsr(i915, false);
>
> /* Calc sr entries for one plane configs */
> - if (HAS_FW_BLC(dev_priv) && enabled) {
> + if (HAS_FW_BLC(i915) && enabled) {
> /* self-refresh has much higher latency */
> static const int sr_latency_ns = 6000;
> const struct drm_display_mode *adjusted_mode =
> @@ -2374,7 +2374,7 @@ static void i9xx_update_wm(struct intel_crtc *unused_crtc)
> int cpp;
> int entries;
>
> - if (IS_I915GM(dev_priv) || IS_I945GM(dev_priv))
> + if (IS_I915GM(i915) || IS_I945GM(i915))
> cpp = 4;
> else
> cpp = fb->format->cpp[0];
> @@ -2387,7 +2387,7 @@ static void i9xx_update_wm(struct intel_crtc *unused_crtc)
> if (srwm < 0)
> srwm = 1;
>
> - if (IS_I945G(dev_priv) || IS_I945GM(dev_priv))
> + if (IS_I945G(i915) || IS_I945GM(i915))
> I915_WRITE(FW_BLC_SELF,
> FW_BLC_SELF_FIFO_MASK | (srwm & 0xff));
> else
> @@ -2408,25 +2408,25 @@ static void i9xx_update_wm(struct intel_crtc *unused_crtc)
> I915_WRITE(FW_BLC2, fwater_hi);
>
> if (enabled)
> - intel_set_memory_cxsr(dev_priv, true);
> + intel_set_memory_cxsr(i915, true);
> }
>
> static void i845_update_wm(struct intel_crtc *unused_crtc)
> {
> - struct drm_i915_private *dev_priv = to_i915(unused_crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(unused_crtc->base.dev);
> struct intel_crtc *crtc;
> const struct drm_display_mode *adjusted_mode;
> u32 fwater_lo;
> int planea_wm;
>
> - crtc = single_enabled_crtc(dev_priv);
> + crtc = single_enabled_crtc(i915);
> if (crtc == NULL)
> return;
>
> adjusted_mode = &crtc->config->base.adjusted_mode;
> planea_wm = intel_calculate_wm(adjusted_mode->crtc_clock,
> &i845_wm_info,
> - dev_priv->display.get_fifo_size(dev_priv, PLANE_A),
> + i915->display.get_fifo_size(i915, PLANE_A),
> 4, pessimal_latency_ns);
> fwater_lo = I915_READ(FW_BLC) & ~0xfff;
> fwater_lo |= (3<<8) | planea_wm;
> @@ -2586,24 +2586,24 @@ static u32 ilk_compute_fbc_wm(const struct intel_crtc_state *cstate,
> }
>
> static unsigned int
> -ilk_display_fifo_size(const struct drm_i915_private *dev_priv)
> +ilk_display_fifo_size(const struct drm_i915_private *i915)
> {
> - if (INTEL_GEN(dev_priv) >= 8)
> + if (INTEL_GEN(i915) >= 8)
> return 3072;
> - else if (INTEL_GEN(dev_priv) >= 7)
> + else if (INTEL_GEN(i915) >= 7)
> return 768;
> else
> return 512;
> }
>
> static unsigned int
> -ilk_plane_wm_reg_max(const struct drm_i915_private *dev_priv,
> +ilk_plane_wm_reg_max(const struct drm_i915_private *i915,
> int level, bool is_sprite)
> {
> - if (INTEL_GEN(dev_priv) >= 8)
> + if (INTEL_GEN(i915) >= 8)
> /* BDW primary/sprite plane watermarks */
> return level == 0 ? 255 : 2047;
> - else if (INTEL_GEN(dev_priv) >= 7)
> + else if (INTEL_GEN(i915) >= 7)
> /* IVB/HSW primary/sprite plane watermarks */
> return level == 0 ? 127 : 1023;
> else if (!is_sprite)
> @@ -2615,30 +2615,30 @@ ilk_plane_wm_reg_max(const struct drm_i915_private *dev_priv,
> }
>
> static unsigned int
> -ilk_cursor_wm_reg_max(const struct drm_i915_private *dev_priv, int level)
> +ilk_cursor_wm_reg_max(const struct drm_i915_private *i915, int level)
> {
> - if (INTEL_GEN(dev_priv) >= 7)
> + if (INTEL_GEN(i915) >= 7)
> return level == 0 ? 63 : 255;
> else
> return level == 0 ? 31 : 63;
> }
>
> -static unsigned int ilk_fbc_wm_reg_max(const struct drm_i915_private *dev_priv)
> +static unsigned int ilk_fbc_wm_reg_max(const struct drm_i915_private *i915)
> {
> - if (INTEL_GEN(dev_priv) >= 8)
> + if (INTEL_GEN(i915) >= 8)
> return 31;
> else
> return 15;
> }
>
> /* Calculate the maximum primary/sprite plane watermark */
> -static unsigned int ilk_plane_wm_max(const struct drm_i915_private *dev_priv,
> +static unsigned int ilk_plane_wm_max(const struct drm_i915_private *i915,
> int level,
> const struct intel_wm_config *config,
> enum intel_ddb_partitioning ddb_partitioning,
> bool is_sprite)
> {
> - unsigned int fifo_size = ilk_display_fifo_size(dev_priv);
> + unsigned int fifo_size = ilk_display_fifo_size(i915);
>
> /* if sprites aren't enabled, sprites get nothing */
> if (is_sprite && !config->sprites_enabled)
> @@ -2646,14 +2646,14 @@ static unsigned int ilk_plane_wm_max(const struct drm_i915_private *dev_priv,
>
> /* HSW allows LP1+ watermarks even with multiple pipes */
> if (level == 0 || config->num_pipes_active > 1) {
> - fifo_size /= INTEL_INFO(dev_priv)->num_pipes;
> + fifo_size /= INTEL_INFO(i915)->num_pipes;
>
> /*
> * For some reason the non self refresh
> * FIFO size is only half of the self
> * refresh FIFO size on ILK/SNB.
> */
> - if (INTEL_GEN(dev_priv) <= 6)
> + if (INTEL_GEN(i915) <= 6)
> fifo_size /= 2;
> }
>
> @@ -2669,11 +2669,11 @@ static unsigned int ilk_plane_wm_max(const struct drm_i915_private *dev_priv,
> }
>
> /* clamp to max that the registers can hold */
> - return min(fifo_size, ilk_plane_wm_reg_max(dev_priv, level, is_sprite));
> + return min(fifo_size, ilk_plane_wm_reg_max(i915, level, is_sprite));
> }
>
> /* Calculate the maximum cursor plane watermark */
> -static unsigned int ilk_cursor_wm_max(const struct drm_i915_private *dev_priv,
> +static unsigned int ilk_cursor_wm_max(const struct drm_i915_private *i915,
> int level,
> const struct intel_wm_config *config)
> {
> @@ -2682,29 +2682,29 @@ static unsigned int ilk_cursor_wm_max(const struct drm_i915_private *dev_priv,
> return 64;
>
> /* otherwise just report max that registers can hold */
> - return ilk_cursor_wm_reg_max(dev_priv, level);
> + return ilk_cursor_wm_reg_max(i915, level);
> }
>
> -static void ilk_compute_wm_maximums(const struct drm_i915_private *dev_priv,
> +static void ilk_compute_wm_maximums(const struct drm_i915_private *i915,
> int level,
> const struct intel_wm_config *config,
> enum intel_ddb_partitioning ddb_partitioning,
> struct ilk_wm_maximums *max)
> {
> - max->pri = ilk_plane_wm_max(dev_priv, level, config, ddb_partitioning, false);
> - max->spr = ilk_plane_wm_max(dev_priv, level, config, ddb_partitioning, true);
> - max->cur = ilk_cursor_wm_max(dev_priv, level, config);
> - max->fbc = ilk_fbc_wm_reg_max(dev_priv);
> + max->pri = ilk_plane_wm_max(i915, level, config, ddb_partitioning, false);
> + max->spr = ilk_plane_wm_max(i915, level, config, ddb_partitioning, true);
> + max->cur = ilk_cursor_wm_max(i915, level, config);
> + max->fbc = ilk_fbc_wm_reg_max(i915);
> }
>
> -static void ilk_compute_wm_reg_maximums(const struct drm_i915_private *dev_priv,
> +static void ilk_compute_wm_reg_maximums(const struct drm_i915_private *i915,
> int level,
> struct ilk_wm_maximums *max)
> {
> - max->pri = ilk_plane_wm_reg_max(dev_priv, level, false);
> - max->spr = ilk_plane_wm_reg_max(dev_priv, level, true);
> - max->cur = ilk_cursor_wm_reg_max(dev_priv, level);
> - max->fbc = ilk_fbc_wm_reg_max(dev_priv);
> + max->pri = ilk_plane_wm_reg_max(i915, level, false);
> + max->spr = ilk_plane_wm_reg_max(i915, level, true);
> + max->cur = ilk_cursor_wm_reg_max(i915, level);
> + max->fbc = ilk_fbc_wm_reg_max(i915);
> }
>
> static bool ilk_validate_wm_level(int level,
> @@ -2748,7 +2748,7 @@ static bool ilk_validate_wm_level(int level,
> return ret;
> }
>
> -static void ilk_compute_wm_level(const struct drm_i915_private *dev_priv,
> +static void ilk_compute_wm_level(const struct drm_i915_private *i915,
> const struct intel_crtc *intel_crtc,
> int level,
> struct intel_crtc_state *cstate,
> @@ -2757,9 +2757,9 @@ static void ilk_compute_wm_level(const struct drm_i915_private *dev_priv,
> const struct intel_plane_state *curstate,
> struct intel_wm_level *result)
> {
> - u16 pri_latency = dev_priv->wm.pri_latency[level];
> - u16 spr_latency = dev_priv->wm.spr_latency[level];
> - u16 cur_latency = dev_priv->wm.cur_latency[level];
> + u16 pri_latency = i915->wm.pri_latency[level];
> + u16 spr_latency = i915->wm.spr_latency[level];
> + u16 cur_latency = i915->wm.cur_latency[level];
>
> /* WM1+ latency values stored in 0.5us units */
> if (level > 0) {
> @@ -2811,19 +2811,19 @@ hsw_compute_linetime_wm(const struct intel_crtc_state *cstate)
> PIPE_WM_LINETIME_TIME(linetime);
> }
>
> -static void intel_read_wm_latency(struct drm_i915_private *dev_priv,
> +static void intel_read_wm_latency(struct drm_i915_private *i915,
> u16 wm[8])
> {
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct intel_uncore *uncore = &i915->uncore;
>
> - if (INTEL_GEN(dev_priv) >= 9) {
> + if (INTEL_GEN(i915) >= 9) {
> u32 val;
> int ret, i;
> - int level, max_level = ilk_wm_max_level(dev_priv);
> + int level, max_level = ilk_wm_max_level(i915);
>
> /* read the first set of memory latencies[0:3] */
> val = 0; /* data0 to be programmed to 0 for first set */
> - ret = sandybridge_pcode_read(dev_priv,
> + ret = sandybridge_pcode_read(i915,
> GEN9_PCODE_READ_MEM_LATENCY,
> &val, NULL);
>
> @@ -2842,7 +2842,7 @@ static void intel_read_wm_latency(struct drm_i915_private *dev_priv,
>
> /* read the second set of memory latencies[4:7] */
> val = 1; /* data0 to be programmed to 1 for second set */
> - ret = sandybridge_pcode_read(dev_priv,
> + ret = sandybridge_pcode_read(i915,
> GEN9_PCODE_READ_MEM_LATENCY,
> &val, NULL);
> if (ret) {
> @@ -2893,10 +2893,10 @@ static void intel_read_wm_latency(struct drm_i915_private *dev_priv,
> * any underrun. If not able to get Dimm info assume 16GB dimm
> * to avoid any underrun.
> */
> - if (dev_priv->dram_info.is_16gb_dimm)
> + if (i915->dram_info.is_16gb_dimm)
> wm[0] += 1;
>
> - } else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) {
> + } else if (IS_HASWELL(i915) || IS_BROADWELL(i915)) {
> u64 sskpd = intel_uncore_read64(uncore, MCH_SSKPD);
>
> wm[0] = (sskpd >> 56) & 0xFF;
> @@ -2906,14 +2906,14 @@ static void intel_read_wm_latency(struct drm_i915_private *dev_priv,
> wm[2] = (sskpd >> 12) & 0xFF;
> wm[3] = (sskpd >> 20) & 0x1FF;
> wm[4] = (sskpd >> 32) & 0x1FF;
> - } else if (INTEL_GEN(dev_priv) >= 6) {
> + } else if (INTEL_GEN(i915) >= 6) {
> u32 sskpd = intel_uncore_read(uncore, MCH_SSKPD);
>
> wm[0] = (sskpd >> SSKPD_WM0_SHIFT) & SSKPD_WM_MASK;
> wm[1] = (sskpd >> SSKPD_WM1_SHIFT) & SSKPD_WM_MASK;
> wm[2] = (sskpd >> SSKPD_WM2_SHIFT) & SSKPD_WM_MASK;
> wm[3] = (sskpd >> SSKPD_WM3_SHIFT) & SSKPD_WM_MASK;
> - } else if (INTEL_GEN(dev_priv) >= 5) {
> + } else if (INTEL_GEN(i915) >= 5) {
> u32 mltr = intel_uncore_read(uncore, MLTR_ILK);
>
> /* ILK primary LP0 latency is 700 ns */
> @@ -2921,44 +2921,44 @@ static void intel_read_wm_latency(struct drm_i915_private *dev_priv,
> wm[1] = (mltr >> MLTR_WM1_SHIFT) & ILK_SRLT_MASK;
> wm[2] = (mltr >> MLTR_WM2_SHIFT) & ILK_SRLT_MASK;
> } else {
> - MISSING_CASE(INTEL_DEVID(dev_priv));
> + MISSING_CASE(INTEL_DEVID(i915));
> }
> }
>
> -static void intel_fixup_spr_wm_latency(struct drm_i915_private *dev_priv,
> +static void intel_fixup_spr_wm_latency(struct drm_i915_private *i915,
> u16 wm[5])
> {
> /* ILK sprite LP0 latency is 1300 ns */
> - if (IS_GEN(dev_priv, 5))
> + if (IS_GEN(i915, 5))
> wm[0] = 13;
> }
>
> -static void intel_fixup_cur_wm_latency(struct drm_i915_private *dev_priv,
> +static void intel_fixup_cur_wm_latency(struct drm_i915_private *i915,
> u16 wm[5])
> {
> /* ILK cursor LP0 latency is 1300 ns */
> - if (IS_GEN(dev_priv, 5))
> + if (IS_GEN(i915, 5))
> wm[0] = 13;
> }
>
> -int ilk_wm_max_level(const struct drm_i915_private *dev_priv)
> +int ilk_wm_max_level(const struct drm_i915_private *i915)
> {
> /* how many WM levels are we expecting */
> - if (INTEL_GEN(dev_priv) >= 9)
> + if (INTEL_GEN(i915) >= 9)
> return 7;
> - else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> + else if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> return 4;
> - else if (INTEL_GEN(dev_priv) >= 6)
> + else if (INTEL_GEN(i915) >= 6)
> return 3;
> else
> return 2;
> }
>
> -static void intel_print_wm_latency(struct drm_i915_private *dev_priv,
> +static void intel_print_wm_latency(struct drm_i915_private *i915,
> const char *name,
> const u16 wm[8])
> {
> - int level, max_level = ilk_wm_max_level(dev_priv);
> + int level, max_level = ilk_wm_max_level(i915);
>
> for (level = 0; level <= max_level; level++) {
> unsigned int latency = wm[level];
> @@ -2973,7 +2973,7 @@ static void intel_print_wm_latency(struct drm_i915_private *dev_priv,
> * - latencies are in us on gen9.
> * - before then, WM1+ latency values are in 0.5us units
> */
> - if (INTEL_GEN(dev_priv) >= 9)
> + if (INTEL_GEN(i915) >= 9)
> latency *= 10;
> else if (level > 0)
> latency *= 5;
> @@ -2984,10 +2984,10 @@ static void intel_print_wm_latency(struct drm_i915_private *dev_priv,
> }
> }
>
> -static bool ilk_increase_wm_latency(struct drm_i915_private *dev_priv,
> +static bool ilk_increase_wm_latency(struct drm_i915_private *i915,
> u16 wm[5], u16 min)
> {
> - int level, max_level = ilk_wm_max_level(dev_priv);
> + int level, max_level = ilk_wm_max_level(i915);
>
> if (wm[0] >= min)
> return false;
> @@ -2999,7 +2999,7 @@ static bool ilk_increase_wm_latency(struct drm_i915_private *dev_priv,
> return true;
> }
>
> -static void snb_wm_latency_quirk(struct drm_i915_private *dev_priv)
> +static void snb_wm_latency_quirk(struct drm_i915_private *i915)
> {
> bool changed;
>
> @@ -3007,20 +3007,20 @@ static void snb_wm_latency_quirk(struct drm_i915_private *dev_priv)
> * The BIOS provided WM memory latency values are often
> * inadequate for high resolution displays. Adjust them.
> */
> - changed = ilk_increase_wm_latency(dev_priv, dev_priv->wm.pri_latency, 12) |
> - ilk_increase_wm_latency(dev_priv, dev_priv->wm.spr_latency, 12) |
> - ilk_increase_wm_latency(dev_priv, dev_priv->wm.cur_latency, 12);
> + changed = ilk_increase_wm_latency(i915, i915->wm.pri_latency, 12) |
> + ilk_increase_wm_latency(i915, i915->wm.spr_latency, 12) |
> + ilk_increase_wm_latency(i915, i915->wm.cur_latency, 12);
>
> if (!changed)
> return;
>
> DRM_DEBUG_KMS("WM latency values increased to avoid potential underruns\n");
> - intel_print_wm_latency(dev_priv, "Primary", dev_priv->wm.pri_latency);
> - intel_print_wm_latency(dev_priv, "Sprite", dev_priv->wm.spr_latency);
> - intel_print_wm_latency(dev_priv, "Cursor", dev_priv->wm.cur_latency);
> + intel_print_wm_latency(i915, "Primary", i915->wm.pri_latency);
> + intel_print_wm_latency(i915, "Sprite", i915->wm.spr_latency);
> + intel_print_wm_latency(i915, "Cursor", i915->wm.cur_latency);
> }
>
> -static void snb_wm_lp3_irq_quirk(struct drm_i915_private *dev_priv)
> +static void snb_wm_lp3_irq_quirk(struct drm_i915_private *i915)
> {
> /*
> * On some SNB machines (Thinkpad X220 Tablet at least)
> @@ -3033,50 +3033,50 @@ static void snb_wm_lp3_irq_quirk(struct drm_i915_private *dev_priv)
> * interrupts only. To play it safe we disable LP3
> * watermarks entirely.
> */
> - if (dev_priv->wm.pri_latency[3] == 0 &&
> - dev_priv->wm.spr_latency[3] == 0 &&
> - dev_priv->wm.cur_latency[3] == 0)
> + if (i915->wm.pri_latency[3] == 0 &&
> + i915->wm.spr_latency[3] == 0 &&
> + i915->wm.cur_latency[3] == 0)
> return;
>
> - dev_priv->wm.pri_latency[3] = 0;
> - dev_priv->wm.spr_latency[3] = 0;
> - dev_priv->wm.cur_latency[3] = 0;
> + i915->wm.pri_latency[3] = 0;
> + i915->wm.spr_latency[3] = 0;
> + i915->wm.cur_latency[3] = 0;
>
> DRM_DEBUG_KMS("LP3 watermarks disabled due to potential for lost interrupts\n");
> - intel_print_wm_latency(dev_priv, "Primary", dev_priv->wm.pri_latency);
> - intel_print_wm_latency(dev_priv, "Sprite", dev_priv->wm.spr_latency);
> - intel_print_wm_latency(dev_priv, "Cursor", dev_priv->wm.cur_latency);
> + intel_print_wm_latency(i915, "Primary", i915->wm.pri_latency);
> + intel_print_wm_latency(i915, "Sprite", i915->wm.spr_latency);
> + intel_print_wm_latency(i915, "Cursor", i915->wm.cur_latency);
> }
>
> -static void ilk_setup_wm_latency(struct drm_i915_private *dev_priv)
> +static void ilk_setup_wm_latency(struct drm_i915_private *i915)
> {
> - intel_read_wm_latency(dev_priv, dev_priv->wm.pri_latency);
> + intel_read_wm_latency(i915, i915->wm.pri_latency);
>
> - memcpy(dev_priv->wm.spr_latency, dev_priv->wm.pri_latency,
> - sizeof(dev_priv->wm.pri_latency));
> - memcpy(dev_priv->wm.cur_latency, dev_priv->wm.pri_latency,
> - sizeof(dev_priv->wm.pri_latency));
> + memcpy(i915->wm.spr_latency, i915->wm.pri_latency,
> + sizeof(i915->wm.pri_latency));
> + memcpy(i915->wm.cur_latency, i915->wm.pri_latency,
> + sizeof(i915->wm.pri_latency));
>
> - intel_fixup_spr_wm_latency(dev_priv, dev_priv->wm.spr_latency);
> - intel_fixup_cur_wm_latency(dev_priv, dev_priv->wm.cur_latency);
> + intel_fixup_spr_wm_latency(i915, i915->wm.spr_latency);
> + intel_fixup_cur_wm_latency(i915, i915->wm.cur_latency);
>
> - intel_print_wm_latency(dev_priv, "Primary", dev_priv->wm.pri_latency);
> - intel_print_wm_latency(dev_priv, "Sprite", dev_priv->wm.spr_latency);
> - intel_print_wm_latency(dev_priv, "Cursor", dev_priv->wm.cur_latency);
> + intel_print_wm_latency(i915, "Primary", i915->wm.pri_latency);
> + intel_print_wm_latency(i915, "Sprite", i915->wm.spr_latency);
> + intel_print_wm_latency(i915, "Cursor", i915->wm.cur_latency);
>
> - if (IS_GEN(dev_priv, 6)) {
> - snb_wm_latency_quirk(dev_priv);
> - snb_wm_lp3_irq_quirk(dev_priv);
> + if (IS_GEN(i915, 6)) {
> + snb_wm_latency_quirk(i915);
> + snb_wm_lp3_irq_quirk(i915);
> }
> }
>
> -static void skl_setup_wm_latency(struct drm_i915_private *dev_priv)
> +static void skl_setup_wm_latency(struct drm_i915_private *i915)
> {
> - intel_read_wm_latency(dev_priv, dev_priv->wm.skl_latency);
> - intel_print_wm_latency(dev_priv, "Gen9 Plane", dev_priv->wm.skl_latency);
> + intel_read_wm_latency(i915, i915->wm.skl_latency);
> + intel_print_wm_latency(i915, "Gen9 Plane", i915->wm.skl_latency);
> }
>
> -static bool ilk_validate_pipe_wm(const struct drm_i915_private *dev_priv,
> +static bool ilk_validate_pipe_wm(const struct drm_i915_private *i915,
> struct intel_pipe_wm *pipe_wm)
> {
> /* LP0 watermark maximums depend on this pipe alone */
> @@ -3088,7 +3088,7 @@ static bool ilk_validate_pipe_wm(const struct drm_i915_private *dev_priv,
> struct ilk_wm_maximums max;
>
> /* LP0 watermarks always use 1/2 DDB partitioning */
> - ilk_compute_wm_maximums(dev_priv, 0, &config, INTEL_DDB_PART_1_2, &max);
> + ilk_compute_wm_maximums(i915, 0, &config, INTEL_DDB_PART_1_2, &max);
>
> /* At least LP0 must be valid */
> if (!ilk_validate_wm_level(0, &max, &pipe_wm->wm[0])) {
> @@ -3106,13 +3106,13 @@ static int ilk_compute_pipe_wm(struct intel_crtc_state *cstate)
> struct intel_crtc *intel_crtc = to_intel_crtc(cstate->base.crtc);
> struct intel_pipe_wm *pipe_wm;
> struct drm_device *dev = state->dev;
> - const struct drm_i915_private *dev_priv = to_i915(dev);
> + const struct drm_i915_private *i915 = to_i915(dev);
> struct drm_plane *plane;
> const struct drm_plane_state *plane_state;
> const struct intel_plane_state *pristate = NULL;
> const struct intel_plane_state *sprstate = NULL;
> const struct intel_plane_state *curstate = NULL;
> - int level, max_level = ilk_wm_max_level(dev_priv), usable_level;
> + int level, max_level = ilk_wm_max_level(i915), usable_level;
> struct ilk_wm_maximums max;
>
> pipe_wm = &cstate->wm.ilk.optimal;
> @@ -3139,7 +3139,7 @@ static int ilk_compute_pipe_wm(struct intel_crtc_state *cstate)
> usable_level = max_level;
>
> /* ILK/SNB: LP2+ watermarks only w/o sprites */
> - if (INTEL_GEN(dev_priv) <= 6 && pipe_wm->sprites_enabled)
> + if (INTEL_GEN(i915) <= 6 && pipe_wm->sprites_enabled)
> usable_level = 1;
>
> /* ILK/SNB/IVB: LP1+ watermarks only w/o scaling */
> @@ -3147,21 +3147,21 @@ static int ilk_compute_pipe_wm(struct intel_crtc_state *cstate)
> usable_level = 0;
>
> memset(&pipe_wm->wm, 0, sizeof(pipe_wm->wm));
> - ilk_compute_wm_level(dev_priv, intel_crtc, 0, cstate,
> + ilk_compute_wm_level(i915, intel_crtc, 0, cstate,
> pristate, sprstate, curstate, &pipe_wm->wm[0]);
>
> - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> + if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> pipe_wm->linetime = hsw_compute_linetime_wm(cstate);
>
> - if (!ilk_validate_pipe_wm(dev_priv, pipe_wm))
> + if (!ilk_validate_pipe_wm(i915, pipe_wm))
> return -EINVAL;
>
> - ilk_compute_wm_reg_maximums(dev_priv, 1, &max);
> + ilk_compute_wm_reg_maximums(i915, 1, &max);
>
> for (level = 1; level <= usable_level; level++) {
> struct intel_wm_level *wm = &pipe_wm->wm[level];
>
> - ilk_compute_wm_level(dev_priv, intel_crtc, level, cstate,
> + ilk_compute_wm_level(i915, intel_crtc, level, cstate,
> pristate, sprstate, curstate, wm);
>
> /*
> @@ -3186,14 +3186,14 @@ static int ilk_compute_pipe_wm(struct intel_crtc_state *cstate)
> static int ilk_compute_intermediate_wm(struct intel_crtc_state *newstate)
> {
> struct intel_crtc *intel_crtc = to_intel_crtc(newstate->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(intel_crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(intel_crtc->base.dev);
> struct intel_pipe_wm *a = &newstate->wm.ilk.intermediate;
> struct intel_atomic_state *intel_state =
> to_intel_atomic_state(newstate->base.state);
> const struct intel_crtc_state *oldstate =
> intel_atomic_get_old_crtc_state(intel_state, intel_crtc);
> const struct intel_pipe_wm *b = &oldstate->wm.ilk.optimal;
> - int level, max_level = ilk_wm_max_level(dev_priv);
> + int level, max_level = ilk_wm_max_level(i915);
>
> /*
> * Start with the final, target watermarks, then combine with the
> @@ -3226,7 +3226,7 @@ static int ilk_compute_intermediate_wm(struct intel_crtc_state *newstate)
> * there's no safe way to transition from the old state to
> * the new state, so we need to fail the atomic transaction.
> */
> - if (!ilk_validate_pipe_wm(dev_priv, a))
> + if (!ilk_validate_pipe_wm(i915, a))
> return -EINVAL;
>
> /*
> @@ -3242,7 +3242,7 @@ static int ilk_compute_intermediate_wm(struct intel_crtc_state *newstate)
> /*
> * Merge the watermarks from all active pipes for a specific level.
> */
> -static void ilk_merge_wm_level(struct drm_i915_private *dev_priv,
> +static void ilk_merge_wm_level(struct drm_i915_private *i915,
> int level,
> struct intel_wm_level *ret_wm)
> {
> @@ -3250,7 +3250,7 @@ static void ilk_merge_wm_level(struct drm_i915_private *dev_priv,
>
> ret_wm->enable = true;
>
> - for_each_intel_crtc(&dev_priv->drm, intel_crtc) {
> + for_each_intel_crtc(&i915->drm, intel_crtc) {
> const struct intel_pipe_wm *active = &intel_crtc->wm.active.ilk;
> const struct intel_wm_level *wm = &active->wm[level];
>
> @@ -3275,27 +3275,27 @@ static void ilk_merge_wm_level(struct drm_i915_private *dev_priv,
> /*
> * Merge all low power watermarks for all active pipes.
> */
> -static void ilk_wm_merge(struct drm_i915_private *dev_priv,
> +static void ilk_wm_merge(struct drm_i915_private *i915,
> const struct intel_wm_config *config,
> const struct ilk_wm_maximums *max,
> struct intel_pipe_wm *merged)
> {
> - int level, max_level = ilk_wm_max_level(dev_priv);
> + int level, max_level = ilk_wm_max_level(i915);
> int last_enabled_level = max_level;
>
> /* ILK/SNB/IVB: LP1+ watermarks only w/ single pipe */
> - if ((INTEL_GEN(dev_priv) <= 6 || IS_IVYBRIDGE(dev_priv)) &&
> + if ((INTEL_GEN(i915) <= 6 || IS_IVYBRIDGE(i915)) &&
> config->num_pipes_active > 1)
> last_enabled_level = 0;
>
> /* ILK: FBC WM must be disabled always */
> - merged->fbc_wm_enabled = INTEL_GEN(dev_priv) >= 6;
> + merged->fbc_wm_enabled = INTEL_GEN(i915) >= 6;
>
> /* merge each WM1+ level */
> for (level = 1; level <= max_level; level++) {
> struct intel_wm_level *wm = &merged->wm[level];
>
> - ilk_merge_wm_level(dev_priv, level, wm);
> + ilk_merge_wm_level(i915, level, wm);
>
> if (level > last_enabled_level)
> wm->enable = false;
> @@ -3320,8 +3320,8 @@ static void ilk_wm_merge(struct drm_i915_private *dev_priv,
> * What we should check here is whether FBC can be
> * enabled sometime later.
> */
> - if (IS_GEN(dev_priv, 5) && !merged->fbc_wm_enabled &&
> - intel_fbc_is_active(dev_priv)) {
> + if (IS_GEN(i915, 5) && !merged->fbc_wm_enabled &&
> + intel_fbc_is_active(i915)) {
> for (level = 2; level <= max_level; level++) {
> struct intel_wm_level *wm = &merged->wm[level];
>
> @@ -3337,16 +3337,16 @@ static int ilk_wm_lp_to_level(int wm_lp, const struct intel_pipe_wm *pipe_wm)
> }
>
> /* The value we need to program into the WM_LPx latency field */
> -static unsigned int ilk_wm_lp_latency(struct drm_i915_private *dev_priv,
> +static unsigned int ilk_wm_lp_latency(struct drm_i915_private *i915,
> int level)
> {
> - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> + if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> return 2 * level;
> else
> - return dev_priv->wm.pri_latency[level];
> + return i915->wm.pri_latency[level];
> }
>
> -static void ilk_compute_wm_results(struct drm_i915_private *dev_priv,
> +static void ilk_compute_wm_results(struct drm_i915_private *i915,
> const struct intel_pipe_wm *merged,
> enum intel_ddb_partitioning partitioning,
> struct ilk_wm_values *results)
> @@ -3370,14 +3370,14 @@ static void ilk_compute_wm_results(struct drm_i915_private *dev_priv,
> * disabled. Doing otherwise could cause underruns.
> */
> results->wm_lp[wm_lp - 1] =
> - (ilk_wm_lp_latency(dev_priv, level) << WM1_LP_LATENCY_SHIFT) |
> + (ilk_wm_lp_latency(i915, level) << WM1_LP_LATENCY_SHIFT) |
> (r->pri_val << WM1_LP_SR_SHIFT) |
> r->cur_val;
>
> if (r->enable)
> results->wm_lp[wm_lp - 1] |= WM1_LP_SR_EN;
>
> - if (INTEL_GEN(dev_priv) >= 8)
> + if (INTEL_GEN(i915) >= 8)
> results->wm_lp[wm_lp - 1] |=
> r->fbc_val << WM1_LP_FBC_SHIFT_BDW;
> else
> @@ -3388,7 +3388,7 @@ static void ilk_compute_wm_results(struct drm_i915_private *dev_priv,
> * Always set WM1S_LP_EN when spr_val != 0, even if the
> * level is disabled. Doing otherwise could cause underruns.
> */
> - if (INTEL_GEN(dev_priv) <= 6 && r->spr_val) {
> + if (INTEL_GEN(i915) <= 6 && r->spr_val) {
> WARN_ON(wm_lp != 1);
> results->wm_lp_spr[wm_lp - 1] = WM1S_LP_EN | r->spr_val;
> } else
> @@ -3396,7 +3396,7 @@ static void ilk_compute_wm_results(struct drm_i915_private *dev_priv,
> }
>
> /* LP0 register values */
> - for_each_intel_crtc(&dev_priv->drm, intel_crtc) {
> + for_each_intel_crtc(&i915->drm, intel_crtc) {
> enum pipe pipe = intel_crtc->pipe;
> const struct intel_wm_level *r =
> &intel_crtc->wm.active.ilk.wm[0];
> @@ -3416,11 +3416,11 @@ static void ilk_compute_wm_results(struct drm_i915_private *dev_priv,
> /* Find the result with the highest level enabled. Check for enable_fbc_wm in
> * case both are at the same level. Prefer r1 in case they're the same. */
> static struct intel_pipe_wm *
> -ilk_find_best_result(struct drm_i915_private *dev_priv,
> +ilk_find_best_result(struct drm_i915_private *i915,
> struct intel_pipe_wm *r1,
> struct intel_pipe_wm *r2)
> {
> - int level, max_level = ilk_wm_max_level(dev_priv);
> + int level, max_level = ilk_wm_max_level(i915);
> int level1 = 0, level2 = 0;
>
> for (level = 1; level <= max_level; level++) {
> @@ -3450,7 +3450,7 @@ ilk_find_best_result(struct drm_i915_private *dev_priv,
> #define WM_DIRTY_FBC (1 << 24)
> #define WM_DIRTY_DDB (1 << 25)
>
> -static unsigned int ilk_compute_wm_dirty(struct drm_i915_private *dev_priv,
> +static unsigned int ilk_compute_wm_dirty(struct drm_i915_private *i915,
> const struct ilk_wm_values *old,
> const struct ilk_wm_values *new)
> {
> @@ -3458,7 +3458,7 @@ static unsigned int ilk_compute_wm_dirty(struct drm_i915_private *dev_priv,
> enum pipe pipe;
> int wm_lp;
>
> - for_each_pipe(dev_priv, pipe) {
> + for_each_pipe(i915, pipe) {
> if (old->wm_linetime[pipe] != new->wm_linetime[pipe]) {
> dirty |= WM_DIRTY_LINETIME(pipe);
> /* Must disable LP1+ watermarks too */
> @@ -3502,10 +3502,10 @@ static unsigned int ilk_compute_wm_dirty(struct drm_i915_private *dev_priv,
> return dirty;
> }
>
> -static bool _ilk_disable_lp_wm(struct drm_i915_private *dev_priv,
> +static bool _ilk_disable_lp_wm(struct drm_i915_private *i915,
> unsigned int dirty)
> {
> - struct ilk_wm_values *previous = &dev_priv->wm.hw;
> + struct ilk_wm_values *previous = &i915->wm.hw;
> bool changed = false;
>
> if (dirty & WM_DIRTY_LP(3) && previous->wm_lp[2] & WM1_LP_SR_EN) {
> @@ -3536,18 +3536,18 @@ static bool _ilk_disable_lp_wm(struct drm_i915_private *dev_priv,
> * The spec says we shouldn't write when we don't need, because every write
> * causes WMs to be re-evaluated, expending some power.
> */
> -static void ilk_write_wm_values(struct drm_i915_private *dev_priv,
> +static void ilk_write_wm_values(struct drm_i915_private *i915,
> struct ilk_wm_values *results)
> {
> - struct ilk_wm_values *previous = &dev_priv->wm.hw;
> + struct ilk_wm_values *previous = &i915->wm.hw;
> unsigned int dirty;
> u32 val;
>
> - dirty = ilk_compute_wm_dirty(dev_priv, previous, results);
> + dirty = ilk_compute_wm_dirty(i915, previous, results);
> if (!dirty)
> return;
>
> - _ilk_disable_lp_wm(dev_priv, dirty);
> + _ilk_disable_lp_wm(i915, dirty);
>
> if (dirty & WM_DIRTY_PIPE(PIPE_A))
> I915_WRITE(WM0_PIPEA_ILK, results->wm_pipe[0]);
> @@ -3564,7 +3564,7 @@ static void ilk_write_wm_values(struct drm_i915_private *dev_priv,
> I915_WRITE(PIPE_WM_LINETIME(PIPE_C), results->wm_linetime[2]);
>
> if (dirty & WM_DIRTY_DDB) {
> - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) {
> + if (IS_HASWELL(i915) || IS_BROADWELL(i915)) {
> val = I915_READ(WM_MISC);
> if (results->partitioning == INTEL_DDB_PART_1_2)
> val &= ~WM_MISC_DATA_PARTITION_5_6;
> @@ -3594,7 +3594,7 @@ static void ilk_write_wm_values(struct drm_i915_private *dev_priv,
> previous->wm_lp_spr[0] != results->wm_lp_spr[0])
> I915_WRITE(WM1S_LP_ILK, results->wm_lp_spr[0]);
>
> - if (INTEL_GEN(dev_priv) >= 7) {
> + if (INTEL_GEN(i915) >= 7) {
> if (dirty & WM_DIRTY_LP(2) && previous->wm_lp_spr[1] != results->wm_lp_spr[1])
> I915_WRITE(WM2S_LP_IVB, results->wm_lp_spr[1]);
> if (dirty & WM_DIRTY_LP(3) && previous->wm_lp_spr[2] != results->wm_lp_spr[2])
> @@ -3608,17 +3608,17 @@ static void ilk_write_wm_values(struct drm_i915_private *dev_priv,
> if (dirty & WM_DIRTY_LP(3) && previous->wm_lp[2] != results->wm_lp[2])
> I915_WRITE(WM3_LP_ILK, results->wm_lp[2]);
>
> - dev_priv->wm.hw = *results;
> + i915->wm.hw = *results;
> }
>
> bool ilk_disable_lp_wm(struct drm_device *dev)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
>
> - return _ilk_disable_lp_wm(dev_priv, WM_DIRTY_LP_ALL);
> + return _ilk_disable_lp_wm(i915, WM_DIRTY_LP_ALL);
> }
>
> -static u8 intel_enabled_dbuf_slices_num(struct drm_i915_private *dev_priv)
> +static u8 intel_enabled_dbuf_slices_num(struct drm_i915_private *i915)
> {
> u8 enabled_slices;
>
> @@ -3626,7 +3626,7 @@ static u8 intel_enabled_dbuf_slices_num(struct drm_i915_private *dev_priv)
> enabled_slices = 1;
>
> /* Gen prior to GEN11 have only one DBuf slice */
> - if (INTEL_GEN(dev_priv) < 11)
> + if (INTEL_GEN(i915) < 11)
> return enabled_slices;
>
> /*
> @@ -3644,16 +3644,16 @@ static u8 intel_enabled_dbuf_slices_num(struct drm_i915_private *dev_priv)
> * FIXME: We still don't have the proper code detect if we need to apply the WA,
> * so assume we'll always need it in order to avoid underruns.
> */
> -static bool skl_needs_memory_bw_wa(struct drm_i915_private *dev_priv)
> +static bool skl_needs_memory_bw_wa(struct drm_i915_private *i915)
> {
> - return IS_GEN9_BC(dev_priv) || IS_BROXTON(dev_priv);
> + return IS_GEN9_BC(i915) || IS_BROXTON(i915);
> }
>
> static bool
> -intel_has_sagv(struct drm_i915_private *dev_priv)
> +intel_has_sagv(struct drm_i915_private *i915)
> {
> - return (IS_GEN9_BC(dev_priv) || INTEL_GEN(dev_priv) >= 10) &&
> - dev_priv->sagv_status != I915_SAGV_NOT_CONTROLLED;
> + return (IS_GEN9_BC(i915) || INTEL_GEN(i915) >= 10) &&
> + i915->sagv_status != I915_SAGV_NOT_CONTROLLED;
> }
>
> /*
> @@ -3668,18 +3668,18 @@ intel_has_sagv(struct drm_i915_private *dev_priv)
> * - We're not using an interlaced display configuration
> */
> int
> -intel_enable_sagv(struct drm_i915_private *dev_priv)
> +intel_enable_sagv(struct drm_i915_private *i915)
> {
> int ret;
>
> - if (!intel_has_sagv(dev_priv))
> + if (!intel_has_sagv(i915))
> return 0;
>
> - if (dev_priv->sagv_status == I915_SAGV_ENABLED)
> + if (i915->sagv_status == I915_SAGV_ENABLED)
> return 0;
>
> DRM_DEBUG_KMS("Enabling SAGV\n");
> - ret = sandybridge_pcode_write(dev_priv, GEN9_PCODE_SAGV_CONTROL,
> + ret = sandybridge_pcode_write(i915, GEN9_PCODE_SAGV_CONTROL,
> GEN9_SAGV_ENABLE);
>
> /* We don't need to wait for SAGV when enabling */
> @@ -3688,33 +3688,33 @@ intel_enable_sagv(struct drm_i915_private *dev_priv)
> * Some skl systems, pre-release machines in particular,
> * don't actually have SAGV.
> */
> - if (IS_SKYLAKE(dev_priv) && ret == -ENXIO) {
> + if (IS_SKYLAKE(i915) && ret == -ENXIO) {
> DRM_DEBUG_DRIVER("No SAGV found on system, ignoring\n");
> - dev_priv->sagv_status = I915_SAGV_NOT_CONTROLLED;
> + i915->sagv_status = I915_SAGV_NOT_CONTROLLED;
> return 0;
> } else if (ret < 0) {
> DRM_ERROR("Failed to enable SAGV\n");
> return ret;
> }
>
> - dev_priv->sagv_status = I915_SAGV_ENABLED;
> + i915->sagv_status = I915_SAGV_ENABLED;
> return 0;
> }
>
> int
> -intel_disable_sagv(struct drm_i915_private *dev_priv)
> +intel_disable_sagv(struct drm_i915_private *i915)
> {
> int ret;
>
> - if (!intel_has_sagv(dev_priv))
> + if (!intel_has_sagv(i915))
> return 0;
>
> - if (dev_priv->sagv_status == I915_SAGV_DISABLED)
> + if (i915->sagv_status == I915_SAGV_DISABLED)
> return 0;
>
> DRM_DEBUG_KMS("Disabling SAGV\n");
> /* bspec says to keep retrying for at least 1 ms */
> - ret = skl_pcode_request(dev_priv, GEN9_PCODE_SAGV_CONTROL,
> + ret = skl_pcode_request(i915, GEN9_PCODE_SAGV_CONTROL,
> GEN9_SAGV_DISABLE,
> GEN9_SAGV_IS_DISABLED, GEN9_SAGV_IS_DISABLED,
> 1);
> @@ -3722,23 +3722,23 @@ intel_disable_sagv(struct drm_i915_private *dev_priv)
> * Some skl systems, pre-release machines in particular,
> * don't actually have SAGV.
> */
> - if (IS_SKYLAKE(dev_priv) && ret == -ENXIO) {
> + if (IS_SKYLAKE(i915) && ret == -ENXIO) {
> DRM_DEBUG_DRIVER("No SAGV found on system, ignoring\n");
> - dev_priv->sagv_status = I915_SAGV_NOT_CONTROLLED;
> + i915->sagv_status = I915_SAGV_NOT_CONTROLLED;
> return 0;
> } else if (ret < 0) {
> DRM_ERROR("Failed to disable SAGV (%d)\n", ret);
> return ret;
> }
>
> - dev_priv->sagv_status = I915_SAGV_DISABLED;
> + i915->sagv_status = I915_SAGV_DISABLED;
> return 0;
> }
>
> bool intel_can_enable_sagv(struct drm_atomic_state *state)
> {
> struct drm_device *dev = state->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_atomic_state *intel_state = to_intel_atomic_state(state);
> struct intel_crtc *crtc;
> struct intel_plane *plane;
> @@ -3747,12 +3747,12 @@ bool intel_can_enable_sagv(struct drm_atomic_state *state)
> int level, latency;
> int sagv_block_time_us;
>
> - if (!intel_has_sagv(dev_priv))
> + if (!intel_has_sagv(i915))
> return false;
>
> - if (IS_GEN(dev_priv, 9))
> + if (IS_GEN(i915, 9))
> sagv_block_time_us = 30;
> - else if (IS_GEN(dev_priv, 10))
> + else if (IS_GEN(i915, 10))
> sagv_block_time_us = 20;
> else
> sagv_block_time_us = 10;
> @@ -3772,7 +3772,7 @@ bool intel_can_enable_sagv(struct drm_atomic_state *state)
>
> /* Since we're now guaranteed to only have one active CRTC... */
> pipe = ffs(intel_state->active_crtcs) - 1;
> - crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> + crtc = intel_get_crtc_for_pipe(i915, pipe);
> cstate = to_intel_crtc_state(crtc->base.state);
>
> if (crtc->base.state->adjusted_mode.flags & DRM_MODE_FLAG_INTERLACE)
> @@ -3787,13 +3787,13 @@ bool intel_can_enable_sagv(struct drm_atomic_state *state)
> continue;
>
> /* Find the highest enabled wm level for this plane */
> - for (level = ilk_wm_max_level(dev_priv);
> + for (level = ilk_wm_max_level(i915);
> !wm->wm[level].plane_en; --level)
> { }
>
> - latency = dev_priv->wm.skl_latency[level];
> + latency = i915->wm.skl_latency[level];
>
> - if (skl_needs_memory_bw_wa(dev_priv) &&
> + if (skl_needs_memory_bw_wa(i915) &&
> plane->base.state->fb->modifier ==
> I915_FORMAT_MOD_X_TILED)
> latency += 15;
> @@ -3810,7 +3810,7 @@ bool intel_can_enable_sagv(struct drm_atomic_state *state)
> return true;
> }
>
> -static u16 intel_get_ddb_size(struct drm_i915_private *dev_priv,
> +static u16 intel_get_ddb_size(struct drm_i915_private *i915,
> const struct intel_crtc_state *cstate,
> const u64 total_data_rate,
> const int num_active,
> @@ -3818,11 +3818,11 @@ static u16 intel_get_ddb_size(struct drm_i915_private *dev_priv,
> {
> const struct drm_display_mode *adjusted_mode;
> u64 total_data_bw;
> - u16 ddb_size = INTEL_INFO(dev_priv)->ddb_size;
> + u16 ddb_size = INTEL_INFO(i915)->ddb_size;
>
> WARN_ON(ddb_size == 0);
>
> - if (INTEL_GEN(dev_priv) < 11)
> + if (INTEL_GEN(i915) < 11)
> return ddb_size - 4; /* 4 blocks for bypass path allocation */
>
> adjusted_mode = &cstate->base.adjusted_mode;
> @@ -3847,7 +3847,7 @@ static u16 intel_get_ddb_size(struct drm_i915_private *dev_priv,
> }
>
> static void
> -skl_ddb_get_pipe_allocation_limits(struct drm_i915_private *dev_priv,
> +skl_ddb_get_pipe_allocation_limits(struct drm_i915_private *i915,
> const struct intel_crtc_state *cstate,
> const u64 total_data_rate,
> struct skl_ddb_allocation *ddb,
> @@ -3867,16 +3867,16 @@ skl_ddb_get_pipe_allocation_limits(struct drm_i915_private *dev_priv,
> if (WARN_ON(!state) || !cstate->base.active) {
> alloc->start = 0;
> alloc->end = 0;
> - *num_active = hweight32(dev_priv->active_crtcs);
> + *num_active = hweight32(i915->active_crtcs);
> return;
> }
>
> if (intel_state->active_pipe_changes)
> *num_active = hweight32(intel_state->active_crtcs);
> else
> - *num_active = hweight32(dev_priv->active_crtcs);
> + *num_active = hweight32(i915->active_crtcs);
>
> - ddb_size = intel_get_ddb_size(dev_priv, cstate, total_data_rate,
> + ddb_size = intel_get_ddb_size(i915, cstate, total_data_rate,
> *num_active, ddb);
>
> /*
> @@ -3939,8 +3939,8 @@ static unsigned int
> skl_cursor_allocation(const struct intel_crtc_state *crtc_state,
> int num_active)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> - int level, max_level = ilk_wm_max_level(dev_priv);
> + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> + int level, max_level = ilk_wm_max_level(i915);
> struct skl_wm_level wm = {};
> int ret, min_ddb_alloc = 0;
> struct skl_wm_params wp;
> @@ -3963,7 +3963,7 @@ skl_cursor_allocation(const struct intel_crtc_state *crtc_state,
> return max(num_active == 1 ? 32 : 8, min_ddb_alloc);
> }
>
> -static void skl_ddb_entry_init_from_hw(struct drm_i915_private *dev_priv,
> +static void skl_ddb_entry_init_from_hw(struct drm_i915_private *i915,
> struct skl_ddb_entry *entry, u32 reg)
> {
>
> @@ -3975,7 +3975,7 @@ static void skl_ddb_entry_init_from_hw(struct drm_i915_private *dev_priv,
> }
>
> static void
> -skl_ddb_get_hw_plane_state(struct drm_i915_private *dev_priv,
> +skl_ddb_get_hw_plane_state(struct drm_i915_private *i915,
> const enum pipe pipe,
> const enum plane_id plane_id,
> struct skl_ddb_entry *ddb_y,
> @@ -3987,7 +3987,7 @@ skl_ddb_get_hw_plane_state(struct drm_i915_private *dev_priv,
> /* Cursor doesn't support NV12/planar, so no extra calculation needed */
> if (plane_id == PLANE_CURSOR) {
> val = I915_READ(CUR_BUF_CFG(pipe));
> - skl_ddb_entry_init_from_hw(dev_priv, ddb_y, val);
> + skl_ddb_entry_init_from_hw(i915, ddb_y, val);
> return;
> }
>
> @@ -3999,9 +3999,9 @@ skl_ddb_get_hw_plane_state(struct drm_i915_private *dev_priv,
> val & PLANE_CTL_ORDER_RGBX,
> val & PLANE_CTL_ALPHA_MASK);
>
> - if (INTEL_GEN(dev_priv) >= 11) {
> + if (INTEL_GEN(i915) >= 11) {
> val = I915_READ(PLANE_BUF_CFG(pipe, plane_id));
> - skl_ddb_entry_init_from_hw(dev_priv, ddb_y, val);
> + skl_ddb_entry_init_from_hw(i915, ddb_y, val);
> } else {
> val = I915_READ(PLANE_BUF_CFG(pipe, plane_id));
> val2 = I915_READ(PLANE_NV12_BUF_CFG(pipe, plane_id));
> @@ -4009,8 +4009,8 @@ skl_ddb_get_hw_plane_state(struct drm_i915_private *dev_priv,
> if (is_planar_yuv_format(fourcc))
> swap(val, val2);
>
> - skl_ddb_entry_init_from_hw(dev_priv, ddb_y, val);
> - skl_ddb_entry_init_from_hw(dev_priv, ddb_uv, val2);
> + skl_ddb_entry_init_from_hw(i915, ddb_y, val);
> + skl_ddb_entry_init_from_hw(i915, ddb_uv, val2);
> }
> }
>
> @@ -4018,30 +4018,30 @@ void skl_pipe_ddb_get_hw_state(struct intel_crtc *crtc,
> struct skl_ddb_entry *ddb_y,
> struct skl_ddb_entry *ddb_uv)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum intel_display_power_domain power_domain;
> enum pipe pipe = crtc->pipe;
> intel_wakeref_t wakeref;
> enum plane_id plane_id;
>
> power_domain = POWER_DOMAIN_PIPE(pipe);
> - wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain);
> + wakeref = intel_display_power_get_if_enabled(i915, power_domain);
> if (!wakeref)
> return;
>
> for_each_plane_id_on_crtc(crtc, plane_id)
> - skl_ddb_get_hw_plane_state(dev_priv, pipe,
> + skl_ddb_get_hw_plane_state(i915, pipe,
> plane_id,
> &ddb_y[plane_id],
> &ddb_uv[plane_id]);
>
> - intel_display_power_put(dev_priv, power_domain, wakeref);
> + intel_display_power_put(i915, power_domain, wakeref);
> }
>
> -void skl_ddb_get_hw_state(struct drm_i915_private *dev_priv,
> +void skl_ddb_get_hw_state(struct drm_i915_private *i915,
> struct skl_ddb_allocation *ddb /* out */)
> {
> - ddb->enabled_slices = intel_enabled_dbuf_slices_num(dev_priv);
> + ddb->enabled_slices = intel_enabled_dbuf_slices_num(i915);
> }
>
> /*
> @@ -4138,7 +4138,7 @@ skl_pipe_downscale_amount(const struct intel_crtc_state *crtc_state)
> int skl_check_pipe_max_pixel_rate(struct intel_crtc *intel_crtc,
> struct intel_crtc_state *cstate)
> {
> - struct drm_i915_private *dev_priv = to_i915(intel_crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(intel_crtc->base.dev);
> struct drm_crtc_state *crtc_state = &cstate->base;
> struct drm_atomic_state *state = crtc_state->state;
> struct drm_plane *plane;
> @@ -4181,7 +4181,7 @@ int skl_check_pipe_max_pixel_rate(struct intel_crtc *intel_crtc,
> crtc_clock = crtc_state->adjusted_mode.crtc_clock;
> dotclk = to_intel_atomic_state(state)->cdclk.logical.cdclk;
>
> - if (IS_GEMINILAKE(dev_priv) || INTEL_GEN(dev_priv) >= 10)
> + if (IS_GEMINILAKE(i915) || INTEL_GEN(i915) >= 10)
> dotclk *= 2;
>
> pipe_max_pixel_rate = div_round_up_u32_fixed16(dotclk, pipe_downscale);
> @@ -4341,7 +4341,7 @@ skl_allocate_pipe_ddb(struct intel_crtc_state *cstate,
> {
> struct drm_atomic_state *state = cstate->base.state;
> struct drm_crtc *crtc = cstate->base.crtc;
> - struct drm_i915_private *dev_priv = to_i915(crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc->dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> struct skl_ddb_entry *alloc = &cstate->wm.skl.ddb;
> u16 alloc_size, start = 0;
> @@ -4367,7 +4367,7 @@ skl_allocate_pipe_ddb(struct intel_crtc_state *cstate,
> return 0;
> }
>
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> total_data_rate =
> icl_get_total_relative_data_rate(cstate,
> plane_data_rate);
> @@ -4378,7 +4378,7 @@ skl_allocate_pipe_ddb(struct intel_crtc_state *cstate,
> uv_plane_data_rate);
>
>
> - skl_ddb_get_pipe_allocation_limits(dev_priv, cstate, total_data_rate,
> + skl_ddb_get_pipe_allocation_limits(i915, cstate, total_data_rate,
> ddb, alloc, &num_active);
> alloc_size = skl_ddb_entry_size(alloc);
> if (alloc_size == 0)
> @@ -4398,7 +4398,7 @@ skl_allocate_pipe_ddb(struct intel_crtc_state *cstate,
> * Find the highest watermark level for which we can satisfy the block
> * requirement of active planes.
> */
> - for (level = ilk_wm_max_level(dev_priv); level >= 0; level--) {
> + for (level = ilk_wm_max_level(i915); level >= 0; level--) {
> blocks = 0;
> for_each_plane_id_on_crtc(intel_crtc, plane_id) {
> const struct skl_plane_wm *wm =
> @@ -4484,7 +4484,7 @@ skl_allocate_pipe_ddb(struct intel_crtc_state *cstate,
> continue;
>
> /* Gen11+ uses a separate plane for UV watermarks */
> - WARN_ON(INTEL_GEN(dev_priv) >= 11 && uv_total[plane_id]);
> + WARN_ON(INTEL_GEN(i915) >= 11 && uv_total[plane_id]);
>
> /* Leave disabled planes at (0,0) */
> if (total[plane_id]) {
> @@ -4506,7 +4506,7 @@ skl_allocate_pipe_ddb(struct intel_crtc_state *cstate,
> * all levels as "enabled." Go back now and disable the ones
> * that aren't actually possible.
> */
> - for (level++; level <= ilk_wm_max_level(dev_priv); level++) {
> + for (level++; level <= ilk_wm_max_level(i915); level++) {
> for_each_plane_id_on_crtc(intel_crtc, plane_id) {
> struct skl_plane_wm *wm =
> &cstate->wm.skl.optimal.planes[plane_id];
> @@ -4531,7 +4531,7 @@ skl_allocate_pipe_ddb(struct intel_crtc_state *cstate,
> * Wa_1408961008:icl, ehl
> * Underruns with WM1+ disabled
> */
> - if (IS_GEN(dev_priv, 11) &&
> + if (IS_GEN(i915, 11) &&
> level == 1 && wm->wm[0].plane_en) {
> wm->wm[level].plane_res_b = wm->wm[0].plane_res_b;
> wm->wm[level].plane_res_l = wm->wm[0].plane_res_l;
> @@ -4562,7 +4562,7 @@ skl_allocate_pipe_ddb(struct intel_crtc_state *cstate,
> * 2xcdclk is 1350 MHz and the pixel rate should never exceed that.
> */
> static uint_fixed_16_16_t
> -skl_wm_method1(const struct drm_i915_private *dev_priv, u32 pixel_rate,
> +skl_wm_method1(const struct drm_i915_private *i915, u32 pixel_rate,
> u8 cpp, u32 latency, u32 dbuf_block_size)
> {
> u32 wm_intermediate_val;
> @@ -4574,7 +4574,7 @@ skl_wm_method1(const struct drm_i915_private *dev_priv, u32 pixel_rate,
> wm_intermediate_val = latency * pixel_rate * cpp;
> ret = div_fixed16(wm_intermediate_val, 1000 * dbuf_block_size);
>
> - if (INTEL_GEN(dev_priv) >= 10)
> + if (INTEL_GEN(i915) >= 10)
> ret = add_fixed16_u32(ret, 1);
>
> return ret;
> @@ -4648,7 +4648,7 @@ skl_compute_wm_params(const struct intel_crtc_state *crtc_state,
> int color_plane)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> u32 interm_pbpl;
>
> /* only planar format has two planes */
> @@ -4673,7 +4673,7 @@ skl_compute_wm_params(const struct intel_crtc_state *crtc_state,
> wp->cpp = format->cpp[color_plane];
> wp->plane_pixel_rate = plane_pixel_rate;
>
> - if (INTEL_GEN(dev_priv) >= 11 &&
> + if (INTEL_GEN(i915) >= 11 &&
> modifier == I915_FORMAT_MOD_Yf_TILED && wp->cpp == 1)
> wp->dbuf_block_size = 256;
> else
> @@ -4698,7 +4698,7 @@ skl_compute_wm_params(const struct intel_crtc_state *crtc_state,
> wp->y_min_scanlines = 4;
> }
>
> - if (skl_needs_memory_bw_wa(dev_priv))
> + if (skl_needs_memory_bw_wa(i915))
> wp->y_min_scanlines *= 2;
>
> wp->plane_bytes_per_line = wp->width * wp->cpp;
> @@ -4707,12 +4707,12 @@ skl_compute_wm_params(const struct intel_crtc_state *crtc_state,
> wp->y_min_scanlines,
> wp->dbuf_block_size);
>
> - if (INTEL_GEN(dev_priv) >= 10)
> + if (INTEL_GEN(i915) >= 10)
> interm_pbpl++;
>
> wp->plane_blocks_per_line = div_fixed16(interm_pbpl,
> wp->y_min_scanlines);
> - } else if (wp->x_tiled && IS_GEN(dev_priv, 9)) {
> + } else if (wp->x_tiled && IS_GEN(i915, 9)) {
> interm_pbpl = DIV_ROUND_UP(wp->plane_bytes_per_line,
> wp->dbuf_block_size);
> wp->plane_blocks_per_line = u32_to_fixed16(interm_pbpl);
> @@ -4758,9 +4758,9 @@ skl_compute_plane_wm_params(const struct intel_crtc_state *crtc_state,
> wp, color_plane);
> }
>
> -static bool skl_wm_has_lines(struct drm_i915_private *dev_priv, int level)
> +static bool skl_wm_has_lines(struct drm_i915_private *i915, int level)
> {
> - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> return true;
>
> /* The number of lines are ignored for the level 0 watermark. */
> @@ -4773,8 +4773,8 @@ static void skl_compute_plane_wm(const struct intel_crtc_state *cstate,
> const struct skl_wm_level *result_prev,
> struct skl_wm_level *result /* out */)
> {
> - struct drm_i915_private *dev_priv = to_i915(cstate->base.crtc->dev);
> - u32 latency = dev_priv->wm.skl_latency[level];
> + struct drm_i915_private *i915 = to_i915(cstate->base.crtc->dev);
> + u32 latency = i915->wm.skl_latency[level];
> uint_fixed_16_16_t method1, method2;
> uint_fixed_16_16_t selected_result;
> u32 res_blocks, res_lines, min_ddb_alloc = 0;
> @@ -4789,14 +4789,14 @@ static void skl_compute_plane_wm(const struct intel_crtc_state *cstate,
> * WaIncreaseLatencyIPCEnabled: kbl,cfl
> * Display WA #1141: kbl,cfl
> */
> - if ((IS_KABYLAKE(dev_priv) || IS_COFFEELAKE(dev_priv)) ||
> - dev_priv->ipc_enabled)
> + if ((IS_KABYLAKE(i915) || IS_COFFEELAKE(i915)) ||
> + i915->ipc_enabled)
> latency += 4;
>
> - if (skl_needs_memory_bw_wa(dev_priv) && wp->x_tiled)
> + if (skl_needs_memory_bw_wa(i915) && wp->x_tiled)
> latency += 15;
>
> - method1 = skl_wm_method1(dev_priv, wp->plane_pixel_rate,
> + method1 = skl_wm_method1(i915, wp->plane_pixel_rate,
> wp->cpp, latency, wp->dbuf_block_size);
> method2 = skl_wm_method2(wp->plane_pixel_rate,
> cstate->base.adjusted_mode.crtc_htotal,
> @@ -4811,8 +4811,8 @@ static void skl_compute_plane_wm(const struct intel_crtc_state *cstate,
> (wp->plane_bytes_per_line / wp->dbuf_block_size < 1)) {
> selected_result = method2;
> } else if (latency >= wp->linetime_us) {
> - if (IS_GEN(dev_priv, 9) &&
> - !IS_GEMINILAKE(dev_priv))
> + if (IS_GEN(i915, 9) &&
> + !IS_GEMINILAKE(i915))
> selected_result = min_fixed16(method1, method2);
> else
> selected_result = method2;
> @@ -4825,7 +4825,7 @@ static void skl_compute_plane_wm(const struct intel_crtc_state *cstate,
> res_lines = div_round_up_fixed16(selected_result,
> wp->plane_blocks_per_line);
>
> - if (IS_GEN9_BC(dev_priv) || IS_BROXTON(dev_priv)) {
> + if (IS_GEN9_BC(i915) || IS_BROXTON(i915)) {
> /* Display WA #1125: skl,bxt,kbl */
> if (level == 0 && wp->rc_surface)
> res_blocks +=
> @@ -4852,7 +4852,7 @@ static void skl_compute_plane_wm(const struct intel_crtc_state *cstate,
> }
> }
>
> - if (INTEL_GEN(dev_priv) >= 11) {
> + if (INTEL_GEN(i915) >= 11) {
> if (wp->y_tiled) {
> int extra_lines;
>
> @@ -4870,7 +4870,7 @@ static void skl_compute_plane_wm(const struct intel_crtc_state *cstate,
> }
> }
>
> - if (!skl_wm_has_lines(dev_priv, level))
> + if (!skl_wm_has_lines(i915, level))
> res_lines = 0;
>
> if (res_lines > 31) {
> @@ -4897,8 +4897,8 @@ skl_compute_wm_levels(const struct intel_crtc_state *cstate,
> const struct skl_wm_params *wm_params,
> struct skl_wm_level *levels)
> {
> - struct drm_i915_private *dev_priv = to_i915(cstate->base.crtc->dev);
> - int level, max_level = ilk_wm_max_level(dev_priv);
> + struct drm_i915_private *i915 = to_i915(cstate->base.crtc->dev);
> + int level, max_level = ilk_wm_max_level(i915);
> struct skl_wm_level *result_prev = &levels[0];
>
> for (level = 0; level <= max_level; level++) {
> @@ -4915,7 +4915,7 @@ static u32
> skl_compute_linetime_wm(const struct intel_crtc_state *cstate)
> {
> struct drm_atomic_state *state = cstate->base.state;
> - struct drm_i915_private *dev_priv = to_i915(state->dev);
> + struct drm_i915_private *i915 = to_i915(state->dev);
> uint_fixed_16_16_t linetime_us;
> u32 linetime_wm;
>
> @@ -4923,7 +4923,7 @@ skl_compute_linetime_wm(const struct intel_crtc_state *cstate)
> linetime_wm = fixed16_to_u32_round_up(mul_u32_fixed16(8, linetime_us));
>
> /* Display WA #1135: BXT:ALL GLK:ALL */
> - if (IS_GEN9_LP(dev_priv) && dev_priv->ipc_enabled)
> + if (IS_GEN9_LP(i915) && i915->ipc_enabled)
> linetime_wm /= 2;
>
> return linetime_wm;
> @@ -4934,21 +4934,21 @@ static void skl_compute_transition_wm(const struct intel_crtc_state *cstate,
> struct skl_plane_wm *wm)
> {
> struct drm_device *dev = cstate->base.crtc->dev;
> - const struct drm_i915_private *dev_priv = to_i915(dev);
> + const struct drm_i915_private *i915 = to_i915(dev);
> u16 trans_min, trans_y_tile_min;
> const u16 trans_amount = 10; /* This is configurable amount */
> u16 wm0_sel_res_b, trans_offset_b, res_blocks;
>
> /* Transition WM are not recommended by HW team for GEN9 */
> - if (INTEL_GEN(dev_priv) <= 9)
> + if (INTEL_GEN(i915) <= 9)
> return;
>
> /* Transition WM don't make any sense if ipc is disabled */
> - if (!dev_priv->ipc_enabled)
> + if (!i915->ipc_enabled)
> return;
>
> trans_min = 14;
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> trans_min = 4;
>
> trans_offset_b = trans_min + trans_amount;
> @@ -4974,7 +4974,7 @@ static void skl_compute_transition_wm(const struct intel_crtc_state *cstate,
> res_blocks = wm0_sel_res_b + trans_offset_b;
>
> /* WA BUG:1938466 add one block for non y-tile planes */
> - if (IS_CNL_REVID(dev_priv, CNL_REVID_A0, CNL_REVID_A0))
> + if (IS_CNL_REVID(i915, CNL_REVID_A0, CNL_REVID_A0))
> res_blocks += 1;
>
> }
> @@ -5093,7 +5093,7 @@ static int icl_build_plane_wm(struct intel_crtc_state *crtc_state,
>
> static int skl_build_pipe_wm(struct intel_crtc_state *cstate)
> {
> - struct drm_i915_private *dev_priv = to_i915(cstate->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(cstate->base.crtc->dev);
> struct skl_pipe_wm *pipe_wm = &cstate->wm.skl.optimal;
> struct drm_crtc_state *crtc_state = &cstate->base;
> struct drm_plane *plane;
> @@ -5110,7 +5110,7 @@ static int skl_build_pipe_wm(struct intel_crtc_state *cstate)
> const struct intel_plane_state *intel_pstate =
> to_intel_plane_state(pstate);
>
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> ret = icl_build_plane_wm(cstate, intel_pstate);
> else
> ret = skl_build_plane_wm(cstate, intel_pstate);
> @@ -5123,7 +5123,7 @@ static int skl_build_pipe_wm(struct intel_crtc_state *cstate)
> return 0;
> }
>
> -static void skl_ddb_entry_write(struct drm_i915_private *dev_priv,
> +static void skl_ddb_entry_write(struct drm_i915_private *i915,
> i915_reg_t reg,
> const struct skl_ddb_entry *entry)
> {
> @@ -5133,7 +5133,7 @@ static void skl_ddb_entry_write(struct drm_i915_private *dev_priv,
> I915_WRITE_FW(reg, 0);
> }
>
> -static void skl_write_wm_level(struct drm_i915_private *dev_priv,
> +static void skl_write_wm_level(struct drm_i915_private *i915,
> i915_reg_t reg,
> const struct skl_wm_level *level)
> {
> @@ -5152,8 +5152,8 @@ static void skl_write_wm_level(struct drm_i915_private *dev_priv,
> void skl_write_plane_wm(struct intel_plane *plane,
> const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> - int level, max_level = ilk_wm_max_level(dev_priv);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> + int level, max_level = ilk_wm_max_level(i915);
> enum plane_id plane_id = plane->id;
> enum pipe pipe = plane->pipe;
> const struct skl_plane_wm *wm =
> @@ -5164,14 +5164,14 @@ void skl_write_plane_wm(struct intel_plane *plane,
> &crtc_state->wm.skl.plane_ddb_uv[plane_id];
>
> for (level = 0; level <= max_level; level++) {
> - skl_write_wm_level(dev_priv, PLANE_WM(pipe, plane_id, level),
> + skl_write_wm_level(i915, PLANE_WM(pipe, plane_id, level),
> &wm->wm[level]);
> }
> - skl_write_wm_level(dev_priv, PLANE_WM_TRANS(pipe, plane_id),
> + skl_write_wm_level(i915, PLANE_WM_TRANS(pipe, plane_id),
> &wm->trans_wm);
>
> - if (INTEL_GEN(dev_priv) >= 11) {
> - skl_ddb_entry_write(dev_priv,
> + if (INTEL_GEN(i915) >= 11) {
> + skl_ddb_entry_write(i915,
> PLANE_BUF_CFG(pipe, plane_id), ddb_y);
> return;
> }
> @@ -5179,17 +5179,17 @@ void skl_write_plane_wm(struct intel_plane *plane,
> if (wm->is_planar)
> swap(ddb_y, ddb_uv);
>
> - skl_ddb_entry_write(dev_priv,
> + skl_ddb_entry_write(i915,
> PLANE_BUF_CFG(pipe, plane_id), ddb_y);
> - skl_ddb_entry_write(dev_priv,
> + skl_ddb_entry_write(i915,
> PLANE_NV12_BUF_CFG(pipe, plane_id), ddb_uv);
> }
>
> void skl_write_cursor_wm(struct intel_plane *plane,
> const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> - int level, max_level = ilk_wm_max_level(dev_priv);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> + int level, max_level = ilk_wm_max_level(i915);
> enum plane_id plane_id = plane->id;
> enum pipe pipe = plane->pipe;
> const struct skl_plane_wm *wm =
> @@ -5198,12 +5198,12 @@ void skl_write_cursor_wm(struct intel_plane *plane,
> &crtc_state->wm.skl.plane_ddb_y[plane_id];
>
> for (level = 0; level <= max_level; level++) {
> - skl_write_wm_level(dev_priv, CUR_WM(pipe, level),
> + skl_write_wm_level(i915, CUR_WM(pipe, level),
> &wm->wm[level]);
> }
> - skl_write_wm_level(dev_priv, CUR_WM_TRANS(pipe), &wm->trans_wm);
> + skl_write_wm_level(i915, CUR_WM_TRANS(pipe), &wm->trans_wm);
>
> - skl_ddb_entry_write(dev_priv, CUR_BUF_CFG(pipe), ddb);
> + skl_ddb_entry_write(i915, CUR_BUF_CFG(pipe), ddb);
> }
>
> bool skl_wm_level_equals(const struct skl_wm_level *l1,
> @@ -5215,11 +5215,11 @@ bool skl_wm_level_equals(const struct skl_wm_level *l1,
> l1->plane_res_b == l2->plane_res_b;
> }
>
> -static bool skl_plane_wm_equals(struct drm_i915_private *dev_priv,
> +static bool skl_plane_wm_equals(struct drm_i915_private *i915,
> const struct skl_plane_wm *wm1,
> const struct skl_plane_wm *wm2)
> {
> - int level, max_level = ilk_wm_max_level(dev_priv);
> + int level, max_level = ilk_wm_max_level(i915);
>
> for (level = 0; level <= max_level; level++) {
> if (!skl_wm_level_equals(&wm1->wm[level], &wm2->wm[level]) ||
> @@ -5234,11 +5234,11 @@ static bool skl_pipe_wm_equals(struct intel_crtc *crtc,
> const struct skl_pipe_wm *wm1,
> const struct skl_pipe_wm *wm2)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum plane_id plane_id;
>
> for_each_plane_id_on_crtc(crtc, plane_id) {
> - if (!skl_plane_wm_equals(dev_priv,
> + if (!skl_plane_wm_equals(i915,
> &wm1->planes[plane_id],
> &wm2->planes[plane_id]))
> return false;
> @@ -5287,10 +5287,10 @@ skl_ddb_add_affected_planes(const struct intel_crtc_state *old_crtc_state,
> {
> struct intel_atomic_state *state = to_intel_atomic_state(new_crtc_state->base.state);
> struct intel_crtc *crtc = to_intel_crtc(new_crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> struct intel_plane *plane;
>
> - for_each_intel_plane_on_crtc(&dev_priv->drm, crtc, plane) {
> + for_each_intel_plane_on_crtc(&i915->drm, crtc, plane) {
> struct intel_plane_state *plane_state;
> enum plane_id plane_id = plane->id;
>
> @@ -5313,14 +5313,14 @@ skl_ddb_add_affected_planes(const struct intel_crtc_state *old_crtc_state,
> static int
> skl_compute_ddb(struct intel_atomic_state *state)
> {
> - const struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> + const struct drm_i915_private *i915 = to_i915(state->base.dev);
> struct skl_ddb_allocation *ddb = &state->wm_results.ddb;
> struct intel_crtc_state *old_crtc_state;
> struct intel_crtc_state *new_crtc_state;
> struct intel_crtc *crtc;
> int ret, i;
>
> - memcpy(ddb, &dev_priv->wm.skl_hw.ddb, sizeof(*ddb));
> + memcpy(ddb, &i915->wm.skl_hw.ddb, sizeof(*ddb));
>
> for_each_oldnew_intel_crtc_in_state(state, crtc, old_crtc_state,
> new_crtc_state, i) {
> @@ -5345,7 +5345,7 @@ static char enast(bool enable)
> static void
> skl_print_wm_changes(struct intel_atomic_state *state)
> {
> - struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> + struct drm_i915_private *i915 = to_i915(state->base.dev);
> const struct intel_crtc_state *old_crtc_state;
> const struct intel_crtc_state *new_crtc_state;
> struct intel_plane *plane;
> @@ -5362,7 +5362,7 @@ skl_print_wm_changes(struct intel_atomic_state *state)
> old_pipe_wm = &old_crtc_state->wm.skl.optimal;
> new_pipe_wm = &new_crtc_state->wm.skl.optimal;
>
> - for_each_intel_plane_on_crtc(&dev_priv->drm, crtc, plane) {
> + for_each_intel_plane_on_crtc(&i915->drm, crtc, plane) {
> enum plane_id plane_id = plane->id;
> const struct skl_ddb_entry *old, *new;
>
> @@ -5378,14 +5378,14 @@ skl_print_wm_changes(struct intel_atomic_state *state)
> skl_ddb_entry_size(old), skl_ddb_entry_size(new));
> }
>
> - for_each_intel_plane_on_crtc(&dev_priv->drm, crtc, plane) {
> + for_each_intel_plane_on_crtc(&i915->drm, crtc, plane) {
> enum plane_id plane_id = plane->id;
> const struct skl_plane_wm *old_wm, *new_wm;
>
> old_wm = &old_pipe_wm->planes[plane_id];
> new_wm = &new_pipe_wm->planes[plane_id];
>
> - if (skl_plane_wm_equals(dev_priv, old_wm, new_wm))
> + if (skl_plane_wm_equals(i915, old_wm, new_wm))
> continue;
>
> DRM_DEBUG_KMS("[PLANE:%d:%s] level %cwm0,%cwm1,%cwm2,%cwm3,%cwm4,%cwm5,%cwm6,%cwm7,%ctwm"
> @@ -5460,7 +5460,7 @@ static int
> skl_ddb_add_affected_pipes(struct intel_atomic_state *state, bool *changed)
> {
> struct drm_device *dev = state->base.dev;
> - const struct drm_i915_private *dev_priv = to_i915(dev);
> + const struct drm_i915_private *i915 = to_i915(dev);
> struct intel_crtc *crtc;
> struct intel_crtc_state *crtc_state;
> u32 realloc_pipes = pipes_modified(state);
> @@ -5470,14 +5470,14 @@ skl_ddb_add_affected_pipes(struct intel_atomic_state *state, bool *changed)
> * When we distrust bios wm we always need to recompute to set the
> * expected DDB allocations for each CRTC.
> */
> - if (dev_priv->wm.distrust_bios_wm)
> + if (i915->wm.distrust_bios_wm)
> (*changed) = true;
>
> /*
> * If this transaction isn't actually touching any CRTC's, don't
> * bother with watermark calculation. Note that if we pass this
> * test, we're guaranteed to hold at least one CRTC state mutex,
> - * which means we can safely use values like dev_priv->active_crtcs
> + * which means we can safely use values like i915->active_crtcs
> * since any racing commits that want to update them would need to
> * hold _all_ CRTC state mutexes.
> */
> @@ -5493,7 +5493,7 @@ skl_ddb_add_affected_pipes(struct intel_atomic_state *state, bool *changed)
> * pretend that all pipes switched active status so that we'll
> * ensure a full DDB recompute.
> */
> - if (dev_priv->wm.distrust_bios_wm) {
> + if (i915->wm.distrust_bios_wm) {
> ret = drm_modeset_lock(&dev->mode_config.connection_mutex,
> state->base.acquire_ctx);
> if (ret)
> @@ -5508,7 +5508,7 @@ skl_ddb_add_affected_pipes(struct intel_atomic_state *state, bool *changed)
> * on the first commit too.
> */
> if (!state->modeset)
> - state->active_crtcs = dev_priv->active_crtcs;
> + state->active_crtcs = i915->active_crtcs;
> }
>
> /*
> @@ -5567,14 +5567,14 @@ skl_ddb_add_affected_pipes(struct intel_atomic_state *state, bool *changed)
> static int skl_wm_add_affected_planes(struct intel_atomic_state *state,
> struct intel_crtc *crtc)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> const struct intel_crtc_state *old_crtc_state =
> intel_atomic_get_old_crtc_state(state, crtc);
> struct intel_crtc_state *new_crtc_state =
> intel_atomic_get_new_crtc_state(state, crtc);
> struct intel_plane *plane;
>
> - for_each_intel_plane_on_crtc(&dev_priv->drm, crtc, plane) {
> + for_each_intel_plane_on_crtc(&i915->drm, crtc, plane) {
> struct intel_plane_state *plane_state;
> enum plane_id plane_id = plane->id;
>
> @@ -5587,7 +5587,7 @@ static int skl_wm_add_affected_planes(struct intel_atomic_state *state,
> * with the software state.
> */
> if (!drm_atomic_crtc_needs_modeset(&new_crtc_state->base) &&
> - skl_plane_wm_equals(dev_priv,
> + skl_plane_wm_equals(i915,
> &old_crtc_state->wm.skl.optimal.planes[plane_id],
> &new_crtc_state->wm.skl.optimal.planes[plane_id]))
> continue;
> @@ -5654,7 +5654,7 @@ static void skl_atomic_update_crtc_wm(struct intel_atomic_state *state,
> struct intel_crtc_state *cstate)
> {
> struct intel_crtc *crtc = to_intel_crtc(cstate->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> + struct drm_i915_private *i915 = to_i915(state->base.dev);
> struct skl_pipe_wm *pipe_wm = &cstate->wm.skl.optimal;
> enum pipe pipe = crtc->pipe;
>
> @@ -5669,27 +5669,27 @@ static void skl_initial_wm(struct intel_atomic_state *state,
> {
> struct intel_crtc *intel_crtc = to_intel_crtc(cstate->base.crtc);
> struct drm_device *dev = intel_crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct skl_ddb_values *results = &state->wm_results;
>
> if ((results->dirty_pipes & drm_crtc_mask(&intel_crtc->base)) == 0)
> return;
>
> - mutex_lock(&dev_priv->wm.wm_mutex);
> + mutex_lock(&i915->wm.wm_mutex);
>
> if (cstate->base.active_changed)
> skl_atomic_update_crtc_wm(state, cstate);
>
> - mutex_unlock(&dev_priv->wm.wm_mutex);
> + mutex_unlock(&i915->wm.wm_mutex);
> }
>
> -static void ilk_compute_wm_config(struct drm_i915_private *dev_priv,
> +static void ilk_compute_wm_config(struct drm_i915_private *i915,
> struct intel_wm_config *config)
> {
> struct intel_crtc *crtc;
>
> /* Compute the currently _active_ config */
> - for_each_intel_crtc(&dev_priv->drm, crtc) {
> + for_each_intel_crtc(&i915->drm, crtc) {
> const struct intel_pipe_wm *wm = &crtc->wm.active.ilk;
>
> if (!wm->pipe_enabled)
> @@ -5701,7 +5701,7 @@ static void ilk_compute_wm_config(struct drm_i915_private *dev_priv,
> }
> }
>
> -static void ilk_program_watermarks(struct drm_i915_private *dev_priv)
> +static void ilk_program_watermarks(struct drm_i915_private *i915)
> {
> struct intel_pipe_wm lp_wm_1_2 = {}, lp_wm_5_6 = {}, *best_lp_wm;
> struct ilk_wm_maximums max;
> @@ -5709,18 +5709,18 @@ static void ilk_program_watermarks(struct drm_i915_private *dev_priv)
> struct ilk_wm_values results = {};
> enum intel_ddb_partitioning partitioning;
>
> - ilk_compute_wm_config(dev_priv, &config);
> + ilk_compute_wm_config(i915, &config);
>
> - ilk_compute_wm_maximums(dev_priv, 1, &config, INTEL_DDB_PART_1_2, &max);
> - ilk_wm_merge(dev_priv, &config, &max, &lp_wm_1_2);
> + ilk_compute_wm_maximums(i915, 1, &config, INTEL_DDB_PART_1_2, &max);
> + ilk_wm_merge(i915, &config, &max, &lp_wm_1_2);
>
> /* 5/6 split only in single pipe config on IVB+ */
> - if (INTEL_GEN(dev_priv) >= 7 &&
> + if (INTEL_GEN(i915) >= 7 &&
> config.num_pipes_active == 1 && config.sprites_enabled) {
> - ilk_compute_wm_maximums(dev_priv, 1, &config, INTEL_DDB_PART_5_6, &max);
> - ilk_wm_merge(dev_priv, &config, &max, &lp_wm_5_6);
> + ilk_compute_wm_maximums(i915, 1, &config, INTEL_DDB_PART_5_6, &max);
> + ilk_wm_merge(i915, &config, &max, &lp_wm_5_6);
>
> - best_lp_wm = ilk_find_best_result(dev_priv, &lp_wm_1_2, &lp_wm_5_6);
> + best_lp_wm = ilk_find_best_result(i915, &lp_wm_1_2, &lp_wm_5_6);
> } else {
> best_lp_wm = &lp_wm_1_2;
> }
> @@ -5728,35 +5728,35 @@ static void ilk_program_watermarks(struct drm_i915_private *dev_priv)
> partitioning = (best_lp_wm == &lp_wm_1_2) ?
> INTEL_DDB_PART_1_2 : INTEL_DDB_PART_5_6;
>
> - ilk_compute_wm_results(dev_priv, best_lp_wm, partitioning, &results);
> + ilk_compute_wm_results(i915, best_lp_wm, partitioning, &results);
>
> - ilk_write_wm_values(dev_priv, &results);
> + ilk_write_wm_values(i915, &results);
> }
>
> static void ilk_initial_watermarks(struct intel_atomic_state *state,
> struct intel_crtc_state *cstate)
> {
> - struct drm_i915_private *dev_priv = to_i915(cstate->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(cstate->base.crtc->dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(cstate->base.crtc);
>
> - mutex_lock(&dev_priv->wm.wm_mutex);
> + mutex_lock(&i915->wm.wm_mutex);
> intel_crtc->wm.active.ilk = cstate->wm.ilk.intermediate;
> - ilk_program_watermarks(dev_priv);
> - mutex_unlock(&dev_priv->wm.wm_mutex);
> + ilk_program_watermarks(i915);
> + mutex_unlock(&i915->wm.wm_mutex);
> }
>
> static void ilk_optimize_watermarks(struct intel_atomic_state *state,
> struct intel_crtc_state *cstate)
> {
> - struct drm_i915_private *dev_priv = to_i915(cstate->base.crtc->dev);
> + struct drm_i915_private *i915 = to_i915(cstate->base.crtc->dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(cstate->base.crtc);
>
> - mutex_lock(&dev_priv->wm.wm_mutex);
> + mutex_lock(&i915->wm.wm_mutex);
> if (cstate->wm.need_postvbl_update) {
> intel_crtc->wm.active.ilk = cstate->wm.ilk.optimal;
> - ilk_program_watermarks(dev_priv);
> + ilk_program_watermarks(i915);
> }
> - mutex_unlock(&dev_priv->wm.wm_mutex);
> + mutex_unlock(&i915->wm.wm_mutex);
> }
>
> static inline void skl_wm_level_from_reg_val(u32 val,
> @@ -5772,13 +5772,13 @@ static inline void skl_wm_level_from_reg_val(u32 val,
> void skl_pipe_wm_get_hw_state(struct intel_crtc *crtc,
> struct skl_pipe_wm *out)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
> int level, max_level;
> enum plane_id plane_id;
> u32 val;
>
> - max_level = ilk_wm_max_level(dev_priv);
> + max_level = ilk_wm_max_level(i915);
>
> for_each_plane_id_on_crtc(crtc, plane_id) {
> struct skl_plane_wm *wm = &out->planes[plane_id];
> @@ -5806,15 +5806,15 @@ void skl_pipe_wm_get_hw_state(struct intel_crtc *crtc,
> out->linetime = I915_READ(PIPE_WM_LINETIME(pipe));
> }
>
> -void skl_wm_get_hw_state(struct drm_i915_private *dev_priv)
> +void skl_wm_get_hw_state(struct drm_i915_private *i915)
> {
> - struct skl_ddb_values *hw = &dev_priv->wm.skl_hw;
> - struct skl_ddb_allocation *ddb = &dev_priv->wm.skl_hw.ddb;
> + struct skl_ddb_values *hw = &i915->wm.skl_hw;
> + struct skl_ddb_allocation *ddb = &i915->wm.skl_hw.ddb;
> struct intel_crtc *crtc;
> struct intel_crtc_state *cstate;
>
> - skl_ddb_get_hw_state(dev_priv, ddb);
> - for_each_intel_crtc(&dev_priv->drm, crtc) {
> + skl_ddb_get_hw_state(i915, ddb);
> + for_each_intel_crtc(&i915->drm, crtc) {
> cstate = to_intel_crtc_state(crtc->base.state);
>
> skl_pipe_wm_get_hw_state(crtc, &cstate->wm.skl.optimal);
> @@ -5823,17 +5823,17 @@ void skl_wm_get_hw_state(struct drm_i915_private *dev_priv)
> hw->dirty_pipes |= drm_crtc_mask(&crtc->base);
> }
>
> - if (dev_priv->active_crtcs) {
> + if (i915->active_crtcs) {
> /* Fully recompute DDB on first atomic commit */
> - dev_priv->wm.distrust_bios_wm = true;
> + i915->wm.distrust_bios_wm = true;
> }
> }
>
> static void ilk_pipe_wm_get_hw_state(struct intel_crtc *crtc)
> {
> struct drm_device *dev = crtc->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct ilk_wm_values *hw = &dev_priv->wm.hw;
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct ilk_wm_values *hw = &i915->wm.hw;
> struct intel_crtc_state *cstate = to_intel_crtc_state(crtc->base.state);
> struct intel_pipe_wm *active = &cstate->wm.ilk.optimal;
> enum pipe pipe = crtc->pipe;
> @@ -5844,7 +5844,7 @@ static void ilk_pipe_wm_get_hw_state(struct intel_crtc *crtc)
> };
>
> hw->wm_pipe[pipe] = I915_READ(wm0_pipe_reg[pipe]);
> - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> + if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> hw->wm_linetime[pipe] = I915_READ(PIPE_WM_LINETIME(pipe));
>
> memset(active, 0, sizeof(*active));
> @@ -5866,7 +5866,7 @@ static void ilk_pipe_wm_get_hw_state(struct intel_crtc *crtc)
> active->wm[0].cur_val = tmp & WM0_PIPE_CURSOR_MASK;
> active->linetime = hw->wm_linetime[pipe];
> } else {
> - int level, max_level = ilk_wm_max_level(dev_priv);
> + int level, max_level = ilk_wm_max_level(i915);
>
> /*
> * For inactive pipes, all watermark levels
> @@ -5885,7 +5885,7 @@ static void ilk_pipe_wm_get_hw_state(struct intel_crtc *crtc)
> #define _FW_WM_VLV(value, plane) \
> (((value) & DSPFW_ ## plane ## _MASK_VLV) >> DSPFW_ ## plane ## _SHIFT)
>
> -static void g4x_read_wm_values(struct drm_i915_private *dev_priv,
> +static void g4x_read_wm_values(struct drm_i915_private *i915,
> struct g4x_wm_values *wm)
> {
> u32 tmp;
> @@ -5911,13 +5911,13 @@ static void g4x_read_wm_values(struct drm_i915_private *dev_priv,
> wm->hpll.plane = _FW_WM(tmp, HPLL_SR);
> }
>
> -static void vlv_read_wm_values(struct drm_i915_private *dev_priv,
> +static void vlv_read_wm_values(struct drm_i915_private *i915,
> struct vlv_wm_values *wm)
> {
> enum pipe pipe;
> u32 tmp;
>
> - for_each_pipe(dev_priv, pipe) {
> + for_each_pipe(i915, pipe) {
> tmp = I915_READ(VLV_DDL(pipe));
>
> wm->ddl[pipe].plane[PLANE_PRIMARY] =
> @@ -5944,7 +5944,7 @@ static void vlv_read_wm_values(struct drm_i915_private *dev_priv,
> tmp = I915_READ(DSPFW3);
> wm->sr.cursor = _FW_WM(tmp, CURSOR_SR);
>
> - if (IS_CHERRYVIEW(dev_priv)) {
> + if (IS_CHERRYVIEW(i915)) {
> tmp = I915_READ(DSPFW7_CHV);
> wm->pipe[PIPE_B].plane[PLANE_SPRITE1] = _FW_WM_VLV(tmp, SPRITED);
> wm->pipe[PIPE_B].plane[PLANE_SPRITE0] = _FW_WM_VLV(tmp, SPRITEC);
> @@ -5987,16 +5987,16 @@ static void vlv_read_wm_values(struct drm_i915_private *dev_priv,
> #undef _FW_WM
> #undef _FW_WM_VLV
>
> -void g4x_wm_get_hw_state(struct drm_i915_private *dev_priv)
> +void g4x_wm_get_hw_state(struct drm_i915_private *i915)
> {
> - struct g4x_wm_values *wm = &dev_priv->wm.g4x;
> + struct g4x_wm_values *wm = &i915->wm.g4x;
> struct intel_crtc *crtc;
>
> - g4x_read_wm_values(dev_priv, wm);
> + g4x_read_wm_values(i915, wm);
>
> wm->cxsr = I915_READ(FW_BLC_SELF) & FW_BLC_SELF_EN;
>
> - for_each_intel_crtc(&dev_priv->drm, crtc) {
> + for_each_intel_crtc(&i915->drm, crtc) {
> struct intel_crtc_state *crtc_state =
> to_intel_crtc_state(crtc->base.state);
> struct g4x_wm_state *active = &crtc->wm.active.g4x;
> @@ -6071,16 +6071,16 @@ void g4x_wm_get_hw_state(struct drm_i915_private *dev_priv)
> yesno(wm->cxsr), yesno(wm->hpll_en), yesno(wm->fbc_en));
> }
>
> -void g4x_wm_sanitize(struct drm_i915_private *dev_priv)
> +void g4x_wm_sanitize(struct drm_i915_private *i915)
> {
> struct intel_plane *plane;
> struct intel_crtc *crtc;
>
> - mutex_lock(&dev_priv->wm.wm_mutex);
> + mutex_lock(&i915->wm.wm_mutex);
>
> - for_each_intel_plane(&dev_priv->drm, plane) {
> + for_each_intel_plane(&i915->drm, plane) {
> struct intel_crtc *crtc =
> - intel_get_crtc_for_pipe(dev_priv, plane->pipe);
> + intel_get_crtc_for_pipe(i915, plane->pipe);
> struct intel_crtc_state *crtc_state =
> to_intel_crtc_state(crtc->base.state);
> struct intel_plane_state *plane_state =
> @@ -6113,7 +6113,7 @@ void g4x_wm_sanitize(struct drm_i915_private *dev_priv)
> }
> }
>
> - for_each_intel_crtc(&dev_priv->drm, crtc) {
> + for_each_intel_crtc(&i915->drm, crtc) {
> struct intel_crtc_state *crtc_state =
> to_intel_crtc_state(crtc->base.state);
>
> @@ -6122,26 +6122,26 @@ void g4x_wm_sanitize(struct drm_i915_private *dev_priv)
> crtc->wm.active.g4x = crtc_state->wm.g4x.optimal;
> }
>
> - g4x_program_watermarks(dev_priv);
> + g4x_program_watermarks(i915);
>
> - mutex_unlock(&dev_priv->wm.wm_mutex);
> + mutex_unlock(&i915->wm.wm_mutex);
> }
>
> -void vlv_wm_get_hw_state(struct drm_i915_private *dev_priv)
> +void vlv_wm_get_hw_state(struct drm_i915_private *i915)
> {
> - struct vlv_wm_values *wm = &dev_priv->wm.vlv;
> + struct vlv_wm_values *wm = &i915->wm.vlv;
> struct intel_crtc *crtc;
> u32 val;
>
> - vlv_read_wm_values(dev_priv, wm);
> + vlv_read_wm_values(i915, wm);
>
> wm->cxsr = I915_READ(FW_BLC_SELF_VLV) & FW_CSPWRDWNEN;
> wm->level = VLV_WM_LEVEL_PM2;
>
> - if (IS_CHERRYVIEW(dev_priv)) {
> - vlv_punit_get(dev_priv);
> + if (IS_CHERRYVIEW(i915)) {
> + vlv_punit_get(i915);
>
> - val = vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM);
> + val = vlv_punit_read(i915, PUNIT_REG_DSPSSPM);
> if (val & DSP_MAXFIFO_PM5_ENABLE)
> wm->level = VLV_WM_LEVEL_PM5;
>
> @@ -6154,25 +6154,25 @@ void vlv_wm_get_hw_state(struct drm_i915_private *dev_priv)
> * HIGH/LOW bits so that we don't actually change
> * the current state.
> */
> - val = vlv_punit_read(dev_priv, PUNIT_REG_DDR_SETUP2);
> + val = vlv_punit_read(i915, PUNIT_REG_DDR_SETUP2);
> val |= FORCE_DDR_FREQ_REQ_ACK;
> - vlv_punit_write(dev_priv, PUNIT_REG_DDR_SETUP2, val);
> + vlv_punit_write(i915, PUNIT_REG_DDR_SETUP2, val);
>
> - if (wait_for((vlv_punit_read(dev_priv, PUNIT_REG_DDR_SETUP2) &
> + if (wait_for((vlv_punit_read(i915, PUNIT_REG_DDR_SETUP2) &
> FORCE_DDR_FREQ_REQ_ACK) == 0, 3)) {
> DRM_DEBUG_KMS("Punit not acking DDR DVFS request, "
> "assuming DDR DVFS is disabled\n");
> - dev_priv->wm.max_level = VLV_WM_LEVEL_PM5;
> + i915->wm.max_level = VLV_WM_LEVEL_PM5;
> } else {
> - val = vlv_punit_read(dev_priv, PUNIT_REG_DDR_SETUP2);
> + val = vlv_punit_read(i915, PUNIT_REG_DDR_SETUP2);
> if ((val & FORCE_DDR_HIGH_FREQ) == 0)
> wm->level = VLV_WM_LEVEL_DDR_DVFS;
> }
>
> - vlv_punit_put(dev_priv);
> + vlv_punit_put(i915);
> }
>
> - for_each_intel_crtc(&dev_priv->drm, crtc) {
> + for_each_intel_crtc(&i915->drm, crtc) {
> struct intel_crtc_state *crtc_state =
> to_intel_crtc_state(crtc->base.state);
> struct vlv_wm_state *active = &crtc->wm.active.vlv;
> @@ -6224,16 +6224,16 @@ void vlv_wm_get_hw_state(struct drm_i915_private *dev_priv)
> wm->sr.plane, wm->sr.cursor, wm->level, wm->cxsr);
> }
>
> -void vlv_wm_sanitize(struct drm_i915_private *dev_priv)
> +void vlv_wm_sanitize(struct drm_i915_private *i915)
> {
> struct intel_plane *plane;
> struct intel_crtc *crtc;
>
> - mutex_lock(&dev_priv->wm.wm_mutex);
> + mutex_lock(&i915->wm.wm_mutex);
>
> - for_each_intel_plane(&dev_priv->drm, plane) {
> + for_each_intel_plane(&i915->drm, plane) {
> struct intel_crtc *crtc =
> - intel_get_crtc_for_pipe(dev_priv, plane->pipe);
> + intel_get_crtc_for_pipe(i915, plane->pipe);
> struct intel_crtc_state *crtc_state =
> to_intel_crtc_state(crtc->base.state);
> struct intel_plane_state *plane_state =
> @@ -6259,7 +6259,7 @@ void vlv_wm_sanitize(struct drm_i915_private *dev_priv)
> }
> }
>
> - for_each_intel_crtc(&dev_priv->drm, crtc) {
> + for_each_intel_crtc(&i915->drm, crtc) {
> struct intel_crtc_state *crtc_state =
> to_intel_crtc_state(crtc->base.state);
>
> @@ -6268,16 +6268,16 @@ void vlv_wm_sanitize(struct drm_i915_private *dev_priv)
> crtc->wm.active.vlv = crtc_state->wm.vlv.optimal;
> }
>
> - vlv_program_watermarks(dev_priv);
> + vlv_program_watermarks(i915);
>
> - mutex_unlock(&dev_priv->wm.wm_mutex);
> + mutex_unlock(&i915->wm.wm_mutex);
> }
>
> /*
> * FIXME should probably kill this and improve
> * the real watermark readout/sanitation instead
> */
> -static void ilk_init_lp_watermarks(struct drm_i915_private *dev_priv)
> +static void ilk_init_lp_watermarks(struct drm_i915_private *i915)
> {
> I915_WRITE(WM3_LP_ILK, I915_READ(WM3_LP_ILK) & ~WM1_LP_SR_EN);
> I915_WRITE(WM2_LP_ILK, I915_READ(WM2_LP_ILK) & ~WM1_LP_SR_EN);
> @@ -6289,14 +6289,14 @@ static void ilk_init_lp_watermarks(struct drm_i915_private *dev_priv)
> */
> }
>
> -void ilk_wm_get_hw_state(struct drm_i915_private *dev_priv)
> +void ilk_wm_get_hw_state(struct drm_i915_private *i915)
> {
> - struct ilk_wm_values *hw = &dev_priv->wm.hw;
> + struct ilk_wm_values *hw = &i915->wm.hw;
> struct intel_crtc *crtc;
>
> - ilk_init_lp_watermarks(dev_priv);
> + ilk_init_lp_watermarks(i915);
>
> - for_each_intel_crtc(&dev_priv->drm, crtc)
> + for_each_intel_crtc(&i915->drm, crtc)
> ilk_pipe_wm_get_hw_state(crtc);
>
> hw->wm_lp[0] = I915_READ(WM1_LP_ILK);
> @@ -6304,15 +6304,15 @@ void ilk_wm_get_hw_state(struct drm_i915_private *dev_priv)
> hw->wm_lp[2] = I915_READ(WM3_LP_ILK);
>
> hw->wm_lp_spr[0] = I915_READ(WM1S_LP_ILK);
> - if (INTEL_GEN(dev_priv) >= 7) {
> + if (INTEL_GEN(i915) >= 7) {
> hw->wm_lp_spr[1] = I915_READ(WM2S_LP_IVB);
> hw->wm_lp_spr[2] = I915_READ(WM3S_LP_IVB);
> }
>
> - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> + if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> hw->partitioning = (I915_READ(WM_MISC) & WM_MISC_DATA_PARTITION_5_6) ?
> INTEL_DDB_PART_5_6 : INTEL_DDB_PART_1_2;
> - else if (IS_IVYBRIDGE(dev_priv))
> + else if (IS_IVYBRIDGE(i915))
> hw->partitioning = (I915_READ(DISP_ARB_CTL2) & DISP_DATA_PARTITION_5_6) ?
> INTEL_DDB_PART_5_6 : INTEL_DDB_PART_1_2;
>
> @@ -6355,22 +6355,22 @@ void ilk_wm_get_hw_state(struct drm_i915_private *dev_priv)
> */
> void intel_update_watermarks(struct intel_crtc *crtc)
> {
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
>
> - if (dev_priv->display.update_wm)
> - dev_priv->display.update_wm(crtc);
> + if (i915->display.update_wm)
> + i915->display.update_wm(crtc);
> }
>
> -void intel_enable_ipc(struct drm_i915_private *dev_priv)
> +void intel_enable_ipc(struct drm_i915_private *i915)
> {
> u32 val;
>
> - if (!HAS_IPC(dev_priv))
> + if (!HAS_IPC(i915))
> return;
>
> val = I915_READ(DISP_ARB_CTL2);
>
> - if (dev_priv->ipc_enabled)
> + if (i915->ipc_enabled)
> val |= DISP_IPC_ENABLE;
> else
> val &= ~DISP_IPC_ENABLE;
> @@ -6378,27 +6378,27 @@ void intel_enable_ipc(struct drm_i915_private *dev_priv)
> I915_WRITE(DISP_ARB_CTL2, val);
> }
>
> -static bool intel_can_enable_ipc(struct drm_i915_private *dev_priv)
> +static bool intel_can_enable_ipc(struct drm_i915_private *i915)
> {
> /* Display WA #0477 WaDisableIPC: skl */
> - if (IS_SKYLAKE(dev_priv))
> + if (IS_SKYLAKE(i915))
> return false;
>
> /* Display WA #1141: SKL:all KBL:all CFL */
> - if (IS_KABYLAKE(dev_priv) || IS_COFFEELAKE(dev_priv))
> - return dev_priv->dram_info.symmetric_memory;
> + if (IS_KABYLAKE(i915) || IS_COFFEELAKE(i915))
> + return i915->dram_info.symmetric_memory;
>
> return true;
> }
>
> -void intel_init_ipc(struct drm_i915_private *dev_priv)
> +void intel_init_ipc(struct drm_i915_private *i915)
> {
> - if (!HAS_IPC(dev_priv))
> + if (!HAS_IPC(i915))
> return;
>
> - dev_priv->ipc_enabled = intel_can_enable_ipc(dev_priv);
> + i915->ipc_enabled = intel_can_enable_ipc(i915);
>
> - intel_enable_ipc(dev_priv);
> + intel_enable_ipc(i915);
> }
>
> /*
> @@ -6430,9 +6430,9 @@ bool ironlake_set_drps(struct drm_i915_private *i915, u8 val)
> return true;
> }
>
> -static void ironlake_enable_drps(struct drm_i915_private *dev_priv)
> +static void ironlake_enable_drps(struct drm_i915_private *i915)
> {
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct intel_uncore *uncore = &i915->uncore;
> u32 rgvmodectl;
> u8 fmax, fmin, fstart, vstart;
>
> @@ -6463,12 +6463,12 @@ static void ironlake_enable_drps(struct drm_i915_private *dev_priv)
> vstart = (intel_uncore_read(uncore, PXVFREQ(fstart)) &
> PXVFREQ_PX_MASK) >> PXVFREQ_PX_SHIFT;
>
> - dev_priv->ips.fmax = fmax; /* IPS callback will increase this */
> - dev_priv->ips.fstart = fstart;
> + i915->ips.fmax = fmax; /* IPS callback will increase this */
> + i915->ips.fstart = fstart;
>
> - dev_priv->ips.max_delay = fstart;
> - dev_priv->ips.min_delay = fmin;
> - dev_priv->ips.cur_delay = fstart;
> + i915->ips.max_delay = fstart;
> + i915->ips.min_delay = fmin;
> + i915->ips.cur_delay = fstart;
>
> DRM_DEBUG_DRIVER("fmax: %d, fmin: %d, fstart: %d\n",
> fmax, fmin, fstart);
> @@ -6492,15 +6492,15 @@ static void ironlake_enable_drps(struct drm_i915_private *dev_priv)
> DRM_ERROR("stuck trying to change perf mode\n");
> mdelay(1);
>
> - ironlake_set_drps(dev_priv, fstart);
> + ironlake_set_drps(i915, fstart);
>
> - dev_priv->ips.last_count1 =
> + i915->ips.last_count1 =
> intel_uncore_read(uncore, DMIEC) +
> intel_uncore_read(uncore, DDREC) +
> intel_uncore_read(uncore, CSIEC);
> - dev_priv->ips.last_time1 = jiffies_to_msecs(jiffies);
> - dev_priv->ips.last_count2 = intel_uncore_read(uncore, GFXEC);
> - dev_priv->ips.last_time2 = ktime_get_raw_ns();
> + i915->ips.last_time1 = jiffies_to_msecs(jiffies);
> + i915->ips.last_count2 = intel_uncore_read(uncore, GFXEC);
> + i915->ips.last_time2 = ktime_get_raw_ns();
>
> spin_unlock_irq(&mchdev_lock);
> }
> @@ -6543,9 +6543,9 @@ static void ironlake_disable_drps(struct drm_i915_private *i915)
> * ourselves, instead of doing a rmw cycle (which might result in us clearing
> * all limits and the gpu stuck at whatever frequency it is at atm).
> */
> -static u32 intel_rps_limits(struct drm_i915_private *dev_priv, u8 val)
> +static u32 intel_rps_limits(struct drm_i915_private *i915, u8 val)
> {
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct intel_rps *rps = &i915->gt_pm.rps;
> u32 limits;
>
> /* Only set the down limit when we've reached the lowest level to avoid
> @@ -6554,7 +6554,7 @@ static u32 intel_rps_limits(struct drm_i915_private *dev_priv, u8 val)
> * the hw runs at the minimal clock before selecting the desired
> * frequency, if the down threshold expires in that window we will not
> * receive a down interrupt. */
> - if (INTEL_GEN(dev_priv) >= 9) {
> + if (INTEL_GEN(i915) >= 9) {
> limits = (rps->max_freq_softlimit) << 23;
> if (val <= rps->min_freq_softlimit)
> limits |= (rps->min_freq_softlimit) << 14;
> @@ -6567,9 +6567,9 @@ static u32 intel_rps_limits(struct drm_i915_private *dev_priv, u8 val)
> return limits;
> }
>
> -static void rps_set_power(struct drm_i915_private *dev_priv, int new_power)
> +static void rps_set_power(struct drm_i915_private *i915, int new_power)
> {
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct intel_rps *rps = &i915->gt_pm.rps;
> u32 threshold_up = 0, threshold_down = 0; /* in % */
> u32 ei_up = 0, ei_down = 0;
>
> @@ -6614,23 +6614,23 @@ static void rps_set_power(struct drm_i915_private *dev_priv, int new_power)
> /* When byt can survive without system hang with dynamic
> * sw freq adjustments, this restriction can be lifted.
> */
> - if (IS_VALLEYVIEW(dev_priv))
> + if (IS_VALLEYVIEW(i915))
> goto skip_hw_write;
>
> I915_WRITE(GEN6_RP_UP_EI,
> - GT_INTERVAL_FROM_US(dev_priv, ei_up));
> + GT_INTERVAL_FROM_US(i915, ei_up));
> I915_WRITE(GEN6_RP_UP_THRESHOLD,
> - GT_INTERVAL_FROM_US(dev_priv,
> + GT_INTERVAL_FROM_US(i915,
> ei_up * threshold_up / 100));
>
> I915_WRITE(GEN6_RP_DOWN_EI,
> - GT_INTERVAL_FROM_US(dev_priv, ei_down));
> + GT_INTERVAL_FROM_US(i915, ei_down));
> I915_WRITE(GEN6_RP_DOWN_THRESHOLD,
> - GT_INTERVAL_FROM_US(dev_priv,
> + GT_INTERVAL_FROM_US(i915,
> ei_down * threshold_down / 100));
>
> I915_WRITE(GEN6_RP_CONTROL,
> - (INTEL_GEN(dev_priv) > 9 ? 0 : GEN6_RP_MEDIA_TURBO) |
> + (INTEL_GEN(i915) > 9 ? 0 : GEN6_RP_MEDIA_TURBO) |
> GEN6_RP_MEDIA_HW_NORMAL_MODE |
> GEN6_RP_MEDIA_IS_GFX |
> GEN6_RP_ENABLE |
> @@ -6643,9 +6643,9 @@ static void rps_set_power(struct drm_i915_private *dev_priv, int new_power)
> rps->power.down_threshold = threshold_down;
> }
>
> -static void gen6_set_rps_thresholds(struct drm_i915_private *dev_priv, u8 val)
> +static void gen6_set_rps_thresholds(struct drm_i915_private *i915, u8 val)
> {
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct intel_rps *rps = &i915->gt_pm.rps;
> int new_power;
>
> new_power = rps->power.mode;
> @@ -6680,7 +6680,7 @@ static void gen6_set_rps_thresholds(struct drm_i915_private *dev_priv, u8 val)
> mutex_lock(&rps->power.mutex);
> if (rps->power.interactive)
> new_power = HIGH_POWER;
> - rps_set_power(dev_priv, new_power);
> + rps_set_power(i915, new_power);
> mutex_unlock(&rps->power.mutex);
> }
>
> @@ -6702,9 +6702,9 @@ void intel_rps_mark_interactive(struct drm_i915_private *i915, bool interactive)
> mutex_unlock(&rps->power.mutex);
> }
>
> -static u32 gen6_rps_pm_mask(struct drm_i915_private *dev_priv, u8 val)
> +static u32 gen6_rps_pm_mask(struct drm_i915_private *i915, u8 val)
> {
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct intel_rps *rps = &i915->gt_pm.rps;
> u32 mask = 0;
>
> /* We use UP_EI_EXPIRED interupts for both up/down in manual mode */
> @@ -6713,28 +6713,28 @@ static u32 gen6_rps_pm_mask(struct drm_i915_private *dev_priv, u8 val)
> if (val < rps->max_freq_softlimit)
> mask |= GEN6_PM_RP_UP_EI_EXPIRED | GEN6_PM_RP_UP_THRESHOLD;
>
> - mask &= dev_priv->pm_rps_events;
> + mask &= i915->pm_rps_events;
>
> - return gen6_sanitize_rps_pm_mask(dev_priv, ~mask);
> + return gen6_sanitize_rps_pm_mask(i915, ~mask);
> }
>
> /* gen6_set_rps is called to update the frequency request, but should also be
> * called when the range (min_delay and max_delay) is modified so that we can
> * update the GEN6_RP_INTERRUPT_LIMITS register accordingly. */
> -static int gen6_set_rps(struct drm_i915_private *dev_priv, u8 val)
> +static int gen6_set_rps(struct drm_i915_private *i915, u8 val)
> {
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct intel_rps *rps = &i915->gt_pm.rps;
>
> /* min/max delay may still have been modified so be sure to
> * write the limits value.
> */
> if (val != rps->cur_freq) {
> - gen6_set_rps_thresholds(dev_priv, val);
> + gen6_set_rps_thresholds(i915, val);
>
> - if (INTEL_GEN(dev_priv) >= 9)
> + if (INTEL_GEN(i915) >= 9)
> I915_WRITE(GEN6_RPNSWREQ,
> GEN9_FREQUENCY(val));
> - else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> + else if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> I915_WRITE(GEN6_RPNSWREQ,
> HSW_FREQUENCY(val));
> else
> @@ -6747,37 +6747,37 @@ static int gen6_set_rps(struct drm_i915_private *dev_priv, u8 val)
> /* Make sure we continue to get interrupts
> * until we hit the minimum or maximum frequencies.
> */
> - I915_WRITE(GEN6_RP_INTERRUPT_LIMITS, intel_rps_limits(dev_priv, val));
> - I915_WRITE(GEN6_PMINTRMSK, gen6_rps_pm_mask(dev_priv, val));
> + I915_WRITE(GEN6_RP_INTERRUPT_LIMITS, intel_rps_limits(i915, val));
> + I915_WRITE(GEN6_PMINTRMSK, gen6_rps_pm_mask(i915, val));
>
> rps->cur_freq = val;
> - trace_intel_gpu_freq_change(intel_gpu_freq(dev_priv, val));
> + trace_intel_gpu_freq_change(intel_gpu_freq(i915, val));
>
> return 0;
> }
>
> -static int valleyview_set_rps(struct drm_i915_private *dev_priv, u8 val)
> +static int valleyview_set_rps(struct drm_i915_private *i915, u8 val)
> {
> int err;
>
> - if (WARN_ONCE(IS_CHERRYVIEW(dev_priv) && (val & 1),
> + if (WARN_ONCE(IS_CHERRYVIEW(i915) && (val & 1),
> "Odd GPU freq value\n"))
> val &= ~1;
>
> - I915_WRITE(GEN6_PMINTRMSK, gen6_rps_pm_mask(dev_priv, val));
> + I915_WRITE(GEN6_PMINTRMSK, gen6_rps_pm_mask(i915, val));
>
> - if (val != dev_priv->gt_pm.rps.cur_freq) {
> - vlv_punit_get(dev_priv);
> - err = vlv_punit_write(dev_priv, PUNIT_REG_GPU_FREQ_REQ, val);
> - vlv_punit_put(dev_priv);
> + if (val != i915->gt_pm.rps.cur_freq) {
> + vlv_punit_get(i915);
> + err = vlv_punit_write(i915, PUNIT_REG_GPU_FREQ_REQ, val);
> + vlv_punit_put(i915);
> if (err)
> return err;
>
> - gen6_set_rps_thresholds(dev_priv, val);
> + gen6_set_rps_thresholds(i915, val);
> }
>
> - dev_priv->gt_pm.rps.cur_freq = val;
> - trace_intel_gpu_freq_change(intel_gpu_freq(dev_priv, val));
> + i915->gt_pm.rps.cur_freq = val;
> + trace_intel_gpu_freq_change(intel_gpu_freq(i915, val));
>
> return 0;
> }
> @@ -6789,9 +6789,9 @@ static int valleyview_set_rps(struct drm_i915_private *dev_priv, u8 val)
> * 2. Request idle freq.
> * 3. Release Forcewake of Media well.
> */
> -static void vlv_set_rps_idle(struct drm_i915_private *dev_priv)
> +static void vlv_set_rps_idle(struct drm_i915_private *i915)
> {
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct intel_rps *rps = &i915->gt_pm.rps;
> u32 val = rps->idle_freq;
> int err;
>
> @@ -6810,28 +6810,28 @@ static void vlv_set_rps_idle(struct drm_i915_private *dev_priv)
> * punit into committing the voltage change) as that takes a lot less
> * power than the render powerwell.
> */
> - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_MEDIA);
> - err = valleyview_set_rps(dev_priv, val);
> - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_MEDIA);
> + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_MEDIA);
> + err = valleyview_set_rps(i915, val);
> + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_MEDIA);
>
> if (err)
> DRM_ERROR("Failed to set RPS for idle\n");
> }
>
> -void gen6_rps_busy(struct drm_i915_private *dev_priv)
> +void gen6_rps_busy(struct drm_i915_private *i915)
> {
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct intel_rps *rps = &i915->gt_pm.rps;
>
> mutex_lock(&rps->lock);
> if (rps->enabled) {
> u8 freq;
>
> - if (dev_priv->pm_rps_events & GEN6_PM_RP_UP_EI_EXPIRED)
> - gen6_rps_reset_ei(dev_priv);
> + if (i915->pm_rps_events & GEN6_PM_RP_UP_EI_EXPIRED)
> + gen6_rps_reset_ei(i915);
> I915_WRITE(GEN6_PMINTRMSK,
> - gen6_rps_pm_mask(dev_priv, rps->cur_freq));
> + gen6_rps_pm_mask(i915, rps->cur_freq));
>
> - gen6_enable_rps_interrupts(dev_priv);
> + gen6_enable_rps_interrupts(i915);
>
> /* Use the user's desired frequency as a guide, but for better
> * performance, jump directly to RPe as our starting frequency.
> @@ -6839,7 +6839,7 @@ void gen6_rps_busy(struct drm_i915_private *dev_priv)
> freq = max(rps->cur_freq,
> rps->efficient_freq);
>
> - if (intel_set_rps(dev_priv,
> + if (intel_set_rps(i915,
> clamp(freq,
> rps->min_freq_softlimit,
> rps->max_freq_softlimit)))
> @@ -6848,26 +6848,26 @@ void gen6_rps_busy(struct drm_i915_private *dev_priv)
> mutex_unlock(&rps->lock);
> }
>
> -void gen6_rps_idle(struct drm_i915_private *dev_priv)
> +void gen6_rps_idle(struct drm_i915_private *i915)
> {
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct intel_rps *rps = &i915->gt_pm.rps;
>
> /* Flush our bottom-half so that it does not race with us
> * setting the idle frequency and so that it is bounded by
> * our rpm wakeref. And then disable the interrupts to stop any
> * futher RPS reclocking whilst we are asleep.
> */
> - gen6_disable_rps_interrupts(dev_priv);
> + gen6_disable_rps_interrupts(i915);
>
> mutex_lock(&rps->lock);
> if (rps->enabled) {
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> - vlv_set_rps_idle(dev_priv);
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> + vlv_set_rps_idle(i915);
> else
> - gen6_set_rps(dev_priv, rps->idle_freq);
> + gen6_set_rps(i915, rps->idle_freq);
> rps->last_adj = 0;
> I915_WRITE(GEN6_PMINTRMSK,
> - gen6_sanitize_rps_pm_mask(dev_priv, ~0));
> + gen6_sanitize_rps_pm_mask(i915, ~0));
> }
> mutex_unlock(&rps->lock);
> }
> @@ -6904,9 +6904,9 @@ void gen6_rps_boost(struct i915_request *rq)
> atomic_inc(&rps->boosts);
> }
>
> -int intel_set_rps(struct drm_i915_private *dev_priv, u8 val)
> +int intel_set_rps(struct drm_i915_private *i915, u8 val)
> {
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct intel_rps *rps = &i915->gt_pm.rps;
> int err;
>
> lockdep_assert_held(&rps->lock);
> @@ -6918,63 +6918,63 @@ int intel_set_rps(struct drm_i915_private *dev_priv, u8 val)
> return 0;
> }
>
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> - err = valleyview_set_rps(dev_priv, val);
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> + err = valleyview_set_rps(i915, val);
> else
> - err = gen6_set_rps(dev_priv, val);
> + err = gen6_set_rps(i915, val);
>
> return err;
> }
>
> -static void gen9_disable_rc6(struct drm_i915_private *dev_priv)
> +static void gen9_disable_rc6(struct drm_i915_private *i915)
> {
> I915_WRITE(GEN6_RC_CONTROL, 0);
> I915_WRITE(GEN9_PG_ENABLE, 0);
> }
>
> -static void gen9_disable_rps(struct drm_i915_private *dev_priv)
> +static void gen9_disable_rps(struct drm_i915_private *i915)
> {
> I915_WRITE(GEN6_RP_CONTROL, 0);
> }
>
> -static void gen6_disable_rc6(struct drm_i915_private *dev_priv)
> +static void gen6_disable_rc6(struct drm_i915_private *i915)
> {
> I915_WRITE(GEN6_RC_CONTROL, 0);
> }
>
> -static void gen6_disable_rps(struct drm_i915_private *dev_priv)
> +static void gen6_disable_rps(struct drm_i915_private *i915)
> {
> I915_WRITE(GEN6_RPNSWREQ, 1 << 31);
> I915_WRITE(GEN6_RP_CONTROL, 0);
> }
>
> -static void cherryview_disable_rc6(struct drm_i915_private *dev_priv)
> +static void cherryview_disable_rc6(struct drm_i915_private *i915)
> {
> I915_WRITE(GEN6_RC_CONTROL, 0);
> }
>
> -static void cherryview_disable_rps(struct drm_i915_private *dev_priv)
> +static void cherryview_disable_rps(struct drm_i915_private *i915)
> {
> I915_WRITE(GEN6_RP_CONTROL, 0);
> }
>
> -static void valleyview_disable_rc6(struct drm_i915_private *dev_priv)
> +static void valleyview_disable_rc6(struct drm_i915_private *i915)
> {
> /* We're doing forcewake before Disabling RC6,
> * This what the BIOS expects when going into suspend */
> - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
>
> I915_WRITE(GEN6_RC_CONTROL, 0);
>
> - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> }
>
> -static void valleyview_disable_rps(struct drm_i915_private *dev_priv)
> +static void valleyview_disable_rps(struct drm_i915_private *i915)
> {
> I915_WRITE(GEN6_RP_CONTROL, 0);
> }
>
> -static bool bxt_check_bios_rc6_setup(struct drm_i915_private *dev_priv)
> +static bool bxt_check_bios_rc6_setup(struct drm_i915_private *i915)
> {
> bool enable_rc6 = true;
> unsigned long rc6_ctx_base;
> @@ -7000,8 +7000,8 @@ static bool bxt_check_bios_rc6_setup(struct drm_i915_private *dev_priv)
> * for this check.
> */
> rc6_ctx_base = I915_READ(RC6_CTX_BASE) & RC6_CTX_BASE_MASK;
> - if (!((rc6_ctx_base >= dev_priv->dsm_reserved.start) &&
> - (rc6_ctx_base + PAGE_SIZE < dev_priv->dsm_reserved.end))) {
> + if (!((rc6_ctx_base >= i915->dsm_reserved.start) &&
> + (rc6_ctx_base + PAGE_SIZE < i915->dsm_reserved.end))) {
> DRM_DEBUG_DRIVER("RC6 Base address not as expected.\n");
> enable_rc6 = false;
> }
> @@ -7062,14 +7062,14 @@ static bool sanitize_rc6(struct drm_i915_private *i915)
> return info->has_rc6;
> }
>
> -static void gen6_init_rps_frequencies(struct drm_i915_private *dev_priv)
> +static void gen6_init_rps_frequencies(struct drm_i915_private *i915)
> {
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct intel_rps *rps = &i915->gt_pm.rps;
>
> /* All of these values are in units of 50MHz */
>
> /* static values from HW: RP0 > RP1 > RPn (min_freq) */
> - if (IS_GEN9_LP(dev_priv)) {
> + if (IS_GEN9_LP(i915)) {
> u32 rp_state_cap = I915_READ(BXT_RP_STATE_CAP);
> rps->rp0_freq = (rp_state_cap >> 16) & 0xff;
> rps->rp1_freq = (rp_state_cap >> 8) & 0xff;
> @@ -7084,11 +7084,11 @@ static void gen6_init_rps_frequencies(struct drm_i915_private *dev_priv)
> rps->max_freq = rps->rp0_freq;
>
> rps->efficient_freq = rps->rp1_freq;
> - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv) ||
> - IS_GEN9_BC(dev_priv) || INTEL_GEN(dev_priv) >= 10) {
> + if (IS_HASWELL(i915) || IS_BROADWELL(i915) ||
> + IS_GEN9_BC(i915) || INTEL_GEN(i915) >= 10) {
> u32 ddcc_status = 0;
>
> - if (sandybridge_pcode_read(dev_priv,
> + if (sandybridge_pcode_read(i915,
> HSW_PCODE_DYNAMIC_DUTY_CYCLE_CONTROL,
> &ddcc_status, NULL) == 0)
> rps->efficient_freq =
> @@ -7098,7 +7098,7 @@ static void gen6_init_rps_frequencies(struct drm_i915_private *dev_priv)
> rps->max_freq);
> }
>
> - if (IS_GEN9_BC(dev_priv) || INTEL_GEN(dev_priv) >= 10) {
> + if (IS_GEN9_BC(i915) || INTEL_GEN(i915) >= 10) {
> /* Store the frequency values in 16.66 MHZ units, which is
> * the natural hardware unit for SKL
> */
> @@ -7110,45 +7110,45 @@ static void gen6_init_rps_frequencies(struct drm_i915_private *dev_priv)
> }
> }
>
> -static void reset_rps(struct drm_i915_private *dev_priv,
> +static void reset_rps(struct drm_i915_private *i915,
> int (*set)(struct drm_i915_private *, u8))
> {
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct intel_rps *rps = &i915->gt_pm.rps;
> u8 freq = rps->cur_freq;
>
> /* force a reset */
> rps->power.mode = -1;
> rps->cur_freq = -1;
>
> - if (set(dev_priv, freq))
> + if (set(i915, freq))
> DRM_ERROR("Failed to reset RPS to initial values\n");
> }
>
> /* See the Gen9_GT_PM_Programming_Guide doc for the below */
> -static void gen9_enable_rps(struct drm_i915_private *dev_priv)
> +static void gen9_enable_rps(struct drm_i915_private *i915)
> {
> - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
>
> /* Program defaults and thresholds for RPS */
> - if (IS_GEN(dev_priv, 9))
> + if (IS_GEN(i915, 9))
> I915_WRITE(GEN6_RC_VIDEO_FREQ,
> - GEN9_FREQUENCY(dev_priv->gt_pm.rps.rp1_freq));
> + GEN9_FREQUENCY(i915->gt_pm.rps.rp1_freq));
>
> /* 1 second timeout*/
> I915_WRITE(GEN6_RP_DOWN_TIMEOUT,
> - GT_INTERVAL_FROM_US(dev_priv, 1000000));
> + GT_INTERVAL_FROM_US(i915, 1000000));
>
> I915_WRITE(GEN6_RP_IDLE_HYSTERSIS, 0xa);
>
> /* Leaning on the below call to gen6_set_rps to program/setup the
> * Up/Down EI & threshold registers, as well as the RP_CONTROL,
> * RP_INTERRUPT_LIMITS & RPNSWREQ registers */
> - reset_rps(dev_priv, gen6_set_rps);
> + reset_rps(i915, gen6_set_rps);
>
> - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> }
>
> -static void gen11_enable_rc6(struct drm_i915_private *dev_priv)
> +static void gen11_enable_rc6(struct drm_i915_private *i915)
> {
> struct intel_engine_cs *engine;
> enum intel_engine_id id;
> @@ -7160,7 +7160,7 @@ static void gen11_enable_rc6(struct drm_i915_private *dev_priv)
> * 1b: Get forcewake during program sequence. Although the driver
> * hasn't enabled a state yet where we need forcewake, BIOS may have.
> */
> - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
>
> /* 2a: Disable RC states. */
> I915_WRITE(GEN6_RC_CONTROL, 0);
> @@ -7171,10 +7171,10 @@ static void gen11_enable_rc6(struct drm_i915_private *dev_priv)
>
> I915_WRITE(GEN6_RC_EVALUATION_INTERVAL, 125000); /* 12500 * 1280ns */
> I915_WRITE(GEN6_RC_IDLE_HYSTERSIS, 25); /* 25 * 1280ns */
> - for_each_engine(engine, dev_priv, id)
> + for_each_engine(engine, i915, id)
> I915_WRITE(RING_MAX_IDLE(engine->mmio_base), 10);
>
> - if (HAS_GUC(dev_priv))
> + if (HAS_GUC(i915))
> I915_WRITE(GUC_MAX_IDLE_COUNT, 0xA);
>
> I915_WRITE(GEN6_RC_SLEEP, 0);
> @@ -7217,10 +7217,10 @@ static void gen11_enable_rc6(struct drm_i915_private *dev_priv)
> GEN9_MEDIA_PG_ENABLE |
> GEN11_MEDIA_SAMPLER_PG_ENABLE);
>
> - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> }
>
> -static void gen9_enable_rc6(struct drm_i915_private *dev_priv)
> +static void gen9_enable_rc6(struct drm_i915_private *i915)
> {
> struct intel_engine_cs *engine;
> enum intel_engine_id id;
> @@ -7231,16 +7231,16 @@ static void gen9_enable_rc6(struct drm_i915_private *dev_priv)
>
> /* 1b: Get forcewake during program sequence. Although the driver
> * hasn't enabled a state yet where we need forcewake, BIOS may have.*/
> - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
>
> /* 2a: Disable RC states. */
> I915_WRITE(GEN6_RC_CONTROL, 0);
>
> /* 2b: Program RC6 thresholds.*/
> - if (INTEL_GEN(dev_priv) >= 10) {
> + if (INTEL_GEN(i915) >= 10) {
> I915_WRITE(GEN6_RC6_WAKE_RATE_LIMIT, 54 << 16 | 85);
> I915_WRITE(GEN10_MEDIA_WAKE_RATE_LIMIT, 150);
> - } else if (IS_SKYLAKE(dev_priv)) {
> + } else if (IS_SKYLAKE(i915)) {
> /*
> * WaRsDoubleRc6WrlWithCoarsePowerGating:skl Doubling WRL only
> * when CPG is enabled
> @@ -7252,10 +7252,10 @@ static void gen9_enable_rc6(struct drm_i915_private *dev_priv)
>
> I915_WRITE(GEN6_RC_EVALUATION_INTERVAL, 125000); /* 12500 * 1280ns */
> I915_WRITE(GEN6_RC_IDLE_HYSTERSIS, 25); /* 25 * 1280ns */
> - for_each_engine(engine, dev_priv, id)
> + for_each_engine(engine, i915, id)
> I915_WRITE(RING_MAX_IDLE(engine->mmio_base), 10);
>
> - if (HAS_GUC(dev_priv))
> + if (HAS_GUC(i915))
> I915_WRITE(GUC_MAX_IDLE_COUNT, 0xA);
>
> I915_WRITE(GEN6_RC_SLEEP, 0);
> @@ -7288,7 +7288,7 @@ static void gen9_enable_rc6(struct drm_i915_private *dev_priv)
> I915_WRITE(GEN6_RC6_THRESHOLD, 37500); /* 37.5/125ms per EI */
>
> /* WaRsUseTimeoutMode:cnl (pre-prod) */
> - if (IS_CNL_REVID(dev_priv, CNL_REVID_A0, CNL_REVID_C0))
> + if (IS_CNL_REVID(i915, CNL_REVID_A0, CNL_REVID_C0))
> rc6_mode = GEN7_RC_CTL_TO_MODE;
> else
> rc6_mode = GEN6_RC_CTL_EI_MODE(1);
> @@ -7302,16 +7302,16 @@ static void gen9_enable_rc6(struct drm_i915_private *dev_priv)
> * 3b: Enable Coarse Power Gating only when RC6 is enabled.
> * WaRsDisableCoarsePowerGating:skl,cnl - Render/Media PG need to be disabled with RC6.
> */
> - if (NEEDS_WaRsDisableCoarsePowerGating(dev_priv))
> + if (NEEDS_WaRsDisableCoarsePowerGating(i915))
> I915_WRITE(GEN9_PG_ENABLE, 0);
> else
> I915_WRITE(GEN9_PG_ENABLE,
> GEN9_RENDER_PG_ENABLE | GEN9_MEDIA_PG_ENABLE);
>
> - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> }
>
> -static void gen8_enable_rc6(struct drm_i915_private *dev_priv)
> +static void gen8_enable_rc6(struct drm_i915_private *i915)
> {
> struct intel_engine_cs *engine;
> enum intel_engine_id id;
> @@ -7321,7 +7321,7 @@ static void gen8_enable_rc6(struct drm_i915_private *dev_priv)
>
> /* 1b: Get forcewake during program sequence. Although the driver
> * hasn't enabled a state yet where we need forcewake, BIOS may have.*/
> - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
>
> /* 2a: Disable RC states. */
> I915_WRITE(GEN6_RC_CONTROL, 0);
> @@ -7330,7 +7330,7 @@ static void gen8_enable_rc6(struct drm_i915_private *dev_priv)
> I915_WRITE(GEN6_RC6_WAKE_RATE_LIMIT, 40 << 16);
> I915_WRITE(GEN6_RC_EVALUATION_INTERVAL, 125000); /* 12500 * 1280ns */
> I915_WRITE(GEN6_RC_IDLE_HYSTERSIS, 25); /* 25 * 1280ns */
> - for_each_engine(engine, dev_priv, id)
> + for_each_engine(engine, i915, id)
> I915_WRITE(RING_MAX_IDLE(engine->mmio_base), 10);
> I915_WRITE(GEN6_RC_SLEEP, 0);
> I915_WRITE(GEN6_RC6_THRESHOLD, 625); /* 800us/1.28 for TO */
> @@ -7342,14 +7342,14 @@ static void gen8_enable_rc6(struct drm_i915_private *dev_priv)
> GEN7_RC_CTL_TO_MODE |
> GEN6_RC_CTL_RC6_ENABLE);
>
> - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> }
>
> -static void gen8_enable_rps(struct drm_i915_private *dev_priv)
> +static void gen8_enable_rps(struct drm_i915_private *i915)
> {
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct intel_rps *rps = &i915->gt_pm.rps;
>
> - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
>
> /* 1 Program defaults and thresholds for RPS*/
> I915_WRITE(GEN6_RPNSWREQ,
> @@ -7380,12 +7380,12 @@ static void gen8_enable_rps(struct drm_i915_private *dev_priv)
> GEN6_RP_UP_BUSY_AVG |
> GEN6_RP_DOWN_IDLE_AVG);
>
> - reset_rps(dev_priv, gen6_set_rps);
> + reset_rps(i915, gen6_set_rps);
>
> - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> }
>
> -static void gen6_enable_rc6(struct drm_i915_private *dev_priv)
> +static void gen6_enable_rc6(struct drm_i915_private *i915)
> {
> struct intel_engine_cs *engine;
> enum intel_engine_id id;
> @@ -7402,7 +7402,7 @@ static void gen6_enable_rc6(struct drm_i915_private *dev_priv)
> I915_WRITE(GTFIFODBG, gtfifodbg);
> }
>
> - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
>
> /* disable the counters and set deterministic thresholds */
> I915_WRITE(GEN6_RC_CONTROL, 0);
> @@ -7413,12 +7413,12 @@ static void gen6_enable_rc6(struct drm_i915_private *dev_priv)
> I915_WRITE(GEN6_RC_EVALUATION_INTERVAL, 125000);
> I915_WRITE(GEN6_RC_IDLE_HYSTERSIS, 25);
>
> - for_each_engine(engine, dev_priv, id)
> + for_each_engine(engine, i915, id)
> I915_WRITE(RING_MAX_IDLE(engine->mmio_base), 10);
>
> I915_WRITE(GEN6_RC_SLEEP, 0);
> I915_WRITE(GEN6_RC1e_THRESHOLD, 1000);
> - if (IS_IVYBRIDGE(dev_priv))
> + if (IS_IVYBRIDGE(i915))
> I915_WRITE(GEN6_RC6_THRESHOLD, 125000);
> else
> I915_WRITE(GEN6_RC6_THRESHOLD, 50000);
> @@ -7427,9 +7427,9 @@ static void gen6_enable_rc6(struct drm_i915_private *dev_priv)
>
> /* We don't use those on Haswell */
> rc6_mask = GEN6_RC_CTL_RC6_ENABLE;
> - if (HAS_RC6p(dev_priv))
> + if (HAS_RC6p(i915))
> rc6_mask |= GEN6_RC_CTL_RC6p_ENABLE;
> - if (HAS_RC6pp(dev_priv))
> + if (HAS_RC6pp(i915))
> rc6_mask |= GEN6_RC_CTL_RC6pp_ENABLE;
> I915_WRITE(GEN6_RC_CONTROL,
> rc6_mask |
> @@ -7437,24 +7437,24 @@ static void gen6_enable_rc6(struct drm_i915_private *dev_priv)
> GEN6_RC_CTL_HW_ENABLE);
>
> rc6vids = 0;
> - ret = sandybridge_pcode_read(dev_priv, GEN6_PCODE_READ_RC6VIDS,
> + ret = sandybridge_pcode_read(i915, GEN6_PCODE_READ_RC6VIDS,
> &rc6vids, NULL);
> - if (IS_GEN(dev_priv, 6) && ret) {
> + if (IS_GEN(i915, 6) && ret) {
> DRM_DEBUG_DRIVER("Couldn't check for BIOS workaround\n");
> - } else if (IS_GEN(dev_priv, 6) && (GEN6_DECODE_RC6_VID(rc6vids & 0xff) < 450)) {
> + } else if (IS_GEN(i915, 6) && (GEN6_DECODE_RC6_VID(rc6vids & 0xff) < 450)) {
> DRM_DEBUG_DRIVER("You should update your BIOS. Correcting minimum rc6 voltage (%dmV->%dmV)\n",
> GEN6_DECODE_RC6_VID(rc6vids & 0xff), 450);
> rc6vids &= 0xffff00;
> rc6vids |= GEN6_ENCODE_RC6_VID(450);
> - ret = sandybridge_pcode_write(dev_priv, GEN6_PCODE_WRITE_RC6VIDS, rc6vids);
> + ret = sandybridge_pcode_write(i915, GEN6_PCODE_WRITE_RC6VIDS, rc6vids);
> if (ret)
> DRM_ERROR("Couldn't fix incorrect rc6 voltage\n");
> }
>
> - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> }
>
> -static void gen6_enable_rps(struct drm_i915_private *dev_priv)
> +static void gen6_enable_rps(struct drm_i915_private *i915)
> {
> /* Here begins a magic sequence of register writes to enable
> * auto-downclocking.
> @@ -7462,20 +7462,20 @@ static void gen6_enable_rps(struct drm_i915_private *dev_priv)
> * Perhaps there might be some value in exposing these to
> * userspace...
> */
> - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
>
> /* Power down if completely idle for over 50ms */
> I915_WRITE(GEN6_RP_DOWN_TIMEOUT, 50000);
> I915_WRITE(GEN6_RP_IDLE_HYSTERSIS, 10);
>
> - reset_rps(dev_priv, gen6_set_rps);
> + reset_rps(i915, gen6_set_rps);
>
> - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> }
>
> -static void gen6_update_ring_freq(struct drm_i915_private *dev_priv)
> +static void gen6_update_ring_freq(struct drm_i915_private *i915)
> {
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct intel_rps *rps = &i915->gt_pm.rps;
> const int min_freq = 15;
> const int scaling_factor = 180;
> unsigned int gpu_freq;
> @@ -7509,7 +7509,7 @@ static void gen6_update_ring_freq(struct drm_i915_private *dev_priv)
>
> min_gpu_freq = rps->min_freq;
> max_gpu_freq = rps->max_freq;
> - if (IS_GEN9_BC(dev_priv) || INTEL_GEN(dev_priv) >= 10) {
> + if (IS_GEN9_BC(i915) || INTEL_GEN(i915) >= 10) {
> /* Convert GT frequency to 50 HZ units */
> min_gpu_freq /= GEN9_FREQ_SCALER;
> max_gpu_freq /= GEN9_FREQ_SCALER;
> @@ -7524,16 +7524,16 @@ static void gen6_update_ring_freq(struct drm_i915_private *dev_priv)
> const int diff = max_gpu_freq - gpu_freq;
> unsigned int ia_freq = 0, ring_freq = 0;
>
> - if (IS_GEN9_BC(dev_priv) || INTEL_GEN(dev_priv) >= 10) {
> + if (IS_GEN9_BC(i915) || INTEL_GEN(i915) >= 10) {
> /*
> * ring_freq = 2 * GT. ring_freq is in 100MHz units
> * No floor required for ring frequency on SKL.
> */
> ring_freq = gpu_freq;
> - } else if (INTEL_GEN(dev_priv) >= 8) {
> + } else if (INTEL_GEN(i915) >= 8) {
> /* max(2 * GT, DDR). NB: GT is 50MHz units */
> ring_freq = max(min_ring_freq, gpu_freq);
> - } else if (IS_HASWELL(dev_priv)) {
> + } else if (IS_HASWELL(i915)) {
> ring_freq = mult_frac(gpu_freq, 5, 4);
> ring_freq = max(min_ring_freq, ring_freq);
> /* leave ia_freq as the default, chosen by cpufreq */
> @@ -7552,7 +7552,7 @@ static void gen6_update_ring_freq(struct drm_i915_private *dev_priv)
> ia_freq = DIV_ROUND_CLOSEST(ia_freq, 100);
> }
>
> - sandybridge_pcode_write(dev_priv,
> + sandybridge_pcode_write(i915,
> GEN6_PCODE_WRITE_MIN_FREQ_TABLE,
> ia_freq << GEN6_PCODE_FREQ_IA_RATIO_SHIFT |
> ring_freq << GEN6_PCODE_FREQ_RING_RATIO_SHIFT |
> @@ -7560,13 +7560,13 @@ static void gen6_update_ring_freq(struct drm_i915_private *dev_priv)
> }
> }
>
> -static int cherryview_rps_max_freq(struct drm_i915_private *dev_priv)
> +static int cherryview_rps_max_freq(struct drm_i915_private *i915)
> {
> u32 val, rp0;
>
> - val = vlv_punit_read(dev_priv, FB_GFX_FMAX_AT_VMAX_FUSE);
> + val = vlv_punit_read(i915, FB_GFX_FMAX_AT_VMAX_FUSE);
>
> - switch (RUNTIME_INFO(dev_priv)->sseu.eu_total) {
> + switch (RUNTIME_INFO(i915)->sseu.eu_total) {
> case 8:
> /* (2 * 4) config */
> rp0 = (val >> FB_GFX_FMAX_AT_VMAX_2SS4EU_FUSE_SHIFT);
> @@ -7588,53 +7588,53 @@ static int cherryview_rps_max_freq(struct drm_i915_private *dev_priv)
> return rp0;
> }
>
> -static int cherryview_rps_rpe_freq(struct drm_i915_private *dev_priv)
> +static int cherryview_rps_rpe_freq(struct drm_i915_private *i915)
> {
> u32 val, rpe;
>
> - val = vlv_punit_read(dev_priv, PUNIT_GPU_DUTYCYCLE_REG);
> + val = vlv_punit_read(i915, PUNIT_GPU_DUTYCYCLE_REG);
> rpe = (val >> PUNIT_GPU_DUTYCYCLE_RPE_FREQ_SHIFT) & PUNIT_GPU_DUTYCYCLE_RPE_FREQ_MASK;
>
> return rpe;
> }
>
> -static int cherryview_rps_guar_freq(struct drm_i915_private *dev_priv)
> +static int cherryview_rps_guar_freq(struct drm_i915_private *i915)
> {
> u32 val, rp1;
>
> - val = vlv_punit_read(dev_priv, FB_GFX_FMAX_AT_VMAX_FUSE);
> + val = vlv_punit_read(i915, FB_GFX_FMAX_AT_VMAX_FUSE);
> rp1 = (val & FB_GFX_FREQ_FUSE_MASK);
>
> return rp1;
> }
>
> -static u32 cherryview_rps_min_freq(struct drm_i915_private *dev_priv)
> +static u32 cherryview_rps_min_freq(struct drm_i915_private *i915)
> {
> u32 val, rpn;
>
> - val = vlv_punit_read(dev_priv, FB_GFX_FMIN_AT_VMIN_FUSE);
> + val = vlv_punit_read(i915, FB_GFX_FMIN_AT_VMIN_FUSE);
> rpn = ((val >> FB_GFX_FMIN_AT_VMIN_FUSE_SHIFT) &
> FB_GFX_FREQ_FUSE_MASK);
>
> return rpn;
> }
>
> -static int valleyview_rps_guar_freq(struct drm_i915_private *dev_priv)
> +static int valleyview_rps_guar_freq(struct drm_i915_private *i915)
> {
> u32 val, rp1;
>
> - val = vlv_nc_read(dev_priv, IOSF_NC_FB_GFX_FREQ_FUSE);
> + val = vlv_nc_read(i915, IOSF_NC_FB_GFX_FREQ_FUSE);
>
> rp1 = (val & FB_GFX_FGUARANTEED_FREQ_FUSE_MASK) >> FB_GFX_FGUARANTEED_FREQ_FUSE_SHIFT;
>
> return rp1;
> }
>
> -static int valleyview_rps_max_freq(struct drm_i915_private *dev_priv)
> +static int valleyview_rps_max_freq(struct drm_i915_private *i915)
> {
> u32 val, rp0;
>
> - val = vlv_nc_read(dev_priv, IOSF_NC_FB_GFX_FREQ_FUSE);
> + val = vlv_nc_read(i915, IOSF_NC_FB_GFX_FREQ_FUSE);
>
> rp0 = (val & FB_GFX_MAX_FREQ_FUSE_MASK) >> FB_GFX_MAX_FREQ_FUSE_SHIFT;
> /* Clamp to max */
> @@ -7643,23 +7643,23 @@ static int valleyview_rps_max_freq(struct drm_i915_private *dev_priv)
> return rp0;
> }
>
> -static int valleyview_rps_rpe_freq(struct drm_i915_private *dev_priv)
> +static int valleyview_rps_rpe_freq(struct drm_i915_private *i915)
> {
> u32 val, rpe;
>
> - val = vlv_nc_read(dev_priv, IOSF_NC_FB_GFX_FMAX_FUSE_LO);
> + val = vlv_nc_read(i915, IOSF_NC_FB_GFX_FMAX_FUSE_LO);
> rpe = (val & FB_FMAX_VMIN_FREQ_LO_MASK) >> FB_FMAX_VMIN_FREQ_LO_SHIFT;
> - val = vlv_nc_read(dev_priv, IOSF_NC_FB_GFX_FMAX_FUSE_HI);
> + val = vlv_nc_read(i915, IOSF_NC_FB_GFX_FMAX_FUSE_HI);
> rpe |= (val & FB_FMAX_VMIN_FREQ_HI_MASK) << 5;
>
> return rpe;
> }
>
> -static int valleyview_rps_min_freq(struct drm_i915_private *dev_priv)
> +static int valleyview_rps_min_freq(struct drm_i915_private *i915)
> {
> u32 val;
>
> - val = vlv_punit_read(dev_priv, PUNIT_REG_GPU_LFM) & 0xff;
> + val = vlv_punit_read(i915, PUNIT_REG_GPU_LFM) & 0xff;
> /*
> * According to the BYT Punit GPU turbo HAS 1.1.6.3 the minimum value
> * for the minimum frequency in GPLL mode is 0xc1. Contrary to this on
> @@ -7671,24 +7671,24 @@ static int valleyview_rps_min_freq(struct drm_i915_private *dev_priv)
> }
>
> /* Check that the pctx buffer wasn't move under us. */
> -static void valleyview_check_pctx(struct drm_i915_private *dev_priv)
> +static void valleyview_check_pctx(struct drm_i915_private *i915)
> {
> unsigned long pctx_addr = I915_READ(VLV_PCBR) & ~4095;
>
> - WARN_ON(pctx_addr != dev_priv->dsm.start +
> - dev_priv->vlv_pctx->stolen->start);
> + WARN_ON(pctx_addr != i915->dsm.start +
> + i915->vlv_pctx->stolen->start);
> }
>
>
> /* Check that the pcbr address is not empty. */
> -static void cherryview_check_pctx(struct drm_i915_private *dev_priv)
> +static void cherryview_check_pctx(struct drm_i915_private *i915)
> {
> unsigned long pctx_addr = I915_READ(VLV_PCBR) & ~4095;
>
> WARN_ON((pctx_addr >> VLV_PCBR_ADDR_SHIFT) == 0);
> }
>
> -static void cherryview_setup_pctx(struct drm_i915_private *dev_priv)
> +static void cherryview_setup_pctx(struct drm_i915_private *i915)
> {
> resource_size_t pctx_paddr, paddr;
> resource_size_t pctx_size = 32*1024;
> @@ -7697,7 +7697,7 @@ static void cherryview_setup_pctx(struct drm_i915_private *dev_priv)
> pcbr = I915_READ(VLV_PCBR);
> if ((pcbr >> VLV_PCBR_ADDR_SHIFT) == 0) {
> DRM_DEBUG_DRIVER("BIOS didn't set up PCBR, fixing up\n");
> - paddr = dev_priv->dsm.end + 1 - pctx_size;
> + paddr = i915->dsm.end + 1 - pctx_size;
> GEM_BUG_ON(paddr > U32_MAX);
>
> pctx_paddr = (paddr & (~4095));
> @@ -7707,7 +7707,7 @@ static void cherryview_setup_pctx(struct drm_i915_private *dev_priv)
> DRM_DEBUG_DRIVER("PCBR: 0x%08x\n", I915_READ(VLV_PCBR));
> }
>
> -static void valleyview_setup_pctx(struct drm_i915_private *dev_priv)
> +static void valleyview_setup_pctx(struct drm_i915_private *i915)
> {
> struct drm_i915_gem_object *pctx;
> resource_size_t pctx_paddr;
> @@ -7719,8 +7719,8 @@ static void valleyview_setup_pctx(struct drm_i915_private *dev_priv)
> /* BIOS set it up already, grab the pre-alloc'd space */
> resource_size_t pcbr_offset;
>
> - pcbr_offset = (pcbr & (~4095)) - dev_priv->dsm.start;
> - pctx = i915_gem_object_create_stolen_for_preallocated(dev_priv,
> + pcbr_offset = (pcbr & (~4095)) - i915->dsm.start;
> + pctx = i915_gem_object_create_stolen_for_preallocated(i915,
> pcbr_offset,
> I915_GTT_OFFSET_NONE,
> pctx_size);
> @@ -7737,148 +7737,148 @@ static void valleyview_setup_pctx(struct drm_i915_private *dev_priv)
> * overlap with other ranges, such as the frame buffer, protected
> * memory, or any other relevant ranges.
> */
> - pctx = i915_gem_object_create_stolen(dev_priv, pctx_size);
> + pctx = i915_gem_object_create_stolen(i915, pctx_size);
> if (!pctx) {
> DRM_DEBUG("not enough stolen space for PCTX, disabling\n");
> goto out;
> }
>
> GEM_BUG_ON(range_overflows_t(u64,
> - dev_priv->dsm.start,
> + i915->dsm.start,
> pctx->stolen->start,
> U32_MAX));
> - pctx_paddr = dev_priv->dsm.start + pctx->stolen->start;
> + pctx_paddr = i915->dsm.start + pctx->stolen->start;
> I915_WRITE(VLV_PCBR, pctx_paddr);
>
> out:
> DRM_DEBUG_DRIVER("PCBR: 0x%08x\n", I915_READ(VLV_PCBR));
> - dev_priv->vlv_pctx = pctx;
> + i915->vlv_pctx = pctx;
> }
>
> -static void valleyview_cleanup_pctx(struct drm_i915_private *dev_priv)
> +static void valleyview_cleanup_pctx(struct drm_i915_private *i915)
> {
> struct drm_i915_gem_object *pctx;
>
> - pctx = fetch_and_zero(&dev_priv->vlv_pctx);
> + pctx = fetch_and_zero(&i915->vlv_pctx);
> if (pctx)
> i915_gem_object_put(pctx);
> }
>
> -static void vlv_init_gpll_ref_freq(struct drm_i915_private *dev_priv)
> +static void vlv_init_gpll_ref_freq(struct drm_i915_private *i915)
> {
> - dev_priv->gt_pm.rps.gpll_ref_freq =
> - vlv_get_cck_clock(dev_priv, "GPLL ref",
> + i915->gt_pm.rps.gpll_ref_freq =
> + vlv_get_cck_clock(i915, "GPLL ref",
> CCK_GPLL_CLOCK_CONTROL,
> - dev_priv->czclk_freq);
> + i915->czclk_freq);
>
> DRM_DEBUG_DRIVER("GPLL reference freq: %d kHz\n",
> - dev_priv->gt_pm.rps.gpll_ref_freq);
> + i915->gt_pm.rps.gpll_ref_freq);
> }
>
> -static void valleyview_init_gt_powersave(struct drm_i915_private *dev_priv)
> +static void valleyview_init_gt_powersave(struct drm_i915_private *i915)
> {
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct intel_rps *rps = &i915->gt_pm.rps;
> u32 val;
>
> - valleyview_setup_pctx(dev_priv);
> + valleyview_setup_pctx(i915);
>
> - vlv_iosf_sb_get(dev_priv,
> + vlv_iosf_sb_get(i915,
> BIT(VLV_IOSF_SB_PUNIT) |
> BIT(VLV_IOSF_SB_NC) |
> BIT(VLV_IOSF_SB_CCK));
>
> - vlv_init_gpll_ref_freq(dev_priv);
> + vlv_init_gpll_ref_freq(i915);
>
> - val = vlv_punit_read(dev_priv, PUNIT_REG_GPU_FREQ_STS);
> + val = vlv_punit_read(i915, PUNIT_REG_GPU_FREQ_STS);
> switch ((val >> 6) & 3) {
> case 0:
> case 1:
> - dev_priv->mem_freq = 800;
> + i915->mem_freq = 800;
> break;
> case 2:
> - dev_priv->mem_freq = 1066;
> + i915->mem_freq = 1066;
> break;
> case 3:
> - dev_priv->mem_freq = 1333;
> + i915->mem_freq = 1333;
> break;
> }
> - DRM_DEBUG_DRIVER("DDR speed: %d MHz\n", dev_priv->mem_freq);
> + DRM_DEBUG_DRIVER("DDR speed: %d MHz\n", i915->mem_freq);
>
> - rps->max_freq = valleyview_rps_max_freq(dev_priv);
> + rps->max_freq = valleyview_rps_max_freq(i915);
> rps->rp0_freq = rps->max_freq;
> DRM_DEBUG_DRIVER("max GPU freq: %d MHz (%u)\n",
> - intel_gpu_freq(dev_priv, rps->max_freq),
> + intel_gpu_freq(i915, rps->max_freq),
> rps->max_freq);
>
> - rps->efficient_freq = valleyview_rps_rpe_freq(dev_priv);
> + rps->efficient_freq = valleyview_rps_rpe_freq(i915);
> DRM_DEBUG_DRIVER("RPe GPU freq: %d MHz (%u)\n",
> - intel_gpu_freq(dev_priv, rps->efficient_freq),
> + intel_gpu_freq(i915, rps->efficient_freq),
> rps->efficient_freq);
>
> - rps->rp1_freq = valleyview_rps_guar_freq(dev_priv);
> + rps->rp1_freq = valleyview_rps_guar_freq(i915);
> DRM_DEBUG_DRIVER("RP1(Guar Freq) GPU freq: %d MHz (%u)\n",
> - intel_gpu_freq(dev_priv, rps->rp1_freq),
> + intel_gpu_freq(i915, rps->rp1_freq),
> rps->rp1_freq);
>
> - rps->min_freq = valleyview_rps_min_freq(dev_priv);
> + rps->min_freq = valleyview_rps_min_freq(i915);
> DRM_DEBUG_DRIVER("min GPU freq: %d MHz (%u)\n",
> - intel_gpu_freq(dev_priv, rps->min_freq),
> + intel_gpu_freq(i915, rps->min_freq),
> rps->min_freq);
>
> - vlv_iosf_sb_put(dev_priv,
> + vlv_iosf_sb_put(i915,
> BIT(VLV_IOSF_SB_PUNIT) |
> BIT(VLV_IOSF_SB_NC) |
> BIT(VLV_IOSF_SB_CCK));
> }
>
> -static void cherryview_init_gt_powersave(struct drm_i915_private *dev_priv)
> +static void cherryview_init_gt_powersave(struct drm_i915_private *i915)
> {
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct intel_rps *rps = &i915->gt_pm.rps;
> u32 val;
>
> - cherryview_setup_pctx(dev_priv);
> + cherryview_setup_pctx(i915);
>
> - vlv_iosf_sb_get(dev_priv,
> + vlv_iosf_sb_get(i915,
> BIT(VLV_IOSF_SB_PUNIT) |
> BIT(VLV_IOSF_SB_NC) |
> BIT(VLV_IOSF_SB_CCK));
>
> - vlv_init_gpll_ref_freq(dev_priv);
> + vlv_init_gpll_ref_freq(i915);
>
> - val = vlv_cck_read(dev_priv, CCK_FUSE_REG);
> + val = vlv_cck_read(i915, CCK_FUSE_REG);
>
> switch ((val >> 2) & 0x7) {
> case 3:
> - dev_priv->mem_freq = 2000;
> + i915->mem_freq = 2000;
> break;
> default:
> - dev_priv->mem_freq = 1600;
> + i915->mem_freq = 1600;
> break;
> }
> - DRM_DEBUG_DRIVER("DDR speed: %d MHz\n", dev_priv->mem_freq);
> + DRM_DEBUG_DRIVER("DDR speed: %d MHz\n", i915->mem_freq);
>
> - rps->max_freq = cherryview_rps_max_freq(dev_priv);
> + rps->max_freq = cherryview_rps_max_freq(i915);
> rps->rp0_freq = rps->max_freq;
> DRM_DEBUG_DRIVER("max GPU freq: %d MHz (%u)\n",
> - intel_gpu_freq(dev_priv, rps->max_freq),
> + intel_gpu_freq(i915, rps->max_freq),
> rps->max_freq);
>
> - rps->efficient_freq = cherryview_rps_rpe_freq(dev_priv);
> + rps->efficient_freq = cherryview_rps_rpe_freq(i915);
> DRM_DEBUG_DRIVER("RPe GPU freq: %d MHz (%u)\n",
> - intel_gpu_freq(dev_priv, rps->efficient_freq),
> + intel_gpu_freq(i915, rps->efficient_freq),
> rps->efficient_freq);
>
> - rps->rp1_freq = cherryview_rps_guar_freq(dev_priv);
> + rps->rp1_freq = cherryview_rps_guar_freq(i915);
> DRM_DEBUG_DRIVER("RP1(Guar) GPU freq: %d MHz (%u)\n",
> - intel_gpu_freq(dev_priv, rps->rp1_freq),
> + intel_gpu_freq(i915, rps->rp1_freq),
> rps->rp1_freq);
>
> - rps->min_freq = cherryview_rps_min_freq(dev_priv);
> + rps->min_freq = cherryview_rps_min_freq(i915);
> DRM_DEBUG_DRIVER("min GPU freq: %d MHz (%u)\n",
> - intel_gpu_freq(dev_priv, rps->min_freq),
> + intel_gpu_freq(i915, rps->min_freq),
> rps->min_freq);
>
> - vlv_iosf_sb_put(dev_priv,
> + vlv_iosf_sb_put(i915,
> BIT(VLV_IOSF_SB_PUNIT) |
> BIT(VLV_IOSF_SB_NC) |
> BIT(VLV_IOSF_SB_CCK));
> @@ -7888,12 +7888,12 @@ static void cherryview_init_gt_powersave(struct drm_i915_private *dev_priv)
> "Odd GPU freq values\n");
> }
>
> -static void valleyview_cleanup_gt_powersave(struct drm_i915_private *dev_priv)
> +static void valleyview_cleanup_gt_powersave(struct drm_i915_private *i915)
> {
> - valleyview_cleanup_pctx(dev_priv);
> + valleyview_cleanup_pctx(i915);
> }
>
> -static void cherryview_enable_rc6(struct drm_i915_private *dev_priv)
> +static void cherryview_enable_rc6(struct drm_i915_private *i915)
> {
> struct intel_engine_cs *engine;
> enum intel_engine_id id;
> @@ -7907,11 +7907,11 @@ static void cherryview_enable_rc6(struct drm_i915_private *dev_priv)
> I915_WRITE(GTFIFODBG, gtfifodbg);
> }
>
> - cherryview_check_pctx(dev_priv);
> + cherryview_check_pctx(i915);
>
> /* 1a & 1b: Get forcewake during program sequence. Although the driver
> * hasn't enabled a state yet where we need forcewake, BIOS may have.*/
> - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
>
> /* Disable RC states. */
> I915_WRITE(GEN6_RC_CONTROL, 0);
> @@ -7921,7 +7921,7 @@ static void cherryview_enable_rc6(struct drm_i915_private *dev_priv)
> I915_WRITE(GEN6_RC_EVALUATION_INTERVAL, 125000); /* 12500 * 1280ns */
> I915_WRITE(GEN6_RC_IDLE_HYSTERSIS, 25); /* 25 * 1280ns */
>
> - for_each_engine(engine, dev_priv, id)
> + for_each_engine(engine, i915, id)
> I915_WRITE(RING_MAX_IDLE(engine->mmio_base), 10);
> I915_WRITE(GEN6_RC_SLEEP, 0);
>
> @@ -7943,14 +7943,14 @@ static void cherryview_enable_rc6(struct drm_i915_private *dev_priv)
> rc6_mode = GEN7_RC_CTL_TO_MODE;
> I915_WRITE(GEN6_RC_CONTROL, rc6_mode);
>
> - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> }
>
> -static void cherryview_enable_rps(struct drm_i915_private *dev_priv)
> +static void cherryview_enable_rps(struct drm_i915_private *i915)
> {
> u32 val;
>
> - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
>
> /* 1: Program defaults and thresholds for RPS*/
> I915_WRITE(GEN6_RP_DOWN_TIMEOUT, 1000000);
> @@ -7970,14 +7970,14 @@ static void cherryview_enable_rps(struct drm_i915_private *dev_priv)
> GEN6_RP_DOWN_IDLE_AVG);
>
> /* Setting Fixed Bias */
> - vlv_punit_get(dev_priv);
> + vlv_punit_get(i915);
>
> val = VLV_OVERRIDE_EN | VLV_SOC_TDP_EN | CHV_BIAS_CPU_50_SOC_50;
> - vlv_punit_write(dev_priv, VLV_TURBO_SOC_OVERRIDE, val);
> + vlv_punit_write(i915, VLV_TURBO_SOC_OVERRIDE, val);
>
> - val = vlv_punit_read(dev_priv, PUNIT_REG_GPU_FREQ_STS);
> + val = vlv_punit_read(i915, PUNIT_REG_GPU_FREQ_STS);
>
> - vlv_punit_put(dev_priv);
> + vlv_punit_put(i915);
>
> /* RPS code assumes GPLL is used */
> WARN_ONCE((val & GPLLENABLE) == 0, "GPLL not enabled\n");
> @@ -7985,18 +7985,18 @@ static void cherryview_enable_rps(struct drm_i915_private *dev_priv)
> DRM_DEBUG_DRIVER("GPLL enabled? %s\n", yesno(val & GPLLENABLE));
> DRM_DEBUG_DRIVER("GPU status: 0x%08x\n", val);
>
> - reset_rps(dev_priv, valleyview_set_rps);
> + reset_rps(i915, valleyview_set_rps);
>
> - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> }
>
> -static void valleyview_enable_rc6(struct drm_i915_private *dev_priv)
> +static void valleyview_enable_rc6(struct drm_i915_private *i915)
> {
> struct intel_engine_cs *engine;
> enum intel_engine_id id;
> u32 gtfifodbg;
>
> - valleyview_check_pctx(dev_priv);
> + valleyview_check_pctx(i915);
>
> gtfifodbg = I915_READ(GTFIFODBG);
> if (gtfifodbg) {
> @@ -8005,7 +8005,7 @@ static void valleyview_enable_rc6(struct drm_i915_private *dev_priv)
> I915_WRITE(GTFIFODBG, gtfifodbg);
> }
>
> - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
>
> /* Disable RC states. */
> I915_WRITE(GEN6_RC_CONTROL, 0);
> @@ -8014,7 +8014,7 @@ static void valleyview_enable_rc6(struct drm_i915_private *dev_priv)
> I915_WRITE(GEN6_RC_EVALUATION_INTERVAL, 125000);
> I915_WRITE(GEN6_RC_IDLE_HYSTERSIS, 25);
>
> - for_each_engine(engine, dev_priv, id)
> + for_each_engine(engine, i915, id)
> I915_WRITE(RING_MAX_IDLE(engine->mmio_base), 10);
>
> I915_WRITE(GEN6_RC6_THRESHOLD, 0x557);
> @@ -8030,14 +8030,14 @@ static void valleyview_enable_rc6(struct drm_i915_private *dev_priv)
> I915_WRITE(GEN6_RC_CONTROL,
> GEN7_RC_CTL_TO_MODE | VLV_RC_CTL_CTX_RST_PARALLEL);
>
> - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> }
>
> -static void valleyview_enable_rps(struct drm_i915_private *dev_priv)
> +static void valleyview_enable_rps(struct drm_i915_private *i915)
> {
> u32 val;
>
> - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
>
> I915_WRITE(GEN6_RP_DOWN_TIMEOUT, 1000000);
> I915_WRITE(GEN6_RP_UP_THRESHOLD, 59400);
> @@ -8055,15 +8055,15 @@ static void valleyview_enable_rps(struct drm_i915_private *dev_priv)
> GEN6_RP_UP_BUSY_AVG |
> GEN6_RP_DOWN_IDLE_CONT);
>
> - vlv_punit_get(dev_priv);
> + vlv_punit_get(i915);
>
> /* Setting Fixed Bias */
> val = VLV_OVERRIDE_EN | VLV_SOC_TDP_EN | VLV_BIAS_CPU_125_SOC_875;
> - vlv_punit_write(dev_priv, VLV_TURBO_SOC_OVERRIDE, val);
> + vlv_punit_write(i915, VLV_TURBO_SOC_OVERRIDE, val);
>
> - val = vlv_punit_read(dev_priv, PUNIT_REG_GPU_FREQ_STS);
> + val = vlv_punit_read(i915, PUNIT_REG_GPU_FREQ_STS);
>
> - vlv_punit_put(dev_priv);
> + vlv_punit_put(i915);
>
> /* RPS code assumes GPLL is used */
> WARN_ONCE((val & GPLLENABLE) == 0, "GPLL not enabled\n");
> @@ -8071,9 +8071,9 @@ static void valleyview_enable_rps(struct drm_i915_private *dev_priv)
> DRM_DEBUG_DRIVER("GPLL enabled? %s\n", yesno(val & GPLLENABLE));
> DRM_DEBUG_DRIVER("GPU status: 0x%08x\n", val);
>
> - reset_rps(dev_priv, valleyview_set_rps);
> + reset_rps(i915, valleyview_set_rps);
>
> - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> }
>
> static unsigned long intel_pxfreq(u32 vidfreq)
> @@ -8105,7 +8105,7 @@ static const struct cparams {
> { 0, 800, 231, 23784 },
> };
>
> -static unsigned long __i915_chipset_val(struct drm_i915_private *dev_priv)
> +static unsigned long __i915_chipset_val(struct drm_i915_private *i915)
> {
> u64 total_count, diff, ret;
> u32 count1, count2, count3, m = 0, c = 0;
> @@ -8114,7 +8114,7 @@ static unsigned long __i915_chipset_val(struct drm_i915_private *dev_priv)
>
> lockdep_assert_held(&mchdev_lock);
>
> - diff1 = now - dev_priv->ips.last_time1;
> + diff1 = now - i915->ips.last_time1;
>
> /* Prevent division-by-zero if we are asking too fast.
> * Also, we don't get interesting results if we are polling
> @@ -8122,7 +8122,7 @@ static unsigned long __i915_chipset_val(struct drm_i915_private *dev_priv)
> * in such cases.
> */
> if (diff1 <= 10)
> - return dev_priv->ips.chipset_power;
> + return i915->ips.chipset_power;
>
> count1 = I915_READ(DMIEC);
> count2 = I915_READ(DDREC);
> @@ -8131,16 +8131,16 @@ static unsigned long __i915_chipset_val(struct drm_i915_private *dev_priv)
> total_count = count1 + count2 + count3;
>
> /* FIXME: handle per-counter overflow */
> - if (total_count < dev_priv->ips.last_count1) {
> - diff = ~0UL - dev_priv->ips.last_count1;
> + if (total_count < i915->ips.last_count1) {
> + diff = ~0UL - i915->ips.last_count1;
> diff += total_count;
> } else {
> - diff = total_count - dev_priv->ips.last_count1;
> + diff = total_count - i915->ips.last_count1;
> }
>
> for (i = 0; i < ARRAY_SIZE(cparams); i++) {
> - if (cparams[i].i == dev_priv->ips.c_m &&
> - cparams[i].t == dev_priv->ips.r_t) {
> + if (cparams[i].i == i915->ips.c_m &&
> + cparams[i].t == i915->ips.r_t) {
> m = cparams[i].m;
> c = cparams[i].c;
> break;
> @@ -8151,25 +8151,25 @@ static unsigned long __i915_chipset_val(struct drm_i915_private *dev_priv)
> ret = ((m * diff) + c);
> ret = div_u64(ret, 10);
>
> - dev_priv->ips.last_count1 = total_count;
> - dev_priv->ips.last_time1 = now;
> + i915->ips.last_count1 = total_count;
> + i915->ips.last_time1 = now;
>
> - dev_priv->ips.chipset_power = ret;
> + i915->ips.chipset_power = ret;
>
> return ret;
> }
>
> -unsigned long i915_chipset_val(struct drm_i915_private *dev_priv)
> +unsigned long i915_chipset_val(struct drm_i915_private *i915)
> {
> intel_wakeref_t wakeref;
> unsigned long val = 0;
>
> - if (!IS_GEN(dev_priv, 5))
> + if (!IS_GEN(i915, 5))
> return 0;
>
> - with_intel_runtime_pm(dev_priv, wakeref) {
> + with_intel_runtime_pm(i915, wakeref) {
> spin_lock_irq(&mchdev_lock);
> - val = __i915_chipset_val(dev_priv);
> + val = __i915_chipset_val(i915);
> spin_unlock_irq(&mchdev_lock);
> }
>
> @@ -8202,18 +8202,18 @@ static int _pxvid_to_vd(u8 pxvid)
> return (pxvid + 2) * 125;
> }
>
> -static u32 pvid_to_extvid(struct drm_i915_private *dev_priv, u8 pxvid)
> +static u32 pvid_to_extvid(struct drm_i915_private *i915, u8 pxvid)
> {
> const int vd = _pxvid_to_vd(pxvid);
> const int vm = vd - 1125;
>
> - if (INTEL_INFO(dev_priv)->is_mobile)
> + if (INTEL_INFO(i915)->is_mobile)
> return vm > 0 ? vm : 0;
>
> return vd;
> }
>
> -static void __i915_update_gfx_val(struct drm_i915_private *dev_priv)
> +static void __i915_update_gfx_val(struct drm_i915_private *i915)
> {
> u64 now, diff, diffms;
> u32 count;
> @@ -8221,7 +8221,7 @@ static void __i915_update_gfx_val(struct drm_i915_private *dev_priv)
> lockdep_assert_held(&mchdev_lock);
>
> now = ktime_get_raw_ns();
> - diffms = now - dev_priv->ips.last_time2;
> + diffms = now - i915->ips.last_time2;
> do_div(diffms, NSEC_PER_MSEC);
>
> /* Don't divide by 0 */
> @@ -8230,50 +8230,50 @@ static void __i915_update_gfx_val(struct drm_i915_private *dev_priv)
>
> count = I915_READ(GFXEC);
>
> - if (count < dev_priv->ips.last_count2) {
> - diff = ~0UL - dev_priv->ips.last_count2;
> + if (count < i915->ips.last_count2) {
> + diff = ~0UL - i915->ips.last_count2;
> diff += count;
> } else {
> - diff = count - dev_priv->ips.last_count2;
> + diff = count - i915->ips.last_count2;
> }
>
> - dev_priv->ips.last_count2 = count;
> - dev_priv->ips.last_time2 = now;
> + i915->ips.last_count2 = count;
> + i915->ips.last_time2 = now;
>
> /* More magic constants... */
> diff = diff * 1181;
> diff = div_u64(diff, diffms * 10);
> - dev_priv->ips.gfx_power = diff;
> + i915->ips.gfx_power = diff;
> }
>
> -void i915_update_gfx_val(struct drm_i915_private *dev_priv)
> +void i915_update_gfx_val(struct drm_i915_private *i915)
> {
> intel_wakeref_t wakeref;
>
> - if (!IS_GEN(dev_priv, 5))
> + if (!IS_GEN(i915, 5))
> return;
>
> - with_intel_runtime_pm(dev_priv, wakeref) {
> + with_intel_runtime_pm(i915, wakeref) {
> spin_lock_irq(&mchdev_lock);
> - __i915_update_gfx_val(dev_priv);
> + __i915_update_gfx_val(i915);
> spin_unlock_irq(&mchdev_lock);
> }
> }
>
> -static unsigned long __i915_gfx_val(struct drm_i915_private *dev_priv)
> +static unsigned long __i915_gfx_val(struct drm_i915_private *i915)
> {
> unsigned long t, corr, state1, corr2, state2;
> u32 pxvid, ext_v;
>
> lockdep_assert_held(&mchdev_lock);
>
> - pxvid = I915_READ(PXVFREQ(dev_priv->gt_pm.rps.cur_freq));
> + pxvid = I915_READ(PXVFREQ(i915->gt_pm.rps.cur_freq));
> pxvid = (pxvid >> 24) & 0x7f;
> - ext_v = pvid_to_extvid(dev_priv, pxvid);
> + ext_v = pvid_to_extvid(i915, pxvid);
>
> state1 = ext_v;
>
> - t = i915_mch_val(dev_priv);
> + t = i915_mch_val(i915);
>
> /* Revel in the empirically derived constants */
>
> @@ -8287,27 +8287,27 @@ static unsigned long __i915_gfx_val(struct drm_i915_private *dev_priv)
>
> corr = corr * ((150142 * state1) / 10000 - 78642);
> corr /= 100000;
> - corr2 = (corr * dev_priv->ips.corr);
> + corr2 = (corr * i915->ips.corr);
>
> state2 = (corr2 * state1) / 10000;
> state2 /= 100; /* convert to mW */
>
> - __i915_update_gfx_val(dev_priv);
> + __i915_update_gfx_val(i915);
>
> - return dev_priv->ips.gfx_power + state2;
> + return i915->ips.gfx_power + state2;
> }
>
> -unsigned long i915_gfx_val(struct drm_i915_private *dev_priv)
> +unsigned long i915_gfx_val(struct drm_i915_private *i915)
> {
> intel_wakeref_t wakeref;
> unsigned long val = 0;
>
> - if (!IS_GEN(dev_priv, 5))
> + if (!IS_GEN(i915, 5))
> return 0;
>
> - with_intel_runtime_pm(dev_priv, wakeref) {
> + with_intel_runtime_pm(i915, wakeref) {
> spin_lock_irq(&mchdev_lock);
> - val = __i915_gfx_val(dev_priv);
> + val = __i915_gfx_val(i915);
> spin_unlock_irq(&mchdev_lock);
> }
>
> @@ -8471,11 +8471,11 @@ ips_ping_for_i915_load(void)
> }
> }
>
> -void intel_gpu_ips_init(struct drm_i915_private *dev_priv)
> +void intel_gpu_ips_init(struct drm_i915_private *i915)
> {
> /* We only register the i915 ips part with intel-ips once everything is
> * set up, to avoid intel-ips sneaking in and reading bogus values. */
> - rcu_assign_pointer(i915_mch_dev, dev_priv);
> + rcu_assign_pointer(i915_mch_dev, i915);
>
> ips_ping_for_i915_load();
> }
> @@ -8485,7 +8485,7 @@ void intel_gpu_ips_teardown(void)
> rcu_assign_pointer(i915_mch_dev, NULL);
> }
>
> -static void intel_init_emon(struct drm_i915_private *dev_priv)
> +static void intel_init_emon(struct drm_i915_private *i915)
> {
> u32 lcfuse;
> u8 pxw[16];
> @@ -8552,40 +8552,40 @@ static void intel_init_emon(struct drm_i915_private *dev_priv)
>
> lcfuse = I915_READ(LCFUSE02);
>
> - dev_priv->ips.corr = (lcfuse & LCFUSE_HIV_MASK);
> + i915->ips.corr = (lcfuse & LCFUSE_HIV_MASK);
> }
>
> -void intel_init_gt_powersave(struct drm_i915_private *dev_priv)
> +void intel_init_gt_powersave(struct drm_i915_private *i915)
> {
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct intel_rps *rps = &i915->gt_pm.rps;
>
> /*
> * RPM depends on RC6 to save restore the GT HW context, so make RC6 a
> * requirement.
> */
> - if (!sanitize_rc6(dev_priv)) {
> + if (!sanitize_rc6(i915)) {
> DRM_INFO("RC6 disabled, disabling runtime PM support\n");
> - pm_runtime_get(&dev_priv->drm.pdev->dev);
> + pm_runtime_get(&i915->drm.pdev->dev);
> }
>
> /* Initialize RPS limits (for userspace) */
> - if (IS_CHERRYVIEW(dev_priv))
> - cherryview_init_gt_powersave(dev_priv);
> - else if (IS_VALLEYVIEW(dev_priv))
> - valleyview_init_gt_powersave(dev_priv);
> - else if (INTEL_GEN(dev_priv) >= 6)
> - gen6_init_rps_frequencies(dev_priv);
> + if (IS_CHERRYVIEW(i915))
> + cherryview_init_gt_powersave(i915);
> + else if (IS_VALLEYVIEW(i915))
> + valleyview_init_gt_powersave(i915);
> + else if (INTEL_GEN(i915) >= 6)
> + gen6_init_rps_frequencies(i915);
>
> /* Derive initial user preferences/limits from the hardware limits */
> rps->max_freq_softlimit = rps->max_freq;
> rps->min_freq_softlimit = rps->min_freq;
>
> /* After setting max-softlimit, find the overclock max freq */
> - if (IS_GEN(dev_priv, 6) ||
> - IS_IVYBRIDGE(dev_priv) || IS_HASWELL(dev_priv)) {
> + if (IS_GEN(i915, 6) ||
> + IS_IVYBRIDGE(i915) || IS_HASWELL(i915)) {
> u32 params = 0;
>
> - sandybridge_pcode_read(dev_priv, GEN6_READ_OC_PARAMS,
> + sandybridge_pcode_read(i915, GEN6_READ_OC_PARAMS,
> ¶ms, NULL);
> if (params & BIT(31)) { /* OC supported */
> DRM_DEBUG_DRIVER("Overclocking supported, max: %dMHz, overclock: %dMHz\n",
> @@ -8601,25 +8601,25 @@ void intel_init_gt_powersave(struct drm_i915_private *dev_priv)
> rps->cur_freq = rps->idle_freq;
> }
>
> -void intel_cleanup_gt_powersave(struct drm_i915_private *dev_priv)
> +void intel_cleanup_gt_powersave(struct drm_i915_private *i915)
> {
> - if (IS_VALLEYVIEW(dev_priv))
> - valleyview_cleanup_gt_powersave(dev_priv);
> + if (IS_VALLEYVIEW(i915))
> + valleyview_cleanup_gt_powersave(i915);
>
> - if (!HAS_RC6(dev_priv))
> - pm_runtime_put(&dev_priv->drm.pdev->dev);
> + if (!HAS_RC6(i915))
> + pm_runtime_put(&i915->drm.pdev->dev);
> }
>
> -void intel_sanitize_gt_powersave(struct drm_i915_private *dev_priv)
> +void intel_sanitize_gt_powersave(struct drm_i915_private *i915)
> {
> - dev_priv->gt_pm.rps.enabled = true; /* force RPS disabling */
> - dev_priv->gt_pm.rc6.enabled = true; /* force RC6 disabling */
> - intel_disable_gt_powersave(dev_priv);
> + i915->gt_pm.rps.enabled = true; /* force RPS disabling */
> + i915->gt_pm.rc6.enabled = true; /* force RC6 disabling */
> + intel_disable_gt_powersave(i915);
>
> - if (INTEL_GEN(dev_priv) >= 11)
> - gen11_reset_rps_interrupts(dev_priv);
> - else if (INTEL_GEN(dev_priv) >= 6)
> - gen6_reset_rps_interrupts(dev_priv);
> + if (INTEL_GEN(i915) >= 11)
> + gen11_reset_rps_interrupts(i915);
> + else if (INTEL_GEN(i915) >= 6)
> + gen6_reset_rps_interrupts(i915);
> }
>
> static inline void intel_disable_llc_pstate(struct drm_i915_private *i915)
> @@ -8634,56 +8634,56 @@ static inline void intel_disable_llc_pstate(struct drm_i915_private *i915)
> i915->gt_pm.llc_pstate.enabled = false;
> }
>
> -static void intel_disable_rc6(struct drm_i915_private *dev_priv)
> +static void intel_disable_rc6(struct drm_i915_private *i915)
> {
> - lockdep_assert_held(&dev_priv->gt_pm.rps.lock);
> + lockdep_assert_held(&i915->gt_pm.rps.lock);
>
> - if (!dev_priv->gt_pm.rc6.enabled)
> + if (!i915->gt_pm.rc6.enabled)
> return;
>
> - if (INTEL_GEN(dev_priv) >= 9)
> - gen9_disable_rc6(dev_priv);
> - else if (IS_CHERRYVIEW(dev_priv))
> - cherryview_disable_rc6(dev_priv);
> - else if (IS_VALLEYVIEW(dev_priv))
> - valleyview_disable_rc6(dev_priv);
> - else if (INTEL_GEN(dev_priv) >= 6)
> - gen6_disable_rc6(dev_priv);
> + if (INTEL_GEN(i915) >= 9)
> + gen9_disable_rc6(i915);
> + else if (IS_CHERRYVIEW(i915))
> + cherryview_disable_rc6(i915);
> + else if (IS_VALLEYVIEW(i915))
> + valleyview_disable_rc6(i915);
> + else if (INTEL_GEN(i915) >= 6)
> + gen6_disable_rc6(i915);
>
> - dev_priv->gt_pm.rc6.enabled = false;
> + i915->gt_pm.rc6.enabled = false;
> }
>
> -static void intel_disable_rps(struct drm_i915_private *dev_priv)
> +static void intel_disable_rps(struct drm_i915_private *i915)
> {
> - lockdep_assert_held(&dev_priv->gt_pm.rps.lock);
> + lockdep_assert_held(&i915->gt_pm.rps.lock);
>
> - if (!dev_priv->gt_pm.rps.enabled)
> + if (!i915->gt_pm.rps.enabled)
> return;
>
> - if (INTEL_GEN(dev_priv) >= 9)
> - gen9_disable_rps(dev_priv);
> - else if (IS_CHERRYVIEW(dev_priv))
> - cherryview_disable_rps(dev_priv);
> - else if (IS_VALLEYVIEW(dev_priv))
> - valleyview_disable_rps(dev_priv);
> - else if (INTEL_GEN(dev_priv) >= 6)
> - gen6_disable_rps(dev_priv);
> - else if (IS_IRONLAKE_M(dev_priv))
> - ironlake_disable_drps(dev_priv);
> + if (INTEL_GEN(i915) >= 9)
> + gen9_disable_rps(i915);
> + else if (IS_CHERRYVIEW(i915))
> + cherryview_disable_rps(i915);
> + else if (IS_VALLEYVIEW(i915))
> + valleyview_disable_rps(i915);
> + else if (INTEL_GEN(i915) >= 6)
> + gen6_disable_rps(i915);
> + else if (IS_IRONLAKE_M(i915))
> + ironlake_disable_drps(i915);
>
> - dev_priv->gt_pm.rps.enabled = false;
> + i915->gt_pm.rps.enabled = false;
> }
>
> -void intel_disable_gt_powersave(struct drm_i915_private *dev_priv)
> +void intel_disable_gt_powersave(struct drm_i915_private *i915)
> {
> - mutex_lock(&dev_priv->gt_pm.rps.lock);
> + mutex_lock(&i915->gt_pm.rps.lock);
>
> - intel_disable_rc6(dev_priv);
> - intel_disable_rps(dev_priv);
> - if (HAS_LLC(dev_priv))
> - intel_disable_llc_pstate(dev_priv);
> + intel_disable_rc6(i915);
> + intel_disable_rps(i915);
> + if (HAS_LLC(i915))
> + intel_disable_llc_pstate(i915);
>
> - mutex_unlock(&dev_priv->gt_pm.rps.lock);
> + mutex_unlock(&i915->gt_pm.rps.lock);
> }
>
> static inline void intel_enable_llc_pstate(struct drm_i915_private *i915)
> @@ -8698,51 +8698,51 @@ static inline void intel_enable_llc_pstate(struct drm_i915_private *i915)
> i915->gt_pm.llc_pstate.enabled = true;
> }
>
> -static void intel_enable_rc6(struct drm_i915_private *dev_priv)
> +static void intel_enable_rc6(struct drm_i915_private *i915)
> {
> - lockdep_assert_held(&dev_priv->gt_pm.rps.lock);
> + lockdep_assert_held(&i915->gt_pm.rps.lock);
>
> - if (dev_priv->gt_pm.rc6.enabled)
> + if (i915->gt_pm.rc6.enabled)
> return;
>
> - if (IS_CHERRYVIEW(dev_priv))
> - cherryview_enable_rc6(dev_priv);
> - else if (IS_VALLEYVIEW(dev_priv))
> - valleyview_enable_rc6(dev_priv);
> - else if (INTEL_GEN(dev_priv) >= 11)
> - gen11_enable_rc6(dev_priv);
> - else if (INTEL_GEN(dev_priv) >= 9)
> - gen9_enable_rc6(dev_priv);
> - else if (IS_BROADWELL(dev_priv))
> - gen8_enable_rc6(dev_priv);
> - else if (INTEL_GEN(dev_priv) >= 6)
> - gen6_enable_rc6(dev_priv);
> + if (IS_CHERRYVIEW(i915))
> + cherryview_enable_rc6(i915);
> + else if (IS_VALLEYVIEW(i915))
> + valleyview_enable_rc6(i915);
> + else if (INTEL_GEN(i915) >= 11)
> + gen11_enable_rc6(i915);
> + else if (INTEL_GEN(i915) >= 9)
> + gen9_enable_rc6(i915);
> + else if (IS_BROADWELL(i915))
> + gen8_enable_rc6(i915);
> + else if (INTEL_GEN(i915) >= 6)
> + gen6_enable_rc6(i915);
>
> - dev_priv->gt_pm.rc6.enabled = true;
> + i915->gt_pm.rc6.enabled = true;
> }
>
> -static void intel_enable_rps(struct drm_i915_private *dev_priv)
> +static void intel_enable_rps(struct drm_i915_private *i915)
> {
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct intel_rps *rps = &i915->gt_pm.rps;
>
> lockdep_assert_held(&rps->lock);
>
> if (rps->enabled)
> return;
>
> - if (IS_CHERRYVIEW(dev_priv)) {
> - cherryview_enable_rps(dev_priv);
> - } else if (IS_VALLEYVIEW(dev_priv)) {
> - valleyview_enable_rps(dev_priv);
> - } else if (INTEL_GEN(dev_priv) >= 9) {
> - gen9_enable_rps(dev_priv);
> - } else if (IS_BROADWELL(dev_priv)) {
> - gen8_enable_rps(dev_priv);
> - } else if (INTEL_GEN(dev_priv) >= 6) {
> - gen6_enable_rps(dev_priv);
> - } else if (IS_IRONLAKE_M(dev_priv)) {
> - ironlake_enable_drps(dev_priv);
> - intel_init_emon(dev_priv);
> + if (IS_CHERRYVIEW(i915)) {
> + cherryview_enable_rps(i915);
> + } else if (IS_VALLEYVIEW(i915)) {
> + valleyview_enable_rps(i915);
> + } else if (INTEL_GEN(i915) >= 9) {
> + gen9_enable_rps(i915);
> + } else if (IS_BROADWELL(i915)) {
> + gen8_enable_rps(i915);
> + } else if (INTEL_GEN(i915) >= 6) {
> + gen6_enable_rps(i915);
> + } else if (IS_IRONLAKE_M(i915)) {
> + ironlake_enable_drps(i915);
> + intel_init_emon(i915);
> }
>
> WARN_ON(rps->max_freq < rps->min_freq);
> @@ -8754,25 +8754,25 @@ static void intel_enable_rps(struct drm_i915_private *dev_priv)
> rps->enabled = true;
> }
>
> -void intel_enable_gt_powersave(struct drm_i915_private *dev_priv)
> +void intel_enable_gt_powersave(struct drm_i915_private *i915)
> {
> /* Powersaving is controlled by the host when inside a VM */
> - if (intel_vgpu_active(dev_priv))
> + if (intel_vgpu_active(i915))
> return;
>
> - mutex_lock(&dev_priv->gt_pm.rps.lock);
> + mutex_lock(&i915->gt_pm.rps.lock);
>
> - if (HAS_RC6(dev_priv))
> - intel_enable_rc6(dev_priv);
> - if (HAS_RPS(dev_priv))
> - intel_enable_rps(dev_priv);
> - if (HAS_LLC(dev_priv))
> - intel_enable_llc_pstate(dev_priv);
> + if (HAS_RC6(i915))
> + intel_enable_rc6(i915);
> + if (HAS_RPS(i915))
> + intel_enable_rps(i915);
> + if (HAS_LLC(i915))
> + intel_enable_llc_pstate(i915);
>
> - mutex_unlock(&dev_priv->gt_pm.rps.lock);
> + mutex_unlock(&i915->gt_pm.rps.lock);
> }
>
> -static void ibx_init_clock_gating(struct drm_i915_private *dev_priv)
> +static void ibx_init_clock_gating(struct drm_i915_private *i915)
> {
> /*
> * On Ibex Peak and Cougar Point, we need to disable clock
> @@ -8782,11 +8782,11 @@ static void ibx_init_clock_gating(struct drm_i915_private *dev_priv)
> I915_WRITE(SOUTH_DSPCLK_GATE_D, PCH_DPLSUNIT_CLOCK_GATE_DISABLE);
> }
>
> -static void g4x_disable_trickle_feed(struct drm_i915_private *dev_priv)
> +static void g4x_disable_trickle_feed(struct drm_i915_private *i915)
> {
> enum pipe pipe;
>
> - for_each_pipe(dev_priv, pipe) {
> + for_each_pipe(i915, pipe) {
> I915_WRITE(DSPCNTR(pipe),
> I915_READ(DSPCNTR(pipe)) |
> DISPPLANE_TRICKLE_FEED_DISABLE);
> @@ -8796,7 +8796,7 @@ static void g4x_disable_trickle_feed(struct drm_i915_private *dev_priv)
> }
> }
>
> -static void ilk_init_clock_gating(struct drm_i915_private *dev_priv)
> +static void ilk_init_clock_gating(struct drm_i915_private *i915)
> {
> u32 dspclk_gate = ILK_VRHUNIT_CLOCK_GATE_DISABLE;
>
> @@ -8836,7 +8836,7 @@ static void ilk_init_clock_gating(struct drm_i915_private *dev_priv)
> * The bit 22 of 0x42004
> * The bit 7,8,9 of 0x42020.
> */
> - if (IS_IRONLAKE_M(dev_priv)) {
> + if (IS_IRONLAKE_M(i915)) {
> /* WaFbcAsynchFlipDisableFbcQueue:ilk */
> I915_WRITE(ILK_DISPLAY_CHICKEN1,
> I915_READ(ILK_DISPLAY_CHICKEN1) |
> @@ -8862,12 +8862,12 @@ static void ilk_init_clock_gating(struct drm_i915_private *dev_priv)
> /* WaDisable_RenderCache_OperationalFlush:ilk */
> I915_WRITE(CACHE_MODE_0, _MASKED_BIT_DISABLE(RC_OP_FLUSH_ENABLE));
>
> - g4x_disable_trickle_feed(dev_priv);
> + g4x_disable_trickle_feed(i915);
>
> - ibx_init_clock_gating(dev_priv);
> + ibx_init_clock_gating(i915);
> }
>
> -static void cpt_init_clock_gating(struct drm_i915_private *dev_priv)
> +static void cpt_init_clock_gating(struct drm_i915_private *i915)
> {
> int pipe;
> u32 val;
> @@ -8885,11 +8885,11 @@ static void cpt_init_clock_gating(struct drm_i915_private *dev_priv)
> /* The below fixes the weird display corruption, a few pixels shifted
> * downward, on (only) LVDS of some HP laptops with IVY.
> */
> - for_each_pipe(dev_priv, pipe) {
> + for_each_pipe(i915, pipe) {
> val = I915_READ(TRANS_CHICKEN2(pipe));
> val |= TRANS_CHICKEN2_TIMING_OVERRIDE;
> val &= ~TRANS_CHICKEN2_FDI_POLARITY_REVERSED;
> - if (dev_priv->vbt.fdi_rx_polarity_inverted)
> + if (i915->vbt.fdi_rx_polarity_inverted)
> val |= TRANS_CHICKEN2_FDI_POLARITY_REVERSED;
> val &= ~TRANS_CHICKEN2_FRAME_START_DELAY_MASK;
> val &= ~TRANS_CHICKEN2_DISABLE_DEEP_COLOR_COUNTER;
> @@ -8897,13 +8897,13 @@ static void cpt_init_clock_gating(struct drm_i915_private *dev_priv)
> I915_WRITE(TRANS_CHICKEN2(pipe), val);
> }
> /* WADP0ClockGatingDisable */
> - for_each_pipe(dev_priv, pipe) {
> + for_each_pipe(i915, pipe) {
> I915_WRITE(TRANS_CHICKEN1(pipe),
> TRANS_CHICKEN1_DP0UNIT_GC_DISABLE);
> }
> }
>
> -static void gen6_check_mch_setup(struct drm_i915_private *dev_priv)
> +static void gen6_check_mch_setup(struct drm_i915_private *i915)
> {
> u32 tmp;
>
> @@ -8913,7 +8913,7 @@ static void gen6_check_mch_setup(struct drm_i915_private *dev_priv)
> tmp);
> }
>
> -static void gen6_init_clock_gating(struct drm_i915_private *dev_priv)
> +static void gen6_init_clock_gating(struct drm_i915_private *i915)
> {
> u32 dspclk_gate = ILK_VRHUNIT_CLOCK_GATE_DISABLE;
>
> @@ -9000,14 +9000,14 @@ static void gen6_init_clock_gating(struct drm_i915_private *dev_priv)
> ILK_DPARBUNIT_CLOCK_GATE_ENABLE |
> ILK_DPFDUNIT_CLOCK_GATE_ENABLE);
>
> - g4x_disable_trickle_feed(dev_priv);
> + g4x_disable_trickle_feed(i915);
>
> - cpt_init_clock_gating(dev_priv);
> + cpt_init_clock_gating(i915);
>
> - gen6_check_mch_setup(dev_priv);
> + gen6_check_mch_setup(i915);
> }
>
> -static void gen7_setup_fixed_func_scheduler(struct drm_i915_private *dev_priv)
> +static void gen7_setup_fixed_func_scheduler(struct drm_i915_private *i915)
> {
> u32 reg = I915_READ(GEN7_FF_THREAD_MODE);
>
> @@ -9025,13 +9025,13 @@ static void gen7_setup_fixed_func_scheduler(struct drm_i915_private *dev_priv)
> I915_WRITE(GEN7_FF_THREAD_MODE, reg);
> }
>
> -static void lpt_init_clock_gating(struct drm_i915_private *dev_priv)
> +static void lpt_init_clock_gating(struct drm_i915_private *i915)
> {
> /*
> * TODO: this bit should only be enabled when really needed, then
> * disabled when not needed anymore in order to save power.
> */
> - if (HAS_PCH_LPT_LP(dev_priv))
> + if (HAS_PCH_LPT_LP(i915))
> I915_WRITE(SOUTH_DSPCLK_GATE_D,
> I915_READ(SOUTH_DSPCLK_GATE_D) |
> PCH_LP_PARTITION_LEVEL_DISABLE);
> @@ -9042,9 +9042,9 @@ static void lpt_init_clock_gating(struct drm_i915_private *dev_priv)
> TRANS_CHICKEN1_DP0UNIT_GC_DISABLE);
> }
>
> -static void lpt_suspend_hw(struct drm_i915_private *dev_priv)
> +static void lpt_suspend_hw(struct drm_i915_private *i915)
> {
> - if (HAS_PCH_LPT_LP(dev_priv)) {
> + if (HAS_PCH_LPT_LP(i915)) {
> u32 val = I915_READ(SOUTH_DSPCLK_GATE_D);
>
> val &= ~PCH_LP_PARTITION_LEVEL_DISABLE;
> @@ -9052,7 +9052,7 @@ static void lpt_suspend_hw(struct drm_i915_private *dev_priv)
> }
> }
>
> -static void gen8_set_l3sqc_credits(struct drm_i915_private *dev_priv,
> +static void gen8_set_l3sqc_credits(struct drm_i915_private *i915,
> int general_prio_credits,
> int high_prio_credits)
> {
> @@ -9078,7 +9078,7 @@ static void gen8_set_l3sqc_credits(struct drm_i915_private *dev_priv,
> I915_WRITE(GEN7_MISCCPCTL, misccpctl);
> }
>
> -static void icl_init_clock_gating(struct drm_i915_private *dev_priv)
> +static void icl_init_clock_gating(struct drm_i915_private *i915)
> {
> /* This is not an Wa. Enable to reduce Sampler power */
> I915_WRITE(GEN10_DFR_RATIO_EN_AND_CHICKEN,
> @@ -9089,9 +9089,9 @@ static void icl_init_clock_gating(struct drm_i915_private *dev_priv)
> _MASKED_BIT_ENABLE(GEN11_ENABLE_32_PLANE_MODE));
> }
>
> -static void cnp_init_clock_gating(struct drm_i915_private *dev_priv)
> +static void cnp_init_clock_gating(struct drm_i915_private *i915)
> {
> - if (!HAS_PCH_CNP(dev_priv))
> + if (!HAS_PCH_CNP(i915))
> return;
>
> /* Display WA #1181 WaSouthDisplayDisablePWMCGEGating: cnp */
> @@ -9099,10 +9099,10 @@ static void cnp_init_clock_gating(struct drm_i915_private *dev_priv)
> CNP_PWM_CGE_GATING_DISABLE);
> }
>
> -static void cnl_init_clock_gating(struct drm_i915_private *dev_priv)
> +static void cnl_init_clock_gating(struct drm_i915_private *i915)
> {
> u32 val;
> - cnp_init_clock_gating(dev_priv);
> + cnp_init_clock_gating(i915);
>
> /* This is not an Wa. Enable for better image quality */
> I915_WRITE(_3D_CHICKEN3,
> @@ -9120,7 +9120,7 @@ static void cnl_init_clock_gating(struct drm_i915_private *dev_priv)
> /* ReadHitWriteOnlyDisable:cnl */
> val |= RCCUNIT_CLKGATE_DIS;
> /* WaSarbUnitClockGatingDisable:cnl (pre-prod) */
> - if (IS_CNL_REVID(dev_priv, CNL_REVID_A0, CNL_REVID_B0))
> + if (IS_CNL_REVID(i915, CNL_REVID_A0, CNL_REVID_B0))
> val |= SARBUNIT_CLKGATE_DIS;
> I915_WRITE(SLICE_UNIT_LEVEL_CLKGATE, val);
>
> @@ -9136,27 +9136,27 @@ static void cnl_init_clock_gating(struct drm_i915_private *dev_priv)
> I915_WRITE(UNSLICE_UNIT_LEVEL_CLKGATE, val);
> }
>
> -static void cfl_init_clock_gating(struct drm_i915_private *dev_priv)
> +static void cfl_init_clock_gating(struct drm_i915_private *i915)
> {
> - cnp_init_clock_gating(dev_priv);
> - gen9_init_clock_gating(dev_priv);
> + cnp_init_clock_gating(i915);
> + gen9_init_clock_gating(i915);
>
> /* WaFbcNukeOnHostModify:cfl */
> I915_WRITE(ILK_DPFC_CHICKEN, I915_READ(ILK_DPFC_CHICKEN) |
> ILK_DPFC_NUKE_ON_ANY_MODIFICATION);
> }
>
> -static void kbl_init_clock_gating(struct drm_i915_private *dev_priv)
> +static void kbl_init_clock_gating(struct drm_i915_private *i915)
> {
> - gen9_init_clock_gating(dev_priv);
> + gen9_init_clock_gating(i915);
>
> /* WaDisableSDEUnitClockGating:kbl */
> - if (IS_KBL_REVID(dev_priv, 0, KBL_REVID_B0))
> + if (IS_KBL_REVID(i915, 0, KBL_REVID_B0))
> I915_WRITE(GEN8_UCGCTL6, I915_READ(GEN8_UCGCTL6) |
> GEN8_SDEUNIT_CLOCK_GATE_DISABLE);
>
> /* WaDisableGamClockGating:kbl */
> - if (IS_KBL_REVID(dev_priv, 0, KBL_REVID_B0))
> + if (IS_KBL_REVID(i915, 0, KBL_REVID_B0))
> I915_WRITE(GEN6_UCGCTL1, I915_READ(GEN6_UCGCTL1) |
> GEN6_GAMUNIT_CLOCK_GATE_DISABLE);
>
> @@ -9165,9 +9165,9 @@ static void kbl_init_clock_gating(struct drm_i915_private *dev_priv)
> ILK_DPFC_NUKE_ON_ANY_MODIFICATION);
> }
>
> -static void skl_init_clock_gating(struct drm_i915_private *dev_priv)
> +static void skl_init_clock_gating(struct drm_i915_private *i915)
> {
> - gen9_init_clock_gating(dev_priv);
> + gen9_init_clock_gating(i915);
>
> /* WAC6entrylatency:skl */
> I915_WRITE(FBC_LLC_READ_CTRL, I915_READ(FBC_LLC_READ_CTRL) |
> @@ -9178,10 +9178,10 @@ static void skl_init_clock_gating(struct drm_i915_private *dev_priv)
> ILK_DPFC_NUKE_ON_ANY_MODIFICATION);
> }
>
> -static void bdw_init_clock_gating(struct drm_i915_private *dev_priv)
> +static void bdw_init_clock_gating(struct drm_i915_private *i915)
> {
> /* The GTT cache must be disabled if the system is using 2M pages. */
> - bool can_use_gtt_cache = !HAS_PAGE_SIZES(dev_priv,
> + bool can_use_gtt_cache = !HAS_PAGE_SIZES(i915,
> I915_GTT_PAGE_SIZE_2M);
> enum pipe pipe;
>
> @@ -9193,7 +9193,7 @@ static void bdw_init_clock_gating(struct drm_i915_private *dev_priv)
> I915_READ(CHICKEN_PAR1_1) | DPA_MASK_VBLANK_SRD);
>
> /* WaPsrDPRSUnmaskVBlankInSRD:bdw */
> - for_each_pipe(dev_priv, pipe) {
> + for_each_pipe(i915, pipe) {
> I915_WRITE(CHICKEN_PIPESL_1(pipe),
> I915_READ(CHICKEN_PIPESL_1(pipe)) |
> BDW_DPRS_MASK_VBLANK_SRD);
> @@ -9213,7 +9213,7 @@ static void bdw_init_clock_gating(struct drm_i915_private *dev_priv)
> GEN8_SDEUNIT_CLOCK_GATE_DISABLE);
>
> /* WaProgramL3SqcReg1Default:bdw */
> - gen8_set_l3sqc_credits(dev_priv, 30, 2);
> + gen8_set_l3sqc_credits(i915, 30, 2);
>
> /* WaGttCachingOffByDefault:bdw */
> I915_WRITE(HSW_GTT_CACHE_EN, can_use_gtt_cache ? GTT_CACHE_EN_ALL : 0);
> @@ -9222,7 +9222,7 @@ static void bdw_init_clock_gating(struct drm_i915_private *dev_priv)
> I915_WRITE(CHICKEN_PAR2_1, I915_READ(CHICKEN_PAR2_1)
> | KVM_CONFIG_CHANGE_NOTIFICATION_SELECT);
>
> - lpt_init_clock_gating(dev_priv);
> + lpt_init_clock_gating(i915);
>
> /* WaDisableDopClockGating:bdw
> *
> @@ -9233,7 +9233,7 @@ static void bdw_init_clock_gating(struct drm_i915_private *dev_priv)
> I915_READ(GEN6_UCGCTL1) | GEN6_EU_TCUNIT_CLOCK_GATE_DISABLE);
> }
>
> -static void hsw_init_clock_gating(struct drm_i915_private *dev_priv)
> +static void hsw_init_clock_gating(struct drm_i915_private *i915)
> {
> /* L3 caching of data atomics doesn't work -- disable it. */
> I915_WRITE(HSW_SCRATCH1, HSW_SCRATCH1_L3_DATA_ATOMICS_DISABLE);
> @@ -9278,10 +9278,10 @@ static void hsw_init_clock_gating(struct drm_i915_private *dev_priv)
> /* WaSwitchSolVfFArbitrationPriority:hsw */
> I915_WRITE(GAM_ECOCHK, I915_READ(GAM_ECOCHK) | HSW_ECOCHK_ARB_PRIO_SOL);
>
> - lpt_init_clock_gating(dev_priv);
> + lpt_init_clock_gating(i915);
> }
>
> -static void ivb_init_clock_gating(struct drm_i915_private *dev_priv)
> +static void ivb_init_clock_gating(struct drm_i915_private *i915)
> {
> u32 snpcr;
>
> @@ -9297,7 +9297,7 @@ static void ivb_init_clock_gating(struct drm_i915_private *dev_priv)
> CHICKEN3_DGMG_DONE_FIX_DISABLE);
>
> /* WaDisablePSDDualDispatchEnable:ivb */
> - if (IS_IVB_GT1(dev_priv))
> + if (IS_IVB_GT1(i915))
> I915_WRITE(GEN7_HALF_SLICE_CHICKEN1,
> _MASKED_BIT_ENABLE(GEN7_PSD_SINGLE_PORT_DISPATCH_ENABLE));
>
> @@ -9313,7 +9313,7 @@ static void ivb_init_clock_gating(struct drm_i915_private *dev_priv)
> GEN7_WA_FOR_GEN7_L3_CONTROL);
> I915_WRITE(GEN7_L3_CHICKEN_MODE_REGISTER,
> GEN7_WA_L3_CHICKEN_MODE);
> - if (IS_IVB_GT1(dev_priv))
> + if (IS_IVB_GT1(i915))
> I915_WRITE(GEN7_ROW_CHICKEN2,
> _MASKED_BIT_ENABLE(DOP_CLOCK_GATING_DISABLE));
> else {
> @@ -9340,9 +9340,9 @@ static void ivb_init_clock_gating(struct drm_i915_private *dev_priv)
> I915_READ(GEN7_SQ_CHICKEN_MBCUNIT_CONFIG) |
> GEN7_SQ_CHICKEN_MBCUNIT_SQINTMOB);
>
> - g4x_disable_trickle_feed(dev_priv);
> + g4x_disable_trickle_feed(i915);
>
> - gen7_setup_fixed_func_scheduler(dev_priv);
> + gen7_setup_fixed_func_scheduler(i915);
>
> if (0) { /* causes HiZ corruption on ivb:gt1 */
> /* enable HiZ Raw Stall Optimization */
> @@ -9370,13 +9370,13 @@ static void ivb_init_clock_gating(struct drm_i915_private *dev_priv)
> snpcr |= GEN6_MBC_SNPCR_MED;
> I915_WRITE(GEN6_MBCUNIT_SNPCR, snpcr);
>
> - if (!HAS_PCH_NOP(dev_priv))
> - cpt_init_clock_gating(dev_priv);
> + if (!HAS_PCH_NOP(i915))
> + cpt_init_clock_gating(i915);
>
> - gen6_check_mch_setup(dev_priv);
> + gen6_check_mch_setup(i915);
> }
>
> -static void vlv_init_clock_gating(struct drm_i915_private *dev_priv)
> +static void vlv_init_clock_gating(struct drm_i915_private *i915)
> {
> /* WaDisableEarlyCull:vlv */
> I915_WRITE(_3D_CHICKEN3,
> @@ -9409,7 +9409,7 @@ static void vlv_init_clock_gating(struct drm_i915_private *dev_priv)
> I915_READ(GEN7_SQ_CHICKEN_MBCUNIT_CONFIG) |
> GEN7_SQ_CHICKEN_MBCUNIT_SQINTMOB);
>
> - gen7_setup_fixed_func_scheduler(dev_priv);
> + gen7_setup_fixed_func_scheduler(i915);
>
> /*
> * According to the spec, bit 13 (RCZUNIT) must be set on IVB.
> @@ -9456,7 +9456,7 @@ static void vlv_init_clock_gating(struct drm_i915_private *dev_priv)
> I915_WRITE(VLV_GUNIT_CLOCK_GATE, GCFG_DIS);
> }
>
> -static void chv_init_clock_gating(struct drm_i915_private *dev_priv)
> +static void chv_init_clock_gating(struct drm_i915_private *i915)
> {
> /* WaVSRefCountFullforceMissDisable:chv */
> /* WaDSRefCountFullforceMissDisable:chv */
> @@ -9481,7 +9481,7 @@ static void chv_init_clock_gating(struct drm_i915_private *dev_priv)
> * See gfxspecs/Related Documents/Performance Guide/
> * LSQC Setting Recommendations.
> */
> - gen8_set_l3sqc_credits(dev_priv, 38, 2);
> + gen8_set_l3sqc_credits(i915, 38, 2);
>
> /*
> * GTT cache may not work with big pages, so if those
> @@ -9490,7 +9490,7 @@ static void chv_init_clock_gating(struct drm_i915_private *dev_priv)
> I915_WRITE(HSW_GTT_CACHE_EN, GTT_CACHE_EN_ALL);
> }
>
> -static void g4x_init_clock_gating(struct drm_i915_private *dev_priv)
> +static void g4x_init_clock_gating(struct drm_i915_private *i915)
> {
> u32 dspclk_gate;
>
> @@ -9502,7 +9502,7 @@ static void g4x_init_clock_gating(struct drm_i915_private *dev_priv)
> dspclk_gate = VRHUNIT_CLOCK_GATE_DISABLE |
> OVRUNIT_CLOCK_GATE_DISABLE |
> OVCUNIT_CLOCK_GATE_DISABLE;
> - if (IS_GM45(dev_priv))
> + if (IS_GM45(i915))
> dspclk_gate |= DSSUNIT_CLOCK_GATE_DISABLE;
> I915_WRITE(DSPCLK_GATE_D, dspclk_gate);
>
> @@ -9513,12 +9513,12 @@ static void g4x_init_clock_gating(struct drm_i915_private *dev_priv)
> /* WaDisable_RenderCache_OperationalFlush:g4x */
> I915_WRITE(CACHE_MODE_0, _MASKED_BIT_DISABLE(RC_OP_FLUSH_ENABLE));
>
> - g4x_disable_trickle_feed(dev_priv);
> + g4x_disable_trickle_feed(i915);
> }
>
> -static void i965gm_init_clock_gating(struct drm_i915_private *dev_priv)
> +static void i965gm_init_clock_gating(struct drm_i915_private *i915)
> {
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct intel_uncore *uncore = &i915->uncore;
>
> intel_uncore_write(uncore, RENCLK_GATE_D1, I965_RCC_CLOCK_GATE_DISABLE);
> intel_uncore_write(uncore, RENCLK_GATE_D2, 0);
> @@ -9535,7 +9535,7 @@ static void i965gm_init_clock_gating(struct drm_i915_private *dev_priv)
> _MASKED_BIT_DISABLE(RC_OP_FLUSH_ENABLE));
> }
>
> -static void i965g_init_clock_gating(struct drm_i915_private *dev_priv)
> +static void i965g_init_clock_gating(struct drm_i915_private *i915)
> {
> I915_WRITE(RENCLK_GATE_D1, I965_RCZ_CLOCK_GATE_DISABLE |
> I965_RCC_CLOCK_GATE_DISABLE |
> @@ -9550,7 +9550,7 @@ static void i965g_init_clock_gating(struct drm_i915_private *dev_priv)
> I915_WRITE(CACHE_MODE_0, _MASKED_BIT_DISABLE(RC_OP_FLUSH_ENABLE));
> }
>
> -static void gen3_init_clock_gating(struct drm_i915_private *dev_priv)
> +static void gen3_init_clock_gating(struct drm_i915_private *i915)
> {
> u32 dstate = I915_READ(D_STATE);
>
> @@ -9558,7 +9558,7 @@ static void gen3_init_clock_gating(struct drm_i915_private *dev_priv)
> DSTATE_DOT_CLOCK_GATING;
> I915_WRITE(D_STATE, dstate);
>
> - if (IS_PINEVIEW(dev_priv))
> + if (IS_PINEVIEW(i915))
> I915_WRITE(ECOSKPD, _MASKED_BIT_ENABLE(ECO_GATING_CX_ONLY));
>
> /* IIR "flip pending" means done if this bit is set */
> @@ -9574,7 +9574,7 @@ static void gen3_init_clock_gating(struct drm_i915_private *dev_priv)
> _MASKED_BIT_ENABLE(MI_ARB_DISPLAY_TRICKLE_FEED_DISABLE));
> }
>
> -static void i85x_init_clock_gating(struct drm_i915_private *dev_priv)
> +static void i85x_init_clock_gating(struct drm_i915_private *i915)
> {
> I915_WRITE(RENCLK_GATE_D1, SV_CLOCK_GATE_DISABLE);
>
> @@ -9586,168 +9586,168 @@ static void i85x_init_clock_gating(struct drm_i915_private *dev_priv)
> _MASKED_BIT_ENABLE(MEM_DISPLAY_TRICKLE_FEED_DISABLE));
> }
>
> -static void i830_init_clock_gating(struct drm_i915_private *dev_priv)
> +static void i830_init_clock_gating(struct drm_i915_private *i915)
> {
> I915_WRITE(MEM_MODE,
> _MASKED_BIT_ENABLE(MEM_DISPLAY_A_TRICKLE_FEED_DISABLE) |
> _MASKED_BIT_ENABLE(MEM_DISPLAY_B_TRICKLE_FEED_DISABLE));
> }
>
> -void intel_init_clock_gating(struct drm_i915_private *dev_priv)
> +void intel_init_clock_gating(struct drm_i915_private *i915)
> {
> - dev_priv->display.init_clock_gating(dev_priv);
> + i915->display.init_clock_gating(i915);
> }
>
> -void intel_suspend_hw(struct drm_i915_private *dev_priv)
> +void intel_suspend_hw(struct drm_i915_private *i915)
> {
> - if (HAS_PCH_LPT(dev_priv))
> - lpt_suspend_hw(dev_priv);
> + if (HAS_PCH_LPT(i915))
> + lpt_suspend_hw(i915);
> }
>
> -static void nop_init_clock_gating(struct drm_i915_private *dev_priv)
> +static void nop_init_clock_gating(struct drm_i915_private *i915)
> {
> DRM_DEBUG_KMS("No clock gating settings or workarounds applied.\n");
> }
>
> /**
> * intel_init_clock_gating_hooks - setup the clock gating hooks
> - * @dev_priv: device private
> + * @i915: device private
> *
> * Setup the hooks that configure which clocks of a given platform can be
> * gated and also apply various GT and display specific workarounds for these
> * platforms. Note that some GT specific workarounds are applied separately
> * when GPU contexts or batchbuffers start their execution.
> */
> -void intel_init_clock_gating_hooks(struct drm_i915_private *dev_priv)
> -{
> - if (IS_GEN(dev_priv, 11))
> - dev_priv->display.init_clock_gating = icl_init_clock_gating;
> - else if (IS_CANNONLAKE(dev_priv))
> - dev_priv->display.init_clock_gating = cnl_init_clock_gating;
> - else if (IS_COFFEELAKE(dev_priv))
> - dev_priv->display.init_clock_gating = cfl_init_clock_gating;
> - else if (IS_SKYLAKE(dev_priv))
> - dev_priv->display.init_clock_gating = skl_init_clock_gating;
> - else if (IS_KABYLAKE(dev_priv))
> - dev_priv->display.init_clock_gating = kbl_init_clock_gating;
> - else if (IS_BROXTON(dev_priv))
> - dev_priv->display.init_clock_gating = bxt_init_clock_gating;
> - else if (IS_GEMINILAKE(dev_priv))
> - dev_priv->display.init_clock_gating = glk_init_clock_gating;
> - else if (IS_BROADWELL(dev_priv))
> - dev_priv->display.init_clock_gating = bdw_init_clock_gating;
> - else if (IS_CHERRYVIEW(dev_priv))
> - dev_priv->display.init_clock_gating = chv_init_clock_gating;
> - else if (IS_HASWELL(dev_priv))
> - dev_priv->display.init_clock_gating = hsw_init_clock_gating;
> - else if (IS_IVYBRIDGE(dev_priv))
> - dev_priv->display.init_clock_gating = ivb_init_clock_gating;
> - else if (IS_VALLEYVIEW(dev_priv))
> - dev_priv->display.init_clock_gating = vlv_init_clock_gating;
> - else if (IS_GEN(dev_priv, 6))
> - dev_priv->display.init_clock_gating = gen6_init_clock_gating;
> - else if (IS_GEN(dev_priv, 5))
> - dev_priv->display.init_clock_gating = ilk_init_clock_gating;
> - else if (IS_G4X(dev_priv))
> - dev_priv->display.init_clock_gating = g4x_init_clock_gating;
> - else if (IS_I965GM(dev_priv))
> - dev_priv->display.init_clock_gating = i965gm_init_clock_gating;
> - else if (IS_I965G(dev_priv))
> - dev_priv->display.init_clock_gating = i965g_init_clock_gating;
> - else if (IS_GEN(dev_priv, 3))
> - dev_priv->display.init_clock_gating = gen3_init_clock_gating;
> - else if (IS_I85X(dev_priv) || IS_I865G(dev_priv))
> - dev_priv->display.init_clock_gating = i85x_init_clock_gating;
> - else if (IS_GEN(dev_priv, 2))
> - dev_priv->display.init_clock_gating = i830_init_clock_gating;
> +void intel_init_clock_gating_hooks(struct drm_i915_private *i915)
> +{
> + if (IS_GEN(i915, 11))
> + i915->display.init_clock_gating = icl_init_clock_gating;
> + else if (IS_CANNONLAKE(i915))
> + i915->display.init_clock_gating = cnl_init_clock_gating;
> + else if (IS_COFFEELAKE(i915))
> + i915->display.init_clock_gating = cfl_init_clock_gating;
> + else if (IS_SKYLAKE(i915))
> + i915->display.init_clock_gating = skl_init_clock_gating;
> + else if (IS_KABYLAKE(i915))
> + i915->display.init_clock_gating = kbl_init_clock_gating;
> + else if (IS_BROXTON(i915))
> + i915->display.init_clock_gating = bxt_init_clock_gating;
> + else if (IS_GEMINILAKE(i915))
> + i915->display.init_clock_gating = glk_init_clock_gating;
> + else if (IS_BROADWELL(i915))
> + i915->display.init_clock_gating = bdw_init_clock_gating;
> + else if (IS_CHERRYVIEW(i915))
> + i915->display.init_clock_gating = chv_init_clock_gating;
> + else if (IS_HASWELL(i915))
> + i915->display.init_clock_gating = hsw_init_clock_gating;
> + else if (IS_IVYBRIDGE(i915))
> + i915->display.init_clock_gating = ivb_init_clock_gating;
> + else if (IS_VALLEYVIEW(i915))
> + i915->display.init_clock_gating = vlv_init_clock_gating;
> + else if (IS_GEN(i915, 6))
> + i915->display.init_clock_gating = gen6_init_clock_gating;
> + else if (IS_GEN(i915, 5))
> + i915->display.init_clock_gating = ilk_init_clock_gating;
> + else if (IS_G4X(i915))
> + i915->display.init_clock_gating = g4x_init_clock_gating;
> + else if (IS_I965GM(i915))
> + i915->display.init_clock_gating = i965gm_init_clock_gating;
> + else if (IS_I965G(i915))
> + i915->display.init_clock_gating = i965g_init_clock_gating;
> + else if (IS_GEN(i915, 3))
> + i915->display.init_clock_gating = gen3_init_clock_gating;
> + else if (IS_I85X(i915) || IS_I865G(i915))
> + i915->display.init_clock_gating = i85x_init_clock_gating;
> + else if (IS_GEN(i915, 2))
> + i915->display.init_clock_gating = i830_init_clock_gating;
> else {
> - MISSING_CASE(INTEL_DEVID(dev_priv));
> - dev_priv->display.init_clock_gating = nop_init_clock_gating;
> + MISSING_CASE(INTEL_DEVID(i915));
> + i915->display.init_clock_gating = nop_init_clock_gating;
> }
> }
>
> /* Set up chip specific power management-related functions */
> -void intel_init_pm(struct drm_i915_private *dev_priv)
> +void intel_init_pm(struct drm_i915_private *i915)
> {
> /* For cxsr */
> - if (IS_PINEVIEW(dev_priv))
> - i915_pineview_get_mem_freq(dev_priv);
> - else if (IS_GEN(dev_priv, 5))
> - i915_ironlake_get_mem_freq(dev_priv);
> + if (IS_PINEVIEW(i915))
> + i915_pineview_get_mem_freq(i915);
> + else if (IS_GEN(i915, 5))
> + i915_ironlake_get_mem_freq(i915);
>
> /* For FIFO watermark updates */
> - if (INTEL_GEN(dev_priv) >= 9) {
> - skl_setup_wm_latency(dev_priv);
> - dev_priv->display.initial_watermarks = skl_initial_wm;
> - dev_priv->display.atomic_update_watermarks = skl_atomic_update_crtc_wm;
> - dev_priv->display.compute_global_watermarks = skl_compute_wm;
> - } else if (HAS_PCH_SPLIT(dev_priv)) {
> - ilk_setup_wm_latency(dev_priv);
> -
> - if ((IS_GEN(dev_priv, 5) && dev_priv->wm.pri_latency[1] &&
> - dev_priv->wm.spr_latency[1] && dev_priv->wm.cur_latency[1]) ||
> - (!IS_GEN(dev_priv, 5) && dev_priv->wm.pri_latency[0] &&
> - dev_priv->wm.spr_latency[0] && dev_priv->wm.cur_latency[0])) {
> - dev_priv->display.compute_pipe_wm = ilk_compute_pipe_wm;
> - dev_priv->display.compute_intermediate_wm =
> + if (INTEL_GEN(i915) >= 9) {
> + skl_setup_wm_latency(i915);
> + i915->display.initial_watermarks = skl_initial_wm;
> + i915->display.atomic_update_watermarks = skl_atomic_update_crtc_wm;
> + i915->display.compute_global_watermarks = skl_compute_wm;
> + } else if (HAS_PCH_SPLIT(i915)) {
> + ilk_setup_wm_latency(i915);
> +
> + if ((IS_GEN(i915, 5) && i915->wm.pri_latency[1] &&
> + i915->wm.spr_latency[1] && i915->wm.cur_latency[1]) ||
> + (!IS_GEN(i915, 5) && i915->wm.pri_latency[0] &&
> + i915->wm.spr_latency[0] && i915->wm.cur_latency[0])) {
> + i915->display.compute_pipe_wm = ilk_compute_pipe_wm;
> + i915->display.compute_intermediate_wm =
> ilk_compute_intermediate_wm;
> - dev_priv->display.initial_watermarks =
> + i915->display.initial_watermarks =
> ilk_initial_watermarks;
> - dev_priv->display.optimize_watermarks =
> + i915->display.optimize_watermarks =
> ilk_optimize_watermarks;
> } else {
> DRM_DEBUG_KMS("Failed to read display plane latency. "
> "Disable CxSR\n");
> }
> - } else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> - vlv_setup_wm_latency(dev_priv);
> - dev_priv->display.compute_pipe_wm = vlv_compute_pipe_wm;
> - dev_priv->display.compute_intermediate_wm = vlv_compute_intermediate_wm;
> - dev_priv->display.initial_watermarks = vlv_initial_watermarks;
> - dev_priv->display.optimize_watermarks = vlv_optimize_watermarks;
> - dev_priv->display.atomic_update_watermarks = vlv_atomic_update_fifo;
> - } else if (IS_G4X(dev_priv)) {
> - g4x_setup_wm_latency(dev_priv);
> - dev_priv->display.compute_pipe_wm = g4x_compute_pipe_wm;
> - dev_priv->display.compute_intermediate_wm = g4x_compute_intermediate_wm;
> - dev_priv->display.initial_watermarks = g4x_initial_watermarks;
> - dev_priv->display.optimize_watermarks = g4x_optimize_watermarks;
> - } else if (IS_PINEVIEW(dev_priv)) {
> - if (!intel_get_cxsr_latency(!IS_MOBILE(dev_priv),
> - dev_priv->is_ddr3,
> - dev_priv->fsb_freq,
> - dev_priv->mem_freq)) {
> + } else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> + vlv_setup_wm_latency(i915);
> + i915->display.compute_pipe_wm = vlv_compute_pipe_wm;
> + i915->display.compute_intermediate_wm = vlv_compute_intermediate_wm;
> + i915->display.initial_watermarks = vlv_initial_watermarks;
> + i915->display.optimize_watermarks = vlv_optimize_watermarks;
> + i915->display.atomic_update_watermarks = vlv_atomic_update_fifo;
> + } else if (IS_G4X(i915)) {
> + g4x_setup_wm_latency(i915);
> + i915->display.compute_pipe_wm = g4x_compute_pipe_wm;
> + i915->display.compute_intermediate_wm = g4x_compute_intermediate_wm;
> + i915->display.initial_watermarks = g4x_initial_watermarks;
> + i915->display.optimize_watermarks = g4x_optimize_watermarks;
> + } else if (IS_PINEVIEW(i915)) {
> + if (!intel_get_cxsr_latency(!IS_MOBILE(i915),
> + i915->is_ddr3,
> + i915->fsb_freq,
> + i915->mem_freq)) {
> DRM_INFO("failed to find known CxSR latency "
> "(found ddr%s fsb freq %d, mem freq %d), "
> "disabling CxSR\n",
> - (dev_priv->is_ddr3 == 1) ? "3" : "2",
> - dev_priv->fsb_freq, dev_priv->mem_freq);
> + (i915->is_ddr3 == 1) ? "3" : "2",
> + i915->fsb_freq, i915->mem_freq);
> /* Disable CxSR and never update its watermark again */
> - intel_set_memory_cxsr(dev_priv, false);
> - dev_priv->display.update_wm = NULL;
> + intel_set_memory_cxsr(i915, false);
> + i915->display.update_wm = NULL;
> } else
> - dev_priv->display.update_wm = pineview_update_wm;
> - } else if (IS_GEN(dev_priv, 4)) {
> - dev_priv->display.update_wm = i965_update_wm;
> - } else if (IS_GEN(dev_priv, 3)) {
> - dev_priv->display.update_wm = i9xx_update_wm;
> - dev_priv->display.get_fifo_size = i9xx_get_fifo_size;
> - } else if (IS_GEN(dev_priv, 2)) {
> - if (INTEL_INFO(dev_priv)->num_pipes == 1) {
> - dev_priv->display.update_wm = i845_update_wm;
> - dev_priv->display.get_fifo_size = i845_get_fifo_size;
> + i915->display.update_wm = pineview_update_wm;
> + } else if (IS_GEN(i915, 4)) {
> + i915->display.update_wm = i965_update_wm;
> + } else if (IS_GEN(i915, 3)) {
> + i915->display.update_wm = i9xx_update_wm;
> + i915->display.get_fifo_size = i9xx_get_fifo_size;
> + } else if (IS_GEN(i915, 2)) {
> + if (INTEL_INFO(i915)->num_pipes == 1) {
> + i915->display.update_wm = i845_update_wm;
> + i915->display.get_fifo_size = i845_get_fifo_size;
> } else {
> - dev_priv->display.update_wm = i9xx_update_wm;
> - dev_priv->display.get_fifo_size = i830_get_fifo_size;
> + i915->display.update_wm = i9xx_update_wm;
> + i915->display.get_fifo_size = i830_get_fifo_size;
> }
> } else {
> DRM_ERROR("unexpected fall-through in intel_init_pm\n");
> }
> }
>
> -static int byt_gpu_freq(struct drm_i915_private *dev_priv, int val)
> +static int byt_gpu_freq(struct drm_i915_private *i915, int val)
> {
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct intel_rps *rps = &i915->gt_pm.rps;
>
> /*
> * N = val - 0xb7
> @@ -9756,16 +9756,16 @@ static int byt_gpu_freq(struct drm_i915_private *dev_priv, int val)
> return DIV_ROUND_CLOSEST(rps->gpll_ref_freq * (val - 0xb7), 1000);
> }
>
> -static int byt_freq_opcode(struct drm_i915_private *dev_priv, int val)
> +static int byt_freq_opcode(struct drm_i915_private *i915, int val)
> {
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct intel_rps *rps = &i915->gt_pm.rps;
>
> return DIV_ROUND_CLOSEST(1000 * val, rps->gpll_ref_freq) + 0xb7;
> }
>
> -static int chv_gpu_freq(struct drm_i915_private *dev_priv, int val)
> +static int chv_gpu_freq(struct drm_i915_private *i915, int val)
> {
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct intel_rps *rps = &i915->gt_pm.rps;
>
> /*
> * N = val / 2
> @@ -9774,52 +9774,52 @@ static int chv_gpu_freq(struct drm_i915_private *dev_priv, int val)
> return DIV_ROUND_CLOSEST(rps->gpll_ref_freq * val, 2 * 2 * 1000);
> }
>
> -static int chv_freq_opcode(struct drm_i915_private *dev_priv, int val)
> +static int chv_freq_opcode(struct drm_i915_private *i915, int val)
> {
> - struct intel_rps *rps = &dev_priv->gt_pm.rps;
> + struct intel_rps *rps = &i915->gt_pm.rps;
>
> /* CHV needs even values */
> return DIV_ROUND_CLOSEST(2 * 1000 * val, rps->gpll_ref_freq) * 2;
> }
>
> -int intel_gpu_freq(struct drm_i915_private *dev_priv, int val)
> +int intel_gpu_freq(struct drm_i915_private *i915, int val)
> {
> - if (INTEL_GEN(dev_priv) >= 9)
> + if (INTEL_GEN(i915) >= 9)
> return DIV_ROUND_CLOSEST(val * GT_FREQUENCY_MULTIPLIER,
> GEN9_FREQ_SCALER);
> - else if (IS_CHERRYVIEW(dev_priv))
> - return chv_gpu_freq(dev_priv, val);
> - else if (IS_VALLEYVIEW(dev_priv))
> - return byt_gpu_freq(dev_priv, val);
> + else if (IS_CHERRYVIEW(i915))
> + return chv_gpu_freq(i915, val);
> + else if (IS_VALLEYVIEW(i915))
> + return byt_gpu_freq(i915, val);
> else
> return val * GT_FREQUENCY_MULTIPLIER;
> }
>
> -int intel_freq_opcode(struct drm_i915_private *dev_priv, int val)
> +int intel_freq_opcode(struct drm_i915_private *i915, int val)
> {
> - if (INTEL_GEN(dev_priv) >= 9)
> + if (INTEL_GEN(i915) >= 9)
> return DIV_ROUND_CLOSEST(val * GEN9_FREQ_SCALER,
> GT_FREQUENCY_MULTIPLIER);
> - else if (IS_CHERRYVIEW(dev_priv))
> - return chv_freq_opcode(dev_priv, val);
> - else if (IS_VALLEYVIEW(dev_priv))
> - return byt_freq_opcode(dev_priv, val);
> + else if (IS_CHERRYVIEW(i915))
> + return chv_freq_opcode(i915, val);
> + else if (IS_VALLEYVIEW(i915))
> + return byt_freq_opcode(i915, val);
> else
> return DIV_ROUND_CLOSEST(val, GT_FREQUENCY_MULTIPLIER);
> }
>
> -void intel_pm_setup(struct drm_i915_private *dev_priv)
> +void intel_pm_setup(struct drm_i915_private *i915)
> {
> - mutex_init(&dev_priv->gt_pm.rps.lock);
> - mutex_init(&dev_priv->gt_pm.rps.power.mutex);
> + mutex_init(&i915->gt_pm.rps.lock);
> + mutex_init(&i915->gt_pm.rps.power.mutex);
>
> - atomic_set(&dev_priv->gt_pm.rps.num_waiters, 0);
> + atomic_set(&i915->gt_pm.rps.num_waiters, 0);
>
> - dev_priv->runtime_pm.suspended = false;
> - atomic_set(&dev_priv->runtime_pm.wakeref_count, 0);
> + i915->runtime_pm.suspended = false;
> + atomic_set(&i915->runtime_pm.wakeref_count, 0);
> }
>
> -static u64 vlv_residency_raw(struct drm_i915_private *dev_priv,
> +static u64 vlv_residency_raw(struct drm_i915_private *i915,
> const i915_reg_t reg)
> {
> u32 lower, upper, tmp;
> @@ -9829,7 +9829,7 @@ static u64 vlv_residency_raw(struct drm_i915_private *dev_priv,
> * The register accessed do not need forcewake. We borrow
> * uncore lock to prevent concurrent access to range reg.
> */
> - lockdep_assert_held(&dev_priv->uncore.lock);
> + lockdep_assert_held(&i915->uncore.lock);
>
> /*
> * vlv and chv residency counters are 40 bits in width.
> @@ -9865,17 +9865,17 @@ static u64 vlv_residency_raw(struct drm_i915_private *dev_priv,
> return lower | (u64)upper << 8;
> }
>
> -u64 intel_rc6_residency_ns(struct drm_i915_private *dev_priv,
> +u64 intel_rc6_residency_ns(struct drm_i915_private *i915,
> const i915_reg_t reg)
> {
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct intel_uncore *uncore = &i915->uncore;
> u64 time_hw, prev_hw, overflow_hw;
> unsigned int fw_domains;
> unsigned long flags;
> unsigned int i;
> u32 mul, div;
>
> - if (!HAS_RC6(dev_priv))
> + if (!HAS_RC6(i915))
> return 0;
>
> /*
> @@ -9887,7 +9887,7 @@ u64 intel_rc6_residency_ns(struct drm_i915_private *dev_priv,
> */
> i = (i915_mmio_reg_offset(reg) -
> i915_mmio_reg_offset(GEN6_GT_GFX_RC6_LOCKED)) / sizeof(u32);
> - if (WARN_ON_ONCE(i >= ARRAY_SIZE(dev_priv->gt_pm.rc6.cur_residency)))
> + if (WARN_ON_ONCE(i >= ARRAY_SIZE(i915->gt_pm.rc6.cur_residency)))
> return 0;
>
> fw_domains = intel_uncore_forcewake_for_reg(uncore, reg, FW_REG_READ);
> @@ -9896,14 +9896,14 @@ u64 intel_rc6_residency_ns(struct drm_i915_private *dev_priv,
> intel_uncore_forcewake_get__locked(uncore, fw_domains);
>
> /* On VLV and CHV, residency time is in CZ units rather than 1.28us */
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> mul = 1000000;
> - div = dev_priv->czclk_freq;
> + div = i915->czclk_freq;
> overflow_hw = BIT_ULL(40);
> - time_hw = vlv_residency_raw(dev_priv, reg);
> + time_hw = vlv_residency_raw(i915, reg);
> } else {
> /* 833.33ns units on Gen9LP, 1.28us elsewhere. */
> - if (IS_GEN9_LP(dev_priv)) {
> + if (IS_GEN9_LP(i915)) {
> mul = 10000;
> div = 12;
> } else {
> @@ -9921,8 +9921,8 @@ u64 intel_rc6_residency_ns(struct drm_i915_private *dev_priv,
> * But relying on a sufficient frequency of queries otherwise counters
> * can still wrap.
> */
> - prev_hw = dev_priv->gt_pm.rc6.prev_hw_residency[i];
> - dev_priv->gt_pm.rc6.prev_hw_residency[i] = time_hw;
> + prev_hw = i915->gt_pm.rc6.prev_hw_residency[i];
> + i915->gt_pm.rc6.prev_hw_residency[i] = time_hw;
>
> /* RC6 delta from last sample. */
> if (time_hw >= prev_hw)
> @@ -9931,8 +9931,8 @@ u64 intel_rc6_residency_ns(struct drm_i915_private *dev_priv,
> time_hw += overflow_hw - prev_hw;
>
> /* Add delta to RC6 extended raw driver copy. */
> - time_hw += dev_priv->gt_pm.rc6.cur_residency[i];
> - dev_priv->gt_pm.rc6.cur_residency[i] = time_hw;
> + time_hw += i915->gt_pm.rc6.cur_residency[i];
> + i915->gt_pm.rc6.cur_residency[i] = time_hw;
>
> intel_uncore_forcewake_put__locked(uncore, fw_domains);
> spin_unlock_irqrestore(&uncore->lock, flags);
> @@ -9940,19 +9940,19 @@ u64 intel_rc6_residency_ns(struct drm_i915_private *dev_priv,
> return mul_u64_u32_div(time_hw, mul, div);
> }
>
> -u64 intel_rc6_residency_us(struct drm_i915_private *dev_priv,
> +u64 intel_rc6_residency_us(struct drm_i915_private *i915,
> i915_reg_t reg)
> {
> - return DIV_ROUND_UP_ULL(intel_rc6_residency_ns(dev_priv, reg), 1000);
> + return DIV_ROUND_UP_ULL(intel_rc6_residency_ns(i915, reg), 1000);
> }
>
> -u32 intel_get_cagf(struct drm_i915_private *dev_priv, u32 rpstat)
> +u32 intel_get_cagf(struct drm_i915_private *i915, u32 rpstat)
> {
> u32 cagf;
>
> - if (INTEL_GEN(dev_priv) >= 9)
> + if (INTEL_GEN(i915) >= 9)
> cagf = (rpstat & GEN9_CAGF_MASK) >> GEN9_CAGF_SHIFT;
> - else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> + else if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> cagf = (rpstat & HSW_CAGF_MASK) >> HSW_CAGF_SHIFT;
> else
> cagf = (rpstat & GEN6_CAGF_MASK) >> GEN6_CAGF_SHIFT;
> diff --git a/drivers/gpu/drm/i915/intel_pm.h b/drivers/gpu/drm/i915/intel_pm.h
> index 1b489fa399e1..e2c4edf263f7 100644
> --- a/drivers/gpu/drm/i915/intel_pm.h
> +++ b/drivers/gpu/drm/i915/intel_pm.h
> @@ -22,39 +22,39 @@ struct skl_ddb_entry;
> struct skl_pipe_wm;
> struct skl_wm_level;
>
> -void intel_init_clock_gating(struct drm_i915_private *dev_priv);
> -void intel_suspend_hw(struct drm_i915_private *dev_priv);
> -int ilk_wm_max_level(const struct drm_i915_private *dev_priv);
> +void intel_init_clock_gating(struct drm_i915_private *i915);
> +void intel_suspend_hw(struct drm_i915_private *i915);
> +int ilk_wm_max_level(const struct drm_i915_private *i915);
> void intel_update_watermarks(struct intel_crtc *crtc);
> -void intel_init_pm(struct drm_i915_private *dev_priv);
> -void intel_init_clock_gating_hooks(struct drm_i915_private *dev_priv);
> -void intel_pm_setup(struct drm_i915_private *dev_priv);
> -void intel_gpu_ips_init(struct drm_i915_private *dev_priv);
> +void intel_init_pm(struct drm_i915_private *i915);
> +void intel_init_clock_gating_hooks(struct drm_i915_private *i915);
> +void intel_pm_setup(struct drm_i915_private *i915);
> +void intel_gpu_ips_init(struct drm_i915_private *i915);
> void intel_gpu_ips_teardown(void);
> -void intel_init_gt_powersave(struct drm_i915_private *dev_priv);
> -void intel_cleanup_gt_powersave(struct drm_i915_private *dev_priv);
> -void intel_sanitize_gt_powersave(struct drm_i915_private *dev_priv);
> -void intel_enable_gt_powersave(struct drm_i915_private *dev_priv);
> -void intel_disable_gt_powersave(struct drm_i915_private *dev_priv);
> -void gen6_rps_busy(struct drm_i915_private *dev_priv);
> -void gen6_rps_idle(struct drm_i915_private *dev_priv);
> +void intel_init_gt_powersave(struct drm_i915_private *i915);
> +void intel_cleanup_gt_powersave(struct drm_i915_private *i915);
> +void intel_sanitize_gt_powersave(struct drm_i915_private *i915);
> +void intel_enable_gt_powersave(struct drm_i915_private *i915);
> +void intel_disable_gt_powersave(struct drm_i915_private *i915);
> +void gen6_rps_busy(struct drm_i915_private *i915);
> +void gen6_rps_idle(struct drm_i915_private *i915);
> void gen6_rps_boost(struct i915_request *rq);
> -void g4x_wm_get_hw_state(struct drm_i915_private *dev_priv);
> -void vlv_wm_get_hw_state(struct drm_i915_private *dev_priv);
> -void ilk_wm_get_hw_state(struct drm_i915_private *dev_priv);
> -void skl_wm_get_hw_state(struct drm_i915_private *dev_priv);
> +void g4x_wm_get_hw_state(struct drm_i915_private *i915);
> +void vlv_wm_get_hw_state(struct drm_i915_private *i915);
> +void ilk_wm_get_hw_state(struct drm_i915_private *i915);
> +void skl_wm_get_hw_state(struct drm_i915_private *i915);
> void skl_pipe_ddb_get_hw_state(struct intel_crtc *crtc,
> struct skl_ddb_entry *ddb_y,
> struct skl_ddb_entry *ddb_uv);
> -void skl_ddb_get_hw_state(struct drm_i915_private *dev_priv,
> +void skl_ddb_get_hw_state(struct drm_i915_private *i915,
> struct skl_ddb_allocation *ddb /* out */);
> void skl_pipe_wm_get_hw_state(struct intel_crtc *crtc,
> struct skl_pipe_wm *out);
> -void g4x_wm_sanitize(struct drm_i915_private *dev_priv);
> -void vlv_wm_sanitize(struct drm_i915_private *dev_priv);
> +void g4x_wm_sanitize(struct drm_i915_private *i915);
> +void vlv_wm_sanitize(struct drm_i915_private *i915);
> bool intel_can_enable_sagv(struct drm_atomic_state *state);
> -int intel_enable_sagv(struct drm_i915_private *dev_priv);
> -int intel_disable_sagv(struct drm_i915_private *dev_priv);
> +int intel_enable_sagv(struct drm_i915_private *i915);
> +int intel_disable_sagv(struct drm_i915_private *i915);
> bool skl_wm_level_equals(const struct skl_wm_level *l1,
> const struct skl_wm_level *l2);
> bool skl_ddb_allocation_overlaps(const struct skl_ddb_entry *ddb,
> @@ -67,24 +67,24 @@ void skl_write_cursor_wm(struct intel_plane *plane,
> bool ilk_disable_lp_wm(struct drm_device *dev);
> int skl_check_pipe_max_pixel_rate(struct intel_crtc *intel_crtc,
> struct intel_crtc_state *cstate);
> -void intel_init_ipc(struct drm_i915_private *dev_priv);
> -void intel_enable_ipc(struct drm_i915_private *dev_priv);
> +void intel_init_ipc(struct drm_i915_private *i915);
> +void intel_enable_ipc(struct drm_i915_private *i915);
>
> -int intel_gpu_freq(struct drm_i915_private *dev_priv, int val);
> -int intel_freq_opcode(struct drm_i915_private *dev_priv, int val);
> -u64 intel_rc6_residency_ns(struct drm_i915_private *dev_priv, i915_reg_t reg);
> -u64 intel_rc6_residency_us(struct drm_i915_private *dev_priv, i915_reg_t reg);
> +int intel_gpu_freq(struct drm_i915_private *i915, int val);
> +int intel_freq_opcode(struct drm_i915_private *i915, int val);
> +u64 intel_rc6_residency_ns(struct drm_i915_private *i915, i915_reg_t reg);
> +u64 intel_rc6_residency_us(struct drm_i915_private *i915, i915_reg_t reg);
>
> -u32 intel_get_cagf(struct drm_i915_private *dev_priv, u32 rpstat1);
> +u32 intel_get_cagf(struct drm_i915_private *i915, u32 rpstat1);
>
> -unsigned long i915_chipset_val(struct drm_i915_private *dev_priv);
> -unsigned long i915_mch_val(struct drm_i915_private *dev_priv);
> -unsigned long i915_gfx_val(struct drm_i915_private *dev_priv);
> -void i915_update_gfx_val(struct drm_i915_private *dev_priv);
> +unsigned long i915_chipset_val(struct drm_i915_private *i915);
> +unsigned long i915_mch_val(struct drm_i915_private *i915);
> +unsigned long i915_gfx_val(struct drm_i915_private *i915);
> +void i915_update_gfx_val(struct drm_i915_private *i915);
>
> -bool ironlake_set_drps(struct drm_i915_private *dev_priv, u8 val);
> -int intel_set_rps(struct drm_i915_private *dev_priv, u8 val);
> +bool ironlake_set_drps(struct drm_i915_private *i915, u8 val);
> +int intel_set_rps(struct drm_i915_private *i915, u8 val);
> void intel_rps_mark_interactive(struct drm_i915_private *i915, bool interactive);
> -bool intel_set_memory_cxsr(struct drm_i915_private *dev_priv, bool enable);
> +bool intel_set_memory_cxsr(struct drm_i915_private *i915, bool enable);
>
> #endif /* __INTEL_PM_H__ */
> diff --git a/drivers/gpu/drm/i915/intel_psr.c b/drivers/gpu/drm/i915/intel_psr.c
> index 01ca502099df..072eea00b17a 100644
> --- a/drivers/gpu/drm/i915/intel_psr.c
> +++ b/drivers/gpu/drm/i915/intel_psr.c
> @@ -71,14 +71,14 @@ static bool psr_global_enabled(u32 debug)
> }
> }
>
> -static bool intel_psr2_enabled(struct drm_i915_private *dev_priv,
> +static bool intel_psr2_enabled(struct drm_i915_private *i915,
> const struct intel_crtc_state *crtc_state)
> {
> /* Cannot enable DSC and PSR2 simultaneously */
> WARN_ON(crtc_state->dsc_params.compression_enable &&
> crtc_state->has_psr2);
>
> - switch (dev_priv->psr.debug & I915_PSR_DEBUG_MODE_MASK) {
> + switch (i915->psr.debug & I915_PSR_DEBUG_MODE_MASK) {
> case I915_PSR_DEBUG_DISABLE:
> case I915_PSR_DEBUG_FORCE_PSR1:
> return false;
> @@ -104,20 +104,20 @@ static int edp_psr_shift(enum transcoder cpu_transcoder)
> }
> }
>
> -void intel_psr_irq_control(struct drm_i915_private *dev_priv, u32 debug)
> +void intel_psr_irq_control(struct drm_i915_private *i915, u32 debug)
> {
> u32 debug_mask, mask;
> enum transcoder cpu_transcoder;
> u32 transcoders = BIT(TRANSCODER_EDP);
>
> - if (INTEL_GEN(dev_priv) >= 8)
> + if (INTEL_GEN(i915) >= 8)
> transcoders |= BIT(TRANSCODER_A) |
> BIT(TRANSCODER_B) |
> BIT(TRANSCODER_C);
>
> debug_mask = 0;
> mask = 0;
> - for_each_cpu_transcoder_masked(dev_priv, cpu_transcoder, transcoders) {
> + for_each_cpu_transcoder_masked(i915, cpu_transcoder, transcoders) {
> int shift = edp_psr_shift(cpu_transcoder);
>
> mask |= EDP_PSR_ERROR(shift);
> @@ -168,26 +168,26 @@ static void psr_event_print(u32 val, bool psr2_enabled)
> DRM_DEBUG_KMS("\tPSR disabled\n");
> }
>
> -void intel_psr_irq_handler(struct drm_i915_private *dev_priv, u32 psr_iir)
> +void intel_psr_irq_handler(struct drm_i915_private *i915, u32 psr_iir)
> {
> u32 transcoders = BIT(TRANSCODER_EDP);
> enum transcoder cpu_transcoder;
> ktime_t time_ns = ktime_get();
> u32 mask = 0;
>
> - if (INTEL_GEN(dev_priv) >= 8)
> + if (INTEL_GEN(i915) >= 8)
> transcoders |= BIT(TRANSCODER_A) |
> BIT(TRANSCODER_B) |
> BIT(TRANSCODER_C);
>
> - for_each_cpu_transcoder_masked(dev_priv, cpu_transcoder, transcoders) {
> + for_each_cpu_transcoder_masked(i915, cpu_transcoder, transcoders) {
> int shift = edp_psr_shift(cpu_transcoder);
>
> if (psr_iir & EDP_PSR_ERROR(shift)) {
> DRM_WARN("[transcoder %s] PSR aux error\n",
> transcoder_name(cpu_transcoder));
>
> - dev_priv->psr.irq_aux_error = true;
> + i915->psr.irq_aux_error = true;
>
> /*
> * If this interruption is not masked it will keep
> @@ -201,19 +201,19 @@ void intel_psr_irq_handler(struct drm_i915_private *dev_priv, u32 psr_iir)
> }
>
> if (psr_iir & EDP_PSR_PRE_ENTRY(shift)) {
> - dev_priv->psr.last_entry_attempt = time_ns;
> + i915->psr.last_entry_attempt = time_ns;
> DRM_DEBUG_KMS("[transcoder %s] PSR entry attempt in 2 vblanks\n",
> transcoder_name(cpu_transcoder));
> }
>
> if (psr_iir & EDP_PSR_POST_EXIT(shift)) {
> - dev_priv->psr.last_exit = time_ns;
> + i915->psr.last_exit = time_ns;
> DRM_DEBUG_KMS("[transcoder %s] PSR exit completed\n",
> transcoder_name(cpu_transcoder));
>
> - if (INTEL_GEN(dev_priv) >= 9) {
> + if (INTEL_GEN(i915) >= 9) {
> u32 val = I915_READ(PSR_EVENT(cpu_transcoder));
> - bool psr2_enabled = dev_priv->psr.psr2_enabled;
> + bool psr2_enabled = i915->psr.psr2_enabled;
>
> I915_WRITE(PSR_EVENT(cpu_transcoder), val);
> psr_event_print(val, psr2_enabled);
> @@ -225,7 +225,7 @@ void intel_psr_irq_handler(struct drm_i915_private *dev_priv, u32 psr_iir)
> mask |= I915_READ(EDP_PSR_IMR);
> I915_WRITE(EDP_PSR_IMR, mask);
>
> - schedule_work(&dev_priv->psr.work);
> + schedule_work(&i915->psr.work);
> }
> }
>
> @@ -279,7 +279,7 @@ static u16 intel_dp_get_su_x_granulartiy(struct intel_dp *intel_dp)
>
> void intel_psr_init_dpcd(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> to_i915(dp_to_dig_port(intel_dp)->base.base.dev);
>
> drm_dp_dpcd_read(&intel_dp->aux, DP_PSR_SUPPORT, intel_dp->psr_dpcd,
> @@ -300,14 +300,14 @@ void intel_psr_init_dpcd(struct intel_dp *intel_dp)
> return;
> }
>
> - dev_priv->psr.sink_support = true;
> - dev_priv->psr.sink_sync_latency =
> + i915->psr.sink_support = true;
> + i915->psr.sink_sync_latency =
> intel_dp_get_sink_sync_latency(intel_dp);
>
> - WARN_ON(dev_priv->psr.dp);
> - dev_priv->psr.dp = intel_dp;
> + WARN_ON(i915->psr.dp);
> + i915->psr.dp = intel_dp;
>
> - if (INTEL_GEN(dev_priv) >= 9 &&
> + if (INTEL_GEN(i915) >= 9 &&
> (intel_dp->psr_dpcd[0] == DP_PSR2_WITH_Y_COORD_IS_SUPPORTED)) {
> bool y_req = intel_dp->psr_dpcd[1] &
> DP_PSR2_SU_Y_COORDINATE_REQUIRED;
> @@ -324,14 +324,14 @@ void intel_psr_init_dpcd(struct intel_dp *intel_dp)
> * Y-coordinate requirement panels we would need to enable
> * GTC first.
> */
> - dev_priv->psr.sink_psr2_support = y_req && alpm;
> + i915->psr.sink_psr2_support = y_req && alpm;
> DRM_DEBUG_KMS("PSR2 %ssupported\n",
> - dev_priv->psr.sink_psr2_support ? "" : "not ");
> + i915->psr.sink_psr2_support ? "" : "not ");
>
> - if (dev_priv->psr.sink_psr2_support) {
> - dev_priv->psr.colorimetry_support =
> + if (i915->psr.sink_psr2_support) {
> + i915->psr.colorimetry_support =
> intel_dp_get_colorimetry_status(intel_dp);
> - dev_priv->psr.su_x_granularity =
> + i915->psr.su_x_granularity =
> intel_dp_get_su_x_granulartiy(intel_dp);
> }
> }
> @@ -341,15 +341,15 @@ static void intel_psr_setup_vsc(struct intel_dp *intel_dp,
> const struct intel_crtc_state *crtc_state)
> {
> struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp);
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> struct dp_sdp psr_vsc;
>
> - if (dev_priv->psr.psr2_enabled) {
> + if (i915->psr.psr2_enabled) {
> /* Prepare VSC Header for SU as per EDP 1.4 spec, Table 6.11 */
> memset(&psr_vsc, 0, sizeof(psr_vsc));
> psr_vsc.sdp_header.HB0 = 0;
> psr_vsc.sdp_header.HB1 = 0x7;
> - if (dev_priv->psr.colorimetry_support) {
> + if (i915->psr.colorimetry_support) {
> psr_vsc.sdp_header.HB2 = 0x5;
> psr_vsc.sdp_header.HB3 = 0x13;
> } else {
> @@ -372,7 +372,7 @@ static void intel_psr_setup_vsc(struct intel_dp *intel_dp,
>
> static void hsw_psr_setup_aux(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> u32 aux_clock_divider, aux_ctl;
> int i;
> static const u8 aux_msg[] = {
> @@ -405,19 +405,19 @@ static void hsw_psr_setup_aux(struct intel_dp *intel_dp)
>
> static void intel_psr_enable_sink(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> u8 dpcd_val = DP_PSR_ENABLE;
>
> /* Enable ALPM at sink for psr2 */
> - if (dev_priv->psr.psr2_enabled) {
> + if (i915->psr.psr2_enabled) {
> drm_dp_dpcd_writeb(&intel_dp->aux, DP_RECEIVER_ALPM_CONFIG,
> DP_ALPM_ENABLE);
> dpcd_val |= DP_PSR_ENABLE_PSR2 | DP_PSR_IRQ_HPD_WITH_CRC_ERRORS;
> } else {
> - if (dev_priv->psr.link_standby)
> + if (i915->psr.link_standby)
> dpcd_val |= DP_PSR_MAIN_LINK_ACTIVE;
>
> - if (INTEL_GEN(dev_priv) >= 8)
> + if (INTEL_GEN(i915) >= 8)
> dpcd_val |= DP_PSR_CRC_VERIFICATION;
> }
>
> @@ -428,26 +428,26 @@ static void intel_psr_enable_sink(struct intel_dp *intel_dp)
>
> static u32 intel_psr1_get_tp_time(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> u32 val = 0;
>
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> val |= EDP_PSR_TP4_TIME_0US;
>
> - if (dev_priv->vbt.psr.tp1_wakeup_time_us == 0)
> + if (i915->vbt.psr.tp1_wakeup_time_us == 0)
> val |= EDP_PSR_TP1_TIME_0us;
> - else if (dev_priv->vbt.psr.tp1_wakeup_time_us <= 100)
> + else if (i915->vbt.psr.tp1_wakeup_time_us <= 100)
> val |= EDP_PSR_TP1_TIME_100us;
> - else if (dev_priv->vbt.psr.tp1_wakeup_time_us <= 500)
> + else if (i915->vbt.psr.tp1_wakeup_time_us <= 500)
> val |= EDP_PSR_TP1_TIME_500us;
> else
> val |= EDP_PSR_TP1_TIME_2500us;
>
> - if (dev_priv->vbt.psr.tp2_tp3_wakeup_time_us == 0)
> + if (i915->vbt.psr.tp2_tp3_wakeup_time_us == 0)
> val |= EDP_PSR_TP2_TP3_TIME_0us;
> - else if (dev_priv->vbt.psr.tp2_tp3_wakeup_time_us <= 100)
> + else if (i915->vbt.psr.tp2_tp3_wakeup_time_us <= 100)
> val |= EDP_PSR_TP2_TP3_TIME_100us;
> - else if (dev_priv->vbt.psr.tp2_tp3_wakeup_time_us <= 500)
> + else if (i915->vbt.psr.tp2_tp3_wakeup_time_us <= 500)
> val |= EDP_PSR_TP2_TP3_TIME_500us;
> else
> val |= EDP_PSR_TP2_TP3_TIME_2500us;
> @@ -463,31 +463,31 @@ static u32 intel_psr1_get_tp_time(struct intel_dp *intel_dp)
>
> static void hsw_activate_psr1(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> u32 max_sleep_time = 0x1f;
> u32 val = EDP_PSR_ENABLE;
>
> /* Let's use 6 as the minimum to cover all known cases including the
> * off-by-one issue that HW has in some cases.
> */
> - int idle_frames = max(6, dev_priv->vbt.psr.idle_frames);
> + int idle_frames = max(6, i915->vbt.psr.idle_frames);
>
> /* sink_sync_latency of 8 means source has to wait for more than 8
> * frames, we'll go with 9 frames for now
> */
> - idle_frames = max(idle_frames, dev_priv->psr.sink_sync_latency + 1);
> + idle_frames = max(idle_frames, i915->psr.sink_sync_latency + 1);
> val |= idle_frames << EDP_PSR_IDLE_FRAME_SHIFT;
>
> val |= max_sleep_time << EDP_PSR_MAX_SLEEP_TIME_SHIFT;
> - if (IS_HASWELL(dev_priv))
> + if (IS_HASWELL(i915))
> val |= EDP_PSR_MIN_LINK_ENTRY_TIME_8_LINES;
>
> - if (dev_priv->psr.link_standby)
> + if (i915->psr.link_standby)
> val |= EDP_PSR_LINK_STANDBY;
>
> val |= intel_psr1_get_tp_time(intel_dp);
>
> - if (INTEL_GEN(dev_priv) >= 8)
> + if (INTEL_GEN(i915) >= 8)
> val |= EDP_PSR_CRC_ENABLE;
>
> val |= I915_READ(EDP_PSR_CTL) & EDP_PSR_RESTORE_PSR_ACTIVE_CTX_MASK;
> @@ -496,29 +496,29 @@ static void hsw_activate_psr1(struct intel_dp *intel_dp)
>
> static void hsw_activate_psr2(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> u32 val;
>
> /* Let's use 6 as the minimum to cover all known cases including the
> * off-by-one issue that HW has in some cases.
> */
> - int idle_frames = max(6, dev_priv->vbt.psr.idle_frames);
> + int idle_frames = max(6, i915->vbt.psr.idle_frames);
>
> - idle_frames = max(idle_frames, dev_priv->psr.sink_sync_latency + 1);
> + idle_frames = max(idle_frames, i915->psr.sink_sync_latency + 1);
> val = idle_frames << EDP_PSR2_IDLE_FRAME_SHIFT;
>
> val |= EDP_PSR2_ENABLE | EDP_SU_TRACK_ENABLE;
> - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> val |= EDP_Y_COORDINATE_ENABLE;
>
> - val |= EDP_PSR2_FRAME_BEFORE_SU(dev_priv->psr.sink_sync_latency + 1);
> + val |= EDP_PSR2_FRAME_BEFORE_SU(i915->psr.sink_sync_latency + 1);
>
> - if (dev_priv->vbt.psr.psr2_tp2_tp3_wakeup_time_us >= 0 &&
> - dev_priv->vbt.psr.psr2_tp2_tp3_wakeup_time_us <= 50)
> + if (i915->vbt.psr.psr2_tp2_tp3_wakeup_time_us >= 0 &&
> + i915->vbt.psr.psr2_tp2_tp3_wakeup_time_us <= 50)
> val |= EDP_PSR2_TP2_TIME_50us;
> - else if (dev_priv->vbt.psr.psr2_tp2_tp3_wakeup_time_us <= 100)
> + else if (i915->vbt.psr.psr2_tp2_tp3_wakeup_time_us <= 100)
> val |= EDP_PSR2_TP2_TIME_100us;
> - else if (dev_priv->vbt.psr.psr2_tp2_tp3_wakeup_time_us <= 500)
> + else if (i915->vbt.psr.psr2_tp2_tp3_wakeup_time_us <= 500)
> val |= EDP_PSR2_TP2_TIME_500us;
> else
> val |= EDP_PSR2_TP2_TIME_2500us;
> @@ -535,12 +535,12 @@ static void hsw_activate_psr2(struct intel_dp *intel_dp)
> static bool intel_psr2_config_valid(struct intel_dp *intel_dp,
> struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> int crtc_hdisplay = crtc_state->base.adjusted_mode.crtc_hdisplay;
> int crtc_vdisplay = crtc_state->base.adjusted_mode.crtc_vdisplay;
> int psr_max_h = 0, psr_max_v = 0;
>
> - if (!dev_priv->psr.sink_psr2_support)
> + if (!i915->psr.sink_psr2_support)
> return false;
>
> /*
> @@ -553,10 +553,10 @@ static bool intel_psr2_config_valid(struct intel_dp *intel_dp,
> return false;
> }
>
> - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) {
> + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) {
> psr_max_h = 4096;
> psr_max_v = 2304;
> - } else if (IS_GEN(dev_priv, 9)) {
> + } else if (IS_GEN(i915, 9)) {
> psr_max_h = 3640;
> psr_max_v = 2304;
> }
> @@ -574,9 +574,9 @@ static bool intel_psr2_config_valid(struct intel_dp *intel_dp,
> * only need to validate the SU block width is a multiple of
> * x granularity.
> */
> - if (crtc_hdisplay % dev_priv->psr.su_x_granularity) {
> + if (crtc_hdisplay % i915->psr.su_x_granularity) {
> DRM_DEBUG_KMS("PSR2 not enabled, hdisplay(%d) not multiple of %d\n",
> - crtc_hdisplay, dev_priv->psr.su_x_granularity);
> + crtc_hdisplay, i915->psr.su_x_granularity);
> return false;
> }
>
> @@ -592,15 +592,15 @@ void intel_psr_compute_config(struct intel_dp *intel_dp,
> struct intel_crtc_state *crtc_state)
> {
> struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> const struct drm_display_mode *adjusted_mode =
> &crtc_state->base.adjusted_mode;
> int psr_setup_time;
>
> - if (!CAN_PSR(dev_priv))
> + if (!CAN_PSR(i915))
> return;
>
> - if (intel_dp != dev_priv->psr.dp)
> + if (intel_dp != i915->psr.dp)
> return;
>
> /*
> @@ -615,7 +615,7 @@ void intel_psr_compute_config(struct intel_dp *intel_dp,
> return;
> }
>
> - if (dev_priv->psr.sink_not_reliable) {
> + if (i915->psr.sink_not_reliable) {
> DRM_DEBUG_KMS("PSR sink implementation is not reliable\n");
> return;
> }
> @@ -645,24 +645,24 @@ void intel_psr_compute_config(struct intel_dp *intel_dp,
>
> static void intel_psr_activate(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
>
> - if (INTEL_GEN(dev_priv) >= 9)
> + if (INTEL_GEN(i915) >= 9)
> WARN_ON(I915_READ(EDP_PSR2_CTL) & EDP_PSR2_ENABLE);
> WARN_ON(I915_READ(EDP_PSR_CTL) & EDP_PSR_ENABLE);
> - WARN_ON(dev_priv->psr.active);
> - lockdep_assert_held(&dev_priv->psr.lock);
> + WARN_ON(i915->psr.active);
> + lockdep_assert_held(&i915->psr.lock);
>
> /* psr1 and psr2 are mutually exclusive.*/
> - if (dev_priv->psr.psr2_enabled)
> + if (i915->psr.psr2_enabled)
> hsw_activate_psr2(intel_dp);
> else
> hsw_activate_psr1(intel_dp);
>
> - dev_priv->psr.active = true;
> + i915->psr.active = true;
> }
>
> -static i915_reg_t gen9_chicken_trans_reg(struct drm_i915_private *dev_priv,
> +static i915_reg_t gen9_chicken_trans_reg(struct drm_i915_private *i915,
> enum transcoder cpu_transcoder)
> {
> static const i915_reg_t regs[] = {
> @@ -672,7 +672,7 @@ static i915_reg_t gen9_chicken_trans_reg(struct drm_i915_private *dev_priv,
> [TRANSCODER_EDP] = CHICKEN_TRANS_EDP,
> };
>
> - WARN_ON(INTEL_GEN(dev_priv) < 9);
> + WARN_ON(INTEL_GEN(i915) < 9);
>
> if (WARN_ON(cpu_transcoder >= ARRAY_SIZE(regs) ||
> !regs[cpu_transcoder].reg))
> @@ -684,19 +684,19 @@ static i915_reg_t gen9_chicken_trans_reg(struct drm_i915_private *dev_priv,
> static void intel_psr_enable_source(struct intel_dp *intel_dp,
> const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> u32 mask;
>
> /* Only HSW and BDW have PSR AUX registers that need to be setup. SKL+
> * use hardcoded values PSR AUX transactions
> */
> - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> + if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> hsw_psr_setup_aux(intel_dp);
>
> - if (dev_priv->psr.psr2_enabled && (IS_GEN(dev_priv, 9) &&
> - !IS_GEMINILAKE(dev_priv))) {
> - i915_reg_t reg = gen9_chicken_trans_reg(dev_priv,
> + if (i915->psr.psr2_enabled && (IS_GEN(i915, 9) &&
> + !IS_GEMINILAKE(i915))) {
> + i915_reg_t reg = gen9_chicken_trans_reg(i915,
> cpu_transcoder);
> u32 chicken = I915_READ(reg);
>
> @@ -716,29 +716,29 @@ static void intel_psr_enable_source(struct intel_dp *intel_dp,
> EDP_PSR_DEBUG_MASK_LPSP |
> EDP_PSR_DEBUG_MASK_MAX_SLEEP;
>
> - if (INTEL_GEN(dev_priv) < 11)
> + if (INTEL_GEN(i915) < 11)
> mask |= EDP_PSR_DEBUG_MASK_DISP_REG_WRITE;
>
> I915_WRITE(EDP_PSR_DEBUG, mask);
> }
>
> -static void intel_psr_enable_locked(struct drm_i915_private *dev_priv,
> +static void intel_psr_enable_locked(struct drm_i915_private *i915,
> const struct intel_crtc_state *crtc_state)
> {
> - struct intel_dp *intel_dp = dev_priv->psr.dp;
> + struct intel_dp *intel_dp = i915->psr.dp;
>
> - WARN_ON(dev_priv->psr.enabled);
> + WARN_ON(i915->psr.enabled);
>
> - dev_priv->psr.psr2_enabled = intel_psr2_enabled(dev_priv, crtc_state);
> - dev_priv->psr.busy_frontbuffer_bits = 0;
> - dev_priv->psr.pipe = to_intel_crtc(crtc_state->base.crtc)->pipe;
> + i915->psr.psr2_enabled = intel_psr2_enabled(i915, crtc_state);
> + i915->psr.busy_frontbuffer_bits = 0;
> + i915->psr.pipe = to_intel_crtc(crtc_state->base.crtc)->pipe;
>
> DRM_DEBUG_KMS("Enabling PSR%s\n",
> - dev_priv->psr.psr2_enabled ? "2" : "1");
> + i915->psr.psr2_enabled ? "2" : "1");
> intel_psr_setup_vsc(intel_dp, crtc_state);
> intel_psr_enable_sink(intel_dp);
> intel_psr_enable_source(intel_dp, crtc_state);
> - dev_priv->psr.enabled = true;
> + i915->psr.enabled = true;
>
> intel_psr_activate(intel_dp);
> }
> @@ -753,41 +753,41 @@ static void intel_psr_enable_locked(struct drm_i915_private *dev_priv,
> void intel_psr_enable(struct intel_dp *intel_dp,
> const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
>
> if (!crtc_state->has_psr)
> return;
>
> - if (WARN_ON(!CAN_PSR(dev_priv)))
> + if (WARN_ON(!CAN_PSR(i915)))
> return;
>
> - WARN_ON(dev_priv->drrs.dp);
> + WARN_ON(i915->drrs.dp);
>
> - mutex_lock(&dev_priv->psr.lock);
> + mutex_lock(&i915->psr.lock);
>
> - if (!psr_global_enabled(dev_priv->psr.debug)) {
> + if (!psr_global_enabled(i915->psr.debug)) {
> DRM_DEBUG_KMS("PSR disabled by flag\n");
> goto unlock;
> }
>
> - intel_psr_enable_locked(dev_priv, crtc_state);
> + intel_psr_enable_locked(i915, crtc_state);
>
> unlock:
> - mutex_unlock(&dev_priv->psr.lock);
> + mutex_unlock(&i915->psr.lock);
> }
>
> -static void intel_psr_exit(struct drm_i915_private *dev_priv)
> +static void intel_psr_exit(struct drm_i915_private *i915)
> {
> u32 val;
>
> - if (!dev_priv->psr.active) {
> - if (INTEL_GEN(dev_priv) >= 9)
> + if (!i915->psr.active) {
> + if (INTEL_GEN(i915) >= 9)
> WARN_ON(I915_READ(EDP_PSR2_CTL) & EDP_PSR2_ENABLE);
> WARN_ON(I915_READ(EDP_PSR_CTL) & EDP_PSR_ENABLE);
> return;
> }
>
> - if (dev_priv->psr.psr2_enabled) {
> + if (i915->psr.psr2_enabled) {
> val = I915_READ(EDP_PSR2_CTL);
> WARN_ON(!(val & EDP_PSR2_ENABLE));
> I915_WRITE(EDP_PSR2_CTL, val & ~EDP_PSR2_ENABLE);
> @@ -796,26 +796,26 @@ static void intel_psr_exit(struct drm_i915_private *dev_priv)
> WARN_ON(!(val & EDP_PSR_ENABLE));
> I915_WRITE(EDP_PSR_CTL, val & ~EDP_PSR_ENABLE);
> }
> - dev_priv->psr.active = false;
> + i915->psr.active = false;
> }
>
> static void intel_psr_disable_locked(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> i915_reg_t psr_status;
> u32 psr_status_mask;
>
> - lockdep_assert_held(&dev_priv->psr.lock);
> + lockdep_assert_held(&i915->psr.lock);
>
> - if (!dev_priv->psr.enabled)
> + if (!i915->psr.enabled)
> return;
>
> DRM_DEBUG_KMS("Disabling PSR%s\n",
> - dev_priv->psr.psr2_enabled ? "2" : "1");
> + i915->psr.psr2_enabled ? "2" : "1");
>
> - intel_psr_exit(dev_priv);
> + intel_psr_exit(i915);
>
> - if (dev_priv->psr.psr2_enabled) {
> + if (i915->psr.psr2_enabled) {
> psr_status = EDP_PSR2_STATUS;
> psr_status_mask = EDP_PSR2_STATUS_STATE_MASK;
> } else {
> @@ -824,14 +824,14 @@ static void intel_psr_disable_locked(struct intel_dp *intel_dp)
> }
>
> /* Wait till PSR is idle */
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> psr_status, psr_status_mask, 0, 2000))
> DRM_ERROR("Timed out waiting PSR idle state\n");
>
> /* Disable PSR on Sink */
> drm_dp_dpcd_writeb(&intel_dp->aux, DP_PSR_EN_CFG, 0);
>
> - dev_priv->psr.enabled = false;
> + i915->psr.enabled = false;
> }
>
> /**
> @@ -844,23 +844,23 @@ static void intel_psr_disable_locked(struct intel_dp *intel_dp)
> void intel_psr_disable(struct intel_dp *intel_dp,
> const struct intel_crtc_state *old_crtc_state)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
>
> if (!old_crtc_state->has_psr)
> return;
>
> - if (WARN_ON(!CAN_PSR(dev_priv)))
> + if (WARN_ON(!CAN_PSR(i915)))
> return;
>
> - mutex_lock(&dev_priv->psr.lock);
> + mutex_lock(&i915->psr.lock);
>
> intel_psr_disable_locked(intel_dp);
>
> - mutex_unlock(&dev_priv->psr.lock);
> - cancel_work_sync(&dev_priv->psr.work);
> + mutex_unlock(&i915->psr.lock);
> + cancel_work_sync(&i915->psr.work);
> }
>
> -static void psr_force_hw_tracking_exit(struct drm_i915_private *dev_priv)
> +static void psr_force_hw_tracking_exit(struct drm_i915_private *i915)
> {
> /*
> * Display WA #0884: all
> @@ -871,7 +871,7 @@ static void psr_force_hw_tracking_exit(struct drm_i915_private *dev_priv)
> * but it makes more sense write to the current active
> * pipe.
> */
> - I915_WRITE(CURSURFLIVE(dev_priv->psr.pipe), 0);
> + I915_WRITE(CURSURFLIVE(i915->psr.pipe), 0);
> }
>
> /**
> @@ -886,22 +886,22 @@ static void psr_force_hw_tracking_exit(struct drm_i915_private *dev_priv)
> void intel_psr_update(struct intel_dp *intel_dp,
> const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> - struct i915_psr *psr = &dev_priv->psr;
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> + struct i915_psr *psr = &i915->psr;
> bool enable, psr2_enable;
>
> - if (!CAN_PSR(dev_priv) || READ_ONCE(psr->dp) != intel_dp)
> + if (!CAN_PSR(i915) || READ_ONCE(psr->dp) != intel_dp)
> return;
>
> - mutex_lock(&dev_priv->psr.lock);
> + mutex_lock(&i915->psr.lock);
>
> enable = crtc_state->has_psr && psr_global_enabled(psr->debug);
> - psr2_enable = intel_psr2_enabled(dev_priv, crtc_state);
> + psr2_enable = intel_psr2_enabled(i915, crtc_state);
>
> if (enable == psr->enabled && psr2_enable == psr->psr2_enabled) {
> /* Force a PSR exit when enabling CRC to avoid CRC timeouts */
> if (crtc_state->crc_enabled && psr->enabled)
> - psr_force_hw_tracking_exit(dev_priv);
> + psr_force_hw_tracking_exit(i915);
>
> goto unlock;
> }
> @@ -910,10 +910,10 @@ void intel_psr_update(struct intel_dp *intel_dp,
> intel_psr_disable_locked(intel_dp);
>
> if (enable)
> - intel_psr_enable_locked(dev_priv, crtc_state);
> + intel_psr_enable_locked(i915, crtc_state);
>
> unlock:
> - mutex_unlock(&dev_priv->psr.lock);
> + mutex_unlock(&i915->psr.lock);
> }
>
> /**
> @@ -930,13 +930,13 @@ int intel_psr_wait_for_idle(const struct intel_crtc_state *new_crtc_state,
> u32 *out_value)
> {
> struct intel_crtc *crtc = to_intel_crtc(new_crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
>
> - if (!dev_priv->psr.enabled || !new_crtc_state->has_psr)
> + if (!i915->psr.enabled || !new_crtc_state->has_psr)
> return 0;
>
> /* FIXME: Update this for PSR2 if we need to wait for idle */
> - if (READ_ONCE(dev_priv->psr.psr2_enabled))
> + if (READ_ONCE(i915->psr.psr2_enabled))
> return 0;
>
> /*
> @@ -946,22 +946,22 @@ int intel_psr_wait_for_idle(const struct intel_crtc_state *new_crtc_state,
> * defensive enough to cover everything.
> */
>
> - return __intel_wait_for_register(&dev_priv->uncore, EDP_PSR_STATUS,
> + return __intel_wait_for_register(&i915->uncore, EDP_PSR_STATUS,
> EDP_PSR_STATUS_STATE_MASK,
> EDP_PSR_STATUS_STATE_IDLE, 2, 50,
> out_value);
> }
>
> -static bool __psr_wait_for_idle_locked(struct drm_i915_private *dev_priv)
> +static bool __psr_wait_for_idle_locked(struct drm_i915_private *i915)
> {
> i915_reg_t reg;
> u32 mask;
> int err;
>
> - if (!dev_priv->psr.enabled)
> + if (!i915->psr.enabled)
> return false;
>
> - if (dev_priv->psr.psr2_enabled) {
> + if (i915->psr.psr2_enabled) {
> reg = EDP_PSR2_STATUS;
> mask = EDP_PSR2_STATUS_STATE_MASK;
> } else {
> @@ -969,20 +969,20 @@ static bool __psr_wait_for_idle_locked(struct drm_i915_private *dev_priv)
> mask = EDP_PSR_STATUS_STATE_MASK;
> }
>
> - mutex_unlock(&dev_priv->psr.lock);
> + mutex_unlock(&i915->psr.lock);
>
> - err = intel_wait_for_register(&dev_priv->uncore, reg, mask, 0, 50);
> + err = intel_wait_for_register(&i915->uncore, reg, mask, 0, 50);
> if (err)
> DRM_ERROR("Timed out waiting for PSR Idle for re-enable\n");
>
> /* After the unlocked wait, verify that PSR is still wanted! */
> - mutex_lock(&dev_priv->psr.lock);
> - return err == 0 && dev_priv->psr.enabled;
> + mutex_lock(&i915->psr.lock);
> + return err == 0 && i915->psr.enabled;
> }
>
> -static int intel_psr_fastset_force(struct drm_i915_private *dev_priv)
> +static int intel_psr_fastset_force(struct drm_i915_private *i915)
> {
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_device *dev = &i915->drm;
> struct drm_modeset_acquire_ctx ctx;
> struct drm_atomic_state *state;
> struct drm_crtc *crtc;
> @@ -1032,7 +1032,7 @@ static int intel_psr_fastset_force(struct drm_i915_private *dev_priv)
> return err;
> }
>
> -int intel_psr_debug_set(struct drm_i915_private *dev_priv, u64 val)
> +int intel_psr_debug_set(struct drm_i915_private *i915, u64 val)
> {
> const u32 mode = val & I915_PSR_DEBUG_MODE_MASK;
> u32 old_mode;
> @@ -1044,25 +1044,25 @@ int intel_psr_debug_set(struct drm_i915_private *dev_priv, u64 val)
> return -EINVAL;
> }
>
> - ret = mutex_lock_interruptible(&dev_priv->psr.lock);
> + ret = mutex_lock_interruptible(&i915->psr.lock);
> if (ret)
> return ret;
>
> - old_mode = dev_priv->psr.debug & I915_PSR_DEBUG_MODE_MASK;
> - dev_priv->psr.debug = val;
> - intel_psr_irq_control(dev_priv, dev_priv->psr.debug);
> + old_mode = i915->psr.debug & I915_PSR_DEBUG_MODE_MASK;
> + i915->psr.debug = val;
> + intel_psr_irq_control(i915, i915->psr.debug);
>
> - mutex_unlock(&dev_priv->psr.lock);
> + mutex_unlock(&i915->psr.lock);
>
> if (old_mode != mode)
> - ret = intel_psr_fastset_force(dev_priv);
> + ret = intel_psr_fastset_force(i915);
>
> return ret;
> }
>
> -static void intel_psr_handle_irq(struct drm_i915_private *dev_priv)
> +static void intel_psr_handle_irq(struct drm_i915_private *i915)
> {
> - struct i915_psr *psr = &dev_priv->psr;
> + struct i915_psr *psr = &i915->psr;
>
> intel_psr_disable_locked(psr->dp);
> psr->sink_not_reliable = true;
> @@ -1072,16 +1072,16 @@ static void intel_psr_handle_irq(struct drm_i915_private *dev_priv)
>
> static void intel_psr_work(struct work_struct *work)
> {
> - struct drm_i915_private *dev_priv =
> - container_of(work, typeof(*dev_priv), psr.work);
> + struct drm_i915_private *i915 =
> + container_of(work, typeof(*i915), psr.work);
>
> - mutex_lock(&dev_priv->psr.lock);
> + mutex_lock(&i915->psr.lock);
>
> - if (!dev_priv->psr.enabled)
> + if (!i915->psr.enabled)
> goto unlock;
>
> - if (READ_ONCE(dev_priv->psr.irq_aux_error))
> - intel_psr_handle_irq(dev_priv);
> + if (READ_ONCE(i915->psr.irq_aux_error))
> + intel_psr_handle_irq(i915);
>
> /*
> * We have to make sure PSR is ready for re-enable
> @@ -1089,7 +1089,7 @@ static void intel_psr_work(struct work_struct *work)
> * PSR might take some time to get fully disabled
> * and be ready for re-enable.
> */
> - if (!__psr_wait_for_idle_locked(dev_priv))
> + if (!__psr_wait_for_idle_locked(i915))
> goto unlock;
>
> /*
> @@ -1097,17 +1097,17 @@ static void intel_psr_work(struct work_struct *work)
> * recheck. Since psr_flush first clears this and then reschedules we
> * won't ever miss a flush when bailing out here.
> */
> - if (dev_priv->psr.busy_frontbuffer_bits || dev_priv->psr.active)
> + if (i915->psr.busy_frontbuffer_bits || i915->psr.active)
> goto unlock;
>
> - intel_psr_activate(dev_priv->psr.dp);
> + intel_psr_activate(i915->psr.dp);
> unlock:
> - mutex_unlock(&dev_priv->psr.lock);
> + mutex_unlock(&i915->psr.lock);
> }
>
> /**
> * intel_psr_invalidate - Invalidade PSR
> - * @dev_priv: i915 device
> + * @i915: i915 device
> * @frontbuffer_bits: frontbuffer plane tracking bits
> * @origin: which operation caused the invalidate
> *
> @@ -1118,33 +1118,33 @@ static void intel_psr_work(struct work_struct *work)
> *
> * Dirty frontbuffers relevant to PSR are tracked in busy_frontbuffer_bits."
> */
> -void intel_psr_invalidate(struct drm_i915_private *dev_priv,
> +void intel_psr_invalidate(struct drm_i915_private *i915,
> unsigned frontbuffer_bits, enum fb_op_origin origin)
> {
> - if (!CAN_PSR(dev_priv))
> + if (!CAN_PSR(i915))
> return;
>
> if (origin == ORIGIN_FLIP)
> return;
>
> - mutex_lock(&dev_priv->psr.lock);
> - if (!dev_priv->psr.enabled) {
> - mutex_unlock(&dev_priv->psr.lock);
> + mutex_lock(&i915->psr.lock);
> + if (!i915->psr.enabled) {
> + mutex_unlock(&i915->psr.lock);
> return;
> }
>
> - frontbuffer_bits &= INTEL_FRONTBUFFER_ALL_MASK(dev_priv->psr.pipe);
> - dev_priv->psr.busy_frontbuffer_bits |= frontbuffer_bits;
> + frontbuffer_bits &= INTEL_FRONTBUFFER_ALL_MASK(i915->psr.pipe);
> + i915->psr.busy_frontbuffer_bits |= frontbuffer_bits;
>
> if (frontbuffer_bits)
> - intel_psr_exit(dev_priv);
> + intel_psr_exit(i915);
>
> - mutex_unlock(&dev_priv->psr.lock);
> + mutex_unlock(&i915->psr.lock);
> }
>
> /**
> * intel_psr_flush - Flush PSR
> - * @dev_priv: i915 device
> + * @i915: i915 device
> * @frontbuffer_bits: frontbuffer plane tracking bits
> * @origin: which operation caused the flush
> *
> @@ -1155,55 +1155,55 @@ void intel_psr_invalidate(struct drm_i915_private *dev_priv,
> *
> * Dirty frontbuffers relevant to PSR are tracked in busy_frontbuffer_bits.
> */
> -void intel_psr_flush(struct drm_i915_private *dev_priv,
> +void intel_psr_flush(struct drm_i915_private *i915,
> unsigned frontbuffer_bits, enum fb_op_origin origin)
> {
> - if (!CAN_PSR(dev_priv))
> + if (!CAN_PSR(i915))
> return;
>
> if (origin == ORIGIN_FLIP)
> return;
>
> - mutex_lock(&dev_priv->psr.lock);
> - if (!dev_priv->psr.enabled) {
> - mutex_unlock(&dev_priv->psr.lock);
> + mutex_lock(&i915->psr.lock);
> + if (!i915->psr.enabled) {
> + mutex_unlock(&i915->psr.lock);
> return;
> }
>
> - frontbuffer_bits &= INTEL_FRONTBUFFER_ALL_MASK(dev_priv->psr.pipe);
> - dev_priv->psr.busy_frontbuffer_bits &= ~frontbuffer_bits;
> + frontbuffer_bits &= INTEL_FRONTBUFFER_ALL_MASK(i915->psr.pipe);
> + i915->psr.busy_frontbuffer_bits &= ~frontbuffer_bits;
>
> /* By definition flush = invalidate + flush */
> if (frontbuffer_bits)
> - psr_force_hw_tracking_exit(dev_priv);
> + psr_force_hw_tracking_exit(i915);
>
> - if (!dev_priv->psr.active && !dev_priv->psr.busy_frontbuffer_bits)
> - schedule_work(&dev_priv->psr.work);
> - mutex_unlock(&dev_priv->psr.lock);
> + if (!i915->psr.active && !i915->psr.busy_frontbuffer_bits)
> + schedule_work(&i915->psr.work);
> + mutex_unlock(&i915->psr.lock);
> }
>
> /**
> * intel_psr_init - Init basic PSR work and mutex.
> - * @dev_priv: i915 device private
> + * @i915: i915 device private
> *
> * This function is called only once at driver load to initialize basic
> * PSR stuff.
> */
> -void intel_psr_init(struct drm_i915_private *dev_priv)
> +void intel_psr_init(struct drm_i915_private *i915)
> {
> u32 val;
>
> - if (!HAS_PSR(dev_priv))
> + if (!HAS_PSR(i915))
> return;
>
> - dev_priv->psr_mmio_base = IS_HASWELL(dev_priv) ?
> + i915->psr_mmio_base = IS_HASWELL(i915) ?
> HSW_EDP_PSR_BASE : BDW_EDP_PSR_BASE;
>
> - if (!dev_priv->psr.sink_support)
> + if (!i915->psr.sink_support)
> return;
>
> if (i915_modparams.enable_psr == -1)
> - if (INTEL_GEN(dev_priv) < 9 || !dev_priv->vbt.psr.enable)
> + if (INTEL_GEN(i915) < 9 || !i915->vbt.psr.enable)
> i915_modparams.enable_psr = 0;
>
> /*
> @@ -1218,31 +1218,31 @@ void intel_psr_init(struct drm_i915_private *dev_priv)
> val &= EDP_PSR_ERROR(edp_psr_shift(TRANSCODER_EDP));
> if (val) {
> DRM_DEBUG_KMS("PSR interruption error set\n");
> - dev_priv->psr.sink_not_reliable = true;
> + i915->psr.sink_not_reliable = true;
> }
>
> /* Set link_standby x link_off defaults */
> - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> + if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> /* HSW and BDW require workarounds that we don't implement. */
> - dev_priv->psr.link_standby = false;
> + i915->psr.link_standby = false;
> else
> /* For new platforms let's respect VBT back again */
> - dev_priv->psr.link_standby = dev_priv->vbt.psr.full_link;
> + i915->psr.link_standby = i915->vbt.psr.full_link;
>
> - INIT_WORK(&dev_priv->psr.work, intel_psr_work);
> - mutex_init(&dev_priv->psr.lock);
> + INIT_WORK(&i915->psr.work, intel_psr_work);
> + mutex_init(&i915->psr.lock);
> }
>
> void intel_psr_short_pulse(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> - struct i915_psr *psr = &dev_priv->psr;
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> + struct i915_psr *psr = &i915->psr;
> u8 val;
> const u8 errors = DP_PSR_RFB_STORAGE_ERROR |
> DP_PSR_VSC_SDP_UNCORRECTABLE_ERROR |
> DP_PSR_LINK_CRC_ERROR;
>
> - if (!CAN_PSR(dev_priv) || !intel_dp_is_edp(intel_dp))
> + if (!CAN_PSR(i915) || !intel_dp_is_edp(intel_dp))
> return;
>
> mutex_lock(&psr->lock);
> @@ -1288,15 +1288,15 @@ void intel_psr_short_pulse(struct intel_dp *intel_dp)
>
> bool intel_psr_enabled(struct intel_dp *intel_dp)
> {
> - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> + struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> bool ret;
>
> - if (!CAN_PSR(dev_priv) || !intel_dp_is_edp(intel_dp))
> + if (!CAN_PSR(i915) || !intel_dp_is_edp(intel_dp))
> return false;
>
> - mutex_lock(&dev_priv->psr.lock);
> - ret = (dev_priv->psr.dp == intel_dp && dev_priv->psr.enabled);
> - mutex_unlock(&dev_priv->psr.lock);
> + mutex_lock(&i915->psr.lock);
> + ret = (i915->psr.dp == intel_dp && i915->psr.enabled);
> + mutex_unlock(&i915->psr.lock);
>
> return ret;
> }
> diff --git a/drivers/gpu/drm/i915/intel_psr.h b/drivers/gpu/drm/i915/intel_psr.h
> index dc818826f36d..8a3ddc4ae44d 100644
> --- a/drivers/gpu/drm/i915/intel_psr.h
> +++ b/drivers/gpu/drm/i915/intel_psr.h
> @@ -12,7 +12,7 @@ struct drm_i915_private;
> struct intel_crtc_state;
> struct intel_dp;
>
> -#define CAN_PSR(dev_priv) (HAS_PSR(dev_priv) && dev_priv->psr.sink_support)
> +#define CAN_PSR(i915) (HAS_PSR(i915) && i915->psr.sink_support)
> void intel_psr_init_dpcd(struct intel_dp *intel_dp);
> void intel_psr_enable(struct intel_dp *intel_dp,
> const struct intel_crtc_state *crtc_state);
> @@ -20,18 +20,18 @@ void intel_psr_disable(struct intel_dp *intel_dp,
> const struct intel_crtc_state *old_crtc_state);
> void intel_psr_update(struct intel_dp *intel_dp,
> const struct intel_crtc_state *crtc_state);
> -int intel_psr_debug_set(struct drm_i915_private *dev_priv, u64 value);
> -void intel_psr_invalidate(struct drm_i915_private *dev_priv,
> +int intel_psr_debug_set(struct drm_i915_private *i915, u64 value);
> +void intel_psr_invalidate(struct drm_i915_private *i915,
> unsigned frontbuffer_bits,
> enum fb_op_origin origin);
> -void intel_psr_flush(struct drm_i915_private *dev_priv,
> +void intel_psr_flush(struct drm_i915_private *i915,
> unsigned frontbuffer_bits,
> enum fb_op_origin origin);
> -void intel_psr_init(struct drm_i915_private *dev_priv);
> +void intel_psr_init(struct drm_i915_private *i915);
> void intel_psr_compute_config(struct intel_dp *intel_dp,
> struct intel_crtc_state *crtc_state);
> -void intel_psr_irq_control(struct drm_i915_private *dev_priv, u32 debug);
> -void intel_psr_irq_handler(struct drm_i915_private *dev_priv, u32 psr_iir);
> +void intel_psr_irq_control(struct drm_i915_private *i915, u32 debug);
> +void intel_psr_irq_handler(struct drm_i915_private *i915, u32 psr_iir);
> void intel_psr_short_pulse(struct intel_dp *intel_dp);
> int intel_psr_wait_for_idle(const struct intel_crtc_state *new_crtc_state,
> u32 *out_value);
> diff --git a/drivers/gpu/drm/i915/intel_quirks.h b/drivers/gpu/drm/i915/intel_quirks.h
> index b0fcff142a56..b617ca047695 100644
> --- a/drivers/gpu/drm/i915/intel_quirks.h
> +++ b/drivers/gpu/drm/i915/intel_quirks.h
> @@ -8,6 +8,6 @@
>
> struct drm_i915_private;
>
> -void intel_init_quirks(struct drm_i915_private *dev_priv);
> +void intel_init_quirks(struct drm_i915_private *i915);
>
> #endif /* __INTEL_QUIRKS_H__ */
> diff --git a/drivers/gpu/drm/i915/intel_runtime_pm.h b/drivers/gpu/drm/i915/intel_runtime_pm.h
> index a7acceb13473..1042f1c0b125 100644
> --- a/drivers/gpu/drm/i915/intel_runtime_pm.h
> +++ b/drivers/gpu/drm/i915/intel_runtime_pm.h
> @@ -20,10 +20,10 @@ enum i915_drm_suspend_mode {
> I915_DRM_SUSPEND_HIBERNATE,
> };
>
> -void intel_runtime_pm_init_early(struct drm_i915_private *dev_priv);
> -void intel_runtime_pm_enable(struct drm_i915_private *dev_priv);
> -void intel_runtime_pm_disable(struct drm_i915_private *dev_priv);
> -void intel_runtime_pm_cleanup(struct drm_i915_private *dev_priv);
> +void intel_runtime_pm_init_early(struct drm_i915_private *i915);
> +void intel_runtime_pm_enable(struct drm_i915_private *i915);
> +void intel_runtime_pm_disable(struct drm_i915_private *i915);
> +void intel_runtime_pm_cleanup(struct drm_i915_private *i915);
>
> intel_wakeref_t intel_runtime_pm_get(struct drm_i915_private *i915);
> intel_wakeref_t intel_runtime_pm_get_if_in_use(struct drm_i915_private *i915);
> diff --git a/drivers/gpu/drm/i915/intel_sdvo.c b/drivers/gpu/drm/i915/intel_sdvo.c
> index 0860ae36bb87..b916c79d6b31 100644
> --- a/drivers/gpu/drm/i915/intel_sdvo.c
> +++ b/drivers/gpu/drm/i915/intel_sdvo.c
> @@ -212,18 +212,18 @@ intel_sdvo_create_enhance_property(struct intel_sdvo *intel_sdvo,
> static void intel_sdvo_write_sdvox(struct intel_sdvo *intel_sdvo, u32 val)
> {
> struct drm_device *dev = intel_sdvo->base.base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> u32 bval = val, cval = val;
> int i;
>
> - if (HAS_PCH_SPLIT(dev_priv)) {
> + if (HAS_PCH_SPLIT(i915)) {
> I915_WRITE(intel_sdvo->sdvo_reg, val);
> POSTING_READ(intel_sdvo->sdvo_reg);
> /*
> * HW workaround, need to write this twice for issue
> * that may result in first write getting masked.
> */
> - if (HAS_PCH_IBX(dev_priv)) {
> + if (HAS_PCH_IBX(i915)) {
> I915_WRITE(intel_sdvo->sdvo_reg, val);
> POSTING_READ(intel_sdvo->sdvo_reg);
> }
> @@ -1400,7 +1400,7 @@ static void intel_sdvo_pre_enable(struct intel_encoder *intel_encoder,
> const struct intel_crtc_state *crtc_state,
> const struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(intel_encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(intel_encoder->base.dev);
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> const struct drm_display_mode *adjusted_mode = &crtc_state->base.adjusted_mode;
> const struct intel_sdvo_connector_state *sdvo_state =
> @@ -1481,13 +1481,13 @@ static void intel_sdvo_pre_enable(struct intel_encoder *intel_encoder,
> return;
>
> /* Set the SDVO control regs. */
> - if (INTEL_GEN(dev_priv) >= 4) {
> + if (INTEL_GEN(i915) >= 4) {
> /* The real mode polarity is set by the SDVO commands, using
> * struct intel_sdvo_dtd. */
> sdvox = SDVO_VSYNC_ACTIVE_HIGH | SDVO_HSYNC_ACTIVE_HIGH;
> - if (!HAS_PCH_SPLIT(dev_priv) && crtc_state->limited_color_range)
> + if (!HAS_PCH_SPLIT(i915) && crtc_state->limited_color_range)
> sdvox |= HDMI_COLOR_RANGE_16_235;
> - if (INTEL_GEN(dev_priv) < 5)
> + if (INTEL_GEN(i915) < 5)
> sdvox |= SDVO_BORDER_ENABLE;
> } else {
> sdvox = I915_READ(intel_sdvo->sdvo_reg);
> @@ -1498,15 +1498,15 @@ static void intel_sdvo_pre_enable(struct intel_encoder *intel_encoder,
> sdvox |= (9 << 19) | SDVO_BORDER_ENABLE;
> }
>
> - if (HAS_PCH_CPT(dev_priv))
> + if (HAS_PCH_CPT(i915))
> sdvox |= SDVO_PIPE_SEL_CPT(crtc->pipe);
> else
> sdvox |= SDVO_PIPE_SEL(crtc->pipe);
>
> - if (INTEL_GEN(dev_priv) >= 4) {
> + if (INTEL_GEN(i915) >= 4) {
> /* done in crtc_mode_set as the dpll_md reg must be written early */
> - } else if (IS_I945G(dev_priv) || IS_I945GM(dev_priv) ||
> - IS_G33(dev_priv) || IS_PINEVIEW(dev_priv)) {
> + } else if (IS_I945G(i915) || IS_I945GM(i915) ||
> + IS_G33(i915) || IS_PINEVIEW(i915)) {
> /* done in crtc_mode_set as it lives inside the dpll register */
> } else {
> sdvox |= (crtc_state->pixel_multiplier - 1)
> @@ -1514,7 +1514,7 @@ static void intel_sdvo_pre_enable(struct intel_encoder *intel_encoder,
> }
>
> if (input_dtd.part2.sdvo_flags & SDVO_NEED_TO_STALL &&
> - INTEL_GEN(dev_priv) < 5)
> + INTEL_GEN(i915) < 5)
> sdvox |= SDVO_STALL_SELECT;
> intel_sdvo_write_sdvox(intel_sdvo, sdvox);
> }
> @@ -1531,7 +1531,7 @@ static bool intel_sdvo_connector_get_hw_state(struct intel_connector *connector)
> return active_outputs & intel_sdvo_connector->output_flag;
> }
>
> -bool intel_sdvo_port_enabled(struct drm_i915_private *dev_priv,
> +bool intel_sdvo_port_enabled(struct drm_i915_private *i915,
> i915_reg_t sdvo_reg, enum pipe *pipe)
> {
> u32 val;
> @@ -1539,9 +1539,9 @@ bool intel_sdvo_port_enabled(struct drm_i915_private *dev_priv,
> val = I915_READ(sdvo_reg);
>
> /* asserts want to know the pipe even if the port is disabled */
> - if (HAS_PCH_CPT(dev_priv))
> + if (HAS_PCH_CPT(i915))
> *pipe = (val & SDVO_PIPE_SEL_MASK_CPT) >> SDVO_PIPE_SEL_SHIFT_CPT;
> - else if (IS_CHERRYVIEW(dev_priv))
> + else if (IS_CHERRYVIEW(i915))
> *pipe = (val & SDVO_PIPE_SEL_MASK_CHV) >> SDVO_PIPE_SEL_SHIFT_CHV;
> else
> *pipe = (val & SDVO_PIPE_SEL_MASK) >> SDVO_PIPE_SEL_SHIFT;
> @@ -1552,14 +1552,14 @@ bool intel_sdvo_port_enabled(struct drm_i915_private *dev_priv,
> static bool intel_sdvo_get_hw_state(struct intel_encoder *encoder,
> enum pipe *pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_sdvo *intel_sdvo = to_sdvo(encoder);
> u16 active_outputs = 0;
> bool ret;
>
> intel_sdvo_get_active_outputs(intel_sdvo, &active_outputs);
>
> - ret = intel_sdvo_port_enabled(dev_priv, intel_sdvo->sdvo_reg, pipe);
> + ret = intel_sdvo_port_enabled(i915, intel_sdvo->sdvo_reg, pipe);
>
> return ret || active_outputs;
> }
> @@ -1568,7 +1568,7 @@ static void intel_sdvo_get_config(struct intel_encoder *encoder,
> struct intel_crtc_state *pipe_config)
> {
> struct drm_device *dev = encoder->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_sdvo *intel_sdvo = to_sdvo(encoder);
> struct intel_sdvo_dtd dtd;
> int encoder_pixel_multiplier = 0;
> @@ -1610,7 +1610,7 @@ static void intel_sdvo_get_config(struct intel_encoder *encoder,
> * encoder->get_config we so already have a valid pixel multplier on all
> * other platfroms.
> */
> - if (IS_I915G(dev_priv) || IS_I915GM(dev_priv)) {
> + if (IS_I915G(i915) || IS_I915GM(i915)) {
> pipe_config->pixel_multiplier =
> ((sdvox & SDVO_PORT_MULTIPLY_MASK)
> >> SDVO_PORT_MULTIPLY_SHIFT) + 1;
> @@ -1693,7 +1693,7 @@ static void intel_disable_sdvo(struct intel_encoder *encoder,
> const struct intel_crtc_state *old_crtc_state,
> const struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_sdvo *intel_sdvo = to_sdvo(encoder);
> struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc);
> u32 temp;
> @@ -1716,13 +1716,13 @@ static void intel_disable_sdvo(struct intel_encoder *encoder,
> * to transcoder A after disabling it to allow the
> * matching DP port to be enabled on transcoder A.
> */
> - if (HAS_PCH_IBX(dev_priv) && crtc->pipe == PIPE_B) {
> + if (HAS_PCH_IBX(i915) && crtc->pipe == PIPE_B) {
> /*
> * We get CPU/PCH FIFO underruns on the other pipe when
> * doing the workaround. Sweep them under the rug.
> */
> - intel_set_cpu_fifo_underrun_reporting(dev_priv, PIPE_A, false);
> - intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, false);
> + intel_set_cpu_fifo_underrun_reporting(i915, PIPE_A, false);
> + intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, false);
>
> temp &= ~SDVO_PIPE_SEL_MASK;
> temp |= SDVO_ENABLE | SDVO_PIPE_SEL(PIPE_A);
> @@ -1731,9 +1731,9 @@ static void intel_disable_sdvo(struct intel_encoder *encoder,
> temp &= ~SDVO_ENABLE;
> intel_sdvo_write_sdvox(intel_sdvo, temp);
>
> - intel_wait_for_vblank_if_active(dev_priv, PIPE_A);
> - intel_set_cpu_fifo_underrun_reporting(dev_priv, PIPE_A, true);
> - intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, true);
> + intel_wait_for_vblank_if_active(i915, PIPE_A);
> + intel_set_cpu_fifo_underrun_reporting(i915, PIPE_A, true);
> + intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, true);
> }
> }
>
> @@ -1755,7 +1755,7 @@ static void intel_enable_sdvo(struct intel_encoder *encoder,
> const struct drm_connector_state *conn_state)
> {
> struct drm_device *dev = encoder->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_sdvo *intel_sdvo = to_sdvo(encoder);
> struct intel_crtc *intel_crtc = to_intel_crtc(pipe_config->base.crtc);
> u32 temp;
> @@ -1768,7 +1768,7 @@ static void intel_enable_sdvo(struct intel_encoder *encoder,
> intel_sdvo_write_sdvox(intel_sdvo, temp);
>
> for (i = 0; i < 2; i++)
> - intel_wait_for_vblank(dev_priv, intel_crtc->pipe);
> + intel_wait_for_vblank(i915, intel_crtc->pipe);
>
> success = intel_sdvo_get_trained_inputs(intel_sdvo, &input1, &input2);
> /*
> @@ -1865,17 +1865,17 @@ static bool intel_sdvo_get_capabilities(struct intel_sdvo *intel_sdvo, struct in
>
> static u16 intel_sdvo_get_hotplug_support(struct intel_sdvo *intel_sdvo)
> {
> - struct drm_i915_private *dev_priv = to_i915(intel_sdvo->base.base.dev);
> + struct drm_i915_private *i915 = to_i915(intel_sdvo->base.base.dev);
> u16 hotplug;
>
> - if (!I915_HAS_HOTPLUG(dev_priv))
> + if (!I915_HAS_HOTPLUG(i915))
> return 0;
>
> /*
> * HW Erratum: SDVO Hotplug is broken on all i945G chips, there's noise
> * on the line.
> */
> - if (IS_I945G(dev_priv) || IS_I945GM(dev_priv))
> + if (IS_I945G(i915) || IS_I945GM(i915))
> return 0;
>
> if (!intel_sdvo_get_value(intel_sdvo, SDVO_CMD_GET_HOT_PLUG_SUPPORT,
> @@ -1919,11 +1919,11 @@ intel_sdvo_get_edid(struct drm_connector *connector)
> static struct edid *
> intel_sdvo_get_analog_edid(struct drm_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->dev);
> + struct drm_i915_private *i915 = to_i915(connector->dev);
>
> return drm_get_edid(connector,
> - intel_gmbus_get_adapter(dev_priv,
> - dev_priv->vbt.crt_ddc_pin));
> + intel_gmbus_get_adapter(i915,
> + i915->vbt.crt_ddc_pin));
> }
>
> static enum drm_connector_status
> @@ -2186,7 +2186,7 @@ static void intel_sdvo_get_tv_modes(struct drm_connector *connector)
> static void intel_sdvo_get_lvds_modes(struct drm_connector *connector)
> {
> struct intel_sdvo *intel_sdvo = intel_attached_sdvo(connector);
> - struct drm_i915_private *dev_priv = to_i915(connector->dev);
> + struct drm_i915_private *i915 = to_i915(connector->dev);
> struct drm_display_mode *newmode;
>
> DRM_DEBUG_KMS("[CONNECTOR:%d:%s]\n",
> @@ -2196,9 +2196,9 @@ static void intel_sdvo_get_lvds_modes(struct drm_connector *connector)
> * Fetch modes from VBT. For SDVO prefer the VBT mode since some
> * SDVO->LVDS transcoders can't cope with the EDID mode.
> */
> - if (dev_priv->vbt.sdvo_lvds_vbt_mode != NULL) {
> + if (i915->vbt.sdvo_lvds_vbt_mode != NULL) {
> newmode = drm_mode_duplicate(connector->dev,
> - dev_priv->vbt.sdvo_lvds_vbt_mode);
> + i915->vbt.sdvo_lvds_vbt_mode);
> if (newmode != NULL) {
> /* Guarantee the mode is preferred */
> newmode->type = (DRM_MODE_TYPE_PREFERRED |
> @@ -2488,15 +2488,15 @@ intel_sdvo_guess_ddc_bus(struct intel_sdvo *sdvo)
> * outputs, then LVDS outputs.
> */
> static void
> -intel_sdvo_select_ddc_bus(struct drm_i915_private *dev_priv,
> +intel_sdvo_select_ddc_bus(struct drm_i915_private *i915,
> struct intel_sdvo *sdvo)
> {
> struct sdvo_device_mapping *mapping;
>
> if (sdvo->port == PORT_B)
> - mapping = &dev_priv->vbt.sdvo_mappings[0];
> + mapping = &i915->vbt.sdvo_mappings[0];
> else
> - mapping = &dev_priv->vbt.sdvo_mappings[1];
> + mapping = &i915->vbt.sdvo_mappings[1];
>
> if (mapping->initialized)
> sdvo->ddc_bus = 1 << ((mapping->ddc_pin & 0xf0) >> 4);
> @@ -2505,24 +2505,24 @@ intel_sdvo_select_ddc_bus(struct drm_i915_private *dev_priv,
> }
>
> static void
> -intel_sdvo_select_i2c_bus(struct drm_i915_private *dev_priv,
> +intel_sdvo_select_i2c_bus(struct drm_i915_private *i915,
> struct intel_sdvo *sdvo)
> {
> struct sdvo_device_mapping *mapping;
> u8 pin;
>
> if (sdvo->port == PORT_B)
> - mapping = &dev_priv->vbt.sdvo_mappings[0];
> + mapping = &i915->vbt.sdvo_mappings[0];
> else
> - mapping = &dev_priv->vbt.sdvo_mappings[1];
> + mapping = &i915->vbt.sdvo_mappings[1];
>
> if (mapping->initialized &&
> - intel_gmbus_is_valid_pin(dev_priv, mapping->i2c_pin))
> + intel_gmbus_is_valid_pin(i915, mapping->i2c_pin))
> pin = mapping->i2c_pin;
> else
> pin = GMBUS_PIN_DPB;
>
> - sdvo->i2c = intel_gmbus_get_adapter(dev_priv, pin);
> + sdvo->i2c = intel_gmbus_get_adapter(i915, pin);
>
> /*
> * With gmbus we should be able to drive sdvo i2c at 2MHz, but somehow
> @@ -2546,17 +2546,17 @@ intel_sdvo_is_hdmi_connector(struct intel_sdvo *intel_sdvo, int device)
> }
>
> static u8
> -intel_sdvo_get_slave_addr(struct drm_i915_private *dev_priv,
> +intel_sdvo_get_slave_addr(struct drm_i915_private *i915,
> struct intel_sdvo *sdvo)
> {
> struct sdvo_device_mapping *my_mapping, *other_mapping;
>
> if (sdvo->port == PORT_B) {
> - my_mapping = &dev_priv->vbt.sdvo_mappings[0];
> - other_mapping = &dev_priv->vbt.sdvo_mappings[1];
> + my_mapping = &i915->vbt.sdvo_mappings[0];
> + other_mapping = &i915->vbt.sdvo_mappings[1];
> } else {
> - my_mapping = &dev_priv->vbt.sdvo_mappings[1];
> - other_mapping = &dev_priv->vbt.sdvo_mappings[0];
> + my_mapping = &i915->vbt.sdvo_mappings[1];
> + other_mapping = &i915->vbt.sdvo_mappings[0];
> }
>
> /* If the BIOS described our SDVO device, take advantage of it. */
> @@ -2616,10 +2616,10 @@ static void
> intel_sdvo_add_hdmi_properties(struct intel_sdvo *intel_sdvo,
> struct intel_sdvo_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.base.dev);
>
> intel_attach_force_audio_property(&connector->base.base);
> - if (INTEL_GEN(dev_priv) >= 4 && IS_MOBILE(dev_priv)) {
> + if (INTEL_GEN(i915) >= 4 && IS_MOBILE(i915)) {
> intel_attach_broadcast_rgb_property(&connector->base.base);
> }
> intel_attach_aspect_ratio_property(&connector->base.base);
> @@ -3180,9 +3180,9 @@ static const struct i2c_lock_operations proxy_lock_ops = {
>
> static bool
> intel_sdvo_init_ddc_proxy(struct intel_sdvo *sdvo,
> - struct drm_i915_private *dev_priv)
> + struct drm_i915_private *i915)
> {
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct pci_dev *pdev = i915->drm.pdev;
>
> sdvo->ddc.owner = THIS_MODULE;
> sdvo->ddc.class = I2C_CLASS_DDC;
> @@ -3195,23 +3195,23 @@ intel_sdvo_init_ddc_proxy(struct intel_sdvo *sdvo,
> return i2c_add_adapter(&sdvo->ddc) == 0;
> }
>
> -static void assert_sdvo_port_valid(const struct drm_i915_private *dev_priv,
> +static void assert_sdvo_port_valid(const struct drm_i915_private *i915,
> enum port port)
> {
> - if (HAS_PCH_SPLIT(dev_priv))
> + if (HAS_PCH_SPLIT(i915))
> WARN_ON(port != PORT_B);
> else
> WARN_ON(port != PORT_B && port != PORT_C);
> }
>
> -bool intel_sdvo_init(struct drm_i915_private *dev_priv,
> +bool intel_sdvo_init(struct drm_i915_private *i915,
> i915_reg_t sdvo_reg, enum port port)
> {
> struct intel_encoder *intel_encoder;
> struct intel_sdvo *intel_sdvo;
> int i;
>
> - assert_sdvo_port_valid(dev_priv, port);
> + assert_sdvo_port_valid(i915, port);
>
> intel_sdvo = kzalloc(sizeof(*intel_sdvo), GFP_KERNEL);
> if (!intel_sdvo)
> @@ -3220,9 +3220,9 @@ bool intel_sdvo_init(struct drm_i915_private *dev_priv,
> intel_sdvo->sdvo_reg = sdvo_reg;
> intel_sdvo->port = port;
> intel_sdvo->slave_addr =
> - intel_sdvo_get_slave_addr(dev_priv, intel_sdvo) >> 1;
> - intel_sdvo_select_i2c_bus(dev_priv, intel_sdvo);
> - if (!intel_sdvo_init_ddc_proxy(intel_sdvo, dev_priv))
> + intel_sdvo_get_slave_addr(i915, intel_sdvo) >> 1;
> + intel_sdvo_select_i2c_bus(i915, intel_sdvo);
> + if (!intel_sdvo_init_ddc_proxy(intel_sdvo, i915))
> goto err_i2c_bus;
>
> /* encoder type will be decided later */
> @@ -3230,7 +3230,7 @@ bool intel_sdvo_init(struct drm_i915_private *dev_priv,
> intel_encoder->type = INTEL_OUTPUT_SDVO;
> intel_encoder->power_domain = POWER_DOMAIN_PORT_OTHER;
> intel_encoder->port = port;
> - drm_encoder_init(&dev_priv->drm, &intel_encoder->base,
> + drm_encoder_init(&i915->drm, &intel_encoder->base,
> &intel_sdvo_enc_funcs, 0,
> "SDVO %c", port_name(port));
>
> @@ -3246,7 +3246,7 @@ bool intel_sdvo_init(struct drm_i915_private *dev_priv,
> }
>
> intel_encoder->compute_config = intel_sdvo_compute_config;
> - if (HAS_PCH_SPLIT(dev_priv)) {
> + if (HAS_PCH_SPLIT(i915)) {
> intel_encoder->disable = pch_disable_sdvo;
> intel_encoder->post_disable = pch_post_disable_sdvo;
> } else {
> @@ -3290,7 +3290,7 @@ bool intel_sdvo_init(struct drm_i915_private *dev_priv,
> */
> intel_sdvo->base.cloneable = 0;
>
> - intel_sdvo_select_ddc_bus(dev_priv, intel_sdvo);
> + intel_sdvo_select_ddc_bus(i915, intel_sdvo);
>
> /* Set the input timing to the screen. Assume always input 0. */
> if (!intel_sdvo_set_target_input(intel_sdvo))
> diff --git a/drivers/gpu/drm/i915/intel_sdvo.h b/drivers/gpu/drm/i915/intel_sdvo.h
> index c9e05bcdd141..1ec6766c5ae1 100644
> --- a/drivers/gpu/drm/i915/intel_sdvo.h
> +++ b/drivers/gpu/drm/i915/intel_sdvo.h
> @@ -15,9 +15,9 @@
> struct drm_i915_private;
> enum pipe;
>
> -bool intel_sdvo_port_enabled(struct drm_i915_private *dev_priv,
> +bool intel_sdvo_port_enabled(struct drm_i915_private *i915,
> i915_reg_t sdvo_reg, enum pipe *pipe);
> -bool intel_sdvo_init(struct drm_i915_private *dev_priv,
> +bool intel_sdvo_init(struct drm_i915_private *i915,
> i915_reg_t reg, enum port port);
>
> #endif /* __INTEL_SDVO_H__ */
> diff --git a/drivers/gpu/drm/i915/intel_sprite.c b/drivers/gpu/drm/i915/intel_sprite.c
> index c180815faabd..73871e5e8e19 100644
> --- a/drivers/gpu/drm/i915/intel_sprite.c
> +++ b/drivers/gpu/drm/i915/intel_sprite.c
> @@ -94,12 +94,12 @@ int intel_usecs_to_scanlines(const struct drm_display_mode *adjusted_mode,
> void intel_pipe_update_start(const struct intel_crtc_state *new_crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(new_crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> const struct drm_display_mode *adjusted_mode = &new_crtc_state->base.adjusted_mode;
> long timeout = msecs_to_jiffies_timeout(1);
> int scanline, min, max, vblank_start;
> wait_queue_head_t *wq = drm_crtc_vblank_waitqueue(&crtc->base);
> - bool need_vlv_dsi_wa = (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) &&
> + bool need_vlv_dsi_wa = (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) &&
> intel_crtc_has_type(new_crtc_state, INTEL_OUTPUT_DSI);
> DEFINE_WAIT(wait);
> u32 psr_status;
> @@ -207,7 +207,7 @@ void intel_pipe_update_end(struct intel_crtc_state *new_crtc_state)
> int scanline_end = intel_get_crtc_scanline(crtc);
> u32 end_vbl_count = intel_crtc_get_vblank_counter(crtc);
> ktime_t end_vbl_time = ktime_get();
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
>
> trace_i915_pipe_update_end(crtc, end_vbl_count, scanline_end);
>
> @@ -227,7 +227,7 @@ void intel_pipe_update_end(struct intel_crtc_state *new_crtc_state)
>
> local_irq_enable();
>
> - if (intel_vgpu_active(dev_priv))
> + if (intel_vgpu_active(i915))
> return;
>
> if (crtc->debug.start_vbl_count &&
> @@ -353,7 +353,7 @@ skl_program_scaler(struct intel_plane *plane,
> const struct intel_crtc_state *crtc_state,
> const struct intel_plane_state *plane_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> enum pipe pipe = plane->pipe;
> int scaler_id = plane_state->scaler_id;
> const struct intel_scaler *scaler =
> @@ -375,7 +375,7 @@ skl_program_scaler(struct intel_plane *plane,
>
> /* TODO: handle sub-pixel coordinates */
> if (is_planar_yuv_format(plane_state->base.fb->format->format) &&
> - !icl_is_hdr_plane(dev_priv, plane->id)) {
> + !icl_is_hdr_plane(i915, plane->id)) {
> y_hphase = skl_scaler_calc_phase(1, hscale, false);
> y_vphase = skl_scaler_calc_phase(1, vscale, false);
>
> @@ -415,7 +415,7 @@ icl_program_input_csc(struct intel_plane *plane,
> const struct intel_crtc_state *crtc_state,
> const struct intel_plane_state *plane_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> enum pipe pipe = plane->pipe;
> enum plane_id plane_id = plane->id;
>
> @@ -507,7 +507,7 @@ skl_program_plane(struct intel_plane *plane,
> const struct intel_plane_state *plane_state,
> int color_plane, bool slave, u32 plane_ctl)
> {
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> enum plane_id plane_id = plane->id;
> enum pipe pipe = plane->pipe;
> const struct drm_intel_sprite_colorkey *key = &plane_state->ckey;
> @@ -529,7 +529,7 @@ skl_program_plane(struct intel_plane *plane,
>
> plane_ctl |= skl_plane_ctl_crtc(crtc_state);
>
> - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> plane_color_ctl = plane_state->color_ctl |
> glk_plane_color_ctl_crtc(crtc_state);
>
> @@ -549,7 +549,7 @@ skl_program_plane(struct intel_plane *plane,
> crtc_y = 0;
> }
>
> - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> + spin_lock_irqsave(&i915->uncore.lock, irqflags);
>
> I915_WRITE_FW(PLANE_STRIDE(pipe, plane_id), stride);
> I915_WRITE_FW(PLANE_POS(pipe, plane_id), (crtc_y << 16) | crtc_x);
> @@ -557,7 +557,7 @@ skl_program_plane(struct intel_plane *plane,
> I915_WRITE_FW(PLANE_AUX_DIST(pipe, plane_id),
> (plane_state->color_plane[1].offset - surf_addr) | aux_stride);
>
> - if (icl_is_hdr_plane(dev_priv, plane_id)) {
> + if (icl_is_hdr_plane(i915, plane_id)) {
> u32 cus_ctl = 0;
>
> if (linked) {
> @@ -578,10 +578,10 @@ skl_program_plane(struct intel_plane *plane,
> I915_WRITE_FW(PLANE_CUS_CTL(pipe, plane_id), cus_ctl);
> }
>
> - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> I915_WRITE_FW(PLANE_COLOR_CTL(pipe, plane_id), plane_color_ctl);
>
> - if (fb->format->is_yuv && icl_is_hdr_plane(dev_priv, plane_id))
> + if (fb->format->is_yuv && icl_is_hdr_plane(i915, plane_id))
> icl_program_input_csc(plane, crtc_state, plane_state);
>
> skl_write_plane_wm(plane, crtc_state);
> @@ -592,7 +592,7 @@ skl_program_plane(struct intel_plane *plane,
>
> I915_WRITE_FW(PLANE_OFFSET(pipe, plane_id), (y << 16) | x);
>
> - if (INTEL_GEN(dev_priv) < 11)
> + if (INTEL_GEN(i915) < 11)
> I915_WRITE_FW(PLANE_AUX_OFFSET(pipe, plane_id),
> (plane_state->color_plane[1].y << 16) |
> plane_state->color_plane[1].x);
> @@ -609,7 +609,7 @@ skl_program_plane(struct intel_plane *plane,
> if (!slave && plane_state->scaler_id >= 0)
> skl_program_scaler(plane, crtc_state, plane_state);
>
> - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> + spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
> }
>
> static void
> @@ -641,14 +641,14 @@ static void
> skl_disable_plane(struct intel_plane *plane,
> const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> enum plane_id plane_id = plane->id;
> enum pipe pipe = plane->pipe;
> unsigned long irqflags;
>
> - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> + spin_lock_irqsave(&i915->uncore.lock, irqflags);
>
> - if (icl_is_hdr_plane(dev_priv, plane_id))
> + if (icl_is_hdr_plane(i915, plane_id))
> I915_WRITE_FW(PLANE_CUS_CTL(pipe, plane_id), 0);
>
> skl_write_plane_wm(plane, crtc_state);
> @@ -656,21 +656,21 @@ skl_disable_plane(struct intel_plane *plane,
> I915_WRITE_FW(PLANE_CTL(pipe, plane_id), 0);
> I915_WRITE_FW(PLANE_SURF(pipe, plane_id), 0);
>
> - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> + spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
> }
>
> static bool
> skl_plane_get_hw_state(struct intel_plane *plane,
> enum pipe *pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> enum intel_display_power_domain power_domain;
> enum plane_id plane_id = plane->id;
> intel_wakeref_t wakeref;
> bool ret;
>
> power_domain = POWER_DOMAIN_PIPE(plane->pipe);
> - wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain);
> + wakeref = intel_display_power_get_if_enabled(i915, power_domain);
> if (!wakeref)
> return false;
>
> @@ -678,7 +678,7 @@ skl_plane_get_hw_state(struct intel_plane *plane,
>
> *pipe = plane->pipe;
>
> - intel_display_power_put(dev_priv, power_domain, wakeref);
> + intel_display_power_put(i915, power_domain, wakeref);
>
> return ret;
> }
> @@ -687,7 +687,7 @@ static void
> chv_update_csc(const struct intel_plane_state *plane_state)
> {
> struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> const struct drm_framebuffer *fb = plane_state->base.fb;
> enum plane_id plane_id = plane->id;
> /*
> @@ -746,7 +746,7 @@ static void
> vlv_update_clrc(const struct intel_plane_state *plane_state)
> {
> struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> const struct drm_framebuffer *fb = plane_state->base.fb;
> enum pipe pipe = plane->pipe;
> enum plane_id plane_id = plane->id;
> @@ -863,7 +863,7 @@ vlv_update_plane(struct intel_plane *plane,
> const struct intel_crtc_state *crtc_state,
> const struct intel_plane_state *plane_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> enum pipe pipe = plane->pipe;
> enum plane_id plane_id = plane->id;
> u32 sprsurf_offset = plane_state->color_plane[0].offset;
> @@ -886,7 +886,7 @@ vlv_update_plane(struct intel_plane *plane,
>
> linear_offset = intel_fb_xy_to_linear(x, y, plane_state, 0);
>
> - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> + spin_lock_irqsave(&i915->uncore.lock, irqflags);
>
> I915_WRITE_FW(SPSTRIDE(pipe, plane_id),
> plane_state->color_plane[0].stride);
> @@ -894,7 +894,7 @@ vlv_update_plane(struct intel_plane *plane,
> I915_WRITE_FW(SPSIZE(pipe, plane_id), (crtc_h << 16) | crtc_w);
> I915_WRITE_FW(SPCONSTALPHA(pipe, plane_id), 0);
>
> - if (IS_CHERRYVIEW(dev_priv) && pipe == PIPE_B)
> + if (IS_CHERRYVIEW(i915) && pipe == PIPE_B)
> chv_update_csc(plane_state);
>
> if (key->flags) {
> @@ -917,38 +917,38 @@ vlv_update_plane(struct intel_plane *plane,
>
> vlv_update_clrc(plane_state);
>
> - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> + spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
> }
>
> static void
> vlv_disable_plane(struct intel_plane *plane,
> const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> enum pipe pipe = plane->pipe;
> enum plane_id plane_id = plane->id;
> unsigned long irqflags;
>
> - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> + spin_lock_irqsave(&i915->uncore.lock, irqflags);
>
> I915_WRITE_FW(SPCNTR(pipe, plane_id), 0);
> I915_WRITE_FW(SPSURF(pipe, plane_id), 0);
>
> - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> + spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
> }
>
> static bool
> vlv_plane_get_hw_state(struct intel_plane *plane,
> enum pipe *pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> enum intel_display_power_domain power_domain;
> enum plane_id plane_id = plane->id;
> intel_wakeref_t wakeref;
> bool ret;
>
> power_domain = POWER_DOMAIN_PIPE(plane->pipe);
> - wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain);
> + wakeref = intel_display_power_get_if_enabled(i915, power_domain);
> if (!wakeref)
> return false;
>
> @@ -956,7 +956,7 @@ vlv_plane_get_hw_state(struct intel_plane *plane,
>
> *pipe = plane->pipe;
>
> - intel_display_power_put(dev_priv, power_domain, wakeref);
> + intel_display_power_put(i915, power_domain, wakeref);
>
> return ret;
> }
> @@ -977,7 +977,7 @@ static u32 ivb_sprite_ctl_crtc(const struct intel_crtc_state *crtc_state)
> static u32 ivb_sprite_ctl(const struct intel_crtc_state *crtc_state,
> const struct intel_plane_state *plane_state)
> {
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> to_i915(plane_state->base.plane->dev);
> const struct drm_framebuffer *fb = plane_state->base.fb;
> unsigned int rotation = plane_state->base.rotation;
> @@ -986,7 +986,7 @@ static u32 ivb_sprite_ctl(const struct intel_crtc_state *crtc_state,
>
> sprctl = SPRITE_ENABLE;
>
> - if (IS_IVYBRIDGE(dev_priv))
> + if (IS_IVYBRIDGE(i915))
> sprctl |= SPRITE_TRICKLE_FEED_DISABLE;
>
> switch (fb->format->format) {
> @@ -1038,7 +1038,7 @@ ivb_update_plane(struct intel_plane *plane,
> const struct intel_crtc_state *crtc_state,
> const struct intel_plane_state *plane_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> enum pipe pipe = plane->pipe;
> u32 sprsurf_offset = plane_state->color_plane[0].offset;
> u32 linear_offset;
> @@ -1067,12 +1067,12 @@ ivb_update_plane(struct intel_plane *plane,
>
> linear_offset = intel_fb_xy_to_linear(x, y, plane_state, 0);
>
> - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> + spin_lock_irqsave(&i915->uncore.lock, irqflags);
>
> I915_WRITE_FW(SPRSTRIDE(pipe), plane_state->color_plane[0].stride);
> I915_WRITE_FW(SPRPOS(pipe), (crtc_y << 16) | crtc_x);
> I915_WRITE_FW(SPRSIZE(pipe), (crtc_h << 16) | crtc_w);
> - if (IS_IVYBRIDGE(dev_priv))
> + if (IS_IVYBRIDGE(i915))
> I915_WRITE_FW(SPRSCALE(pipe), sprscale);
>
> if (key->flags) {
> @@ -1083,7 +1083,7 @@ ivb_update_plane(struct intel_plane *plane,
>
> /* HSW consolidates SPRTILEOFF and SPRLINOFF into a single SPROFFSET
> * register */
> - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) {
> + if (IS_HASWELL(i915) || IS_BROADWELL(i915)) {
> I915_WRITE_FW(SPROFFSET(pipe), (y << 16) | x);
> } else {
> I915_WRITE_FW(SPRLINOFF(pipe), linear_offset);
> @@ -1099,39 +1099,39 @@ ivb_update_plane(struct intel_plane *plane,
> I915_WRITE_FW(SPRSURF(pipe),
> intel_plane_ggtt_offset(plane_state) + sprsurf_offset);
>
> - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> + spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
> }
>
> static void
> ivb_disable_plane(struct intel_plane *plane,
> const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> enum pipe pipe = plane->pipe;
> unsigned long irqflags;
>
> - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> + spin_lock_irqsave(&i915->uncore.lock, irqflags);
>
> I915_WRITE_FW(SPRCTL(pipe), 0);
> /* Disable the scaler */
> - if (IS_IVYBRIDGE(dev_priv))
> + if (IS_IVYBRIDGE(i915))
> I915_WRITE_FW(SPRSCALE(pipe), 0);
> I915_WRITE_FW(SPRSURF(pipe), 0);
>
> - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> + spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
> }
>
> static bool
> ivb_plane_get_hw_state(struct intel_plane *plane,
> enum pipe *pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> enum intel_display_power_domain power_domain;
> intel_wakeref_t wakeref;
> bool ret;
>
> power_domain = POWER_DOMAIN_PIPE(plane->pipe);
> - wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain);
> + wakeref = intel_display_power_get_if_enabled(i915, power_domain);
> if (!wakeref)
> return false;
>
> @@ -1139,7 +1139,7 @@ ivb_plane_get_hw_state(struct intel_plane *plane,
>
> *pipe = plane->pipe;
>
> - intel_display_power_put(dev_priv, power_domain, wakeref);
> + intel_display_power_put(i915, power_domain, wakeref);
>
> return ret;
> }
> @@ -1168,7 +1168,7 @@ static u32 g4x_sprite_ctl_crtc(const struct intel_crtc_state *crtc_state)
> static u32 g4x_sprite_ctl(const struct intel_crtc_state *crtc_state,
> const struct intel_plane_state *plane_state)
> {
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> to_i915(plane_state->base.plane->dev);
> const struct drm_framebuffer *fb = plane_state->base.fb;
> unsigned int rotation = plane_state->base.rotation;
> @@ -1177,7 +1177,7 @@ static u32 g4x_sprite_ctl(const struct intel_crtc_state *crtc_state,
>
> dvscntr = DVS_ENABLE;
>
> - if (IS_GEN(dev_priv, 6))
> + if (IS_GEN(i915, 6))
> dvscntr |= DVS_TRICKLE_FEED_DISABLE;
>
> switch (fb->format->format) {
> @@ -1229,7 +1229,7 @@ g4x_update_plane(struct intel_plane *plane,
> const struct intel_crtc_state *crtc_state,
> const struct intel_plane_state *plane_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> enum pipe pipe = plane->pipe;
> u32 dvssurf_offset = plane_state->color_plane[0].offset;
> u32 linear_offset;
> @@ -1258,7 +1258,7 @@ g4x_update_plane(struct intel_plane *plane,
>
> linear_offset = intel_fb_xy_to_linear(x, y, plane_state, 0);
>
> - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> + spin_lock_irqsave(&i915->uncore.lock, irqflags);
>
> I915_WRITE_FW(DVSSTRIDE(pipe), plane_state->color_plane[0].stride);
> I915_WRITE_FW(DVSPOS(pipe), (crtc_y << 16) | crtc_x);
> @@ -1283,38 +1283,38 @@ g4x_update_plane(struct intel_plane *plane,
> I915_WRITE_FW(DVSSURF(pipe),
> intel_plane_ggtt_offset(plane_state) + dvssurf_offset);
>
> - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> + spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
> }
>
> static void
> g4x_disable_plane(struct intel_plane *plane,
> const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> enum pipe pipe = plane->pipe;
> unsigned long irqflags;
>
> - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> + spin_lock_irqsave(&i915->uncore.lock, irqflags);
>
> I915_WRITE_FW(DVSCNTR(pipe), 0);
> /* Disable the scaler */
> I915_WRITE_FW(DVSSCALE(pipe), 0);
> I915_WRITE_FW(DVSSURF(pipe), 0);
>
> - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> + spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
> }
>
> static bool
> g4x_plane_get_hw_state(struct intel_plane *plane,
> enum pipe *pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> enum intel_display_power_domain power_domain;
> intel_wakeref_t wakeref;
> bool ret;
>
> power_domain = POWER_DOMAIN_PIPE(plane->pipe);
> - wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain);
> + wakeref = intel_display_power_get_if_enabled(i915, power_domain);
> if (!wakeref)
> return false;
>
> @@ -1322,7 +1322,7 @@ g4x_plane_get_hw_state(struct intel_plane *plane,
>
> *pipe = plane->pipe;
>
> - intel_display_power_put(dev_priv, power_domain, wakeref);
> + intel_display_power_put(i915, power_domain, wakeref);
>
> return ret;
> }
> @@ -1406,16 +1406,16 @@ g4x_sprite_check(struct intel_crtc_state *crtc_state,
> struct intel_plane_state *plane_state)
> {
> struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> int min_scale = DRM_PLANE_HELPER_NO_SCALING;
> int max_scale = DRM_PLANE_HELPER_NO_SCALING;
> int ret;
>
> if (intel_fb_scalable(plane_state->base.fb)) {
> - if (INTEL_GEN(dev_priv) < 7) {
> + if (INTEL_GEN(i915) < 7) {
> min_scale = 1;
> max_scale = 16 << 16;
> - } else if (IS_IVYBRIDGE(dev_priv)) {
> + } else if (IS_IVYBRIDGE(i915)) {
> min_scale = 1;
> max_scale = 2 << 16;
> }
> @@ -1443,7 +1443,7 @@ g4x_sprite_check(struct intel_crtc_state *crtc_state,
> if (ret)
> return ret;
>
> - if (INTEL_GEN(dev_priv) >= 7)
> + if (INTEL_GEN(i915) >= 7)
> plane_state->ctl = ivb_sprite_ctl(crtc_state, plane_state);
> else
> plane_state->ctl = g4x_sprite_ctl(crtc_state, plane_state);
> @@ -1454,11 +1454,11 @@ g4x_sprite_check(struct intel_crtc_state *crtc_state,
> int chv_plane_check_rotation(const struct intel_plane_state *plane_state)
> {
> struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> unsigned int rotation = plane_state->base.rotation;
>
> /* CHV ignores the mirror bit when the rotate bit is set :( */
> - if (IS_CHERRYVIEW(dev_priv) &&
> + if (IS_CHERRYVIEW(i915) &&
> rotation & DRM_MODE_ROTATE_180 &&
> rotation & DRM_MODE_REFLECT_X) {
> DRM_DEBUG_KMS("Cannot rotate and reflect at the same time\n");
> @@ -1506,7 +1506,7 @@ static int skl_plane_check_fb(const struct intel_crtc_state *crtc_state,
> const struct intel_plane_state *plane_state)
> {
> struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> const struct drm_framebuffer *fb = plane_state->base.fb;
> unsigned int rotation = plane_state->base.rotation;
> struct drm_format_name_buf format_name;
> @@ -1540,7 +1540,7 @@ static int skl_plane_check_fb(const struct intel_crtc_state *crtc_state,
> */
> switch (fb->format->format) {
> case DRM_FORMAT_RGB565:
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> break;
> /* fall through */
> case DRM_FORMAT_C8:
> @@ -1579,7 +1579,7 @@ static int skl_plane_check_fb(const struct intel_crtc_state *crtc_state,
> static int skl_plane_check_dst_coordinates(const struct intel_crtc_state *crtc_state,
> const struct intel_plane_state *plane_state)
> {
> - struct drm_i915_private *dev_priv =
> + struct drm_i915_private *i915 =
> to_i915(plane_state->base.plane->dev);
> int crtc_x = plane_state->base.dst.x1;
> int crtc_w = drm_rect_width(&plane_state->base.dst);
> @@ -1594,7 +1594,7 @@ static int skl_plane_check_dst_coordinates(const struct intel_crtc_state *crtc_s
> * than the cursor ending less than 4 pixels from the left edge of the
> * screen may cause FIFO underflow and display corruption.
> */
> - if ((IS_GEMINILAKE(dev_priv) || IS_CANNONLAKE(dev_priv)) &&
> + if ((IS_GEMINILAKE(i915) || IS_CANNONLAKE(i915)) &&
> (crtc_x + crtc_w < 4 || crtc_x > pipe_src_w - 4)) {
> DRM_DEBUG_KMS("requested plane X %s position %d invalid (valid range %d-%d)\n",
> crtc_x + crtc_w < 4 ? "end" : "start",
> @@ -1627,7 +1627,7 @@ static int skl_plane_check(struct intel_crtc_state *crtc_state,
> struct intel_plane_state *plane_state)
> {
> struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> const struct drm_framebuffer *fb = plane_state->base.fb;
> int min_scale = DRM_PLANE_HELPER_NO_SCALING;
> int max_scale = DRM_PLANE_HELPER_NO_SCALING;
> @@ -1675,23 +1675,23 @@ static int skl_plane_check(struct intel_crtc_state *crtc_state,
>
> plane_state->ctl = skl_plane_ctl(crtc_state, plane_state);
>
> - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> plane_state->color_ctl = glk_plane_color_ctl(crtc_state,
> plane_state);
>
> return 0;
> }
>
> -static bool has_dst_key_in_primary_plane(struct drm_i915_private *dev_priv)
> +static bool has_dst_key_in_primary_plane(struct drm_i915_private *i915)
> {
> - return INTEL_GEN(dev_priv) >= 9;
> + return INTEL_GEN(i915) >= 9;
> }
>
> static void intel_plane_set_ckey(struct intel_plane_state *plane_state,
> const struct drm_intel_sprite_colorkey *set)
> {
> struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> - struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> + struct drm_i915_private *i915 = to_i915(plane->base.dev);
> struct drm_intel_sprite_colorkey *key = &plane_state->ckey;
>
> *key = *set;
> @@ -1708,7 +1708,7 @@ static void intel_plane_set_ckey(struct intel_plane_state *plane_state,
> * On SKL+ we want dst key enabled on
> * the primary and not on the sprite.
> */
> - if (INTEL_GEN(dev_priv) >= 9 && plane->id != PLANE_PRIMARY &&
> + if (INTEL_GEN(i915) >= 9 && plane->id != PLANE_PRIMARY &&
> set->flags & I915_SET_COLORKEY_DESTINATION)
> key->flags = 0;
> }
> @@ -1716,7 +1716,7 @@ static void intel_plane_set_ckey(struct intel_plane_state *plane_state,
> int intel_sprite_set_colorkey_ioctl(struct drm_device *dev, void *data,
> struct drm_file *file_priv)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct drm_intel_sprite_colorkey *set = data;
> struct drm_plane *plane;
> struct drm_plane_state *plane_state;
> @@ -1734,7 +1734,7 @@ int intel_sprite_set_colorkey_ioctl(struct drm_device *dev, void *data,
> if ((set->flags & (I915_SET_COLORKEY_DESTINATION | I915_SET_COLORKEY_SOURCE)) == (I915_SET_COLORKEY_DESTINATION | I915_SET_COLORKEY_SOURCE))
> return -EINVAL;
>
> - if ((IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) &&
> + if ((IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) &&
> set->flags & I915_SET_COLORKEY_DESTINATION)
> return -EINVAL;
>
> @@ -1747,7 +1747,7 @@ int intel_sprite_set_colorkey_ioctl(struct drm_device *dev, void *data,
> * Also multiple planes can't do destination keying on the same
> * pipe simultaneously.
> */
> - if (INTEL_GEN(dev_priv) >= 9 &&
> + if (INTEL_GEN(i915) >= 9 &&
> to_intel_plane(plane)->id >= PLANE_SPRITE1 &&
> set->flags & I915_SET_COLORKEY_DESTINATION)
> return -EINVAL;
> @@ -1771,9 +1771,9 @@ int intel_sprite_set_colorkey_ioctl(struct drm_device *dev, void *data,
> * On some platforms we have to configure
> * the dst colorkey on the primary plane.
> */
> - if (!ret && has_dst_key_in_primary_plane(dev_priv)) {
> + if (!ret && has_dst_key_in_primary_plane(i915)) {
> struct intel_crtc *crtc =
> - intel_get_crtc_for_pipe(dev_priv,
> + intel_get_crtc_for_pipe(i915,
> to_intel_plane(plane)->pipe);
>
> plane_state = drm_atomic_get_plane_state(state,
> @@ -2197,26 +2197,26 @@ static const struct drm_plane_funcs skl_plane_funcs = {
> .format_mod_supported = skl_plane_format_mod_supported,
> };
>
> -static bool skl_plane_has_fbc(struct drm_i915_private *dev_priv,
> +static bool skl_plane_has_fbc(struct drm_i915_private *i915,
> enum pipe pipe, enum plane_id plane_id)
> {
> - if (!HAS_FBC(dev_priv))
> + if (!HAS_FBC(i915))
> return false;
>
> return pipe == PIPE_A && plane_id == PLANE_PRIMARY;
> }
>
> -static bool skl_plane_has_planar(struct drm_i915_private *dev_priv,
> +static bool skl_plane_has_planar(struct drm_i915_private *i915,
> enum pipe pipe, enum plane_id plane_id)
> {
> - if (INTEL_GEN(dev_priv) >= 11)
> + if (INTEL_GEN(i915) >= 11)
> return plane_id <= PLANE_SPRITE3;
>
> /* Display WA #0870: skl, bxt */
> - if (IS_SKYLAKE(dev_priv) || IS_BROXTON(dev_priv))
> + if (IS_SKYLAKE(i915) || IS_BROXTON(i915))
> return false;
>
> - if (IS_GEN(dev_priv, 9) && !IS_GEMINILAKE(dev_priv) && pipe == PIPE_C)
> + if (IS_GEN(i915, 9) && !IS_GEMINILAKE(i915) && pipe == PIPE_C)
> return false;
>
> if (plane_id != PLANE_PRIMARY && plane_id != PLANE_SPRITE0)
> @@ -2225,16 +2225,16 @@ static bool skl_plane_has_planar(struct drm_i915_private *dev_priv,
> return true;
> }
>
> -static bool skl_plane_has_ccs(struct drm_i915_private *dev_priv,
> +static bool skl_plane_has_ccs(struct drm_i915_private *i915,
> enum pipe pipe, enum plane_id plane_id)
> {
> if (plane_id == PLANE_CURSOR)
> return false;
>
> - if (INTEL_GEN(dev_priv) >= 10)
> + if (INTEL_GEN(i915) >= 10)
> return true;
>
> - if (IS_GEMINILAKE(dev_priv))
> + if (IS_GEMINILAKE(i915))
> return pipe != PIPE_C;
>
> return pipe != PIPE_C &&
> @@ -2243,7 +2243,7 @@ static bool skl_plane_has_ccs(struct drm_i915_private *dev_priv,
> }
>
> struct intel_plane *
> -skl_universal_plane_create(struct drm_i915_private *dev_priv,
> +skl_universal_plane_create(struct drm_i915_private *i915,
> enum pipe pipe, enum plane_id plane_id)
> {
> struct intel_plane *plane;
> @@ -2263,9 +2263,9 @@ skl_universal_plane_create(struct drm_i915_private *dev_priv,
> plane->id = plane_id;
> plane->frontbuffer_bit = INTEL_FRONTBUFFER(pipe, plane_id);
>
> - plane->has_fbc = skl_plane_has_fbc(dev_priv, pipe, plane_id);
> + plane->has_fbc = skl_plane_has_fbc(i915, pipe, plane_id);
> if (plane->has_fbc) {
> - struct intel_fbc *fbc = &dev_priv->fbc;
> + struct intel_fbc *fbc = &i915->fbc;
>
> fbc->possible_framebuffer_bits |= plane->frontbuffer_bit;
> }
> @@ -2278,24 +2278,24 @@ skl_universal_plane_create(struct drm_i915_private *dev_priv,
> if (icl_is_nv12_y_plane(plane_id))
> plane->update_slave = icl_update_slave;
>
> - if (skl_plane_has_planar(dev_priv, pipe, plane_id)) {
> - if (icl_is_hdr_plane(dev_priv, plane_id)) {
> + if (skl_plane_has_planar(i915, pipe, plane_id)) {
> + if (icl_is_hdr_plane(i915, plane_id)) {
> formats = icl_hdr_planar_formats;
> num_formats = ARRAY_SIZE(icl_hdr_planar_formats);
> - } else if (INTEL_GEN(dev_priv) >= 11) {
> + } else if (INTEL_GEN(i915) >= 11) {
> formats = icl_planar_formats;
> num_formats = ARRAY_SIZE(icl_planar_formats);
> - } else if (INTEL_GEN(dev_priv) == 10 || IS_GEMINILAKE(dev_priv)) {
> + } else if (INTEL_GEN(i915) == 10 || IS_GEMINILAKE(i915)) {
> formats = glk_planar_formats;
> num_formats = ARRAY_SIZE(glk_planar_formats);
> } else {
> formats = skl_planar_formats;
> num_formats = ARRAY_SIZE(skl_planar_formats);
> }
> - } else if (icl_is_hdr_plane(dev_priv, plane_id)) {
> + } else if (icl_is_hdr_plane(i915, plane_id)) {
> formats = icl_hdr_plane_formats;
> num_formats = ARRAY_SIZE(icl_hdr_plane_formats);
> - } else if (INTEL_GEN(dev_priv) >= 11) {
> + } else if (INTEL_GEN(i915) >= 11) {
> formats = icl_plane_formats;
> num_formats = ARRAY_SIZE(icl_plane_formats);
> } else {
> @@ -2303,7 +2303,7 @@ skl_universal_plane_create(struct drm_i915_private *dev_priv,
> num_formats = ARRAY_SIZE(skl_plane_formats);
> }
>
> - plane->has_ccs = skl_plane_has_ccs(dev_priv, pipe, plane_id);
> + plane->has_ccs = skl_plane_has_ccs(i915, pipe, plane_id);
> if (plane->has_ccs)
> modifiers = skl_plane_format_modifiers_ccs;
> else
> @@ -2316,7 +2316,7 @@ skl_universal_plane_create(struct drm_i915_private *dev_priv,
>
> possible_crtcs = BIT(pipe);
>
> - ret = drm_universal_plane_init(&dev_priv->drm, &plane->base,
> + ret = drm_universal_plane_init(&i915->drm, &plane->base,
> possible_crtcs, &skl_plane_funcs,
> formats, num_formats, modifiers,
> plane_type,
> @@ -2329,7 +2329,7 @@ skl_universal_plane_create(struct drm_i915_private *dev_priv,
> DRM_MODE_ROTATE_0 | DRM_MODE_ROTATE_90 |
> DRM_MODE_ROTATE_180 | DRM_MODE_ROTATE_270;
>
> - if (INTEL_GEN(dev_priv) >= 10)
> + if (INTEL_GEN(i915) >= 10)
> supported_rotations |= DRM_MODE_REFLECT_X;
>
> drm_plane_create_rotation_property(&plane->base,
> @@ -2361,7 +2361,7 @@ skl_universal_plane_create(struct drm_i915_private *dev_priv,
> }
>
> struct intel_plane *
> -intel_sprite_plane_create(struct drm_i915_private *dev_priv,
> +intel_sprite_plane_create(struct drm_i915_private *i915,
> enum pipe pipe, int sprite)
> {
> struct intel_plane *plane;
> @@ -2373,15 +2373,15 @@ intel_sprite_plane_create(struct drm_i915_private *dev_priv,
> int num_formats;
> int ret;
>
> - if (INTEL_GEN(dev_priv) >= 9)
> - return skl_universal_plane_create(dev_priv, pipe,
> + if (INTEL_GEN(i915) >= 9)
> + return skl_universal_plane_create(i915, pipe,
> PLANE_SPRITE0 + sprite);
>
> plane = intel_plane_alloc();
> if (IS_ERR(plane))
> return plane;
>
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> plane->max_stride = i9xx_plane_max_stride;
> plane->update_plane = vlv_update_plane;
> plane->disable_plane = vlv_disable_plane;
> @@ -2393,7 +2393,7 @@ intel_sprite_plane_create(struct drm_i915_private *dev_priv,
> modifiers = i9xx_plane_format_modifiers;
>
> plane_funcs = &vlv_sprite_funcs;
> - } else if (INTEL_GEN(dev_priv) >= 7) {
> + } else if (INTEL_GEN(i915) >= 7) {
> plane->max_stride = g4x_sprite_max_stride;
> plane->update_plane = ivb_update_plane;
> plane->disable_plane = ivb_disable_plane;
> @@ -2413,7 +2413,7 @@ intel_sprite_plane_create(struct drm_i915_private *dev_priv,
> plane->check_plane = g4x_sprite_check;
>
> modifiers = i9xx_plane_format_modifiers;
> - if (IS_GEN(dev_priv, 6)) {
> + if (IS_GEN(i915, 6)) {
> formats = snb_plane_formats;
> num_formats = ARRAY_SIZE(snb_plane_formats);
>
> @@ -2426,7 +2426,7 @@ intel_sprite_plane_create(struct drm_i915_private *dev_priv,
> }
> }
>
> - if (IS_CHERRYVIEW(dev_priv) && pipe == PIPE_B) {
> + if (IS_CHERRYVIEW(i915) && pipe == PIPE_B) {
> supported_rotations =
> DRM_MODE_ROTATE_0 | DRM_MODE_ROTATE_180 |
> DRM_MODE_REFLECT_X;
> @@ -2441,7 +2441,7 @@ intel_sprite_plane_create(struct drm_i915_private *dev_priv,
>
> possible_crtcs = BIT(pipe);
>
> - ret = drm_universal_plane_init(&dev_priv->drm, &plane->base,
> + ret = drm_universal_plane_init(&i915->drm, &plane->base,
> possible_crtcs, plane_funcs,
> formats, num_formats, modifiers,
> DRM_PLANE_TYPE_OVERLAY,
> diff --git a/drivers/gpu/drm/i915/intel_sprite.h b/drivers/gpu/drm/i915/intel_sprite.h
> index 500f6bffb139..401295e05116 100644
> --- a/drivers/gpu/drm/i915/intel_sprite.h
> +++ b/drivers/gpu/drm/i915/intel_sprite.h
> @@ -21,7 +21,7 @@ struct intel_plane_state;
> bool is_planar_yuv_format(u32 pixelformat);
> int intel_usecs_to_scanlines(const struct drm_display_mode *adjusted_mode,
> int usecs);
> -struct intel_plane *intel_sprite_plane_create(struct drm_i915_private *dev_priv,
> +struct intel_plane *intel_sprite_plane_create(struct drm_i915_private *i915,
> enum pipe pipe, int plane);
> int intel_sprite_set_colorkey_ioctl(struct drm_device *dev, void *data,
> struct drm_file *file_priv);
> @@ -31,7 +31,7 @@ int intel_plane_check_stride(const struct intel_plane_state *plane_state);
> int intel_plane_check_src_coordinates(struct intel_plane_state *plane_state);
> int chv_plane_check_rotation(const struct intel_plane_state *plane_state);
> struct intel_plane *
> -skl_universal_plane_create(struct drm_i915_private *dev_priv,
> +skl_universal_plane_create(struct drm_i915_private *i915,
> enum pipe pipe, enum plane_id plane_id);
>
> static inline bool icl_is_nv12_y_plane(enum plane_id id)
> @@ -49,10 +49,10 @@ static inline u8 icl_hdr_plane_mask(void)
> BIT(PLANE_SPRITE0) | BIT(PLANE_SPRITE1);
> }
>
> -static inline bool icl_is_hdr_plane(struct drm_i915_private *dev_priv,
> +static inline bool icl_is_hdr_plane(struct drm_i915_private *i915,
> enum plane_id plane_id)
> {
> - return INTEL_GEN(dev_priv) >= 11 &&
> + return INTEL_GEN(i915) >= 11 &&
> icl_hdr_plane_mask() & BIT(plane_id);
> }
>
> diff --git a/drivers/gpu/drm/i915/intel_tv.c b/drivers/gpu/drm/i915/intel_tv.c
> index 5dc594eafaf2..4e66261bf5af 100644
> --- a/drivers/gpu/drm/i915/intel_tv.c
> +++ b/drivers/gpu/drm/i915/intel_tv.c
> @@ -906,7 +906,7 @@ static struct intel_tv *intel_attached_tv(struct drm_connector *connector)
> static bool
> intel_tv_get_hw_state(struct intel_encoder *encoder, enum pipe *pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> u32 tmp = I915_READ(TV_CTL);
>
> *pipe = (tmp & TV_ENC_PIPE_SEL_MASK) >> TV_ENC_PIPE_SEL_SHIFT;
> @@ -920,10 +920,10 @@ intel_enable_tv(struct intel_encoder *encoder,
> const struct drm_connector_state *conn_state)
> {
> struct drm_device *dev = encoder->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
>
> /* Prevents vblank waits from timing out in intel_tv_detect_type() */
> - intel_wait_for_vblank(dev_priv,
> + intel_wait_for_vblank(i915,
> to_intel_crtc(pipe_config->base.crtc)->pipe);
>
> I915_WRITE(TV_CTL, I915_READ(TV_CTL) | TV_ENC_ENABLE);
> @@ -935,7 +935,7 @@ intel_disable_tv(struct intel_encoder *encoder,
> const struct drm_connector_state *old_conn_state)
> {
> struct drm_device *dev = encoder->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
>
> I915_WRITE(TV_CTL, I915_READ(TV_CTL) & ~TV_ENC_ENABLE);
> }
> @@ -1084,7 +1084,7 @@ static void
> intel_tv_get_config(struct intel_encoder *encoder,
> struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct drm_display_mode *adjusted_mode =
> &pipe_config->base.adjusted_mode;
> struct drm_display_mode mode = {};
> @@ -1158,15 +1158,15 @@ intel_tv_get_config(struct intel_encoder *encoder,
> adjusted_mode->crtc_clock /= 2;
>
> /* pixel counter doesn't work on i965gm TV output */
> - if (IS_I965GM(dev_priv))
> + if (IS_I965GM(i915))
> adjusted_mode->private_flags |=
> I915_MODE_FLAG_USE_SCANLINE_COUNTER;
> }
>
> -static bool intel_tv_source_too_wide(struct drm_i915_private *dev_priv,
> +static bool intel_tv_source_too_wide(struct drm_i915_private *i915,
> int hdisplay)
> {
> - return IS_GEN(dev_priv, 3) && hdisplay > 1024;
> + return IS_GEN(i915, 3) && hdisplay > 1024;
> }
>
> static bool intel_tv_vert_scaling(const struct drm_display_mode *tv_mode,
> @@ -1184,7 +1184,7 @@ intel_tv_compute_config(struct intel_encoder *encoder,
> struct intel_crtc_state *pipe_config,
> struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_tv_connector_state *tv_conn_state =
> to_intel_tv_connector_state(conn_state);
> const struct tv_mode *tv_mode = intel_tv_mode_find(conn_state);
> @@ -1209,7 +1209,7 @@ intel_tv_compute_config(struct intel_encoder *encoder,
> intel_tv_mode_to_mode(adjusted_mode, tv_mode);
> drm_mode_set_crtcinfo(adjusted_mode, 0);
>
> - if (intel_tv_source_too_wide(dev_priv, hdisplay) ||
> + if (intel_tv_source_too_wide(i915, hdisplay) ||
> !intel_tv_vert_scaling(adjusted_mode, conn_state, vdisplay)) {
> int extra, top, bottom;
>
> @@ -1306,7 +1306,7 @@ intel_tv_compute_config(struct intel_encoder *encoder,
> * the active portion. Hence following this formula seems
> * more trouble that it's worth.
> *
> - * if (IS_GEN(dev_priv, 4)) {
> + * if (IS_GEN(i915, 4)) {
> * num = cdclk * (tv_mode->oversample >> !tv_mode->progressive);
> * den = tv_mode->clock;
> * } else {
> @@ -1327,7 +1327,7 @@ intel_tv_compute_config(struct intel_encoder *encoder,
> adjusted_mode->name[0] = '\0';
>
> /* pixel counter doesn't work on i965gm TV output */
> - if (IS_I965GM(dev_priv))
> + if (IS_I965GM(i915))
> adjusted_mode->private_flags |=
> I915_MODE_FLAG_USE_SCANLINE_COUNTER;
>
> @@ -1335,7 +1335,7 @@ intel_tv_compute_config(struct intel_encoder *encoder,
> }
>
> static void
> -set_tv_mode_timings(struct drm_i915_private *dev_priv,
> +set_tv_mode_timings(struct drm_i915_private *i915,
> const struct tv_mode *tv_mode,
> bool burst_ena)
> {
> @@ -1393,7 +1393,7 @@ set_tv_mode_timings(struct drm_i915_private *dev_priv,
> I915_WRITE(TV_V_CTL_7, vctl7);
> }
>
> -static void set_color_conversion(struct drm_i915_private *dev_priv,
> +static void set_color_conversion(struct drm_i915_private *i915,
> const struct color_conversion *color_conversion)
> {
> if (!color_conversion)
> @@ -1417,7 +1417,7 @@ static void intel_tv_pre_enable(struct intel_encoder *encoder,
> const struct intel_crtc_state *pipe_config,
> const struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(pipe_config->base.crtc);
> struct intel_tv *intel_tv = enc_to_tv(encoder);
> const struct intel_tv_connector_state *tv_conn_state =
> @@ -1507,18 +1507,18 @@ static void intel_tv_pre_enable(struct intel_encoder *encoder,
> tv_mode->dda3_inc << TV_SCDDA3_INC_SHIFT;
>
> /* Enable two fixes for the chips that need them. */
> - if (IS_I915GM(dev_priv))
> + if (IS_I915GM(i915))
> tv_ctl |= TV_ENC_C0_FIX | TV_ENC_SDP_FIX;
>
> - set_tv_mode_timings(dev_priv, tv_mode, burst_ena);
> + set_tv_mode_timings(i915, tv_mode, burst_ena);
>
> I915_WRITE(TV_SC_CTL_1, scctl1);
> I915_WRITE(TV_SC_CTL_2, scctl2);
> I915_WRITE(TV_SC_CTL_3, scctl3);
>
> - set_color_conversion(dev_priv, color_conversion);
> + set_color_conversion(i915, color_conversion);
>
> - if (INTEL_GEN(dev_priv) >= 4)
> + if (INTEL_GEN(i915) >= 4)
> I915_WRITE(TV_CLR_KNOBS, 0x00404000);
> else
> I915_WRITE(TV_CLR_KNOBS, 0x00606000);
> @@ -1528,7 +1528,7 @@ static void intel_tv_pre_enable(struct intel_encoder *encoder,
> ((video_levels->black << TV_BLACK_LEVEL_SHIFT) |
> (video_levels->blank << TV_BLANK_LEVEL_SHIFT)));
>
> - assert_pipe_disabled(dev_priv, intel_crtc->pipe);
> + assert_pipe_disabled(i915, intel_crtc->pipe);
>
> /* Filter ctl must be set before TV_WIN_SIZE */
> tv_filter_ctl = TV_AUTO_SCALE;
> @@ -1568,18 +1568,18 @@ intel_tv_detect_type(struct intel_tv *intel_tv,
> struct drm_crtc *crtc = connector->state->crtc;
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> struct drm_device *dev = connector->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> u32 tv_ctl, save_tv_ctl;
> u32 tv_dac, save_tv_dac;
> int type;
>
> /* Disable TV interrupts around load detect or we'll recurse */
> if (connector->polled & DRM_CONNECTOR_POLL_HPD) {
> - spin_lock_irq(&dev_priv->irq_lock);
> - i915_disable_pipestat(dev_priv, 0,
> + spin_lock_irq(&i915->irq_lock);
> + i915_disable_pipestat(i915, 0,
> PIPE_HOTPLUG_INTERRUPT_STATUS |
> PIPE_HOTPLUG_TV_INTERRUPT_STATUS);
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_unlock_irq(&i915->irq_lock);
> }
>
> save_tv_dac = tv_dac = I915_READ(TV_DAC);
> @@ -1605,7 +1605,7 @@ intel_tv_detect_type(struct intel_tv *intel_tv,
> * The TV sense state should be cleared to zero on cantiga platform. Otherwise
> * the TV is misdetected. This is hardware requirement.
> */
> - if (IS_GM45(dev_priv))
> + if (IS_GM45(i915))
> tv_dac &= ~(TVDAC_STATE_CHG_EN | TVDAC_A_SENSE_CTL |
> TVDAC_B_SENSE_CTL | TVDAC_C_SENSE_CTL);
>
> @@ -1613,7 +1613,7 @@ intel_tv_detect_type(struct intel_tv *intel_tv,
> I915_WRITE(TV_DAC, tv_dac);
> POSTING_READ(TV_DAC);
>
> - intel_wait_for_vblank(dev_priv, intel_crtc->pipe);
> + intel_wait_for_vblank(i915, intel_crtc->pipe);
>
> type = -1;
> tv_dac = I915_READ(TV_DAC);
> @@ -1643,15 +1643,15 @@ intel_tv_detect_type(struct intel_tv *intel_tv,
> POSTING_READ(TV_CTL);
>
> /* For unknown reasons the hw barfs if we don't do this vblank wait. */
> - intel_wait_for_vblank(dev_priv, intel_crtc->pipe);
> + intel_wait_for_vblank(i915, intel_crtc->pipe);
>
> /* Restore interrupt config */
> if (connector->polled & DRM_CONNECTOR_POLL_HPD) {
> - spin_lock_irq(&dev_priv->irq_lock);
> - i915_enable_pipestat(dev_priv, 0,
> + spin_lock_irq(&i915->irq_lock);
> + i915_enable_pipestat(i915, 0,
> PIPE_HOTPLUG_INTERRUPT_STATUS |
> PIPE_HOTPLUG_TV_INTERRUPT_STATUS);
> - spin_unlock_irq(&dev_priv->irq_lock);
> + spin_unlock_irq(&i915->irq_lock);
> }
>
> return type;
> @@ -1764,7 +1764,7 @@ intel_tv_set_mode_type(struct drm_display_mode *mode,
> static int
> intel_tv_get_modes(struct drm_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->dev);
> + struct drm_i915_private *i915 = to_i915(connector->dev);
> const struct tv_mode *tv_mode = intel_tv_mode_find(connector->state);
> int i, count = 0;
>
> @@ -1778,7 +1778,7 @@ intel_tv_get_modes(struct drm_connector *connector)
> continue;
>
> /* no vertical scaling with wide sources on gen3 */
> - if (IS_GEN(dev_priv, 3) && input->w > 1024 &&
> + if (IS_GEN(i915, 3) && input->w > 1024 &&
> input->h > intel_tv_mode_vdisplay(tv_mode))
> continue;
>
> @@ -1857,9 +1857,9 @@ static const struct drm_encoder_funcs intel_tv_enc_funcs = {
> };
>
> void
> -intel_tv_init(struct drm_i915_private *dev_priv)
> +intel_tv_init(struct drm_i915_private *i915)
> {
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_device *dev = &i915->drm;
> struct drm_connector *connector;
> struct intel_tv *intel_tv;
> struct intel_encoder *intel_encoder;
> @@ -1872,7 +1872,7 @@ intel_tv_init(struct drm_i915_private *dev_priv)
> if ((I915_READ(TV_CTL) & TV_FUSE_STATE_MASK) == TV_FUSE_STATE_DISABLED)
> return;
>
> - if (!intel_bios_is_tv_present(dev_priv)) {
> + if (!intel_bios_is_tv_present(i915)) {
> DRM_DEBUG_KMS("Integrated TV is not present.\n");
> return;
> }
> @@ -1966,7 +1966,7 @@ intel_tv_init(struct drm_i915_private *dev_priv)
> /* Create TV properties then attach current values */
> for (i = 0; i < ARRAY_SIZE(tv_modes); i++) {
> /* 1080p50/1080p60 not supported on gen3 */
> - if (IS_GEN(dev_priv, 3) &&
> + if (IS_GEN(i915, 3) &&
> tv_modes[i].oversample == 1)
> break;
>
> diff --git a/drivers/gpu/drm/i915/intel_tv.h b/drivers/gpu/drm/i915/intel_tv.h
> index 44518575ec5c..fc0b89563b4f 100644
> --- a/drivers/gpu/drm/i915/intel_tv.h
> +++ b/drivers/gpu/drm/i915/intel_tv.h
> @@ -8,6 +8,6 @@
>
> struct drm_i915_private;
>
> -void intel_tv_init(struct drm_i915_private *dev_priv);
> +void intel_tv_init(struct drm_i915_private *i915);
>
> #endif /* __INTEL_TV_H__ */
> diff --git a/drivers/gpu/drm/i915/intel_uc.c b/drivers/gpu/drm/i915/intel_uc.c
> index a8e7f0ba7c3b..11e2dcf6b917 100644
> --- a/drivers/gpu/drm/i915/intel_uc.c
> +++ b/drivers/gpu/drm/i915/intel_uc.c
> @@ -33,12 +33,12 @@ static void guc_free_load_err_log(struct intel_guc *guc);
>
> /* Reset GuC providing us with fresh state for both GuC and HuC.
> */
> -static int __intel_uc_reset_hw(struct drm_i915_private *dev_priv)
> +static int __intel_uc_reset_hw(struct drm_i915_private *i915)
> {
> int ret;
> u32 guc_status;
>
> - ret = intel_reset_guc(dev_priv);
> + ret = intel_reset_guc(i915);
> if (ret) {
> DRM_ERROR("Failed to reset GuC, ret = %d\n", ret);
> return ret;
> diff --git a/drivers/gpu/drm/i915/intel_uc.h b/drivers/gpu/drm/i915/intel_uc.h
> index 3ea06c87dfcd..4ddd50d990b3 100644
> --- a/drivers/gpu/drm/i915/intel_uc.h
> +++ b/drivers/gpu/drm/i915/intel_uc.h
> @@ -28,20 +28,20 @@
> #include "intel_huc.h"
> #include "i915_params.h"
>
> -void intel_uc_init_early(struct drm_i915_private *dev_priv);
> -void intel_uc_cleanup_early(struct drm_i915_private *dev_priv);
> -void intel_uc_init_mmio(struct drm_i915_private *dev_priv);
> -int intel_uc_init_misc(struct drm_i915_private *dev_priv);
> -void intel_uc_fini_misc(struct drm_i915_private *dev_priv);
> -void intel_uc_sanitize(struct drm_i915_private *dev_priv);
> -int intel_uc_init_hw(struct drm_i915_private *dev_priv);
> -void intel_uc_fini_hw(struct drm_i915_private *dev_priv);
> -int intel_uc_init(struct drm_i915_private *dev_priv);
> -void intel_uc_fini(struct drm_i915_private *dev_priv);
> +void intel_uc_init_early(struct drm_i915_private *i915);
> +void intel_uc_cleanup_early(struct drm_i915_private *i915);
> +void intel_uc_init_mmio(struct drm_i915_private *i915);
> +int intel_uc_init_misc(struct drm_i915_private *i915);
> +void intel_uc_fini_misc(struct drm_i915_private *i915);
> +void intel_uc_sanitize(struct drm_i915_private *i915);
> +int intel_uc_init_hw(struct drm_i915_private *i915);
> +void intel_uc_fini_hw(struct drm_i915_private *i915);
> +int intel_uc_init(struct drm_i915_private *i915);
> +void intel_uc_fini(struct drm_i915_private *i915);
> void intel_uc_reset_prepare(struct drm_i915_private *i915);
> void intel_uc_suspend(struct drm_i915_private *i915);
> void intel_uc_runtime_suspend(struct drm_i915_private *i915);
> -int intel_uc_resume(struct drm_i915_private *dev_priv);
> +int intel_uc_resume(struct drm_i915_private *i915);
>
> static inline bool intel_uc_is_using_guc(struct drm_i915_private *i915)
> {
> diff --git a/drivers/gpu/drm/i915/intel_uc_fw.c b/drivers/gpu/drm/i915/intel_uc_fw.c
> index f342ddd47df8..d156911d1560 100644
> --- a/drivers/gpu/drm/i915/intel_uc_fw.c
> +++ b/drivers/gpu/drm/i915/intel_uc_fw.c
> @@ -32,15 +32,15 @@
> /**
> * intel_uc_fw_fetch - fetch uC firmware
> *
> - * @dev_priv: device private
> + * @i915: device private
> * @uc_fw: uC firmware
> *
> * Fetch uC firmware into GEM obj.
> */
> -void intel_uc_fw_fetch(struct drm_i915_private *dev_priv,
> +void intel_uc_fw_fetch(struct drm_i915_private *i915,
> struct intel_uc_fw *uc_fw)
> {
> - struct pci_dev *pdev = dev_priv->drm.pdev;
> + struct pci_dev *pdev = i915->drm.pdev;
> struct drm_i915_gem_object *obj;
> const struct firmware *fw = NULL;
> struct uc_css_header *css;
> @@ -48,10 +48,10 @@ void intel_uc_fw_fetch(struct drm_i915_private *dev_priv,
> int err;
>
> if (!uc_fw->path) {
> - dev_info(dev_priv->drm.dev,
> + dev_info(i915->drm.dev,
> "%s: No firmware was defined for %s!\n",
> intel_uc_fw_type_repr(uc_fw->type),
> - intel_platform_name(INTEL_INFO(dev_priv)->platform));
> + intel_platform_name(INTEL_INFO(i915)->platform));
> return;
> }
>
> @@ -159,7 +159,7 @@ void intel_uc_fw_fetch(struct drm_i915_private *dev_priv,
> goto fail;
> }
>
> - obj = i915_gem_object_create_shmem_from_data(dev_priv,
> + obj = i915_gem_object_create_shmem_from_data(i915,
> fw->data, fw->size);
> if (IS_ERR(obj)) {
> err = PTR_ERR(obj);
> diff --git a/drivers/gpu/drm/i915/intel_uc_fw.h b/drivers/gpu/drm/i915/intel_uc_fw.h
> index ff98f8661d72..881c3aef074c 100644
> --- a/drivers/gpu/drm/i915/intel_uc_fw.h
> +++ b/drivers/gpu/drm/i915/intel_uc_fw.h
> @@ -142,7 +142,7 @@ static inline u32 intel_uc_fw_get_upload_size(struct intel_uc_fw *uc_fw)
> return uc_fw->header_size + uc_fw->ucode_size;
> }
>
> -void intel_uc_fw_fetch(struct drm_i915_private *dev_priv,
> +void intel_uc_fw_fetch(struct drm_i915_private *i915,
> struct intel_uc_fw *uc_fw);
> void intel_uc_fw_cleanup_fetch(struct intel_uc_fw *uc_fw);
> int intel_uc_fw_upload(struct intel_uc_fw *uc_fw,
> diff --git a/drivers/gpu/drm/i915/intel_uncore.c b/drivers/gpu/drm/i915/intel_uncore.c
> index 85171a8b866a..87dec005165d 100644
> --- a/drivers/gpu/drm/i915/intel_uncore.c
> +++ b/drivers/gpu/drm/i915/intel_uncore.c
> @@ -537,10 +537,10 @@ void intel_uncore_runtime_resume(struct intel_uncore *uncore)
> iosf_mbi_register_pmic_bus_access_notifier(&uncore->pmic_bus_access_nb);
> }
>
> -void intel_uncore_sanitize(struct drm_i915_private *dev_priv)
> +void intel_uncore_sanitize(struct drm_i915_private *i915)
> {
> /* BIOS often leaves RC6 enabled, but disable it for hw init */
> - intel_sanitize_gt_powersave(dev_priv);
> + intel_sanitize_gt_powersave(i915);
> }
>
> static void __intel_uncore_forcewake_get(struct intel_uncore *uncore,
> @@ -647,7 +647,7 @@ void intel_uncore_forcewake_user_put(struct intel_uncore *uncore)
> * @fw_domains: forcewake domains to get reference on
> *
> * See intel_uncore_forcewake_get(). This variant places the onus
> - * on the caller to explicitly handle the dev_priv->uncore.lock spinlock.
> + * on the caller to explicitly handle the i915->uncore.lock spinlock.
> */
> void intel_uncore_forcewake_get__locked(struct intel_uncore *uncore,
> enum forcewake_domains fw_domains)
> @@ -708,7 +708,7 @@ void intel_uncore_forcewake_put(struct intel_uncore *uncore,
> * @fw_domains: forcewake domains to get reference on
> *
> * See intel_uncore_forcewake_put(). This variant places the onus
> - * on the caller to explicitly handle the dev_priv->uncore.lock spinlock.
> + * on the caller to explicitly handle the i915->uncore.lock spinlock.
> */
> void intel_uncore_forcewake_put__locked(struct intel_uncore *uncore,
> enum forcewake_domains fw_domains)
> @@ -821,10 +821,10 @@ find_fw_domain(struct intel_uncore *uncore, u32 offset)
> #define GEN_FW_RANGE(s, e, d) \
> { .start = (s), .end = (e), .domains = (d) }
>
> -#define HAS_FWTABLE(dev_priv) \
> - (INTEL_GEN(dev_priv) >= 9 || \
> - IS_CHERRYVIEW(dev_priv) || \
> - IS_VALLEYVIEW(dev_priv))
> +#define HAS_FWTABLE(i915) \
> + (INTEL_GEN(i915) >= 9 || \
> + IS_CHERRYVIEW(i915) || \
> + IS_VALLEYVIEW(i915))
>
> /* *Must* be sorted by offset ranges! See intel_fw_table_check(). */
> static const struct intel_forcewake_range __vlv_fw_ranges[] = {
> @@ -1461,7 +1461,7 @@ static void intel_uncore_fw_domains_init(struct intel_uncore *uncore)
> static int i915_pmic_bus_access_notifier(struct notifier_block *nb,
> unsigned long action, void *data)
> {
> - struct drm_i915_private *dev_priv = container_of(nb,
> + struct drm_i915_private *i915 = container_of(nb,
> struct drm_i915_private, uncore.pmic_bus_access_nb);
>
> switch (action) {
> @@ -1479,12 +1479,12 @@ static int i915_pmic_bus_access_notifier(struct notifier_block *nb,
> * wake reference -> disable wakeref asserts for the time of
> * the access.
> */
> - disable_rpm_wakeref_asserts(dev_priv);
> - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> - enable_rpm_wakeref_asserts(dev_priv);
> + disable_rpm_wakeref_asserts(i915);
> + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
> + enable_rpm_wakeref_asserts(i915);
> break;
> case MBI_PMIC_BUS_ACCESS_END:
> - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> break;
> }
>
> diff --git a/drivers/gpu/drm/i915/intel_uncore.h b/drivers/gpu/drm/i915/intel_uncore.h
> index d6af3de70121..d30bda2c3444 100644
> --- a/drivers/gpu/drm/i915/intel_uncore.h
> +++ b/drivers/gpu/drm/i915/intel_uncore.h
> @@ -177,7 +177,7 @@ intel_uncore_has_fifo(const struct intel_uncore *uncore)
> return uncore->flags & UNCORE_HAS_FIFO;
> }
>
> -void intel_uncore_sanitize(struct drm_i915_private *dev_priv);
> +void intel_uncore_sanitize(struct drm_i915_private *i915);
> void intel_uncore_init_early(struct intel_uncore *uncore);
> int intel_uncore_init_mmio(struct intel_uncore *uncore);
> void intel_uncore_prune_mmio_domains(struct intel_uncore *uncore);
> @@ -361,7 +361,7 @@ intel_uncore_read64_2x32(struct intel_uncore *uncore,
> *
> * Certain architectures will die if the same cacheline is concurrently accessed
> * by different clients (e.g. on Ivybridge). Access to registers should
> - * therefore generally be serialised, by either the dev_priv->uncore.lock or
> + * therefore generally be serialised, by either the i915->uncore.lock or
> * a more localised lock guarding all access to that bank of registers.
> */
> #define intel_uncore_read_fw(...) __raw_uncore_read32(__VA_ARGS__)
> diff --git a/drivers/gpu/drm/i915/intel_vbt_defs.h b/drivers/gpu/drm/i915/intel_vbt_defs.h
> index 89ef14cafb6b..e2e9883d6b28 100644
> --- a/drivers/gpu/drm/i915/intel_vbt_defs.h
> +++ b/drivers/gpu/drm/i915/intel_vbt_defs.h
> @@ -337,7 +337,7 @@ enum vbt_gmbus_ddi {
> * basically any of the fields to ensure the correct interpretation for the BDB
> * version in question.
> *
> - * When we copy the child device configs to dev_priv->vbt.child_dev, we reserve
> + * When we copy the child device configs to i915->vbt.child_dev, we reserve
> * space for the full structure below, and initialize the tail not actually
> * present in VBT to zeros. Accessing those fields is fine, as long as the
> * default zero is taken into account, again according to the BDB version.
> diff --git a/drivers/gpu/drm/i915/intel_vdsc.c b/drivers/gpu/drm/i915/intel_vdsc.c
> index ffec807b8960..164fd4b56112 100644
> --- a/drivers/gpu/drm/i915/intel_vdsc.c
> +++ b/drivers/gpu/drm/i915/intel_vdsc.c
> @@ -478,7 +478,7 @@ static void intel_configure_pps_for_dsc_encoder(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> const struct drm_dsc_config *vdsc_cfg = &crtc_state->dp_dsc_cfg;
> enum pipe pipe = crtc->pipe;
> enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> @@ -897,7 +897,7 @@ void intel_dsc_enable(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> enum pipe pipe = crtc->pipe;
> i915_reg_t dss_ctl1_reg, dss_ctl2_reg;
> u32 dss_ctl1_val = 0;
> @@ -907,7 +907,7 @@ void intel_dsc_enable(struct intel_encoder *encoder,
> return;
>
> /* Enable Power wells for VDSC/joining */
> - intel_display_power_get(dev_priv,
> + intel_display_power_get(i915,
> intel_dsc_power_domain(crtc_state));
>
> intel_configure_pps_for_dsc_encoder(encoder, crtc_state);
> @@ -933,7 +933,7 @@ void intel_dsc_enable(struct intel_encoder *encoder,
> void intel_dsc_disable(const struct intel_crtc_state *old_crtc_state)
> {
> struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc);
> - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> + struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> enum pipe pipe = crtc->pipe;
> i915_reg_t dss_ctl1_reg, dss_ctl2_reg;
> u32 dss_ctl1_val = 0, dss_ctl2_val = 0;
> @@ -961,6 +961,6 @@ void intel_dsc_disable(const struct intel_crtc_state *old_crtc_state)
> I915_WRITE(dss_ctl2_reg, dss_ctl2_val);
>
> /* Disable Power wells for VDSC/joining */
> - intel_display_power_put_unchecked(dev_priv,
> + intel_display_power_put_unchecked(i915,
> intel_dsc_power_domain(old_crtc_state));
> }
> diff --git a/drivers/gpu/drm/i915/intel_wopcm.c b/drivers/gpu/drm/i915/intel_wopcm.c
> index 7b4ba84b9fb8..646540b97267 100644
> --- a/drivers/gpu/drm/i915/intel_wopcm.c
> +++ b/drivers/gpu/drm/i915/intel_wopcm.c
> @@ -225,7 +225,7 @@ int intel_wopcm_init(struct intel_wopcm *wopcm)
> return 0;
> }
>
> -static inline int write_and_verify(struct drm_i915_private *dev_priv,
> +static inline int write_and_verify(struct drm_i915_private *i915,
> i915_reg_t reg, u32 val, u32 mask,
> u32 locked_bit)
> {
> @@ -252,27 +252,27 @@ static inline int write_and_verify(struct drm_i915_private *dev_priv,
> */
> int intel_wopcm_init_hw(struct intel_wopcm *wopcm)
> {
> - struct drm_i915_private *dev_priv = wopcm_to_i915(wopcm);
> + struct drm_i915_private *i915 = wopcm_to_i915(wopcm);
> u32 huc_agent;
> u32 mask;
> int err;
>
> - if (!USES_GUC(dev_priv))
> + if (!USES_GUC(i915))
> return 0;
>
> - GEM_BUG_ON(!HAS_GUC(dev_priv));
> + GEM_BUG_ON(!HAS_GUC(i915));
> GEM_BUG_ON(!wopcm->guc.size);
> GEM_BUG_ON(!wopcm->guc.base);
>
> - err = write_and_verify(dev_priv, GUC_WOPCM_SIZE, wopcm->guc.size,
> + err = write_and_verify(i915, GUC_WOPCM_SIZE, wopcm->guc.size,
> GUC_WOPCM_SIZE_MASK | GUC_WOPCM_SIZE_LOCKED,
> GUC_WOPCM_SIZE_LOCKED);
> if (err)
> goto err_out;
>
> - huc_agent = USES_HUC(dev_priv) ? HUC_LOADING_AGENT_GUC : 0;
> + huc_agent = USES_HUC(i915) ? HUC_LOADING_AGENT_GUC : 0;
> mask = GUC_WOPCM_OFFSET_MASK | GUC_WOPCM_OFFSET_VALID | huc_agent;
> - err = write_and_verify(dev_priv, DMA_GUC_WOPCM_OFFSET,
> + err = write_and_verify(i915, DMA_GUC_WOPCM_OFFSET,
> wopcm->guc.base | huc_agent, mask,
> GUC_WOPCM_OFFSET_VALID);
> if (err)
> diff --git a/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c b/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
> index 2093d08a7569..2514842236f6 100644
> --- a/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
> +++ b/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
> @@ -147,17 +147,17 @@ fake_dma_object(struct drm_i915_private *i915, u64 size)
>
> static int igt_ppgtt_alloc(void *arg)
> {
> - struct drm_i915_private *dev_priv = arg;
> + struct drm_i915_private *i915 = arg;
> struct i915_ppgtt *ppgtt;
> u64 size, last, limit;
> int err = 0;
>
> /* Allocate a ppggt and try to fill the entire range */
>
> - if (!HAS_PPGTT(dev_priv))
> + if (!HAS_PPGTT(i915))
> return 0;
>
> - ppgtt = __ppgtt_create(dev_priv);
> + ppgtt = __ppgtt_create(i915);
> if (IS_ERR(ppgtt))
> return PTR_ERR(ppgtt);
>
> @@ -208,9 +208,9 @@ static int igt_ppgtt_alloc(void *arg)
> }
>
> err_ppgtt_cleanup:
> - mutex_lock(&dev_priv->drm.struct_mutex);
> + mutex_lock(&i915->drm.struct_mutex);
> i915_vm_put(&ppgtt->vm);
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + mutex_unlock(&i915->drm.struct_mutex);
> return err;
> }
>
> @@ -992,7 +992,7 @@ static int shrink_boom(struct drm_i915_private *i915,
> return err;
> }
>
> -static int exercise_ppgtt(struct drm_i915_private *dev_priv,
> +static int exercise_ppgtt(struct drm_i915_private *i915,
> int (*func)(struct drm_i915_private *i915,
> struct i915_address_space *vm,
> u64 hole_start, u64 hole_end,
> @@ -1003,15 +1003,15 @@ static int exercise_ppgtt(struct drm_i915_private *dev_priv,
> IGT_TIMEOUT(end_time);
> int err;
>
> - if (!HAS_FULL_PPGTT(dev_priv))
> + if (!HAS_FULL_PPGTT(i915))
> return 0;
>
> - file = mock_file(dev_priv);
> + file = mock_file(i915);
> if (IS_ERR(file))
> return PTR_ERR(file);
>
> - mutex_lock(&dev_priv->drm.struct_mutex);
> - ppgtt = i915_ppgtt_create(dev_priv);
> + mutex_lock(&i915->drm.struct_mutex);
> + ppgtt = i915_ppgtt_create(i915);
> if (IS_ERR(ppgtt)) {
> err = PTR_ERR(ppgtt);
> goto out_unlock;
> @@ -1019,13 +1019,13 @@ static int exercise_ppgtt(struct drm_i915_private *dev_priv,
> GEM_BUG_ON(offset_in_page(ppgtt->vm.total));
> GEM_BUG_ON(ppgtt->vm.closed);
>
> - err = func(dev_priv, &ppgtt->vm, 0, ppgtt->vm.total, end_time);
> + err = func(i915, &ppgtt->vm, 0, ppgtt->vm.total, end_time);
>
> i915_vm_put(&ppgtt->vm);
> out_unlock:
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + mutex_unlock(&i915->drm.struct_mutex);
>
> - mock_file_free(dev_priv, file);
> + mock_file_free(i915, file);
> return err;
> }
>
> diff --git a/drivers/gpu/drm/i915/selftests/intel_guc.c b/drivers/gpu/drm/i915/selftests/intel_guc.c
> index 7fd0321e0947..ab3e71c957df 100644
> --- a/drivers/gpu/drm/i915/selftests/intel_guc.c
> +++ b/drivers/gpu/drm/i915/selftests/intel_guc.c
> @@ -107,12 +107,12 @@ static int validate_client(struct intel_guc_client *client,
> int client_priority,
> bool is_preempt_client)
> {
> - struct drm_i915_private *dev_priv = guc_to_i915(client->guc);
> + struct drm_i915_private *i915 = guc_to_i915(client->guc);
> struct i915_gem_context *ctx_owner = is_preempt_client ?
> - dev_priv->preempt_context : dev_priv->kernel_context;
> + i915->preempt_context : i915->kernel_context;
>
> if (client->owner != ctx_owner ||
> - client->engines != INTEL_INFO(dev_priv)->engine_mask ||
> + client->engines != INTEL_INFO(i915)->engine_mask ||
> client->priority != client_priority ||
> client->doorbell_id == GUC_DOORBELL_INVALID)
> return -EINVAL;
> @@ -137,16 +137,16 @@ static bool client_doorbell_in_sync(struct intel_guc_client *client)
> */
> static int igt_guc_clients(void *args)
> {
> - struct drm_i915_private *dev_priv = args;
> + struct drm_i915_private *i915 = args;
> intel_wakeref_t wakeref;
> struct intel_guc *guc;
> int err = 0;
>
> - GEM_BUG_ON(!HAS_GUC(dev_priv));
> - mutex_lock(&dev_priv->drm.struct_mutex);
> - wakeref = intel_runtime_pm_get(dev_priv);
> + GEM_BUG_ON(!HAS_GUC(i915));
> + mutex_lock(&i915->drm.struct_mutex);
> + wakeref = intel_runtime_pm_get(i915);
>
> - guc = &dev_priv->guc;
> + guc = &i915->guc;
> if (!guc) {
> pr_err("No guc object!\n");
> err = -EINVAL;
> @@ -227,8 +227,8 @@ static int igt_guc_clients(void *args)
> guc_clients_create(guc);
> guc_clients_enable(guc);
> unlock:
> - intel_runtime_pm_put(dev_priv, wakeref);
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + intel_runtime_pm_put(i915, wakeref);
> + mutex_unlock(&i915->drm.struct_mutex);
> return err;
> }
>
> @@ -239,17 +239,17 @@ static int igt_guc_clients(void *args)
> */
> static int igt_guc_doorbells(void *arg)
> {
> - struct drm_i915_private *dev_priv = arg;
> + struct drm_i915_private *i915 = arg;
> intel_wakeref_t wakeref;
> struct intel_guc *guc;
> int i, err = 0;
> u16 db_id;
>
> - GEM_BUG_ON(!HAS_GUC(dev_priv));
> - mutex_lock(&dev_priv->drm.struct_mutex);
> - wakeref = intel_runtime_pm_get(dev_priv);
> + GEM_BUG_ON(!HAS_GUC(i915));
> + mutex_lock(&i915->drm.struct_mutex);
> + wakeref = intel_runtime_pm_get(i915);
>
> - guc = &dev_priv->guc;
> + guc = &i915->guc;
> if (!guc) {
> pr_err("No guc object!\n");
> err = -EINVAL;
> @@ -261,10 +261,10 @@ static int igt_guc_doorbells(void *arg)
> goto unlock;
>
> for (i = 0; i < ATTEMPTS; i++) {
> - clients[i] = guc_client_alloc(dev_priv,
> - INTEL_INFO(dev_priv)->engine_mask,
> + clients[i] = guc_client_alloc(i915,
> + INTEL_INFO(i915)->engine_mask,
> i % GUC_CLIENT_PRIORITY_NUM,
> - dev_priv->kernel_context);
> + i915->kernel_context);
>
> if (!clients[i]) {
> pr_err("[%d] No guc client\n", i);
> @@ -340,20 +340,20 @@ static int igt_guc_doorbells(void *arg)
> guc_client_free(clients[i]);
> }
> unlock:
> - intel_runtime_pm_put(dev_priv, wakeref);
> - mutex_unlock(&dev_priv->drm.struct_mutex);
> + intel_runtime_pm_put(i915, wakeref);
> + mutex_unlock(&i915->drm.struct_mutex);
> return err;
> }
>
> -int intel_guc_live_selftest(struct drm_i915_private *dev_priv)
> +int intel_guc_live_selftest(struct drm_i915_private *i915)
> {
> static const struct i915_subtest tests[] = {
> SUBTEST(igt_guc_clients),
> SUBTEST(igt_guc_doorbells),
> };
>
> - if (!USES_GUC_SUBMISSION(dev_priv))
> + if (!USES_GUC_SUBMISSION(i915))
> return 0;
>
> - return i915_subtests(tests, dev_priv);
> + return i915_subtests(tests, i915);
> }
> diff --git a/drivers/gpu/drm/i915/selftests/intel_uncore.c b/drivers/gpu/drm/i915/selftests/intel_uncore.c
> index e0d7ebecb215..b5304cfeda21 100644
> --- a/drivers/gpu/drm/i915/selftests/intel_uncore.c
> +++ b/drivers/gpu/drm/i915/selftests/intel_uncore.c
> @@ -254,15 +254,15 @@ static int live_forcewake_ops(void *arg)
> static int live_forcewake_domains(void *arg)
> {
> #define FW_RANGE 0x40000
> - struct drm_i915_private *dev_priv = arg;
> - struct intel_uncore *uncore = &dev_priv->uncore;
> + struct drm_i915_private *i915 = arg;
> + struct intel_uncore *uncore = &i915->uncore;
> unsigned long *valid;
> u32 offset;
> int err;
>
> - if (!HAS_FPGA_DBG_UNCLAIMED(dev_priv) &&
> - !IS_VALLEYVIEW(dev_priv) &&
> - !IS_CHERRYVIEW(dev_priv))
> + if (!HAS_FPGA_DBG_UNCLAIMED(i915) &&
> + !IS_VALLEYVIEW(i915) &&
> + !IS_CHERRYVIEW(i915))
> return 0;
>
> /*
> diff --git a/drivers/gpu/drm/i915/vlv_dsi.c b/drivers/gpu/drm/i915/vlv_dsi.c
> index e272d826210a..0178bdf0c5a9 100644
> --- a/drivers/gpu/drm/i915/vlv_dsi.c
> +++ b/drivers/gpu/drm/i915/vlv_dsi.c
> @@ -78,19 +78,19 @@ void vlv_dsi_wait_for_fifo_empty(struct intel_dsi *intel_dsi, enum port port)
> {
> struct drm_encoder *encoder = &intel_dsi->base.base;
> struct drm_device *dev = encoder->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> u32 mask;
>
> mask = LP_CTRL_FIFO_EMPTY | HS_CTRL_FIFO_EMPTY |
> LP_DATA_FIFO_EMPTY | HS_DATA_FIFO_EMPTY;
>
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> MIPI_GEN_FIFO_STAT(port), mask, mask,
> 100))
> DRM_ERROR("DPI FIFOs are not empty\n");
> }
>
> -static void write_data(struct drm_i915_private *dev_priv,
> +static void write_data(struct drm_i915_private *i915,
> i915_reg_t reg,
> const u8 *data, u32 len)
> {
> @@ -106,7 +106,7 @@ static void write_data(struct drm_i915_private *dev_priv,
> }
> }
>
> -static void read_data(struct drm_i915_private *dev_priv,
> +static void read_data(struct drm_i915_private *i915,
> i915_reg_t reg,
> u8 *data, u32 len)
> {
> @@ -125,7 +125,7 @@ static ssize_t intel_dsi_host_transfer(struct mipi_dsi_host *host,
> {
> struct intel_dsi_host *intel_dsi_host = to_intel_dsi_host(host);
> struct drm_device *dev = intel_dsi_host->intel_dsi->base.base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> enum port port = intel_dsi_host->port;
> struct mipi_dsi_packet packet;
> ssize_t ret;
> @@ -154,13 +154,13 @@ static ssize_t intel_dsi_host_transfer(struct mipi_dsi_host *host,
>
> /* note: this is never true for reads */
> if (packet.payload_length) {
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> MIPI_GEN_FIFO_STAT(port),
> data_mask, 0,
> 50))
> DRM_ERROR("Timeout waiting for HS/LP DATA FIFO !full\n");
>
> - write_data(dev_priv, data_reg, packet.payload,
> + write_data(i915, data_reg, packet.payload,
> packet.payload_length);
> }
>
> @@ -168,7 +168,7 @@ static ssize_t intel_dsi_host_transfer(struct mipi_dsi_host *host,
> I915_WRITE(MIPI_INTR_STAT(port), GEN_READ_DATA_AVAIL);
> }
>
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> MIPI_GEN_FIFO_STAT(port),
> ctrl_mask, 0,
> 50)) {
> @@ -180,13 +180,13 @@ static ssize_t intel_dsi_host_transfer(struct mipi_dsi_host *host,
> /* ->rx_len is set only for reads */
> if (msg->rx_len) {
> data_mask = GEN_READ_DATA_AVAIL;
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> MIPI_INTR_STAT(port),
> data_mask, data_mask,
> 50))
> DRM_ERROR("Timeout waiting for read data.\n");
>
> - read_data(dev_priv, data_reg, msg->rx_buf, msg->rx_len);
> + read_data(i915, data_reg, msg->rx_buf, msg->rx_len);
> }
>
> /* XXX: fix for reads and writes */
> @@ -221,7 +221,7 @@ static int dpi_send_cmd(struct intel_dsi *intel_dsi, u32 cmd, bool hs,
> {
> struct drm_encoder *encoder = &intel_dsi->base.base;
> struct drm_device *dev = encoder->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> u32 mask;
>
> /* XXX: pipe, hs */
> @@ -240,7 +240,7 @@ static int dpi_send_cmd(struct intel_dsi *intel_dsi, u32 cmd, bool hs,
> I915_WRITE(MIPI_DPI_CONTROL(port), cmd);
>
> mask = SPL_PKT_SENT_INTERRUPT;
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> MIPI_INTR_STAT(port), mask, mask,
> 100))
> DRM_ERROR("Video mode command 0x%08x send failed.\n", cmd);
> @@ -248,25 +248,25 @@ static int dpi_send_cmd(struct intel_dsi *intel_dsi, u32 cmd, bool hs,
> return 0;
> }
>
> -static void band_gap_reset(struct drm_i915_private *dev_priv)
> +static void band_gap_reset(struct drm_i915_private *i915)
> {
> - vlv_flisdsi_get(dev_priv);
> + vlv_flisdsi_get(i915);
>
> - vlv_flisdsi_write(dev_priv, 0x08, 0x0001);
> - vlv_flisdsi_write(dev_priv, 0x0F, 0x0005);
> - vlv_flisdsi_write(dev_priv, 0x0F, 0x0025);
> + vlv_flisdsi_write(i915, 0x08, 0x0001);
> + vlv_flisdsi_write(i915, 0x0F, 0x0005);
> + vlv_flisdsi_write(i915, 0x0F, 0x0025);
> udelay(150);
> - vlv_flisdsi_write(dev_priv, 0x0F, 0x0000);
> - vlv_flisdsi_write(dev_priv, 0x08, 0x0000);
> + vlv_flisdsi_write(i915, 0x0F, 0x0000);
> + vlv_flisdsi_write(i915, 0x08, 0x0000);
>
> - vlv_flisdsi_put(dev_priv);
> + vlv_flisdsi_put(i915);
> }
>
> static int intel_dsi_compute_config(struct intel_encoder *encoder,
> struct intel_crtc_state *pipe_config,
> struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = container_of(encoder, struct intel_dsi,
> base);
> struct intel_connector *intel_connector = intel_dsi->attached_connector;
> @@ -281,7 +281,7 @@ static int intel_dsi_compute_config(struct intel_encoder *encoder,
> if (fixed_mode) {
> intel_fixed_panel_mode(fixed_mode, adjusted_mode);
>
> - if (HAS_GMCH(dev_priv))
> + if (HAS_GMCH(i915))
> intel_gmch_panel_fitting(crtc, pipe_config,
> conn_state->scaling_mode);
> else
> @@ -300,7 +300,7 @@ static int intel_dsi_compute_config(struct intel_encoder *encoder,
> else
> pipe_config->pipe_bpp = 18;
>
> - if (IS_GEN9_LP(dev_priv)) {
> + if (IS_GEN9_LP(i915)) {
> /* Enable Frame time stamp based scanline reporting */
> adjusted_mode->private_flags |=
> I915_MODE_FLAG_GET_SCANLINE_FROM_TIMESTAMP;
> @@ -327,7 +327,7 @@ static int intel_dsi_compute_config(struct intel_encoder *encoder,
>
> static bool glk_dsi_enable_io(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> enum port port;
> u32 tmp;
> @@ -359,7 +359,7 @@ static bool glk_dsi_enable_io(struct intel_encoder *encoder)
>
> /* Wait for Pwr ACK */
> for_each_dsi_port(port, intel_dsi->ports) {
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> MIPI_CTRL(port),
> GLK_MIPIIO_PORT_POWERED,
> GLK_MIPIIO_PORT_POWERED,
> @@ -378,14 +378,14 @@ static bool glk_dsi_enable_io(struct intel_encoder *encoder)
>
> static void glk_dsi_device_ready(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> enum port port;
> u32 val;
>
> /* Wait for MIPI PHY status bit to set */
> for_each_dsi_port(port, intel_dsi->ports) {
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> MIPI_CTRL(port),
> GLK_PHY_STATUS_PORT_READY,
> GLK_PHY_STATUS_PORT_READY,
> @@ -413,7 +413,7 @@ static void glk_dsi_device_ready(struct intel_encoder *encoder)
> I915_WRITE(MIPI_DEVICE_READY(port), val);
>
> /* Wait for ULPS active */
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> MIPI_CTRL(port),
> GLK_ULPS_NOT_ACTIVE,
> 0,
> @@ -440,7 +440,7 @@ static void glk_dsi_device_ready(struct intel_encoder *encoder)
>
> /* Wait for Stop state */
> for_each_dsi_port(port, intel_dsi->ports) {
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> MIPI_CTRL(port),
> GLK_DATA_LANE_STOP_STATE,
> GLK_DATA_LANE_STOP_STATE,
> @@ -450,7 +450,7 @@ static void glk_dsi_device_ready(struct intel_encoder *encoder)
>
> /* Wait for AFE LATCH */
> for_each_dsi_port(port, intel_dsi->ports) {
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> BXT_MIPI_PORT_CTRL(port),
> AFE_LATCHOUT,
> AFE_LATCHOUT,
> @@ -461,7 +461,7 @@ static void glk_dsi_device_ready(struct intel_encoder *encoder)
>
> static void bxt_dsi_device_ready(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> enum port port;
> u32 val;
> @@ -488,21 +488,21 @@ static void bxt_dsi_device_ready(struct intel_encoder *encoder)
>
> static void vlv_dsi_device_ready(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> enum port port;
> u32 val;
>
> DRM_DEBUG_KMS("\n");
>
> - vlv_flisdsi_get(dev_priv);
> + vlv_flisdsi_get(i915);
> /* program rcomp for compliance, reduce from 50 ohms to 45 ohms
> * needed everytime after power gate */
> - vlv_flisdsi_write(dev_priv, 0x04, 0x0004);
> - vlv_flisdsi_put(dev_priv);
> + vlv_flisdsi_write(i915, 0x04, 0x0004);
> + vlv_flisdsi_put(i915);
>
> /* bandgap reset is needed after everytime we do power gate */
> - band_gap_reset(dev_priv);
> + band_gap_reset(i915);
>
> for_each_dsi_port(port, intel_dsi->ports) {
>
> @@ -527,11 +527,11 @@ static void vlv_dsi_device_ready(struct intel_encoder *encoder)
>
> static void intel_dsi_device_ready(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
>
> - if (IS_GEMINILAKE(dev_priv))
> + if (IS_GEMINILAKE(i915))
> glk_dsi_device_ready(encoder);
> - else if (IS_GEN9_LP(dev_priv))
> + else if (IS_GEN9_LP(i915))
> bxt_dsi_device_ready(encoder);
> else
> vlv_dsi_device_ready(encoder);
> @@ -539,7 +539,7 @@ static void intel_dsi_device_ready(struct intel_encoder *encoder)
>
> static void glk_dsi_enter_low_power_mode(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> enum port port;
> u32 val;
> @@ -554,7 +554,7 @@ static void glk_dsi_enter_low_power_mode(struct intel_encoder *encoder)
>
> /* Wait for MIPI PHY status bit to unset */
> for_each_dsi_port(port, intel_dsi->ports) {
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> MIPI_CTRL(port),
> GLK_PHY_STATUS_PORT_READY, 0, 20))
> DRM_ERROR("PHY is not turning OFF\n");
> @@ -562,7 +562,7 @@ static void glk_dsi_enter_low_power_mode(struct intel_encoder *encoder)
>
> /* Wait for Pwr ACK bit to unset */
> for_each_dsi_port(port, intel_dsi->ports) {
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> MIPI_CTRL(port),
> GLK_MIPIIO_PORT_POWERED, 0, 20))
> DRM_ERROR("MIPI IO Port is not powergated\n");
> @@ -571,7 +571,7 @@ static void glk_dsi_enter_low_power_mode(struct intel_encoder *encoder)
>
> static void glk_dsi_disable_mipi_io(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> enum port port;
> u32 tmp;
> @@ -583,7 +583,7 @@ static void glk_dsi_disable_mipi_io(struct intel_encoder *encoder)
>
> /* Wait for MIPI PHY status bit to unset */
> for_each_dsi_port(port, intel_dsi->ports) {
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> MIPI_CTRL(port),
> GLK_PHY_STATUS_PORT_READY, 0, 20))
> DRM_ERROR("PHY is not turning OFF\n");
> @@ -605,14 +605,14 @@ static void glk_dsi_clear_device_ready(struct intel_encoder *encoder)
>
> static void vlv_dsi_clear_device_ready(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> enum port port;
>
> DRM_DEBUG_KMS("\n");
> for_each_dsi_port(port, intel_dsi->ports) {
> /* Common bit for both MIPI Port A & MIPI Port C on VLV/CHV */
> - i915_reg_t port_ctrl = IS_GEN9_LP(dev_priv) ?
> + i915_reg_t port_ctrl = IS_GEN9_LP(i915) ?
> BXT_MIPI_PORT_CTRL(port) : MIPI_PORT_CTRL(PORT_A);
> u32 val;
>
> @@ -632,8 +632,8 @@ static void vlv_dsi_clear_device_ready(struct intel_encoder *encoder)
> * On VLV/CHV, wait till Clock lanes are in LP-00 state for MIPI
> * Port A only. MIPI Port C has no similar bit for checking.
> */
> - if ((IS_GEN9_LP(dev_priv) || port == PORT_A) &&
> - intel_wait_for_register(&dev_priv->uncore,
> + if ((IS_GEN9_LP(i915) || port == PORT_A) &&
> + intel_wait_for_register(&i915->uncore,
> port_ctrl, AFE_LATCHOUT, 0,
> 30))
> DRM_ERROR("DSI LP not going Low\n");
> @@ -651,14 +651,14 @@ static void vlv_dsi_clear_device_ready(struct intel_encoder *encoder)
> static void intel_dsi_port_enable(struct intel_encoder *encoder,
> const struct intel_crtc_state *crtc_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> enum port port;
>
> if (intel_dsi->dual_link == DSI_DUAL_LINK_FRONT_BACK) {
> u32 temp;
> - if (IS_GEN9_LP(dev_priv)) {
> + if (IS_GEN9_LP(i915)) {
> for_each_dsi_port(port, intel_dsi->ports) {
> temp = I915_READ(MIPI_CTRL(port));
> temp &= ~BXT_PIXEL_OVERLAP_CNT_MASK |
> @@ -676,7 +676,7 @@ static void intel_dsi_port_enable(struct intel_encoder *encoder,
> }
>
> for_each_dsi_port(port, intel_dsi->ports) {
> - i915_reg_t port_ctrl = IS_GEN9_LP(dev_priv) ?
> + i915_reg_t port_ctrl = IS_GEN9_LP(i915) ?
> BXT_MIPI_PORT_CTRL(port) : MIPI_PORT_CTRL(port);
> u32 temp;
>
> @@ -688,7 +688,7 @@ static void intel_dsi_port_enable(struct intel_encoder *encoder,
> if (intel_dsi->ports == (BIT(PORT_A) | BIT(PORT_C))) {
> temp |= (intel_dsi->dual_link - 1)
> << DUAL_LINK_MODE_SHIFT;
> - if (IS_BROXTON(dev_priv))
> + if (IS_BROXTON(i915))
> temp |= LANE_CONFIGURATION_DUAL_LINK_A;
> else
> temp |= crtc->pipe ?
> @@ -708,12 +708,12 @@ static void intel_dsi_port_enable(struct intel_encoder *encoder,
> static void intel_dsi_port_disable(struct intel_encoder *encoder)
> {
> struct drm_device *dev = encoder->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> enum port port;
>
> for_each_dsi_port(port, intel_dsi->ports) {
> - i915_reg_t port_ctrl = IS_GEN9_LP(dev_priv) ?
> + i915_reg_t port_ctrl = IS_GEN9_LP(i915) ?
> BXT_MIPI_PORT_CTRL(port) : MIPI_PORT_CTRL(port);
> u32 temp;
>
> @@ -775,7 +775,7 @@ static void intel_dsi_pre_enable(struct intel_encoder *encoder,
> {
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> struct drm_crtc *crtc = pipe_config->base.crtc;
> - struct drm_i915_private *dev_priv = to_i915(crtc->dev);
> + struct drm_i915_private *i915 = to_i915(crtc->dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> int pipe = intel_crtc->pipe;
> enum port port;
> @@ -784,13 +784,13 @@ static void intel_dsi_pre_enable(struct intel_encoder *encoder,
>
> DRM_DEBUG_KMS("\n");
>
> - intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, true);
> + intel_set_cpu_fifo_underrun_reporting(i915, pipe, true);
>
> /*
> * The BIOS may leave the PLL in a wonky state where it doesn't
> * lock. It needs to be fully powered down to fix it.
> */
> - if (IS_GEN9_LP(dev_priv)) {
> + if (IS_GEN9_LP(i915)) {
> bxt_dsi_pll_disable(encoder);
> bxt_dsi_pll_enable(encoder, pipe_config);
> } else {
> @@ -798,7 +798,7 @@ static void intel_dsi_pre_enable(struct intel_encoder *encoder,
> vlv_dsi_pll_enable(encoder, pipe_config);
> }
>
> - if (IS_BROXTON(dev_priv)) {
> + if (IS_BROXTON(i915)) {
> /* Add MIPI IO reset programming for modeset */
> val = I915_READ(BXT_P_CR_GT_DISP_PWRON);
> I915_WRITE(BXT_P_CR_GT_DISP_PWRON,
> @@ -809,7 +809,7 @@ static void intel_dsi_pre_enable(struct intel_encoder *encoder,
> I915_WRITE(BXT_P_DSI_REGULATOR_TX_CTRL, 0);
> }
>
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> u32 val;
>
> /* Disable DPOunit clock gating, can stall pipe */
> @@ -818,7 +818,7 @@ static void intel_dsi_pre_enable(struct intel_encoder *encoder,
> I915_WRITE(DSPCLK_GATE_D, val);
> }
>
> - if (!IS_GEMINILAKE(dev_priv))
> + if (!IS_GEMINILAKE(i915))
> intel_dsi_prepare(encoder, pipe_config);
>
> /* Power on, try both CRC pmic gpio and VBT */
> @@ -830,7 +830,7 @@ static void intel_dsi_pre_enable(struct intel_encoder *encoder,
> /* Deassert reset */
> intel_dsi_vbt_exec_sequence(intel_dsi, MIPI_SEQ_DEASSERT_RESET);
>
> - if (IS_GEMINILAKE(dev_priv)) {
> + if (IS_GEMINILAKE(i915)) {
> glk_cold_boot = glk_dsi_enable_io(encoder);
>
> /* Prepare port in cold boot(s3/s4) scenario */
> @@ -842,7 +842,7 @@ static void intel_dsi_pre_enable(struct intel_encoder *encoder,
> intel_dsi_device_ready(encoder);
>
> /* Prepare port in normal boot scenario */
> - if (IS_GEMINILAKE(dev_priv) && !glk_cold_boot)
> + if (IS_GEMINILAKE(i915) && !glk_cold_boot)
> intel_dsi_prepare(encoder, pipe_config);
>
> /* Send initialization commands in LP mode */
> @@ -901,9 +901,9 @@ static void intel_dsi_disable(struct intel_encoder *encoder,
>
> static void intel_dsi_clear_device_ready(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
>
> - if (IS_GEMINILAKE(dev_priv))
> + if (IS_GEMINILAKE(i915))
> glk_dsi_clear_device_ready(encoder);
> else
> vlv_dsi_clear_device_ready(encoder);
> @@ -913,7 +913,7 @@ static void intel_dsi_post_disable(struct intel_encoder *encoder,
> const struct intel_crtc_state *pipe_config,
> const struct drm_connector_state *conn_state)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> enum port port;
> u32 val;
> @@ -941,7 +941,7 @@ static void intel_dsi_post_disable(struct intel_encoder *encoder,
> /* Transition to LP-00 */
> intel_dsi_clear_device_ready(encoder);
>
> - if (IS_BROXTON(dev_priv)) {
> + if (IS_BROXTON(i915)) {
> /* Power down DSI regulator to save power */
> I915_WRITE(BXT_P_DSI_REGULATOR_CFG, STAP_SELECT);
> I915_WRITE(BXT_P_DSI_REGULATOR_TX_CTRL, HS_IO_CTRL_SELECT);
> @@ -952,7 +952,7 @@ static void intel_dsi_post_disable(struct intel_encoder *encoder,
> val & ~MIPIO_RST_CTRL);
> }
>
> - if (IS_GEN9_LP(dev_priv)) {
> + if (IS_GEN9_LP(i915)) {
> bxt_dsi_pll_disable(encoder);
> } else {
> u32 val;
> @@ -983,7 +983,7 @@ static void intel_dsi_post_disable(struct intel_encoder *encoder,
> static bool intel_dsi_get_hw_state(struct intel_encoder *encoder,
> enum pipe *pipe)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> intel_wakeref_t wakeref;
> enum port port;
> @@ -991,7 +991,7 @@ static bool intel_dsi_get_hw_state(struct intel_encoder *encoder,
>
> DRM_DEBUG_KMS("\n");
>
> - wakeref = intel_display_power_get_if_enabled(dev_priv,
> + wakeref = intel_display_power_get_if_enabled(i915,
> encoder->power_domain);
> if (!wakeref)
> return false;
> @@ -1001,12 +1001,12 @@ static bool intel_dsi_get_hw_state(struct intel_encoder *encoder,
> * configuration, otherwise accessing DSI registers will hang the
> * machine. See BSpec North Display Engine registers/MIPI[BXT].
> */
> - if (IS_GEN9_LP(dev_priv) && !bxt_dsi_pll_is_enabled(dev_priv))
> + if (IS_GEN9_LP(i915) && !bxt_dsi_pll_is_enabled(i915))
> goto out_put_power;
>
> /* XXX: this only works for one DSI output */
> for_each_dsi_port(port, intel_dsi->ports) {
> - i915_reg_t ctrl_reg = IS_GEN9_LP(dev_priv) ?
> + i915_reg_t ctrl_reg = IS_GEN9_LP(i915) ?
> BXT_MIPI_PORT_CTRL(port) : MIPI_PORT_CTRL(port);
> bool enabled = I915_READ(ctrl_reg) & DPI_ENABLE;
>
> @@ -1015,7 +1015,7 @@ static bool intel_dsi_get_hw_state(struct intel_encoder *encoder,
> * bit in port C control register does not get set. As a
> * workaround, check pipe B conf instead.
> */
> - if ((IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) &&
> + if ((IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) &&
> port == PORT_C)
> enabled = I915_READ(PIPECONF(PIPE_B)) & PIPECONF_ENABLE;
>
> @@ -1031,7 +1031,7 @@ static bool intel_dsi_get_hw_state(struct intel_encoder *encoder,
> if (!(I915_READ(MIPI_DEVICE_READY(port)) & DEVICE_READY))
> continue;
>
> - if (IS_GEN9_LP(dev_priv)) {
> + if (IS_GEN9_LP(i915)) {
> u32 tmp = I915_READ(MIPI_CTRL(port));
> tmp &= BXT_PIPE_SELECT_MASK;
> tmp >>= BXT_PIPE_SELECT_SHIFT;
> @@ -1049,7 +1049,7 @@ static bool intel_dsi_get_hw_state(struct intel_encoder *encoder,
> }
>
> out_put_power:
> - intel_display_power_put(dev_priv, encoder->power_domain, wakeref);
> + intel_display_power_put(i915, encoder->power_domain, wakeref);
>
> return active;
> }
> @@ -1058,7 +1058,7 @@ static void bxt_dsi_get_pipe_config(struct intel_encoder *encoder,
> struct intel_crtc_state *pipe_config)
> {
> struct drm_device *dev = encoder->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct drm_display_mode *adjusted_mode =
> &pipe_config->base.adjusted_mode;
> struct drm_display_mode *adjusted_mode_sw;
> @@ -1218,13 +1218,13 @@ static void bxt_dsi_get_pipe_config(struct intel_encoder *encoder,
> static void intel_dsi_get_config(struct intel_encoder *encoder,
> struct intel_crtc_state *pipe_config)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> u32 pclk;
> DRM_DEBUG_KMS("\n");
>
> pipe_config->output_types |= BIT(INTEL_OUTPUT_DSI);
>
> - if (IS_GEN9_LP(dev_priv)) {
> + if (IS_GEN9_LP(i915)) {
> bxt_dsi_get_pipe_config(encoder, pipe_config);
> pclk = bxt_dsi_get_pclk(encoder, pipe_config);
> } else {
> @@ -1255,7 +1255,7 @@ static void set_dsi_timings(struct drm_encoder *encoder,
> const struct drm_display_mode *adjusted_mode)
> {
> struct drm_device *dev = encoder->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(encoder);
> enum port port;
> unsigned int bpp = mipi_dsi_pixel_format_to_bpp(intel_dsi->pixel_format);
> @@ -1290,7 +1290,7 @@ static void set_dsi_timings(struct drm_encoder *encoder,
> hbp = txbyteclkhs(hbp, bpp, lane_count, intel_dsi->burst_mode_ratio);
>
> for_each_dsi_port(port, intel_dsi->ports) {
> - if (IS_GEN9_LP(dev_priv)) {
> + if (IS_GEN9_LP(i915)) {
> /*
> * Program hdisplay and vdisplay on MIPI transcoder.
> * This is different from calculated hactive and
> @@ -1342,7 +1342,7 @@ static void intel_dsi_prepare(struct intel_encoder *intel_encoder,
> {
> struct drm_encoder *encoder = &intel_encoder->base;
> struct drm_device *dev = encoder->dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> struct intel_crtc *intel_crtc = to_intel_crtc(pipe_config->base.crtc);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(encoder);
> const struct drm_display_mode *adjusted_mode = &pipe_config->base.adjusted_mode;
> @@ -1362,7 +1362,7 @@ static void intel_dsi_prepare(struct intel_encoder *intel_encoder,
> }
>
> for_each_dsi_port(port, intel_dsi->ports) {
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> /*
> * escape clock divider, 20MHz, shared for A and C.
> * device ready must be off when doing this! txclkesc?
> @@ -1377,7 +1377,7 @@ static void intel_dsi_prepare(struct intel_encoder *intel_encoder,
> tmp &= ~READ_REQUEST_PRIORITY_MASK;
> I915_WRITE(MIPI_CTRL(port), tmp |
> READ_REQUEST_PRIORITY_HIGH);
> - } else if (IS_GEN9_LP(dev_priv)) {
> + } else if (IS_GEN9_LP(i915)) {
> enum pipe pipe = intel_crtc->pipe;
>
> tmp = I915_READ(MIPI_CTRL(port));
> @@ -1415,7 +1415,7 @@ static void intel_dsi_prepare(struct intel_encoder *intel_encoder,
> if (intel_dsi->clock_stop)
> tmp |= CLOCKSTOP;
>
> - if (IS_GEN9_LP(dev_priv)) {
> + if (IS_GEN9_LP(i915)) {
> tmp |= BXT_DPHY_DEFEATURE_EN;
> if (!is_cmd_mode(intel_dsi))
> tmp |= BXT_DEFEATURE_DPI_FIFO_CTR;
> @@ -1466,7 +1466,7 @@ static void intel_dsi_prepare(struct intel_encoder *intel_encoder,
> I915_WRITE(MIPI_INIT_COUNT(port),
> txclkesc(intel_dsi->escape_clk_div, 100));
>
> - if (IS_GEN9_LP(dev_priv) && (!intel_dsi->dual_link)) {
> + if (IS_GEN9_LP(i915) && (!intel_dsi->dual_link)) {
> /*
> * BXT spec says write MIPI_INIT_COUNT for
> * both the ports, even if only one is
> @@ -1500,7 +1500,7 @@ static void intel_dsi_prepare(struct intel_encoder *intel_encoder,
> */
> I915_WRITE(MIPI_LP_BYTECLK(port), intel_dsi->lp_byte_clk);
>
> - if (IS_GEMINILAKE(dev_priv)) {
> + if (IS_GEMINILAKE(i915)) {
> I915_WRITE(MIPI_TLPX_TIME_COUNT(port),
> intel_dsi->lp_byte_clk);
> /* Shadow of DPHY reg */
> @@ -1533,19 +1533,19 @@ static void intel_dsi_prepare(struct intel_encoder *intel_encoder,
>
> static void intel_dsi_unprepare(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> enum port port;
> u32 val;
>
> - if (IS_GEMINILAKE(dev_priv))
> + if (IS_GEMINILAKE(i915))
> return;
>
> for_each_dsi_port(port, intel_dsi->ports) {
> /* Panel commands can be sent when clock is in LP11 */
> I915_WRITE(MIPI_DEVICE_READY(port), 0x0);
>
> - if (IS_GEN9_LP(dev_priv))
> + if (IS_GEN9_LP(i915))
> bxt_dsi_reset_clocks(encoder, port);
> else
> vlv_dsi_reset_clocks(encoder, port);
> @@ -1594,7 +1594,7 @@ static const struct drm_connector_funcs intel_dsi_connector_funcs = {
> static enum drm_panel_orientation
> vlv_dsi_get_hw_panel_orientation(struct intel_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> struct intel_encoder *encoder = connector->encoder;
> enum intel_display_power_domain power_domain;
> enum drm_panel_orientation orientation;
> @@ -1607,11 +1607,11 @@ vlv_dsi_get_hw_panel_orientation(struct intel_connector *connector)
> if (!encoder->get_hw_state(encoder, &pipe))
> return DRM_MODE_PANEL_ORIENTATION_UNKNOWN;
>
> - crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> + crtc = intel_get_crtc_for_pipe(i915, pipe);
> plane = to_intel_plane(crtc->base.primary);
>
> power_domain = POWER_DOMAIN_PIPE(pipe);
> - wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain);
> + wakeref = intel_display_power_get_if_enabled(i915, power_domain);
> if (!wakeref)
> return DRM_MODE_PANEL_ORIENTATION_UNKNOWN;
>
> @@ -1624,7 +1624,7 @@ vlv_dsi_get_hw_panel_orientation(struct intel_connector *connector)
> else
> orientation = DRM_MODE_PANEL_ORIENTATION_NORMAL;
>
> - intel_display_power_put(dev_priv, power_domain, wakeref);
> + intel_display_power_put(i915, power_domain, wakeref);
>
> return orientation;
> }
> @@ -1632,10 +1632,10 @@ vlv_dsi_get_hw_panel_orientation(struct intel_connector *connector)
> static enum drm_panel_orientation
> vlv_dsi_get_panel_orientation(struct intel_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
> enum drm_panel_orientation orientation;
>
> - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> orientation = vlv_dsi_get_hw_panel_orientation(connector);
> if (orientation != DRM_MODE_PANEL_ORIENTATION_UNKNOWN)
> return orientation;
> @@ -1646,13 +1646,13 @@ vlv_dsi_get_panel_orientation(struct intel_connector *connector)
>
> static void intel_dsi_add_properties(struct intel_connector *connector)
> {
> - struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> + struct drm_i915_private *i915 = to_i915(connector->base.dev);
>
> if (connector->panel.fixed_mode) {
> u32 allowed_scalers;
>
> allowed_scalers = BIT(DRM_MODE_SCALE_ASPECT) | BIT(DRM_MODE_SCALE_FULLSCREEN);
> - if (!HAS_GMCH(dev_priv))
> + if (!HAS_GMCH(i915))
> allowed_scalers |= BIT(DRM_MODE_SCALE_CENTER);
>
> drm_connector_attach_scaling_mode_property(&connector->base,
> @@ -1679,8 +1679,8 @@ static void intel_dsi_add_properties(struct intel_connector *connector)
> static void vlv_dphy_param_init(struct intel_dsi *intel_dsi)
> {
> struct drm_device *dev = intel_dsi->base.base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> - struct mipi_config *mipi_config = dev_priv->vbt.dsi.config;
> + struct drm_i915_private *i915 = to_i915(dev);
> + struct mipi_config *mipi_config = i915->vbt.dsi.config;
> u32 tlpx_ns, extra_byte_count, tlpx_ui;
> u32 ui_num, ui_den;
> u32 prepare_cnt, exit_zero_cnt, clk_zero_cnt, trail_cnt;
> @@ -1727,7 +1727,7 @@ static void vlv_dphy_param_init(struct intel_dsi *intel_dsi)
> * For GEMINILAKE dphy_param_reg will be programmed in terms of
> * HS byte clock count for other platform in HS ddr clock count
> */
> - mul = IS_GEMINILAKE(dev_priv) ? 8 : 2;
> + mul = IS_GEMINILAKE(i915) ? 8 : 2;
> ths_prepare_ns = max(mipi_config->ths_prepare,
> mipi_config->tclk_prepare);
>
> @@ -1837,9 +1837,9 @@ static void vlv_dphy_param_init(struct intel_dsi *intel_dsi)
> intel_dsi_log_params(intel_dsi);
> }
>
> -void vlv_dsi_init(struct drm_i915_private *dev_priv)
> +void vlv_dsi_init(struct drm_i915_private *i915)
> {
> - struct drm_device *dev = &dev_priv->drm;
> + struct drm_device *dev = &i915->drm;
> struct intel_dsi *intel_dsi;
> struct intel_encoder *intel_encoder;
> struct drm_encoder *encoder;
> @@ -1851,13 +1851,13 @@ void vlv_dsi_init(struct drm_i915_private *dev_priv)
> DRM_DEBUG_KMS("\n");
>
> /* There is no detection method for MIPI so rely on VBT */
> - if (!intel_bios_is_dsi_present(dev_priv, &port))
> + if (!intel_bios_is_dsi_present(i915, &port))
> return;
>
> - if (IS_GEN9_LP(dev_priv))
> - dev_priv->mipi_mmio_base = BXT_MIPI_BASE;
> + if (IS_GEN9_LP(i915))
> + i915->mipi_mmio_base = BXT_MIPI_BASE;
> else
> - dev_priv->mipi_mmio_base = VLV_MIPI_BASE;
> + i915->mipi_mmio_base = VLV_MIPI_BASE;
>
> intel_dsi = kzalloc(sizeof(*intel_dsi), GFP_KERNEL);
> if (!intel_dsi)
> @@ -1897,20 +1897,20 @@ void vlv_dsi_init(struct drm_i915_private *dev_priv)
> * On BYT/CHV, pipe A maps to MIPI DSI port A, pipe B maps to MIPI DSI
> * port C. BXT isn't limited like this.
> */
> - if (IS_GEN9_LP(dev_priv))
> + if (IS_GEN9_LP(i915))
> intel_encoder->crtc_mask = BIT(PIPE_A) | BIT(PIPE_B) | BIT(PIPE_C);
> else if (port == PORT_A)
> intel_encoder->crtc_mask = BIT(PIPE_A);
> else
> intel_encoder->crtc_mask = BIT(PIPE_B);
>
> - if (dev_priv->vbt.dsi.config->dual_link)
> + if (i915->vbt.dsi.config->dual_link)
> intel_dsi->ports = BIT(PORT_A) | BIT(PORT_C);
> else
> intel_dsi->ports = BIT(port);
>
> - intel_dsi->dcs_backlight_ports = dev_priv->vbt.dsi.bl_ports;
> - intel_dsi->dcs_cabc_ports = dev_priv->vbt.dsi.cabc_ports;
> + intel_dsi->dcs_backlight_ports = i915->vbt.dsi.bl_ports;
> + intel_dsi->dcs_cabc_ports = i915->vbt.dsi.cabc_ports;
>
> /* Create a DSI host (and a device) for each port. */
> for_each_dsi_port(port, intel_dsi->ports) {
> @@ -1949,8 +1949,8 @@ void vlv_dsi_init(struct drm_i915_private *dev_priv)
> * In case of BYT with CRC PMIC, we need to use GPIO for
> * Panel control.
> */
> - if ((IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) &&
> - (dev_priv->vbt.dsi.config->pwm_blc == PPS_BLC_PMIC)) {
> + if ((IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) &&
> + (i915->vbt.dsi.config->pwm_blc == PPS_BLC_PMIC)) {
> intel_dsi->gpio_panel =
> gpiod_get(dev->dev, "panel", GPIOD_OUT_HIGH);
>
> diff --git a/drivers/gpu/drm/i915/vlv_dsi_pll.c b/drivers/gpu/drm/i915/vlv_dsi_pll.c
> index 99cc3e2e9c2c..26655d5472b4 100644
> --- a/drivers/gpu/drm/i915/vlv_dsi_pll.c
> +++ b/drivers/gpu/drm/i915/vlv_dsi_pll.c
> @@ -53,7 +53,7 @@ static u32 dsi_clk_from_pclk(u32 pclk, enum mipi_dsi_pixel_format fmt,
> return dsi_clk_khz;
> }
>
> -static int dsi_calc_mnp(struct drm_i915_private *dev_priv,
> +static int dsi_calc_mnp(struct drm_i915_private *i915,
> struct intel_crtc_state *config,
> int target_dsi_clk)
> {
> @@ -68,7 +68,7 @@ static int dsi_calc_mnp(struct drm_i915_private *dev_priv,
> return -ECHRNG;
> }
>
> - if (IS_CHERRYVIEW(dev_priv)) {
> + if (IS_CHERRYVIEW(i915)) {
> ref_clk = 100000;
> n = 4;
> m_min = 70;
> @@ -116,7 +116,7 @@ static int dsi_calc_mnp(struct drm_i915_private *dev_priv,
> int vlv_dsi_pll_compute(struct intel_encoder *encoder,
> struct intel_crtc_state *config)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> int ret;
> u32 dsi_clk;
> @@ -124,7 +124,7 @@ int vlv_dsi_pll_compute(struct intel_encoder *encoder,
> dsi_clk = dsi_clk_from_pclk(intel_dsi->pclk, intel_dsi->pixel_format,
> intel_dsi->lane_count);
>
> - ret = dsi_calc_mnp(dev_priv, config, dsi_clk);
> + ret = dsi_calc_mnp(i915, config, dsi_clk);
> if (ret) {
> DRM_DEBUG_KMS("dsi_calc_mnp failed\n");
> return ret;
> @@ -147,15 +147,15 @@ int vlv_dsi_pll_compute(struct intel_encoder *encoder,
> void vlv_dsi_pll_enable(struct intel_encoder *encoder,
> const struct intel_crtc_state *config)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
>
> DRM_DEBUG_KMS("\n");
>
> - vlv_cck_get(dev_priv);
> + vlv_cck_get(i915);
>
> - vlv_cck_write(dev_priv, CCK_REG_DSI_PLL_CONTROL, 0);
> - vlv_cck_write(dev_priv, CCK_REG_DSI_PLL_DIVIDER, config->dsi_pll.div);
> - vlv_cck_write(dev_priv, CCK_REG_DSI_PLL_CONTROL,
> + vlv_cck_write(i915, CCK_REG_DSI_PLL_CONTROL, 0);
> + vlv_cck_write(i915, CCK_REG_DSI_PLL_DIVIDER, config->dsi_pll.div);
> + vlv_cck_write(i915, CCK_REG_DSI_PLL_CONTROL,
> config->dsi_pll.ctrl & ~DSI_PLL_VCO_EN);
>
> /* wait at least 0.5 us after ungating before enabling VCO,
> @@ -163,38 +163,38 @@ void vlv_dsi_pll_enable(struct intel_encoder *encoder,
> */
> usleep_range(10, 50);
>
> - vlv_cck_write(dev_priv, CCK_REG_DSI_PLL_CONTROL, config->dsi_pll.ctrl);
> + vlv_cck_write(i915, CCK_REG_DSI_PLL_CONTROL, config->dsi_pll.ctrl);
>
> - if (wait_for(vlv_cck_read(dev_priv, CCK_REG_DSI_PLL_CONTROL) &
> + if (wait_for(vlv_cck_read(i915, CCK_REG_DSI_PLL_CONTROL) &
> DSI_PLL_LOCK, 20)) {
>
> - vlv_cck_put(dev_priv);
> + vlv_cck_put(i915);
> DRM_ERROR("DSI PLL lock failed\n");
> return;
> }
> - vlv_cck_put(dev_priv);
> + vlv_cck_put(i915);
>
> DRM_DEBUG_KMS("DSI PLL locked\n");
> }
>
> void vlv_dsi_pll_disable(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> u32 tmp;
>
> DRM_DEBUG_KMS("\n");
>
> - vlv_cck_get(dev_priv);
> + vlv_cck_get(i915);
>
> - tmp = vlv_cck_read(dev_priv, CCK_REG_DSI_PLL_CONTROL);
> + tmp = vlv_cck_read(i915, CCK_REG_DSI_PLL_CONTROL);
> tmp &= ~DSI_PLL_VCO_EN;
> tmp |= DSI_PLL_LDO_GATE;
> - vlv_cck_write(dev_priv, CCK_REG_DSI_PLL_CONTROL, tmp);
> + vlv_cck_write(i915, CCK_REG_DSI_PLL_CONTROL, tmp);
>
> - vlv_cck_put(dev_priv);
> + vlv_cck_put(i915);
> }
>
> -bool bxt_dsi_pll_is_enabled(struct drm_i915_private *dev_priv)
> +bool bxt_dsi_pll_is_enabled(struct drm_i915_private *i915)
> {
> bool enabled;
> u32 val;
> @@ -216,7 +216,7 @@ bool bxt_dsi_pll_is_enabled(struct drm_i915_private *dev_priv)
> * causes a system hang.
> */
> val = I915_READ(BXT_DSI_PLL_CTL);
> - if (IS_GEMINILAKE(dev_priv)) {
> + if (IS_GEMINILAKE(i915)) {
> if (!(val & BXT_DSIA_16X_MASK)) {
> DRM_DEBUG_DRIVER("Invalid PLL divider (%08x)\n", val);
> enabled = false;
> @@ -233,7 +233,7 @@ bool bxt_dsi_pll_is_enabled(struct drm_i915_private *dev_priv)
>
> void bxt_dsi_pll_disable(struct intel_encoder *encoder)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> u32 val;
>
> DRM_DEBUG_KMS("\n");
> @@ -246,7 +246,7 @@ void bxt_dsi_pll_disable(struct intel_encoder *encoder)
> * PLL lock should deassert within 200us.
> * Wait up to 1ms before timing out.
> */
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> BXT_DSI_PLL_ENABLE,
> BXT_DSI_PLL_LOCKED,
> 0,
> @@ -257,21 +257,21 @@ void bxt_dsi_pll_disable(struct intel_encoder *encoder)
> u32 vlv_dsi_get_pclk(struct intel_encoder *encoder,
> struct intel_crtc_state *config)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> int bpp = mipi_dsi_pixel_format_to_bpp(intel_dsi->pixel_format);
> u32 dsi_clock, pclk;
> u32 pll_ctl, pll_div;
> u32 m = 0, p = 0, n;
> - int refclk = IS_CHERRYVIEW(dev_priv) ? 100000 : 25000;
> + int refclk = IS_CHERRYVIEW(i915) ? 100000 : 25000;
> int i;
>
> DRM_DEBUG_KMS("\n");
>
> - vlv_cck_get(dev_priv);
> - pll_ctl = vlv_cck_read(dev_priv, CCK_REG_DSI_PLL_CONTROL);
> - pll_div = vlv_cck_read(dev_priv, CCK_REG_DSI_PLL_DIVIDER);
> - vlv_cck_put(dev_priv);
> + vlv_cck_get(i915);
> + pll_ctl = vlv_cck_read(i915, CCK_REG_DSI_PLL_CONTROL);
> + pll_div = vlv_cck_read(i915, CCK_REG_DSI_PLL_DIVIDER);
> + vlv_cck_put(i915);
>
> config->dsi_pll.ctrl = pll_ctl & ~DSI_PLL_LOCK;
> config->dsi_pll.div = pll_div;
> @@ -325,7 +325,7 @@ u32 bxt_dsi_get_pclk(struct intel_encoder *encoder,
> u32 dsi_clk;
> u32 dsi_ratio;
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> int bpp = mipi_dsi_pixel_format_to_bpp(intel_dsi->pixel_format);
>
> config->dsi_pll.ctrl = I915_READ(BXT_DSI_PLL_CTL);
> @@ -343,7 +343,7 @@ u32 bxt_dsi_get_pclk(struct intel_encoder *encoder,
> void vlv_dsi_reset_clocks(struct intel_encoder *encoder, enum port port)
> {
> u32 temp;
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
>
> temp = I915_READ(MIPI_CTRL(port));
> @@ -356,7 +356,7 @@ void vlv_dsi_reset_clocks(struct intel_encoder *encoder, enum port port)
> static void glk_dsi_program_esc_clock(struct drm_device *dev,
> const struct intel_crtc_state *config)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> u32 dsi_rate = 0;
> u32 pll_ratio = 0;
> u32 ddr_clk = 0;
> @@ -404,7 +404,7 @@ static void glk_dsi_program_esc_clock(struct drm_device *dev,
> static void bxt_dsi_program_clocks(struct drm_device *dev, enum port port,
> const struct intel_crtc_state *config)
> {
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
> u32 tmp;
> u32 dsi_rate = 0;
> u32 pll_ratio = 0;
> @@ -457,7 +457,7 @@ static void bxt_dsi_program_clocks(struct drm_device *dev, enum port port,
> int bxt_dsi_pll_compute(struct intel_encoder *encoder,
> struct intel_crtc_state *config)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> u8 dsi_ratio, dsi_ratio_min, dsi_ratio_max;
> u32 dsi_clk;
> @@ -472,7 +472,7 @@ int bxt_dsi_pll_compute(struct intel_encoder *encoder,
> */
> dsi_ratio = DIV_ROUND_UP(dsi_clk * 2, BXT_REF_CLOCK_KHZ);
>
> - if (IS_BROXTON(dev_priv)) {
> + if (IS_BROXTON(i915)) {
> dsi_ratio_min = BXT_DSI_PLL_RATIO_MIN;
> dsi_ratio_max = BXT_DSI_PLL_RATIO_MAX;
> } else {
> @@ -496,7 +496,7 @@ int bxt_dsi_pll_compute(struct intel_encoder *encoder,
> /* As per recommendation from hardware team,
> * Prog PVD ratio =1 if dsi ratio <= 50
> */
> - if (IS_BROXTON(dev_priv) && dsi_ratio <= 50)
> + if (IS_BROXTON(i915) && dsi_ratio <= 50)
> config->dsi_pll.ctrl |= BXT_DSI_PLL_PVD_RATIO_1;
>
> return 0;
> @@ -505,7 +505,7 @@ int bxt_dsi_pll_compute(struct intel_encoder *encoder,
> void bxt_dsi_pll_enable(struct intel_encoder *encoder,
> const struct intel_crtc_state *config)
> {
> - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> + struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> enum port port;
> u32 val;
> @@ -517,7 +517,7 @@ void bxt_dsi_pll_enable(struct intel_encoder *encoder,
> POSTING_READ(BXT_DSI_PLL_CTL);
>
> /* Program TX, RX, Dphy clocks */
> - if (IS_BROXTON(dev_priv)) {
> + if (IS_BROXTON(i915)) {
> for_each_dsi_port(port, intel_dsi->ports)
> bxt_dsi_program_clocks(encoder->base.dev, port, config);
> } else {
> @@ -530,7 +530,7 @@ void bxt_dsi_pll_enable(struct intel_encoder *encoder,
> I915_WRITE(BXT_DSI_PLL_ENABLE, val);
>
> /* Timeout and fail if PLL not locked */
> - if (intel_wait_for_register(&dev_priv->uncore,
> + if (intel_wait_for_register(&i915->uncore,
> BXT_DSI_PLL_ENABLE,
> BXT_DSI_PLL_LOCKED,
> BXT_DSI_PLL_LOCKED,
> @@ -546,10 +546,10 @@ void bxt_dsi_reset_clocks(struct intel_encoder *encoder, enum port port)
> {
> u32 tmp;
> struct drm_device *dev = encoder->base.dev;
> - struct drm_i915_private *dev_priv = to_i915(dev);
> + struct drm_i915_private *i915 = to_i915(dev);
>
> /* Clear old configurations */
> - if (IS_BROXTON(dev_priv)) {
> + if (IS_BROXTON(i915)) {
> tmp = I915_READ(BXT_MIPI_CLOCK_CTL);
> tmp &= ~(BXT_MIPI_TX_ESCLK_FIXDIV_MASK(port));
> tmp &= ~(BXT_MIPI_RX_ESCLK_UPPER_FIXDIV_MASK(port));
>
More information about the Intel-gfx
mailing list